Categories
Nevin Manimala Statistics

Addressing the challenges of reconstructing systematic reviews datasets: a case study and a noisy label filter procedure

Syst Rev. 2024 Feb 17;13(1):69. doi: 10.1186/s13643-024-02472-w.

ABSTRACT

Systematic reviews and meta-analyses typically require significant time and effort. Machine learning models have the potential to enhance screening efficiency in these processes. To effectively evaluate such models, fully labeled datasets-detailing all records screened by humans and their labeling decisions-are imperative. This paper presents the creation of a comprehensive dataset for a systematic review of treatments for Borderline Personality Disorder, as reported by Oud et al. (2018) for running a simulation study. The authors adhered to the PRISMA guidelines and published both the search query and the list of included records, but the complete dataset with all labels was not disclosed. We replicated their search and, facing the absence of initial screening data, introduced a Noisy Label Filter (NLF) procedure using active learning to validate noisy labels. Following the NLF application, no further relevant records were found. A simulation study employing the reconstructed dataset demonstrated that active learning could reduce screening time by 82.30% compared to random reading. The paper discusses potential causes for discrepancies, provides recommendations, and introduces a decision tree to assist in reconstructing datasets for the purpose of running simulation studies.

PMID:38368379 | DOI:10.1186/s13643-024-02472-w

By Nevin Manimala

Portfolio Website for Nevin Manimala