» Articles » PMID: 34039433

Text Mining to Support Abstract Screening for Knowledge Syntheses: a Semi-automated Workflow

Overview
Journal Syst Rev
Publisher Biomed Central
Date 2021 May 27
PMID 34039433
Citations 14
Authors
Affiliations
Soon will be listed here.
Abstract

Background: Current text mining tools supporting abstract screening in systematic reviews are not widely used, in part because they lack sensitivity and precision. We set out to develop an accessible, semi-automated "workflow" to conduct abstract screening for systematic reviews and other knowledge synthesis methods.

Methods: We adopt widely recommended text-mining and machine-learning methods to (1) process title-abstracts into numerical training data; and (2) train a classification model to predict eligible abstracts. The predicted abstracts are screened by human reviewers for ("true") eligibility, and the newly eligible abstracts are used to identify similar abstracts, using near-neighbor methods, which are also screened. These abstracts, as well as their eligibility results, are used to update the classification model, and the above steps are iterated until no new eligible abstracts are identified. The workflow was implemented in R and evaluated using a systematic review of insulin formulations for type-1 diabetes (14,314 abstracts) and a scoping review of knowledge-synthesis methods (17,200 abstracts). Workflow performance was evaluated against the recommended practice of screening abstracts by 2 reviewers, independently. Standard measures were examined: sensitivity (inclusion of all truly eligible abstracts), specificity (exclusion of all truly ineligible abstracts), precision (inclusion of all truly eligible abstracts among all abstracts screened as eligible), F1-score (harmonic average of sensitivity and precision), and accuracy (correctly predicted eligible or ineligible abstracts). Workload reduction was measured as the hours the workflow saved, given only a subset of abstracts needed human screening.

Results: With respect to the systematic and scoping reviews respectively, the workflow attained 88%/89% sensitivity, 99%/99% specificity, 71%/72% precision, an F1-score of 79%/79%, 98%/97% accuracy, 63%/55% workload reduction, with 12%/11% fewer abstracts for full-text retrieval and screening, and 0%/1.5% missed studies in the completed reviews.

Conclusion: The workflow was a sensitive, precise, and efficient alternative to the recommended practice of screening abstracts with 2 reviewers. All eligible studies were identified in the first case, while 6 studies (1.5%) were missed in the second that would likely not impact the review's conclusions. We have described the workflow in language accessible to reviewers with limited exposure to natural language processing and machine learning, and have made the code available to reviewers.

Citing Articles

Trial characteristics, geographic distribution, and selected methodological issues of 1425 infertility trials published from 2012 to 2023: a systematic review.

Feng Q, Li W, Crispin J, Longobardi S, DHooghe T, Mol B Hum Reprod Open. 2025; 2025(1):hoaf004.

PMID: 39980657 PMC: 11842059. DOI: 10.1093/hropen/hoaf004.


Human-Comparable Sensitivity of Large Language Models in Identifying Eligible Studies Through Title and Abstract Screening: 3-Layer Strategy Using GPT-3.5 and GPT-4 for Systematic Reviews.

Matsui K, Utsumi T, Aoki Y, Maruki T, Takeshima M, Takaesu Y J Med Internet Res. 2024; 26:e52758.

PMID: 39151163 PMC: 11364944. DOI: 10.2196/52758.


Machine learning enables automated screening for systematic reviews and meta-analysis in urology.

Menold H, Wieland V, Haney C, Uysal D, Wessels F, Cacciamani G World J Urol. 2024; 42(1):396.

PMID: 38985296 PMC: 11236840. DOI: 10.1007/s00345-024-05078-y.


Automation of systematic reviews of biomedical literature: a scoping review of studies indexed in PubMed.

Toth B, Berek L, Gulacsi L, Pentek M, Zrubka Z Syst Rev. 2024; 13(1):174.

PMID: 38978132 PMC: 11229257. DOI: 10.1186/s13643-024-02592-3.


Title and abstract screening for literature reviews using large language models: an exploratory study in the biomedical domain.

Dennstadt F, Zink J, Putora P, Hastings J, Cihoric N Syst Rev. 2024; 13(1):158.

PMID: 38879534 PMC: 11180407. DOI: 10.1186/s13643-024-02575-4.


References
1.
Cooper M, Ungar W, Zlotkin S . An assessment of inter-rater agreement of the literature filtering process in the development of evidence-based dietary guidelines. Public Health Nutr. 2006; 9(4):494-500. DOI: 10.1079/phn2005877. View

2.
Shemilt I, Khan N, Park S, Thomas J . Use of cost-effectiveness analysis to compare the efficiency of study identification methods in systematic reviews. Syst Rev. 2016; 5(1):140. PMC: 4989498. DOI: 10.1186/s13643-016-0315-4. View

3.
OMara-Eves A, Thomas J, McNaught J, Miwa M, Ananiadou S . Using text mining for study identification in systematic reviews: a systematic review of current approaches. Syst Rev. 2015; 4:5. PMC: 4320539. DOI: 10.1186/2046-4053-4-5. View

4.
LeCun Y, Bengio Y, Hinton G . Deep learning. Nature. 2015; 521(7553):436-44. DOI: 10.1038/nature14539. View

5.
Matwin S, Kouznetsov A, Inkpen D, Frunza O, OBlenis P . A new algorithm for reducing the workload of experts in performing systematic reviews. J Am Med Inform Assoc. 2010; 17(4):446-53. PMC: 2995653. DOI: 10.1136/jamia.2010.004325. View