Automatic Endpoint Detection to Support the Systematic Review Process
Overview
Authors
Affiliations
Preparing a systematic review can take hundreds of hours to complete, but the process of reconciling different results from multiple studies is the bedrock of evidence-based medicine. We introduce a two-step approach to automatically extract three facets - two entities (the agent and object) and the way in which the entities are compared (the endpoint) - from direct comparative sentences in full-text articles. The system does not require a user to predefine entities in advance and thus can be used in domains where entity recognition is difficult or unavailable. As with a systematic review, the tabular summary produced using the automatically extracted facets shows how experimental results differ between studies. Experiments were conducted using a collection of more than 2million sentences from three journals Diabetes, Carcinogenesis and Endocrinology and two machine learning algorithms, support vector machines (SVM) and a general linear model (GLM). F1 and accuracy measures for the SVM and GLM differed by only 0.01 across all three comparison facets in a randomly selected set of test sentences. The system achieved the best performance of 92% for objects, whereas the accuracy for both agent and endpoints was 73%. F1 scores were higher for objects (0.77) than for endpoints (0.51) or agents (0.47). A situated evaluation of Metformin, a drug to treat diabetes, showed system accuracy of 95%, 83% and 79% for the object, endpoint and agent respectively. The situated evaluation had higher F1 scores of 0.88, 0.64 and 0.62 for object, endpoint, and agent respectively. On average, only 5.31% of the sentences in a full-text article are direct comparisons, but the tabular summaries suggest that these sentences provide a rich source of currently underutilized information that can be used to accelerate the systematic review process and identify gaps where future research should be focused.
Toth B, Berek L, Gulacsi L, Pentek M, Zrubka Z Syst Rev. 2024; 13(1):174.
PMID: 38978132 PMC: 11229257. DOI: 10.1186/s13643-024-02592-3.
Using semantics to scale up evidence-based chemical risk-assessments.
Blake C, Flaws J PLoS One. 2021; 16(12):e0260712.
PMID: 34910747 PMC: 8673667. DOI: 10.1371/journal.pone.0260712.
Schmidt L, Finnerty Mutlu A, Elmore R, McGuinness L, Olorisade B, Thomas J F1000Res. 2023; 10:401.
PMID: 34408850 PMC: 8361807. DOI: 10.12688/f1000research.51117.2.
What Does the Evidence Say? Models to Help Make Sense of the Biomedical Literature.
Wallace B IJCAI (U S). 2021; 2019:6416-6420.
PMID: 34025086 PMC: 8136417. DOI: 10.24963/ijcai.2019/899.
Improving Endpoint Detection to Support Automated Systematic Reviews.
Lucic A, Blake C AMIA Annu Symp Proc. 2017; 2016:1900-1909.
PMID: 28269949 PMC: 5333237.