» Articles » PMID: 32562898

Understanding Spatial Language in Radiology: Representation Framework, Annotation, and Spatial Relation Extraction from Chest X-ray Reports Using Deep Learning

Overview
Journal J Biomed Inform
Publisher Elsevier
Date 2020 Jun 21
PMID 32562898
Citations 8
Authors
Affiliations
Soon will be listed here.
Abstract

Radiology reports contain a radiologist's interpretations of images, and these images frequently describe spatial relations. Important radiographic findings are mostly described in reference to an anatomical location through spatial prepositions. Such spatial relationships are also linked to various differential diagnoses and often described through uncertainty phrases. Structured representation of this clinically significant spatial information has the potential to be used in a variety of downstream clinical informatics applications. Our focus is to extract these spatial representations from the reports. For this, we first define a representation framework based on the Spatial Role Labeling (SpRL) scheme, which we refer to as Rad-SpRL. In Rad-SpRL, common radiological entities tied to spatial relations are encoded through four spatial roles: Trajector, Landmark, Diagnosis, and Hedge, all identified in relation to a spatial preposition (or Spatial Indicator). We annotated a total of 2,000 chest X-ray reports following Rad-SpRL. We then propose a deep learning-based natural language processing (NLP) method involving word and character-level encodings to first extract the Spatial Indicators followed by identifying the corresponding spatial roles. Specifically, we use a bidirectional long short-term memory (Bi-LSTM) conditional random field (CRF) neural network as the baseline model. Additionally, we incorporate contextualized word representations from pre-trained language models (BERT and XLNet) for extracting the spatial information. We evaluate both gold and predicted Spatial Indicators to extract the four types of spatial roles. The results are promising, with the highest average F1 measure for Spatial Indicator extraction being 91.29 (XLNet); the highest average overall F1 measure considering all the four spatial roles being 92.9 using gold Indicators (XLNet); and 85.6 using predicted Indicators (BERT pre-trained on MIMIC notes). The corpus is available in Mendeley at http://dx.doi.org/10.17632/yhb26hfz8n.1 and https://github.com/krobertslab/datasets/blob/master/Rad-SpRL.xml.

Citing Articles

A scoping review of large language model based approaches for information extraction from radiology reports.

Reichenpfader D, Muller H, Denecke K NPJ Digit Med. 2024; 7(1):222.

PMID: 39182008 PMC: 11344824. DOI: 10.1038/s41746-024-01219-0.


Event-Based Clinical Finding Extraction from Radiology Reports with Pre-trained Language Model.

Lau W, Lybarger K, Gunn M, Yetisgen M J Digit Imaging. 2022; 36(1):91-104.

PMID: 36253581 PMC: 9576130. DOI: 10.1007/s10278-022-00717-5.


Increasing Women's Knowledge about HPV Using BERT Text Summarization: An Online Randomized Study.

Bitar H, Babour A, Nafa F, Alzamzami O, Alismail S Int J Environ Res Public Health. 2022; 19(13).

PMID: 35805761 PMC: 9265758. DOI: 10.3390/ijerph19138100.


Identifying ARDS using the Hierarchical Attention Network with Sentence Objectives Framework.

Lybarger K, Mabrey L, Thau M, Bhatraju P, Wurfel M, Yetisgen M AMIA Annu Symp Proc. 2022; 2021:823-832.

PMID: 35308902 PMC: 8861765.


Deep Learning-Based Natural Language Processing in Radiology: The Impact of Report Complexity, Disease Prevalence, Dataset Size, and Algorithm Type on Model Performance.

Olthof A, van Ooijen P, Cornelissen L J Med Syst. 2021; 45(10):91.

PMID: 34480231 PMC: 8416876. DOI: 10.1007/s10916-021-01761-4.


References
1.
Demner-Fushman D, Kohli M, Rosenman M, Shooshan S, Rodriguez L, Antani S . Preparing a collection of radiology examinations for distribution and retrieval. J Am Med Inform Assoc. 2015; 23(2):304-10. PMC: 5009925. DOI: 10.1093/jamia/ocv080. View

2.
Langlotz C . RadLex: a new method for indexing online educational materials. Radiographics. 2006; 26(6):1595-7. DOI: 10.1148/rg.266065168. View

3.
Roberts K, Rodriguez L, Shooshan S, Demner-Fushman D . Automatic Extraction and Post-coordination of Spatial Relations in Consumer Language. AMIA Annu Symp Proc. 2016; 2015:1083-92. PMC: 4765706. View

4.
Lee J, Yoon W, Kim S, Kim D, Kim S, So C . BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2019; 36(4):1234-1240. PMC: 7703786. DOI: 10.1093/bioinformatics/btz682. View

5.
Annarumma M, Withey S, Bakewell R, Pesce E, Goh V, Montana G . Automated Triaging of Adult Chest Radiographs with Deep Artificial Neural Networks. Radiology. 2019; 291(1):196-202. PMC: 6438359. DOI: 10.1148/radiol.2018180921. View