» Articles » PMID: 39347975

Assessing ChatGPT's Summarization of Ga PSMA PET/CT Reports for Patients

Overview
Publisher Springer
Date 2024 Sep 30
PMID 39347975
Authors
Affiliations
Soon will be listed here.
Abstract

Purpose: ChatGPT has recently been the subject of many studies, and its responses to medical questions have been successful. We examined ChatGPT-4's evaluation of structured Ga prostate-specific membrane antigen (PSMA) PET/CT reports of newly diagnosed prostate cancer patients.

Methods: Ga PSMA PET/CT reports of 164 patients were entered to ChatGPT-4. ChatGPT-4 was asked to respond the following questions according to the PET/CT reports: 1-Has the cancer in the prostate extended to organs adjacent to the prostate? 2-Has the cancer in the prostate spread to neighboring lymph nodes? 3-Has the cancer in the prostate spread to lymph nodes in distant areas? 4-Has the cancer in the prostate spread to the bones? 5-Has the cancer in the prostate spread to other organs? ChatGPT-4's responses were scored on a Likert-type scale for clarity and accuracy.

Results: The mean scores for clarity were 4.93 ± 0.32, 4.95 ± 0.25, 4.96 ± 0.19, 4.99 ± 0.11, and 4.96 ± 0.30, respectively. The mean scores for accuracy were 4.87 ∓ 0.61, 4.87 ∓ 0.62, 4.79 ± 0.83, 4.96 ± 0.25, and 4.93 ± 0.45, respectively. Patients with distant lymphatic metastases had a lower mean accuracy score than those without (4.28 ± 1.45 vs. 4.94 ± 0.39; p < 0.001). ChatGPT-4's responses in 13 patients (8%) had the potential for harmful information.

Conclusion: ChatGPT-4 successfully interprets structured Ga PSMA PET/CT reports of reports of newly diagnosed prostate cancer patients. However, it is unlikely that ChatGPT-4 evaluations will replace physicians' evaluations today, especially since it can produce fabricated information.

References
1.
Stengel F, Stienen M, Ivanov M, Gandia-Gonzalez M, Raffa G, Ganau M . Can AI pass the written European Board Examination in Neurological Surgery? - Ethical and practical issues. Brain Spine. 2024; 4:102765. PMC: 10951784. DOI: 10.1016/j.bas.2024.102765. View

2.
Chen T, Multala E, Kearns P, Delashaw J, Dumont A, Maraganore D . Assessment of ChatGPT's performance on neurology written board examination questions. BMJ Neurol Open. 2023; 5(2):e000530. PMC: 10626870. DOI: 10.1136/bmjno-2023-000530. View

3.
Ghanem D, Nassar J, El Bachour J, Hanna T . ChatGPT Earns American Board Certification in Hand Surgery. Hand Surg Rehabil. 2024; 43(3):101688. DOI: 10.1016/j.hansur.2024.101688. View

4.
Mallio C, Bernetti C, Sertorio A, Zobel B . ChatGPT in radiology structured reporting: analysis of ChatGPT-3.5 Turbo and GPT-4 in reducing word count and recalling findings. Quant Imaging Med Surg. 2024; 14(2):2096-2102. PMC: 10895108. DOI: 10.21037/qims-23-1300. View

5.
Mallio C, Sertorio A, Bernetti C, Zobel B . Large language models for structured reporting in radiology: performance of GPT-4, ChatGPT-3.5, Perplexity and Bing. Radiol Med. 2023; 128(7):808-812. DOI: 10.1007/s11547-023-01651-4. View