» Articles » PMID: 37973919

A Study of Generative Large Language Model for Medical Research and Healthcare

Abstract

There are enormous enthusiasm and concerns in applying large language models (LLMs) to healthcare. Yet current assumptions are based on general-purpose LLMs such as ChatGPT, which are not developed for medical use. This study develops a generative clinical LLM, GatorTronGPT, using 277 billion words of text including (1) 82 billion words of clinical text from 126 clinical departments and approximately 2 million patients at the University of Florida Health and (2) 195 billion words of diverse general English text. We train GatorTronGPT using a GPT-3 architecture with up to 20 billion parameters and evaluate its utility for biomedical natural language processing (NLP) and healthcare text generation. GatorTronGPT improves biomedical natural language processing. We apply GatorTronGPT to generate 20 billion words of synthetic text. Synthetic NLP models trained using synthetic text generated by GatorTronGPT outperform models trained using real-world clinical text. Physicians' Turing test using 1 (worst) to 9 (best) scale shows that there are no significant differences in linguistic readability (p = 0.22; 6.57 of GatorTronGPT compared with 6.93 of human) and clinical relevance (p = 0.91; 7.0 of GatorTronGPT compared with 6.97 of human) and that physicians cannot differentiate them (p < 0.001). This study provides insights into the opportunities and challenges of LLMs for medical research and healthcare.

Citing Articles

Low responsiveness of machine learning models to critical or deteriorating health conditions.

Pias T, Afrose S, Tuli M, Trisha I, Deng X, Nemeroff C Commun Med (Lond). 2025; 5(1):62.

PMID: 40069422 PMC: 11897252. DOI: 10.1038/s43856-025-00775-0.


Agents for Change: Artificial Intelligent Workflows for Quantitative Clinical Pharmacology and Translational Sciences.

Shahin M, Goswami S, Lobentanzer S, Corrigan B Clin Transl Sci. 2025; 18(3):e70188.

PMID: 40055986 PMC: 11889410. DOI: 10.1111/cts.70188.


Medical foundation large language models for comprehensive text analysis and beyond.

Xie Q, Chen Q, Chen A, Peng C, Hu Y, Lin F NPJ Digit Med. 2025; 8(1):141.

PMID: 40044845 PMC: 11882967. DOI: 10.1038/s41746-025-01533-1.


Analyzing patient perspectives with large language models: a cross-sectional study of sentiment and thematic classification on exception from informed consent.

Kornblith A, Singh C, Innes J, Chang T, Adelgais K, Holsti M Sci Rep. 2025; 15(1):6179.

PMID: 39979559 PMC: 11842787. DOI: 10.1038/s41598-025-89996-w.


InfectA-Chat, an Arabic Large Language Model for Infectious Diseases: Comparative Analysis.

Selcuk Y, Kim E, Ahn I JMIR Med Inform. 2025; 13:e63881.

PMID: 39928922 PMC: 11851044. DOI: 10.2196/63881.


References
1.
Luo R, Sun L, Xia Y, Qin T, Zhang S, Poon H . BioGPT: generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform. 2022; 23(6). DOI: 10.1093/bib/bbac409. View

2.
Yang X, Lyu T, Li Q, Lee C, Bian J, Hogan W . A study of deep learning methods for de-identification of clinical notes in cross-institute settings. BMC Med Inform Decis Mak. 2019; 19(Suppl 5):232. PMC: 6894104. DOI: 10.1186/s12911-019-0935-4. View

3.
Wongpakaran N, Wongpakaran T, Wedding D, Gwet K . A comparison of Cohen's Kappa and Gwet's AC1 when calculating inter-rater reliability coefficients: a study conducted with personality disorder samples. BMC Med Res Methodol. 2013; 13:61. PMC: 3643869. DOI: 10.1186/1471-2288-13-61. View

4.
Kroth P, Morioka-Douglas N, Veres S, Babbott S, Poplau S, Qeadan F . Association of Electronic Health Record Design and Use Factors With Clinician Stress and Burnout. JAMA Netw Open. 2019; 2(8):e199609. PMC: 6704736. DOI: 10.1001/jamanetworkopen.2019.9609. View

5.
Grunebaum A, Chervenak J, Pollet S, Katz A, Chervenak F . The exciting potential for ChatGPT in obstetrics and gynecology. Am J Obstet Gynecol. 2023; 228(6):696-705. DOI: 10.1016/j.ajog.2023.03.009. View