» Articles » PMID: 38199604

Assessing Prognosis in Depression: Comparing Perspectives of AI Models, Mental Health Professionals and the General Public

Overview
Date 2024 Jan 10
PMID 38199604
Authors
Affiliations
Soon will be listed here.
Abstract

Background: Artificial intelligence (AI) has rapidly permeated various sectors, including healthcare, highlighting its potential to facilitate mental health assessments. This study explores the underexplored domain of AI's role in evaluating prognosis and long-term outcomes in depressive disorders, offering insights into how AI large language models (LLMs) compare with human perspectives.

Methods: Using case vignettes, we conducted a comparative analysis involving different LLMs (ChatGPT-3.5, ChatGPT-4, Claude and Bard), mental health professionals (general practitioners, psychiatrists, clinical psychologists and mental health nurses), and the general public that reported previously. We evaluate the LLMs ability to generate prognosis, anticipated outcomes with and without professional intervention, and envisioned long-term positive and negative consequences for individuals with depression.

Results: In most of the examined cases, the four LLMs consistently identified depression as the primary diagnosis and recommended a combined treatment of psychotherapy and antidepressant medication. ChatGPT-3.5 exhibited a significantly pessimistic prognosis distinct from other LLMs, professionals and the public. ChatGPT-4, Claude and Bard aligned closely with mental health professionals and the general public perspectives, all of whom anticipated no improvement or worsening without professional help. Regarding long-term outcomes, ChatGPT 3.5, Claude and Bard consistently projected significantly fewer negative long-term consequences of treatment than ChatGPT-4.

Conclusions: This study underscores the potential of AI to complement the expertise of mental health professionals and promote a collaborative paradigm in mental healthcare. The observation that three of the four LLMs closely mirrored the anticipations of mental health experts in scenarios involving treatment underscores the technology's prospective value in offering professional clinical forecasts. The pessimistic outlook presented by ChatGPT 3.5 is concerning, as it could potentially diminish patients' drive to initiate or continue depression therapy. In summary, although LLMs show potential in enhancing healthcare services, their utilisation requires thorough verification and a seamless integration with human judgement and skills.

Citing Articles

The Feasibility of Large Language Models in Verbal Comprehension Assessment: Mixed Methods Feasibility Study.

Hadar-Shoval D, Lvovsky M, Asraf K, Shimoni Y, Elyoseph Z JMIR Form Res. 2025; 9:e68347.

PMID: 39993720 PMC: 11894350. DOI: 10.2196/68347.


The externalization of internal experiences in psychotherapy through generative artificial intelligence: a theoretical, clinical, and ethical analysis.

Haber Y, Hadar Shoval D, Levkovich I, Yinon D, Gigi K, Pen O Front Digit Health. 2025; 7:1512273.

PMID: 39968063 PMC: 11832678. DOI: 10.3389/fdgth.2025.1512273.


Evaluating Diagnostic Accuracy and Treatment Efficacy in Mental Health: A Comparative Analysis of Large Language Model Tools and Mental Health Professionals.

Levkovich I Eur J Investig Health Psychol Educ. 2025; 15(1).

PMID: 39852192 PMC: 11765082. DOI: 10.3390/ejihpe15010009.


Is Artificial Intelligence the Next Co-Pilot for Primary Care in Diagnosing and Recommending Treatments for Depression?.

Levkovich I Med Sci (Basel). 2025; 13(1.

PMID: 39846703 PMC: 11755475. DOI: 10.3390/medsci13010008.


Developing a Machine Learning-Based Automated Patient Engagement Estimator for Telehealth: Algorithm Development and Validation Study.

Guhan P, Awasthi N, McDonald K, Bussell K, Reeves G, Manocha D JMIR Form Res. 2025; 9:e46390.

PMID: 39832353 PMC: 11791444. DOI: 10.2196/46390.


References
1.
Kennedy S, Lanceley A, Whitten M, Kelly C, Nicholls J . Consent on the labour ward: A qualitative study of the views and experiences of healthcare professionals. Eur J Obstet Gynecol Reprod Biol. 2021; 264:150-154. DOI: 10.1016/j.ejogrb.2021.07.003. View

2.
Cuijpers P, Reijnders M, Huibers M . The Role of Common Factors in Psychotherapy Outcomes. Annu Rev Clin Psychol. 2018; 15:207-231. DOI: 10.1146/annurev-clinpsy-050718-095424. View

3.
Levkovich I, Elyoseph Z . Suicide Risk Assessments Through the Eyes of ChatGPT-3.5 Versus ChatGPT-4: Vignette Study. JMIR Ment Health. 2023; 10:e51232. PMC: 10551796. DOI: 10.2196/51232. View

4.
Wittchen H, Muhlig S, Beesdo K . Mental disorders in primary care. Dialogues Clin Neurosci. 2011; 5(2):115-28. PMC: 3181625. View

5.
Barth J, Munder T, Gerger H, Nuesch E, Trelle S, Znoj H . Comparative Efficacy of Seven Psychotherapeutic Interventions for Patients with Depression: A Network Meta-Analysis. Focus (Am Psychiatr Publ). 2020; 14(2):229-243. PMC: 6519647. DOI: 10.1176/appi.focus.140201. View