» Articles » PMID: 39330905

Enhancing the Interpretability of Malaria and Typhoid Diagnosis with Explainable AI and Large Language Models

Abstract

Malaria and Typhoid fever are prevalent diseases in tropical regions, and both are exacerbated by unclear protocols, drug resistance, and environmental factors. Prompt and accurate diagnosis is crucial to improve accessibility and reduce mortality rates. Traditional diagnosis methods cannot effectively capture the complexities of these diseases due to the presence of similar symptoms. Although machine learning (ML) models offer accurate predictions, they operate as "black boxes" with non-interpretable decision-making processes, making it challenging for healthcare providers to comprehend how the conclusions are reached. This study employs explainable AI (XAI) models such as Local Interpretable Model-agnostic Explanations (LIME), and Large Language Models (LLMs) like GPT to clarify diagnostic results for healthcare workers, building trust and transparency in medical diagnostics by describing which symptoms had the greatest impact on the model's decisions and providing clear, understandable explanations. The models were implemented on Google Colab and Visual Studio Code because of their rich libraries and extensions. Results showed that the Random Forest model outperformed the other tested models; in addition, important features were identified with the LIME plots while ChatGPT 3.5 had a comparative advantage over other LLMs. The study integrates RF, LIME, and GPT in building a mobile app to enhance the interpretability and transparency in malaria and typhoid diagnosis system. Despite its promising results, the system's performance is constrained by the quality of the dataset. Additionally, while LIME and GPT improve transparency, they may introduce complexities in real-time deployment due to computational demands and the need for internet service to maintain relevance and accuracy. The findings suggest that AI-driven diagnostic systems can significantly enhance healthcare delivery in environments with limited resources, and future works can explore the applicability of this framework to other medical conditions and datasets.

References
1.
Ghosh D, Cabrera J . Enriched Random Forest for High Dimensional Genomic Data. IEEE/ACM Trans Comput Biol Bioinform. 2021; 19(5):2817-2828. PMC: 9923687. DOI: 10.1109/TCBB.2021.3089417. View

2.
Fan Y, Lu X, Sun G . IHCP: interpretable hepatitis C prediction system based on black-box machine learning models. BMC Bioinformatics. 2023; 24(1):333. PMC: 10481489. DOI: 10.1186/s12859-023-05456-0. View

3.
Nakamura Y, Hanaoka S, Nomura Y, Nakao T, Miki S, Watadani T . Automatic detection of actionable radiology reports using bidirectional encoder representations from transformers. BMC Med Inform Decis Mak. 2021; 21(1):262. PMC: 8436473. DOI: 10.1186/s12911-021-01623-6. View

4.
Mather R, Hopkins H, Parry C, Dittrich S . Redefining typhoid diagnosis: what would an improved test need to look like?. BMJ Glob Health. 2019; 4(5):e001831. PMC: 6830052. DOI: 10.1136/bmjgh-2019-001831. View

5.
Galan J . Typhoid toxin provides a window into typhoid fever and the biology of Salmonella Typhi. Proc Natl Acad Sci U S A. 2016; 113(23):6338-44. PMC: 4988619. DOI: 10.1073/pnas.1606335113. View