RT: a Retrieving and Chain-of-Thought Framework for Few-shot Medical Named Entity Recognition
Overview
Authors
Affiliations
Objectives: This article aims to enhance the performance of larger language models (LLMs) on the few-shot biomedical named entity recognition (NER) task by developing a simple and effective method called Retrieving and Chain-of-Thought (RT) framework and to evaluate the improvement after applying RT framework.
Materials And Methods: Given the remarkable advancements in retrieval-based language model and Chain-of-Thought across various natural language processing tasks, we propose a pioneering RT framework designed to amalgamate both approaches. The RT approach encompasses dedicated modules for information retrieval and Chain-of-Thought processes. In the retrieval module, RT discerns pertinent examples from demonstrations during instructional tuning for each input sentence. Subsequently, the Chain-of-Thought module employs a systematic reasoning process to identify entities. We conducted a comprehensive comparative analysis of our RT framework against 16 other models for few-shot NER tasks on BC5CDR and NCBI corpora. Additionally, we explored the impacts of negative samples, output formats, and missing data on performance.
Results: Our proposed RT framework outperforms other LMs for few-shot NER tasks with micro-F1 scores of 93.50 and 91.76 on BC5CDR and NCBI corpora, respectively. We found that using both positive and negative samples, Chain-of-Thought (vs Tree-of-Thought) performed better. Additionally, utilization of a partially annotated dataset has a marginal effect of the model performance.
Discussion: This is the first investigation to combine a retrieval-based LLM and Chain-of-Thought methodology to enhance the performance in biomedical few-shot NER. The retrieval-based LLM aids in retrieving the most relevant examples of the input sentence, offering crucial knowledge to predict the entity in the sentence. We also conducted a meticulous examination of our methodology, incorporating an ablation study.
Conclusion: The RT framework with LLM has demonstrated state-of-the-art performance on few-shot NER tasks.
BiomedRAG: A retrieval augmented large language model for biomedicine.
Li M, Kilicoglu H, Xu H, Zhang R J Biomed Inform. 2025; 162:104769.
PMID: 39814274 PMC: 11837810. DOI: 10.1016/j.jbi.2024.104769.
Asad M Cureus. 2024; 16(9):e69710.
PMID: 39308847 PMC: 11415727. DOI: 10.7759/cureus.69710.
Large language models in biomedicine and health: current research landscape and future directions.
Lu Z, Peng Y, Cohen T, Ghassemi M, Weng C, Tian S J Am Med Inform Assoc. 2024; 31(9):1801-1811.
PMID: 39169867 PMC: 11339542. DOI: 10.1093/jamia/ocae202.