MKRAG: Medical Knowledge Retrieval Augmented Generation for Medical Question Answering.

Journal: AMIA ... Annual Symposium proceedings. AMIA Symposium
Published Date:

Abstract

Large Language Models (LLMs), although powerful in general domains, often perform poorly on domain-specific tasks such as medical question answering (QA). In addition, LLMs tend to function as "black-boxes", making it challenging to modify their behavior. To address the problem, our work employs a transparent process of retrieval augmented generation (RAG), aiming to improve LLM responses without the need for fine-tuning or retraining. Specifically, we propose a comprehensive retrieval strategy to extract medical facts from an external knowledge base, and then inject them into the LLM's query prompt. Focusing on medical QA, we evaluate the impact of different retrieval models and the number of facts on LLM performance using the MedQA-SMILE dataset. Notably, our retrieval-augmented Vicuna-7B model exhibited an accuracy improvement from 44.46% to 48.54%. This work underscores the potential of RAG to enhance LLM performance, offering a practical approach to mitigate the challenges posed by black-box LLMs.

Authors

  • Yucheng Shi
    College of Intelligence and Computing, Tianjin University, Tianjin 300072, China.
  • Shaochen Xu
    School of Computing, University of Georgia, Athens, GA, United States.
  • Tianze Yang
    School of Computing, University of Georgia, Athens, GA 30602 USA.
  • Zhengliang Liu
    School of Computing, University of Georgia, Athens, GA, United States.
  • Tianming Liu
    School of Computing, University of Georgia, Athens, GA, United States.
  • Xiang Li
    Department of Radiology, Massachusetts General Hospital and Harvard Medical School, Boston, MA, United States.
  • Ninghao Liu
    School of Computing, University of Georgia, Athens, GA, United States.