Skip to main navigation Skip to search Skip to main content

Knowledge-tuning Large Language Models with Structured Medical Knowledge Bases for Trustworthy Response Generation in Chinese

  • Harbin Institute of Technology

Research output: Contribution to journalArticlepeer-review

Abstract

Large Language Models (LLMs) have demonstrated remarkable success in diverse natural language processing (NLP) tasks in general domains. However, LLMs sometimes generate responses with the hallucination about medical facts due to limited domain knowledge. Such shortcomings pose potential risks in the utilization of LLMs within medical contexts. To address this challenge, we propose knowledge-tuning, which leverages structured medical knowledge bases for the LLMs to grasp domain knowledge efficiently and facilitate trustworthy response generation. We also release cMedKnowQA, a Chinese medical knowledge question-answering dataset constructed from medical knowledge bases to assess the medical knowledge proficiency of LLMs. Experimental results show that the LLMs which are knowledge-tuned with cMedKnowQA can exhibit higher levels of accuracy in response generation compared with vanilla instruction-tuning and offer a new trustworthy way for the domain adaptation of LLMs. We release our code and data at https://github.com/SCIR-HI/Huatuo-Llama-Med-Chinese.

Original languageEnglish
Article number53
JournalACM Transactions on Knowledge Discovery from Data
Volume19
Issue number2
DOIs
StatePublished - 14 Feb 2025

Keywords

  • Large Language Model
  • Medical Knowledge Base
  • Medical Question Answering
  • Trustworthy Response Generation

Fingerprint

Dive into the research topics of 'Knowledge-tuning Large Language Models with Structured Medical Knowledge Bases for Trustworthy Response Generation in Chinese'. Together they form a unique fingerprint.

Cite this