首页|期刊导航|Digital Chinese Medicine|TCMLLM-PR:evaluation of large language models for prescription recommendation in traditional Chinese medicine

TCMLLM-PR:evaluation of large language models for prescription recommendation in traditional Chinese medicineOA

中文摘要

Objective To develop and evaluate a fine-tuned large language model(LLM)for traditional Chinese medicine(TCM)prescription recommendation named TCMLLM-PR.Methods First,we constructed an instruction-tuning dataset containing 68654 samples(ap-proximately 10 million tokens)by integrating data from eight sources,including four TCM textbooks,Pharmacopoeia of the People’s Republic of China 2020(CHP),Chinese Medicine Clinical Cases(CMCC),and hospital clinical records covering lung disease,liver disease,stroke,diabetes,and splenic-stomach disease.Then,we trained TCMLLM-PR using Chat-GLM-6B with P-Tuning v2 technology.The evaluation consisted of three aspects:(i)compari-son with traditional prescription recommendation models(PTM,TCMPR,and PresRecST);(ii)comparison with TCM-specific LLMs(ShenNong,Huatuo,and HuatuoGPT)and general-domain ChatGPT;(iii)assessment of model migration capability across different disease datasets.We employed precision,recall,and F1 score as evaluation metrics.Results The experiments showed that TCMLLM-PR significantly outperformed baseline models on TCM textbooks and CHP datasets,with F1@10 improvements of 31.80%and 59.48%,respectively.In cross-dataset validation,the model performed best when migrating from TCM textbooks to liver disease dataset,achieving an F1@10 of 0.1551.Analysis of real-world cases demonstrated that TCMLLM-PR''s prescription recommendations most closely matched actual doctors’prescriptions.Conclusion This study integrated LLMs into TCM prescription recommendations,leverag-ing a tailored instruction-tuning dataset and developing TCMLLM-PR.This study will pub-licly release the best model parameters of TCMLLM-PR to promote the development of the decision-making process in TCM practices(https://github.com/2020MEAI/TCMLLM).

TIAN Haoyu;YANG Kuo;DONG Xin;ZHAO Chenxi;YE Mingwei;WANG Hongyan;LIU Yiming;HU Minjie;ZHU Qiang;YU Jian;ZHANG Lei;ZHOU Xuezhong

Beijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,ChinaNational Data Center of Traditional Chinese Medicine,China Academy of Chinese Medical Sciences,Beijing 100700,ChinaBeijing Key Lab of Traffic Data Analysis and Mining,School of Computer Science&Technology,Beijing Jiaotong University,Beijing 100044,China

临床医学

Large language modelsInstruction-tuningPrescription recommendationTraditional Chinese medicine(TCM)Assisted decision-making

《Digital Chinese Medicine》 2024 (4)

P.343-355,13

National Key Research and Development Program(2023YFC3502604)National Natural Science Foundation of China(U23B2062 and 82374302).

10.1016/j.dcmed.2025.01.007

评论