TY - JOUR
T1 - Strategies for the Analysis and Elimination of Hallucinations in Artificial Intelligence Generated Medical Knowledge
AU - Chen, Fengxian
AU - Li, Yan
AU - Chen, Yaolong
AU - Bian, Zhaoxiang
AU - Duo, La
AU - Zhou, Qingguo
AU - Zhang, Lu
AU - ADVANCED Working Group
N1 - This study was jointly supported by the National Natural Science Foundation Regional Fund (No. 62266037).
Publisher Copyright:
© 2025 Chinese Cochrane Center, West China Hospital of Sichuan University and John Wiley & Sons Australia, Ltd.
PY - 2025/9/22
Y1 - 2025/9/22
N2 - The application of artificial intelligence (AI) in healthcare has become increasingly widespread, showing significant potential in assisting with diagnosis and treatment. However, generative AI (GAI) models often produce “hallucinations”—plausible but factually incorrect or unsubstantiated outputs—that threaten clinical decision-making and patient safety. This article systematically analyzes the causes of hallucinations across data, training, and inference dimensions and proposes multi-dimensional strategies to mitigate them. Our findings reveal three critical conclusions: The technical optimization through knowledge graphs and multi-stage training significantly reduces hallucinations, while clinical integration through expert feedback loops and multidisciplinary workflows enhances output reliability. Additionally, implementing robust evaluation systems that combine adversarial testing and real-world validation substantially improves factual accuracy in clinical settings. These integrated strategies underscore the importance of harmonizing technical advancements with clinical governance to develop trustworthy, patient-centric AI systems.
AB - The application of artificial intelligence (AI) in healthcare has become increasingly widespread, showing significant potential in assisting with diagnosis and treatment. However, generative AI (GAI) models often produce “hallucinations”—plausible but factually incorrect or unsubstantiated outputs—that threaten clinical decision-making and patient safety. This article systematically analyzes the causes of hallucinations across data, training, and inference dimensions and proposes multi-dimensional strategies to mitigate them. Our findings reveal three critical conclusions: The technical optimization through knowledge graphs and multi-stage training significantly reduces hallucinations, while clinical integration through expert feedback loops and multidisciplinary workflows enhances output reliability. Additionally, implementing robust evaluation systems that combine adversarial testing and real-world validation substantially improves factual accuracy in clinical settings. These integrated strategies underscore the importance of harmonizing technical advancements with clinical governance to develop trustworthy, patient-centric AI systems.
KW - assisted diagnosis and treatment
KW - evaluation system
KW - generative artificial intelligence
KW - multi-stage training
UR - http://www.scopus.com/inward/record.url?scp=105018017251&partnerID=8YFLogxK
U2 - 10.1111/jebm.70075
DO - 10.1111/jebm.70075
M3 - Journal article
C2 - 40983876
AN - SCOPUS:105018017251
SN - 1756-5383
VL - 18
JO - Journal of Evidence-Based Medicine
JF - Journal of Evidence-Based Medicine
IS - 3
M1 - e70075
ER -