Lack of methodological rigor and limited coverage of generative artificial intelligence in existing artificial intelligence reporting guidelines: a scoping review

Xufei Luo, Bingyi Wang, Qianling Shi, Zijun Wang, Honghao Lai, Hui Liu, Yishan Qin, Fengxian Chen, Xuping Song, Long Ge, Lu Zhang, Zhaoxiang Bian, Yaolong Chen*, ADVANCED Working Group, Hongfeng He, Ye Wang, Haodong Li, Huayu Zhang, Di Zhu, Yuanyuan YaoDongrui Peng, Zhewei Li, Jie Zhang, Yishan Qin, Fan Wang, Zhenyu Tang, Yueyan Li, Hanxiang Liu, Jungang Zhao

*Corresponding author for this work

Research output: Contribution to journalJournal articlepeer-review

1 Citation (Scopus)

Abstract

Objectives: This study aimed to systematically map the development methods, scope, and limitations of existing artificial intelligence (AI) reporting guidelines in medicine and to explore their applicability to generative AI (GAI) tools, such as large language models (LLMs). Study Design and Setting: We reported a scoping review adhering to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews. Five information sources were searched, including MEDLINE (via PubMed), Enhancing the QUAlity and Transparency Of health Research (EQUATOR) Network, China National Knowledge Infrastructure, FAIRsharing, and Google Scholar, from inception to December 31, 2024. Two reviewers independently screened records and extracted data using a predefined Excel template. Data included guideline characteristics (eg, development methods, target audience, AI domain), adherence to EQUATOR Network recommendations, and consensus methodologies. Discrepancies were resolved by a third reviewer. Results: Sixty-eight AI reporting guidelines were included; 48.5% focused on general AI, whereas only 7.4% addressed GAI/LLMs. Methodological rigor was limited; 39.7% described development processes, 42.6% involved multidisciplinary experts, and 33.8% followed EQUATOR recommendations. Significant overlap existed, particularly in medical imaging (20.6% of guidelines). GAI-specific guidelines (14.7%) lacked comprehensive coverage and methodological transparency. Conclusion: Existing AI reporting guidelines in medicine have suboptimal methodological rigor, redundancy, and insufficient coverage of GAI applications. Future and updated guidelines should prioritize standardized development processes, multidisciplinary collaboration, and expanded focus on emerging AI technologies like LLMs.

Original languageEnglish
Article number111903
Number of pages10
JournalJournal of Clinical Epidemiology
Volume186
Early online date18 Jul 2025
DOIs
Publication statusPublished - Oct 2025

User-Defined Keywords

  • Artificial intelligence
  • Generative artificial intelligence
  • Large language models
  • Methodological quality
  • Reporting guidelines
  • Scoping review

Fingerprint

Dive into the research topics of 'Lack of methodological rigor and limited coverage of generative artificial intelligence in existing artificial intelligence reporting guidelines: a scoping review'. Together they form a unique fingerprint.

Cite this