Lack of methodological rigor and limited coverage of generative artificial intelligence in existing artificial intelligence reporting guidelines: a scoping review.

  • Abstract
  • References
  • Similar Papers
Abstract
Translate article icon Translate Article Star icon
Take notes icon Take Notes

Lack of methodological rigor and limited coverage of generative artificial intelligence in existing artificial intelligence reporting guidelines: a scoping review.

ReferencesShowing 10 of 33 papers
  • Open Access Icon
  • Cite Count Icon 67
  • 10.2967/jnumed.121.263239
Nuclear Medicine and Artificial Intelligence: Best Practices for Evaluation (the RELAINCE Guidelines).
  • May 26, 2022
  • Journal of nuclear medicine : official publication, Society of Nuclear Medicine
  • Abhinav K Jha + 14 more

  • Open Access Icon
  • Cite Count Icon 1839
  • 10.1136/bmjopen-2016-012799
STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration
  • Nov 1, 2016
  • BMJ open
  • Jérémie F Cohen + 10 more

  • Cite Count Icon 10
  • 10.1093/ehjdh/ztae080
Machine learning based prediction models for cardiovascular disease risk using electronic health records data: systematic review and meta-analysis.
  • Oct 27, 2024
  • European heart journal. Digital health
  • Tianyi Liu + 3 more

  • Cite Count Icon 114
  • 10.1016/j.jpurol.2023.05.018
ChatGPT and large language model (LLM) chatbots: The current state of acceptability and a proposal for guidelines on utilization in academic medicine
  • Jun 2, 2023
  • Journal of Pediatric Urology
  • Jin K Kim + 3 more

  • Open Access Icon
  • PDF Download Icon
  • Cite Count Icon 7
  • 10.1002/imo2.7
STAGER checklist: Standardized testing and assessment guidelines for evaluating generative artificial intelligence reliability
  • Jul 2, 2024
  • iMetaOmics
  • Jinghong Chen + 34 more

  • Cite Count Icon 14
  • 10.1148/radiol.239024
Guidelines for Use of Large Language Models by Authors, Reviewers, and Editors: Considerations for Imaging Journals.
  • Oct 1, 2023
  • Radiology
  • Linda Moy

  • Open Access Icon
  • Cite Count Icon 864
  • 10.1148/ryai.2020200029
Checklist for Artificial Intelligence in Medical Imaging (CLAIM): A Guide for Authors and Reviewers.
  • Mar 1, 2020
  • Radiology: Artificial Intelligence
  • John Mongan + 2 more

  • Cite Count Icon 15
  • 10.3348/kjr.2024.0843
Minimum Reporting Items for Clear Evaluation of Accuracy Reports of Large Language Models in Healthcare (MI-CLEAR-LLM)
  • Sep 12, 2024
  • Korean Journal of Radiology
  • Seong Ho Park + 4 more

  • Open Access Icon
  • Cite Count Icon 207
  • 10.1038/s41591-020-0941-1
Developing specific reporting guidelines for diagnostic accuracy studies assessing AI interventions: The STARD-AI Steering Group.
  • Jun 1, 2020
  • Nature Medicine
  • Viknesh Sounderajah + 17 more

  • Open Access Icon
  • Cite Count Icon 156
  • 10.1038/s41591-021-01517-0
A quality assessment tool for artificial intelligence-centered diagnostic test accuracy studies: QUADAS-AI.
  • Oct 1, 2021
  • Nature Medicine
  • Viknesh Sounderajah + 39 more

Similar Papers
  • Research Article
  • Cite Count Icon 8
  • 10.1287/ijds.2023.0007
How Can IJDS Authors, Reviewers, and Editors Use (and Misuse) Generative AI?
  • Apr 1, 2023
  • INFORMS Journal on Data Science
  • Galit Shmueli + 7 more

How Can <i>IJDS</i> Authors, Reviewers, and Editors Use (and Misuse) Generative AI?

  • Research Article
  • 10.2196/64640
Generative Artificial Intelligence Tools in Medical Research (GAMER): Protocol for a Scoping Review and Development of Reporting Guidelines.
  • Aug 14, 2025
  • JMIR research protocols
  • Xufei Luo + 6 more

The integration of artificial intelligence (AI) has revolutionized medical research, offering innovative solutions for data collection, patient engagement, and information dissemination. Powerful generative AI (GenAI) tools and other similar chatbots have emerged, facilitating user interactions with virtual conversational agents. However, the increasing use of GenAI tools in medical research presents challenges, including ethical concerns, data privacy issues, and the potential for generating false content. These issues necessitate standardization of reporting to ensure transparency and scientific rigor. The development of the Generative Artificial Intelligence Tools in Medical Research (GAMER) reporting guidelines aims to establish comprehensive, standardized guidelines for reporting the use of GenAI tools in medical research. The GAMER guidelines are being developed following the methodology recommended by the Enhancing the Quality and Transparency of Health Research (EQUATOR) Network, involving a scoping review and expert Delphi consensus. The scoping review searched PubMed, Web of Science, Embase, CINAHL, PsycINFO, and Google Scholar (for the first 200 results) using keywords like "generative AI" and "medical research" to identify reporting elements in GenAI-related studies. The Delphi process involves 30-50 experts with ≥3 years of experience in AI applications or medical research, selected based on publication records and expertise across disciplines (eg, clinicians and data scientists) and regions (eg, Asia and Europe). A 7-point-scale survey will establish consensus on checklist items. The testing phase invites authors to apply the GAMER checklist to GenAI-related manuscripts and provide feedback via a questionnaire, while experts assess reliability (κ statistic) and usability (time taken, 7-point Likert scale). The study has been approved by the Ethics Committee of the Institute of Health Data Science at Lanzhou University (HDS-202406-01). The GAMER project was launched in July 2023 by the Evidence-Based Medicine Center of Lanzhou University and the WHO Collaborating Centre for Guideline Implementation and Knowledge Translation, and it concluded in July 2024. The scoping review was completed in November 2023. The Delphi process was conducted from October 2023 to April 2024. The testing phase began in March 2025 and is ongoing. The expected outcome of the GAMER project is a reporting checklist accompanied by relevant terminology, examples, and explanations to guide stakeholders in better reporting the use of GenAI tools. GAMER aims to guide researchers, reviewers, and editors in the transparent and scientific application of GenAI tools in medical research. By providing a standardized reporting checklist, GAMER seeks to enhance the clarity, completeness, and integrity of research involving GenAI tools, thereby promoting collaboration, comparability, and cumulative knowledge generation in AI-driven health care technologies. DERR1-10.2196/64640.

  • Front Matter
  • Cite Count Icon 92
  • 10.1136/bmj.a718
Enhancing the quality and transparency of health research
  • Jul 8, 2008
  • BMJ
  • T Groves

Enhancing the quality and transparency of health research

  • Discussion
  • Cite Count Icon 6
  • 10.1016/j.ebiom.2023.104672
Response to M. Trengove & coll regarding "Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine".
  • Jul 1, 2023
  • eBioMedicine
  • Stefan Harrer

Response to M. Trengove & coll regarding "Attention is not all you need: the complicated case of ethically using large language models in healthcare and medicine".

  • Research Article
  • Cite Count Icon 28
  • 10.5204/mcj.3004
ChatGPT Isn't Magic
  • Oct 2, 2023
  • M/C Journal
  • Tama Leaver + 1 more

during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (see The Effect of Open Access).

  • Research Article
  • Cite Count Icon 16
  • 10.1162/daed_e_01897
Getting AI Right: Introductory Notes on AI &amp; Society
  • May 1, 2022
  • Daedalus
  • James Manyika

Getting AI Right: Introductory Notes on AI &amp; Society

  • Research Article
  • 10.1016/j.arth.2025.05.093
Reporting Guidelines for Artificial Intelligence Use in Orthopaedic Surgery Research.
  • Jun 1, 2025
  • The Journal of arthroplasty
  • Cody C Wyles + 4 more

Reporting Guidelines for Artificial Intelligence Use in Orthopaedic Surgery Research.

  • Discussion
  • Cite Count Icon 15
  • 10.1016/j.jclinepi.2013.01.001
Introducing a new series on effective writing and publishing of scientific papers
  • Feb 26, 2013
  • Journal of Clinical Epidemiology
  • Daniel Kotz + 3 more

Introducing a new series on effective writing and publishing of scientific papers

  • Research Article
  • Cite Count Icon 6
  • 10.2196/47105
Guidelines and Standard Frameworks for AI in Medicine: Protocol for a Systematic Literature Review.
  • Oct 25, 2023
  • JMIR research protocols
  • Kirubel Biruk Shiferaw + 3 more

Applications of artificial intelligence (AI) are pervasive in modern biomedical science. In fact, research results suggesting algorithms and AI models for different target diseases and conditions are continuously increasing. While this situation undoubtedly improves the outcome of AI models, health care providers are increasingly unsure which AI model to use due to multiple alternatives for a specific target and the "black box" nature of AI. Moreover, the fact that studies rarely use guidelines in developing and reporting AI models poses additional challenges in trusting and adapting models for practical implementation. This review protocol describes the planned steps and methods for a review of the synthesized evidence regarding the quality of available guidelines and frameworks to facilitate AI applications in medicine. We will commence a systematic literature search using medical subject headings terms for medicine, guidelines, and machine learning (ML). All available guidelines, standard frameworks, best practices, checklists, and recommendations will be included, irrespective of the study design. The search will be conducted on web-based repositories such as PubMed, Web of Science, and the EQUATOR (Enhancing the Quality and Transparency of Health Research) network. After removing duplicate results, a preliminary scan for titles will be done by 2 reviewers. After the first scan, the reviewers will rescan the selected literature for abstract review, and any incongruities about whether to include the article for full-text review or not will be resolved by the third and fourth reviewer based on the predefined criteria. A Google Scholar (Google LLC) search will also be performed to identify gray literature. The quality of identified guidelines will be evaluated using the Appraisal of Guidelines, Research, and Evaluation (AGREE II) tool. A descriptive summary and narrative synthesis will be carried out, and the details of critical appraisal and subgroup synthesis findings will be presented. The results will be reported using the PRISMA (Preferred Reporting Items for Systematic Review and Meta-Analyses) reporting guidelines. Data analysis is currently underway, and we anticipate finalizing the review by November 2023. Guidelines and recommended frameworks for developing, reporting, and implementing AI studies have been developed by different experts to facilitate the reliable assessment of validity and consistent interpretation of ML models for medical applications. We postulate that a guideline supports the assessment of an ML model only if the quality and reliability of the guideline are high. Assessing the quality and aspects of available guidelines, recommendations, checklists, and frameworks-as will be done in the proposed review-will provide comprehensive insights into current gaps and help to formulate future research directions. DERR1-10.2196/47105.

  • Research Article
  • Cite Count Icon 1
  • 10.3760/cma.j.issn.0254-6450.2019.01.020
Statistical reporting requirements for medical journals: Amplifications and explanations
  • Jan 10, 2019
  • Zhonghua liu xing bing xue za zhi = Zhonghua liuxingbingxue zazhi
  • Ruohua Yan + 1 more

Our study aimed to amplify and explain the items of statistical reporting requirements proposed by medical journals, and to improve the statistical reporting quality of medical articles. Statistical reporting requirements were obtained from the reporting standards published by the International Committee of Medical Journal Editors (ICMJE), the Enhancing the QUAlity and Transparency of Health Research (EQUATOR) network, and the editorial board of Chinese Medical Journal, etc. The items involved in statistical reporting requirements were summarized as issues of study design, statistical analysis, and interpretation of results. Each item was amplified based on cases of original articles. It is noticeable that the statistical reporting requirements of English medical journals generally referring to guidance documents, including "Recommendations for the conduct, reporting, editing, and publication of scholarly work in medical journals" proposed by the ICMJE, or the statements for different study types published by the EQUATOR network, where the statistical reporting of medical articles had been detailed specified. The statistical reporting requirements of Chinese medical journals, however, were usually stated by the editorial boards. Although the formats and contents of statistical analysis had been regulated, the requirements of Chinese medical journals were to some extent insufficient and should be enhanced in accordance with the international standards. In conclusion, the amplification and explanation of statistical reporting requirements were expected to help investigators understand the requirements for statistical reporting in medical researches, so as to effectively improve the quality of medical articles.

  • Abstract
  • 10.1017/cts.2024.1147
577 EQUATOR network implementation science reporting and adherence challenges for learning health systems
  • Apr 1, 2025
  • Journal of Clinical and Translational Science
  • Christopher Carpenter + 7 more

Objectives/Goals: The Standards for Reporting Implementation Studies (StaRI) are the Enhancing the Quality and Transparency of Health Research (EQUATOR) Network 27-item checklist for Implementation Science. This study quantifies StaRI adherence among self-defined Implementation Science studies in published Learning Health Systems (LHS) research. Methods/Study Population: A medical librarian-designed a search strategy identified original Implementation Science research published in one of the top 20 Implementation Science journals between 2017 and 2021. Inclusion criteria included studies or protocols describing the implementation of any intervention in healthcare settings. Exclusion criteria included concept papers, non-implementation research, or editorials. Full-text documents were reviewed by two investigators to abstract and judge StaRI implementation and intervention adherence, partial adherence, or non-adherence. Results/Anticipated Results: A total of 330 documents were screened, 97 met inclusion criteria, and 47 were abstracted including 30 research studies and 17 protocols. Adherence to individual StaRI reporting items ranged from 13% to 100%. Most StaRI items were reported in >60% of manuscripts and protocols. The lowest adherence in research studies was noted around economic evaluation reporting for implementation (16%) or intervention (13%) strategies, harms (13%), contextual changes (30%), or fidelity of either the intervention (34%) or implementation (53%) approach. Subgroup analyses were infrequently contemplated or reported (43%). In protocols, the implications of the implementation strategy (41%) or intervention approach (47%) were not commonly reported. Discussion/Significance of Impact: When leveraging implementation science to report reproducible and sustainable practice change initiatives, LHS researchers will need to include assessments of economics, harms, context, and fidelity in order to attain higher levels of adherence to EQUATOR’s StaRI checklist.

  • Research Article
  • 10.1200/op.2023.19.11_suppl.182
Inclusion of sex and gender instructions to authors and SAGER guidelines recommendations across highest impact oncology journals.
  • Nov 1, 2023
  • JCO Oncology Practice
  • Maria Teresa Bourlon + 4 more

182 Background: The notion of sex and gender is constantly evolving through many disciplines. In medical oncology, disease behavior and treatment modalities have been shown to impact men and women differently. Sex and gender-tailored clinical research in oncology could promote a better understanding of disease progression and interpretation of clinical trials. The Guidelines for Sex and Gender Equity in Research (SAGER) promote the specification of sex and gender in any given manuscript. The Enhancing the Quality and Transparency of Health Research (EQUATOR) Network promotes the wider use of 575 guidelines, which SAGER is a part of. We sought to determine the frequency in which SAGER Guidelines and the EQUATOR network are referenced in the instructions to authors of the top 100 medical oncology journals with the highest impact factor. Methods: We identified the top 100 medical oncology journals with the highest impact factor. For each, we revised their instructions to authors' material and recorded the frequency with which they referenced SAGER guidelines, the EQUATOR network, or sex and gender recommendations. Results: Median value of the (N=100) journal´s impact factor was 7.47 (range 4.8 – 286.1). Overall, 28 journals mentioned SAGER guidelines, 31 mentioned the EQUATOR network, and 37 mentioned sex and gender. Of the 28 journals that mention SAGER guidelines, 27 also mention sex and gender in their instructions to authors. A total of 12 journals mentioned both SAGER and the EQUATOR network in their instructions to authors. We dichotomized the IF value into “journals with a high impact factor” (≥10 IF) (n=31) and “journals with a low impact factor” (&lt;10) (n=69). SAGER was mentioned in 7 (22.5%) high-impact factor journals and 21 (30.4%) low-impact factor journals. The EQUATOR network was mentioned in 8 (25.8%) high-impact factor journals and 23 (33.3%) low-impact factor journals. Sex and gender was mentioned in 12 (38.7%) high-impact factor journals and 25 (36.2%) low-impact factor journals. Conclusions: Most oncology journals still warrant consideration for appropriate sex and gender recommendations. Guidelines like SAGER and EQUATOR are still commonly overlooked in instructions to authors. Consideration for this is paramount in oncology studies, as sex and gender-tailored research could impact disease biology and treatment understanding in oncology.

  • Research Article
  • Cite Count Icon 23
  • 10.1136/bmjopen-2018-024942
Reporting quality of the Delphi technique in reporting guidelines: a protocol for a systematic analysis of the EQUATOR Network Library
  • Apr 1, 2019
  • BMJ Open
  • Masahiro Banno + 2 more

IntroductionReporting guidelines are important tools for improving the quality of medical research. The Enhancing the QUAlity and Transparency Of health Research (EQUATOR) Network’s Library contains a comprehensive and up-to-date database...

  • Supplementary Content
  • Cite Count Icon 1
  • 10.1136/bmjqs-2024-017491
The problem with the existing reporting standards for adverse event and medical error research
  • Feb 11, 2025
  • BMJ Quality & Safety
  • Christopher R Carpenter + 6 more

The Enhancing the Quality and Transparency of Health Research (EQUATOR) Network indexes over 600 reporting guidelines designed to improve the reproducibility of manuscripts across medical fields and study designs. Although...

  • Research Article
  • Cite Count Icon 3
  • 10.1136/bmjopen-2021-059715
Protocol for the development of a reporting guideline for causal and counterfactual prediction models in biomedicine
  • Jun 1, 2022
  • BMJ Open
  • Jie Xu + 6 more

IntroductionWhile there are guidelines for reporting on observational studies (eg, Strengthening the Reporting of Observational Studies in Epidemiology, Reporting of Studies Conducted Using Observational Routinely Collected Health Data Statement), estimation...

More from: Journal of clinical epidemiology
  • New
  • Front Matter
  • 10.1016/j.jclinepi.2025.112044
David Sackett Young Investigator Award, Peer Reviewer of the Year Award, and Peer Reviewer Acknowledgment.
  • Nov 6, 2025
  • Journal of clinical epidemiology

  • Research Article
  • 10.1016/j.jclinepi.2025.111928
Scoping review authors view knowledge user consultations as beneficial but not without challenges: a qualitative study.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Elaine Toomey + 8 more

  • Discussion
  • 10.1016/j.jclinepi.2025.111956
Letter to the editor: The necessity of specifying measurement models: a critical reappraisal specification issues in PRECIOUS.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Xin Meng + 4 more

  • Research Article
  • 10.1016/j.jclinepi.2025.111932
Prevalence and predictors of potentially inappropriate prescribing using codified STOPP-START and Beers criteria: a retrospective cohort study in Ontario's older population.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Lise M Bjerre + 12 more

  • Discussion
  • 10.1016/j.jclinepi.2025.112041
Response to: "Identifying variables that independently predict…" is not a well-defined research task.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Brett P Dyer

  • Addendum
  • 10.1016/j.jclinepi.2025.111969
Corrigendum to GRADE guidance 39: using GRADE-ADOLOPMENT to adopt, adapt or create contextualized recommendations from source guidelines and evidence syntheses [Journal of Clinical Epidemiology 81 (2024) 111494
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Miloslav Klugar + 50 more

  • Research Article
  • 10.1016/j.jclinepi.2025.112040
Methodological review reveals essential gaps and inconsistencies in clinical claims, effects and outcomes in HTA reviews of diagnostic tests.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • Jacqueline Dinnes + 7 more

  • Research Article
  • 10.1016/j.jclinepi.2025.112043
Commentary: "Identifying variables that independently predict…" is not a well-defined research task.
  • Nov 1, 2025
  • Journal of clinical epidemiology
  • John B Carlin

  • Research Article
  • 10.1016/j.jclinepi.2025.112039
Preference-based controlled design: toward increased patients' engagement, efficiency and external validity of cardiovascular clinical trials.
  • Oct 31, 2025
  • Journal of clinical epidemiology
  • Bjorn Redfors + 3 more

  • Research Article
  • 10.1016/j.jclinepi.2025.112038
Most methodological characteristics do not exaggerate effect estimates in nutrition RCTs: findings from a meta-epidemiological study.
  • Oct 31, 2025
  • Journal of clinical epidemiology
  • Gina Bantle + 4 more

Save Icon
Up Arrow
Open/Close
  • Ask R Discovery Star icon
  • Chat PDF Star icon

AI summaries and top papers from 250M+ research sources.

Search IconWhat is the difference between bacteria and viruses?
Open In New Tab Icon
Search IconWhat is the function of the immune system?
Open In New Tab Icon
Search IconCan diabetes be passed down from one generation to the next?
Open In New Tab Icon