Melissa D. McCradden, Alex John London, Judy Wawira Gichoya, Mark Sendak, Lauren Erdman, Ian Stedman, Lauren Oakden-Rayner, Ismail Akrout, James A. Anderson, Lesley-Anne Farmer, Robert Greer, Anna Goldenberg, Yvonne Ho, Shalmali Joshi, Jennie Louise, Muhammad Mamdani, Mjaye L. Mazwi, Abdullahi Mohamud, Lyle J. Palmer, Antonios Peperidis, Stephen R. Pfohl, Mandy Rickard, Carolyn Semmler, Karandeep Singh, Devin Singh, Seyi Soremekun, Lana Tikhomirov, Anton H. van der Vegt, Karin Verspoor, Xiaoxuan Liu
{"title":"CANAIRI: the Collaboration for Translational Artificial Intelligence Trials in healthcare","authors":"Melissa D. McCradden, Alex John London, Judy Wawira Gichoya, Mark Sendak, Lauren Erdman, Ian Stedman, Lauren Oakden-Rayner, Ismail Akrout, James A. Anderson, Lesley-Anne Farmer, Robert Greer, Anna Goldenberg, Yvonne Ho, Shalmali Joshi, Jennie Louise, Muhammad Mamdani, Mjaye L. Mazwi, Abdullahi Mohamud, Lyle J. Palmer, Antonios Peperidis, Stephen R. Pfohl, Mandy Rickard, Carolyn Semmler, Karandeep Singh, Devin Singh, Seyi Soremekun, Lana Tikhomirov, Anton H. van der Vegt, Karin Verspoor, Xiaoxuan Liu","doi":"10.1038/s41591-024-03364-1","DOIUrl":null,"url":null,"abstract":"<p>Over the past few years, authoritative, trustworthy guidance for the clinical translation of artificial intelligence (AI) has formed mainly around two areas: responsible model development and validation; and prospective clinical trials. Initial work focused on building a good model, which generally means a model that demonstrates good performance, addresses an important clinical task, trains on the right data to address that task and could be used for some meaningful goal<sup>1</sup>. The model should then be assessed against sets of unseen cases to ensure it can generalize beyond the test set and can potentially be tested externally. Further practices are emerging around ongoing monitoring to detect model drift and performance changes. Collectively, these practices are characterized as responsible machine learning.</p><p>However, the distinctions between the in silico context and the clinical environment are substantial, which highlights the need for clinical evaluations. In 2020, the SPIRIT-AI and CONSORT-AI reporting guidelines were published to establish the minimum reporting criteria for the conduct of prospective, interventional clinical trials evaluating the impact of an AI model<sup>2,3</sup>. The DECIDE-AI guidelines were published shortly thereafter to address first-in-human feasibility trials of AI tools<sup>4</sup>. Regulatory frameworks emphasize the importance of clinical evidence, but precisely what kind and degree of evidence is needed for the approval of clinical AI applications is a matter of ongoing uncertainty<sup>5</sup>.</p>","PeriodicalId":19037,"journal":{"name":"Nature Medicine","volume":"25 1","pages":""},"PeriodicalIF":58.7000,"publicationDate":"2025-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Nature Medicine","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1038/s41591-024-03364-1","RegionNum":1,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"BIOCHEMISTRY & MOLECULAR BIOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Over the past few years, authoritative, trustworthy guidance for the clinical translation of artificial intelligence (AI) has formed mainly around two areas: responsible model development and validation; and prospective clinical trials. Initial work focused on building a good model, which generally means a model that demonstrates good performance, addresses an important clinical task, trains on the right data to address that task and could be used for some meaningful goal1. The model should then be assessed against sets of unseen cases to ensure it can generalize beyond the test set and can potentially be tested externally. Further practices are emerging around ongoing monitoring to detect model drift and performance changes. Collectively, these practices are characterized as responsible machine learning.
However, the distinctions between the in silico context and the clinical environment are substantial, which highlights the need for clinical evaluations. In 2020, the SPIRIT-AI and CONSORT-AI reporting guidelines were published to establish the minimum reporting criteria for the conduct of prospective, interventional clinical trials evaluating the impact of an AI model2,3. The DECIDE-AI guidelines were published shortly thereafter to address first-in-human feasibility trials of AI tools4. Regulatory frameworks emphasize the importance of clinical evidence, but precisely what kind and degree of evidence is needed for the approval of clinical AI applications is a matter of ongoing uncertainty5.
期刊介绍:
Nature Medicine is a monthly journal publishing original peer-reviewed research in all areas of medicine. The publication focuses on originality, timeliness, interdisciplinary interest, and the impact on improving human health. In addition to research articles, Nature Medicine also publishes commissioned content such as News, Reviews, and Perspectives. This content aims to provide context for the latest advances in translational and clinical research, reaching a wide audience of M.D. and Ph.D. readers. All editorial decisions for the journal are made by a team of full-time professional editors.
Nature Medicine consider all types of clinical research, including:
-Case-reports and small case series
-Clinical trials, whether phase 1, 2, 3 or 4
-Observational studies
-Meta-analyses
-Biomarker studies
-Public and global health studies
Nature Medicine is also committed to facilitating communication between translational and clinical researchers. As such, we consider “hybrid” studies with preclinical and translational findings reported alongside data from clinical studies.