{"title":"Development and Validation of an AI-Based Multimodal Model for Pathological Staging of Gastric Cancer Using CT and Endoscopic Images.","authors":"Chao Zhang, Siyuan Li, Daolai Huang, Bo Wen, Shizhuang Wei, Yaodong Song, Xianghua Wu","doi":"10.1016/j.acra.2024.12.029","DOIUrl":null,"url":null,"abstract":"<p><strong>Rationale and objectives: </strong>Accurate preoperative pathological staging of gastric cancer is crucial for optimal treatment selection and improved patient outcomes. Traditional imaging methods such as CT and endoscopy have limitations in staging accuracy.</p><p><strong>Methods: </strong>This retrospective study included 691 gastric cancer patients treated from March 2017 to March 2024. Enhanced venous-phase CT and endoscopic images, along with postoperative pathological results, were collected. We developed three modeling approaches: (1) nine deep learning models applied to CT images (DeepCT), (2) 11 machine learning algorithms using handcrafted radiomic features from CT images (HandcraftedCT), and (3) ResNet-50-extracted deep features from endoscopic images followed by 11 machine learning algorithms (DeepEndo). The two top-performing models from each approach were combined into the Integrated Multi-Modal Model using a stacking ensemble method. Performance was assessed using ROC-AUC, sensitivity, and specificity.</p><p><strong>Results: </strong>The Integrated Multi-Modal Model achieved an ROC-AUC of 0.933 (95% CI, 0.887-0.979) on the test set, outperforming individual models. Sensitivity and specificity were 0.869 and 0.840, respectively. Various evaluation metrics demonstrated that the final fusion model effectively integrated the strengths of each sub-model, resulting in a balanced and robust performance with reduced false-positive and false-negative rates.</p><p><strong>Conclusion: </strong>The Integrated Multi-Modal Model effectively integrates radiomic and deep learning features from CT and endoscopic images, demonstrating superior performance in preoperative pathological staging of gastric cancer. This multimodal approach enhances predictive accuracy and provides a reliable tool for clinicians to develop individualized treatment plans, thereby improving patient outcomes.</p><p><strong>Data availability: </strong>The data presented in this study are available on request from the corresponding author. The data are not publicly available due to ethical reasons. All code used in this study is based on third-party libraries and all custom code developed for this study is available upon reasonable request from the corresponding author.</p>","PeriodicalId":50928,"journal":{"name":"Academic Radiology","volume":" ","pages":""},"PeriodicalIF":3.8000,"publicationDate":"2025-01-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Academic Radiology","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1016/j.acra.2024.12.029","RegionNum":2,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"RADIOLOGY, NUCLEAR MEDICINE & MEDICAL IMAGING","Score":null,"Total":0}
引用次数: 0
Abstract
Rationale and objectives: Accurate preoperative pathological staging of gastric cancer is crucial for optimal treatment selection and improved patient outcomes. Traditional imaging methods such as CT and endoscopy have limitations in staging accuracy.
Methods: This retrospective study included 691 gastric cancer patients treated from March 2017 to March 2024. Enhanced venous-phase CT and endoscopic images, along with postoperative pathological results, were collected. We developed three modeling approaches: (1) nine deep learning models applied to CT images (DeepCT), (2) 11 machine learning algorithms using handcrafted radiomic features from CT images (HandcraftedCT), and (3) ResNet-50-extracted deep features from endoscopic images followed by 11 machine learning algorithms (DeepEndo). The two top-performing models from each approach were combined into the Integrated Multi-Modal Model using a stacking ensemble method. Performance was assessed using ROC-AUC, sensitivity, and specificity.
Results: The Integrated Multi-Modal Model achieved an ROC-AUC of 0.933 (95% CI, 0.887-0.979) on the test set, outperforming individual models. Sensitivity and specificity were 0.869 and 0.840, respectively. Various evaluation metrics demonstrated that the final fusion model effectively integrated the strengths of each sub-model, resulting in a balanced and robust performance with reduced false-positive and false-negative rates.
Conclusion: The Integrated Multi-Modal Model effectively integrates radiomic and deep learning features from CT and endoscopic images, demonstrating superior performance in preoperative pathological staging of gastric cancer. This multimodal approach enhances predictive accuracy and provides a reliable tool for clinicians to develop individualized treatment plans, thereby improving patient outcomes.
Data availability: The data presented in this study are available on request from the corresponding author. The data are not publicly available due to ethical reasons. All code used in this study is based on third-party libraries and all custom code developed for this study is available upon reasonable request from the corresponding author.
期刊介绍:
Academic Radiology publishes original reports of clinical and laboratory investigations in diagnostic imaging, the diagnostic use of radioactive isotopes, computed tomography, positron emission tomography, magnetic resonance imaging, ultrasound, digital subtraction angiography, image-guided interventions and related techniques. It also includes brief technical reports describing original observations, techniques, and instrumental developments; state-of-the-art reports on clinical issues, new technology and other topics of current medical importance; meta-analyses; scientific studies and opinions on radiologic education; and letters to the Editor.