首页 > 最新文献

Journal of imaging informatics in medicine最新文献

英文 中文
From Liver to Brain: A 2.5D Deep Learning Model for Predicting Hepatic Encephalopathy Using Opportunistic Non-contrast CT in Hepatitis B Related Acute-on-Chronic Liver Failure Patients. 从肝到脑:利用机会性非对比CT预测乙型肝炎相关急慢性肝衰竭患者肝性脑病的2.5D深度学习模型
Pub Date : 2026-01-05 DOI: 10.1007/s10278-025-01802-1
Zonglin Liu, Xueyun Zhang, Ying Chen, Qi Zhang, Zhenxuan Ma, Yue Wu, Yuxin Huang, Yajie Li, Xi Zhao, Wenchao Gu, Jiaxing Wu, Ying Tao, Yuxin Shi, Zhenwei Yao, Yan Ren, Yuxian Huang, Shiman Wu

This study aims to develop a 2.5D deep learning framework leveraging non-contrast CT scans for early prediction of hepatic encephalopathy (HE) in hepatitis B-related acute-on-chronic liver failure (ACLF) patients. This retrospective study enrolled 228 ACLF patients meeting APASL criteria from two centers. Participants were divided into training (n = 102), internal validation (n = 44), and external testing (n = 82) cohorts. Non-contrast CT scans (5 mm slices) from six scanner models were preprocessed to 1 × 1 × 1 mm³ isotropic resolution with windowing (30-110 HU). Liver ROIs were manually segmented by two radiologists. The image center on the maximal cross-sectional slice and its adjacent slices (±1/2/4) were extracted to form 2.5D inputs. Deep learning models (DenseNet121, DenseNet201, ResNet50, InceptionV3) were employed for feature extraction. Multi-instance learning methods, including probability likelihood histograms and bag-of-words, were used for feature fusion. Machine learning classifiers (Logistic Regression, RandomForest, LightGBM) with 5-fold cross validation were built for HE prediction. DenseNet121 demonstrated the best slice-level prediction performance (validation AUC: 0.698). The LightGBM classifier with MIL fusion achieved AUCs of 0.969 (training), 0.886 (validation), and 0.829 (external testing), outperforming other fusion methods. Grad-CAM visualizations confirmed model attention to peri-portal fibrotic regions, demonstrating anatomical relevance. The MIL-based 2.5D deep learning model effectively predicts HE risk using routine non-contrast CT in ACLF patients, providing a non-invasive method for individualized risk assessment.

本研究旨在开发一个2.5D深度学习框架,利用非对比CT扫描对乙型肝炎相关急性慢性肝衰竭(ACLF)患者的肝性脑病(HE)进行早期预测。本回顾性研究纳入了来自两个中心的228例符合APASL标准的ACLF患者。参与者被分为训练组(n = 102)、内部验证组(n = 44)和外部测试组(n = 82)。对6种扫描仪型号的非对比CT扫描(5mm切片)进行预处理,达到1 × 1 × 1 mm³各向同性分辨率,加窗(30-110 HU)。肝脏roi由两名放射科医生手工分割。提取最大截面切片上的图像中心及其相邻切片(±1/2/4),形成2.5D输入。采用深度学习模型(DenseNet121、DenseNet201、ResNet50、InceptionV3)进行特征提取。采用概率似然直方图和词袋等多实例学习方法进行特征融合。建立了具有5倍交叉验证的机器学习分类器(Logistic Regression, RandomForest, LightGBM)用于HE预测。DenseNet121显示出最佳的切片级预测性能(验证AUC: 0.698)。采用MIL融合的LightGBM分类器的auc分别为0.969(训练)、0.886(验证)和0.829(外部测试),优于其他融合方法。Grad-CAM可视化证实了模型对门脉周围纤维化区域的关注,显示了解剖学上的相关性。基于mil的2.5D深度学习模型可有效预测ACLF患者常规非对比CT的HE风险,为个体化风险评估提供了一种无创方法。
{"title":"From Liver to Brain: A 2.5D Deep Learning Model for Predicting Hepatic Encephalopathy Using Opportunistic Non-contrast CT in Hepatitis B Related Acute-on-Chronic Liver Failure Patients.","authors":"Zonglin Liu, Xueyun Zhang, Ying Chen, Qi Zhang, Zhenxuan Ma, Yue Wu, Yuxin Huang, Yajie Li, Xi Zhao, Wenchao Gu, Jiaxing Wu, Ying Tao, Yuxin Shi, Zhenwei Yao, Yan Ren, Yuxian Huang, Shiman Wu","doi":"10.1007/s10278-025-01802-1","DOIUrl":"https://doi.org/10.1007/s10278-025-01802-1","url":null,"abstract":"<p><p>This study aims to develop a 2.5D deep learning framework leveraging non-contrast CT scans for early prediction of hepatic encephalopathy (HE) in hepatitis B-related acute-on-chronic liver failure (ACLF) patients. This retrospective study enrolled 228 ACLF patients meeting APASL criteria from two centers. Participants were divided into training (n = 102), internal validation (n = 44), and external testing (n = 82) cohorts. Non-contrast CT scans (5 mm slices) from six scanner models were preprocessed to 1 × 1 × 1 mm³ isotropic resolution with windowing (30-110 HU). Liver ROIs were manually segmented by two radiologists. The image center on the maximal cross-sectional slice and its adjacent slices (±1/2/4) were extracted to form 2.5D inputs. Deep learning models (DenseNet121, DenseNet201, ResNet50, InceptionV3) were employed for feature extraction. Multi-instance learning methods, including probability likelihood histograms and bag-of-words, were used for feature fusion. Machine learning classifiers (Logistic Regression, RandomForest, LightGBM) with 5-fold cross validation were built for HE prediction. DenseNet121 demonstrated the best slice-level prediction performance (validation AUC: 0.698). The LightGBM classifier with MIL fusion achieved AUCs of 0.969 (training), 0.886 (validation), and 0.829 (external testing), outperforming other fusion methods. Grad-CAM visualizations confirmed model attention to peri-portal fibrotic regions, demonstrating anatomical relevance. The MIL-based 2.5D deep learning model effectively predicts HE risk using routine non-contrast CT in ACLF patients, providing a non-invasive method for individualized risk assessment.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145907323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
TET Loss: A Temperature-Entropy Calibrated Transfer Loss for Reliable Medical Image Classification. TET损失:用于可靠医学图像分类的温度熵校准转移损失。
Pub Date : 2026-01-05 DOI: 10.1007/s10278-025-01816-9
Weichao Pan

Deep learning models for medical image classification often exhibit overconfident predictions and domain mismatch when transferred from natural image pretraining, which undermines their generalization and clinical reliability. This study proposes TET Loss (Temperature-Entropy calibrated Transfer Loss Function), a plug-and-play objective function that combines temperature scaling to moderate logit sharpness with entropy regularization to promote uncertainty-aware learning. TET Loss is model-agnostic and introduces zero inference-time overhead. Across four public benchmarks (BreastMNIST, DermaMNIST, PneumoniaMNIST, and RetinaMNIST), TET Loss consistently enhances CNNs, transformers, and hybrid backbones under short 10-epoch fine-tuning. For example, EfficientViT-M2 improves its F1 score from 53.9 to 66.7% on BreastMNIST, and BiFormer-Tiny increases its F1 from 73.1 to 86.1% with an AUC gain to 94.1%. On PneumoniaMNIST, RMT-T3 with TET Loss reaches an F1 of 96.4% and an AUC of 99.1%, surpassing several medical-specific architectures trained for 50-150 epochs. Grad-CAM visualizations demonstrate tighter lesion localization and fewer spurious activations, reflecting improved interpretability. By calibrating confidence while preserving discriminative learning, TET Loss provides a lightweight and effective pathway toward more reliable and robust medical imaging systems. Our code will be available at https://github.com/JEFfersusu/TET_loss .

医学图像分类的深度学习模型在从自然图像预训练转移时,往往表现出过度自信的预测和领域不匹配,从而破坏了其泛化和临床可靠性。本研究提出了TET Loss(温度-熵校准传递损失函数),这是一个即插即用的目标函数,将温度缩放与熵正则化相结合,以调节logit锐度,以促进不确定性感知学习。TET Loss是模型无关的,并且引入了零推断时间开销。在四个公共基准测试(BreastMNIST, DermaMNIST, PneumoniaMNIST和RetinaMNIST)中,TET Loss在短10 epoch微调下始终增强cnn,变压器和混合主干。例如,efficientviti - m2在BreastMNIST上的F1得分从53.9提高到66.7%,BiFormer-Tiny的F1得分从73.1提高到86.1%,AUC增加到94.1%。在PneumoniaMNIST上,具有TET Loss的RMT-T3的F1达到96.4%,AUC达到99.1%,超过了一些训练了50-150个epoch的医学特定架构。Grad-CAM可视化显示更紧密的病变定位和更少的虚假激活,反映了更好的可解释性。通过校准信心,同时保留判别学习,TET Loss为更可靠和强大的医学成像系统提供了一个轻量级和有效的途径。我们的代码可以在https://github.com/JEFfersusu/TET_loss上找到。
{"title":"TET Loss: A Temperature-Entropy Calibrated Transfer Loss for Reliable Medical Image Classification.","authors":"Weichao Pan","doi":"10.1007/s10278-025-01816-9","DOIUrl":"https://doi.org/10.1007/s10278-025-01816-9","url":null,"abstract":"<p><p>Deep learning models for medical image classification often exhibit overconfident predictions and domain mismatch when transferred from natural image pretraining, which undermines their generalization and clinical reliability. This study proposes TET Loss (Temperature-Entropy calibrated Transfer Loss Function), a plug-and-play objective function that combines temperature scaling to moderate logit sharpness with entropy regularization to promote uncertainty-aware learning. TET Loss is model-agnostic and introduces zero inference-time overhead. Across four public benchmarks (BreastMNIST, DermaMNIST, PneumoniaMNIST, and RetinaMNIST), TET Loss consistently enhances CNNs, transformers, and hybrid backbones under short 10-epoch fine-tuning. For example, EfficientViT-M2 improves its F1 score from 53.9 to 66.7% on BreastMNIST, and BiFormer-Tiny increases its F1 from 73.1 to 86.1% with an AUC gain to 94.1%. On PneumoniaMNIST, RMT-T3 with TET Loss reaches an F1 of 96.4% and an AUC of 99.1%, surpassing several medical-specific architectures trained for 50-150 epochs. Grad-CAM visualizations demonstrate tighter lesion localization and fewer spurious activations, reflecting improved interpretability. By calibrating confidence while preserving discriminative learning, TET Loss provides a lightweight and effective pathway toward more reliable and robust medical imaging systems. Our code will be available at https://github.com/JEFfersusu/TET_loss .</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145907447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Effectiveness of AI-CAD Software for Breast Cancer Detection in Automated Breast Ultrasound. AI-CAD软件在乳腺自动超声中检测乳腺癌的有效性
Pub Date : 2025-12-22 DOI: 10.1007/s10278-025-01786-y
Sung Ui Shin, Mijung Jang, Bo La Yun, Su Min Cho, Ji Eun Park, Juyeon Lee, Hye Shin Ahn, Bohyoung Kim, Sun Mi Kim

To assess the diagnostic performance and clinical usefulness of deep learning-based computer-aided detection (AI-CAD) for automated breast ultrasound (ABUS) across radiologists with varying ABUS experience. This retrospective study included 114 women (228 breasts) who underwent ABUS in 2019. Three radiologists interpreted images with and without AI-CAD. We evaluated sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), accuracy, and area under the curve (AUC), reading time and interobserver agreement in Breast Imaging Reporting & Data System (BI-RADS) categorization and biopsy recommendations. Among 114 women (50.9 ± 10.8 years), 28 were diagnosed with breast cancer. The following performance metrics improved significant with AI-CAD: Reader 1 (least experienced of ABUS; 2 years of ABUS experience), AUC, 0.837 to 0.947 (p = 0.009), and NPV, 95.8% to 98.4% (p = 0.022); Reader 2 (7 years of experience), PPV, 50.0% to 59.5% (p = 0.042); Reader 3 (8 years of experience), PPV, 55.6% to 66.7% (p = 0.034). Reader 1 with AI-CAD achieved a performance comparable or higher than those of more experienced readers without AI. Specifically: compared with Reader 2, specificity (93.5% vs. 88.0%), PPV (65.8% vs. 50.0%), and accuracy (93.0% vs. 87.7%) were higher. Although Reader 3 originally demonstrated higher NPV (98.4% vs. 95.8%) and AUC (0.954 vs. 0.837) without CAD, these differences were no longer significant when Reader 1 used AI-CAD. Across all readers, AI-CAD reduced the mean reading time by an average of 25 s (p < 0.001). Inter-observer agreement after AI-CAD use (BI-RADS κ: 0.279 → 0.363; biopsy recommendation κ: 0.666 → 0.736) showed no statistically significant difference. AI-CAD enhanced diagnostic performance and reading efficiency in ABUS interpretation, demonstrating the most pronounced improvement for the less experienced reader.

评估基于深度学习的计算机辅助检测(AI-CAD)对具有不同ABUS经验的放射科医生的自动乳房超声(ABUS)的诊断性能和临床实用性。这项回顾性研究包括了2019年接受ABUS手术的114名女性(228个乳房)。三名放射科医生在使用和不使用AI-CAD的情况下解释图像。我们评估了乳房成像报告和数据系统(BI-RADS)分类和活检建议的敏感性、特异性、阳性预测值(PPV)、阴性预测值(NPV)、准确性、曲线下面积(AUC)、阅读时间和观察者间一致性。114例女性(50.9±10.8岁)中,28例确诊为乳腺癌。AI-CAD显著改善了以下性能指标:Reader 1 (ABUS经验最少,2年ABUS经验),AUC从0.837到0.947 (p = 0.009), NPV从95.8%到98.4% (p = 0.022);阅读者2(7年工作经验),PPV 50.0% ~ 59.5% (p = 0.042);读者3(8年经验),PPV, 55.6% ~ 66.7% (p = 0.034)。具有AI- cad的阅读器1的性能与没有AI的更有经验的阅读器相当或更高。具体而言:与Reader 2相比,特异性(93.5% vs. 88.0%)、PPV (65.8% vs. 50.0%)和准确性(93.0% vs. 87.7%)更高。虽然Reader 3最初在没有CAD的情况下表现出更高的NPV(98.4%对95.8%)和AUC(0.954对0.837),但当Reader 1使用AI-CAD时,这些差异不再显著。在所有读者中,AI-CAD平均减少了25秒的阅读时间
{"title":"Effectiveness of AI-CAD Software for Breast Cancer Detection in Automated Breast Ultrasound.","authors":"Sung Ui Shin, Mijung Jang, Bo La Yun, Su Min Cho, Ji Eun Park, Juyeon Lee, Hye Shin Ahn, Bohyoung Kim, Sun Mi Kim","doi":"10.1007/s10278-025-01786-y","DOIUrl":"https://doi.org/10.1007/s10278-025-01786-y","url":null,"abstract":"<p><p>To assess the diagnostic performance and clinical usefulness of deep learning-based computer-aided detection (AI-CAD) for automated breast ultrasound (ABUS) across radiologists with varying ABUS experience. This retrospective study included 114 women (228 breasts) who underwent ABUS in 2019. Three radiologists interpreted images with and without AI-CAD. We evaluated sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), accuracy, and area under the curve (AUC), reading time and interobserver agreement in Breast Imaging Reporting & Data System (BI-RADS) categorization and biopsy recommendations. Among 114 women (50.9 ± 10.8 years), 28 were diagnosed with breast cancer. The following performance metrics improved significant with AI-CAD: Reader 1 (least experienced of ABUS; 2 years of ABUS experience), AUC, 0.837 to 0.947 (p = 0.009), and NPV, 95.8% to 98.4% (p = 0.022); Reader 2 (7 years of experience), PPV, 50.0% to 59.5% (p = 0.042); Reader 3 (8 years of experience), PPV, 55.6% to 66.7% (p = 0.034). Reader 1 with AI-CAD achieved a performance comparable or higher than those of more experienced readers without AI. Specifically: compared with Reader 2, specificity (93.5% vs. 88.0%), PPV (65.8% vs. 50.0%), and accuracy (93.0% vs. 87.7%) were higher. Although Reader 3 originally demonstrated higher NPV (98.4% vs. 95.8%) and AUC (0.954 vs. 0.837) without CAD, these differences were no longer significant when Reader 1 used AI-CAD. Across all readers, AI-CAD reduced the mean reading time by an average of 25 s (p < 0.001). Inter-observer agreement after AI-CAD use (BI-RADS κ: 0.279 → 0.363; biopsy recommendation κ: 0.666 → 0.736) showed no statistically significant difference. AI-CAD enhanced diagnostic performance and reading efficiency in ABUS interpretation, demonstrating the most pronounced improvement for the less experienced reader.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145812588","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Improving Chronological Age Estimation in Children Using the Demirjian Method Enhanced with Transformer and Regression Models. 用变压器模型和回归模型改进Demirjian法对儿童实足年龄的估计。
Pub Date : 2025-12-22 DOI: 10.1007/s10278-025-01769-z
Huseyin Simsek, Abdulsamet Aktas, Hamza Osman Ilhan, Nagihan Kara Simsek, Yasin Yasa, Esra Ozcelik, Ayse Betul Oktay

This study presents a two-phase methodology for estimating chronological age in children using panoramic dental images and deep learning-based feature extraction. The dataset comprised 626 panoramic radiographs from children aged 6.0 to 13.8 years (320 males, 306 females; mean age = 9.88 years). Two expert dentists annotated each radiograph according to the Demirjian stages of seven mandibular teeth. In the first phase, three architectures-ResNet-18, EfficientNetV2-M, and Swin V2 Base-were trained separately for males and females to extract high-dimensional feature representations. Images were preprocessed via intensity quantization, histogram equalization, segmentation, and resizing to standardized 224 × 224 pixel inputs. From the fully connected layer of the Swin V2 Base model, 512 features were extracted for each tooth, and the concatenation of seven teeth yielded a 3584-dimensional feature vector per subject. These feature vectors were then used for regression analysis to predict chronological age on a day-level scale. In the second phase, nine machine learning regression models-LightGBM, RandomForest, ExtraTrees, GradientBoosting, XGBoost, KNN, SVR, MLP, and Gaussian Process Regression-were trained using the extracted features. Pairwise t-test analysis revealed ExtraTrees as the most statistically significant model. For this model, RMSE and MAE were 6.98 and 5.18 months for females, and 6.55 and 5.01 months for males. SHAP-based analysis highlighted the second molar (M2) and first premolar (P1) as the most influential features for females, and the first premolar (P1) and second molar (M2) for males. This automated pipeline enhances age prediction accuracy, reduces observer variability, and provides a reliable tool for clinical and forensic dental age estimation. Future work will explore dataset expansion, multimodal integration, and refined model architectures.

本研究提出了一种使用全景牙齿图像和基于深度学习的特征提取来估计儿童实足年龄的两阶段方法。数据集包括626张全景x线片,来自6.0至13.8岁的儿童(320名男性,306名女性,平均年龄= 9.88岁)。两位专家牙医根据七颗下颌牙齿的Demirjian阶段对每张x光片进行了注释。在第一阶段,分别对男性和女性的resnet -18、EfficientNetV2-M和Swin V2 base三个架构进行训练,以提取高维特征表示。图像通过强度量化、直方图均衡化、分割和调整大小到标准化的224 × 224像素输入进行预处理。在Swin V2 Base模型的全连接层中,每颗牙齿提取512个特征,7颗牙齿的拼接得到每个受试者3584维的特征向量。然后将这些特征向量用于回归分析,以预测日水平上的实足年龄。在第二阶段,使用提取的特征训练了9个机器学习回归模型——lightgbm、RandomForest、ExtraTrees、GradientBoosting、XGBoost、KNN、SVR、MLP和高斯过程回归。两两t检验分析显示extratree是最显著的模型。该模型雌性的RMSE和MAE分别为6.98和5.18个月,雄性的RMSE和MAE分别为6.55和5.01个月。基于shap的分析强调,第二臼齿(M2)和第一前臼齿(P1)是女性最具影响力的特征,第一前臼齿(P1)和第二臼齿(M2)是男性最具影响力的特征。这种自动化的管道提高了年龄预测的准确性,减少了观察者的可变性,并为临床和法医牙齿年龄估计提供了可靠的工具。未来的工作将探索数据集扩展、多模态集成和改进的模型架构。
{"title":"Improving Chronological Age Estimation in Children Using the Demirjian Method Enhanced with Transformer and Regression Models.","authors":"Huseyin Simsek, Abdulsamet Aktas, Hamza Osman Ilhan, Nagihan Kara Simsek, Yasin Yasa, Esra Ozcelik, Ayse Betul Oktay","doi":"10.1007/s10278-025-01769-z","DOIUrl":"https://doi.org/10.1007/s10278-025-01769-z","url":null,"abstract":"<p><p>This study presents a two-phase methodology for estimating chronological age in children using panoramic dental images and deep learning-based feature extraction. The dataset comprised 626 panoramic radiographs from children aged 6.0 to 13.8 years (320 males, 306 females; mean age = 9.88 years). Two expert dentists annotated each radiograph according to the Demirjian stages of seven mandibular teeth. In the first phase, three architectures-ResNet-18, EfficientNetV2-M, and Swin V2 Base-were trained separately for males and females to extract high-dimensional feature representations. Images were preprocessed via intensity quantization, histogram equalization, segmentation, and resizing to standardized 224 × 224 pixel inputs. From the fully connected layer of the Swin V2 Base model, 512 features were extracted for each tooth, and the concatenation of seven teeth yielded a 3584-dimensional feature vector per subject. These feature vectors were then used for regression analysis to predict chronological age on a day-level scale. In the second phase, nine machine learning regression models-LightGBM, RandomForest, ExtraTrees, GradientBoosting, XGBoost, KNN, SVR, MLP, and Gaussian Process Regression-were trained using the extracted features. Pairwise t-test analysis revealed ExtraTrees as the most statistically significant model. For this model, RMSE and MAE were 6.98 and 5.18 months for females, and 6.55 and 5.01 months for males. SHAP-based analysis highlighted the second molar (M2) and first premolar (P1) as the most influential features for females, and the first premolar (P1) and second molar (M2) for males. This automated pipeline enhances age prediction accuracy, reduces observer variability, and provides a reliable tool for clinical and forensic dental age estimation. Future work will explore dataset expansion, multimodal integration, and refined model architectures.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145812613","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topological Feature Extraction from Multi-color Channels for Pattern Recognition: An Application to Fundus Image Analysis. 基于多色通道的模式识别拓扑特征提取:在眼底图像分析中的应用。
Pub Date : 2025-12-22 DOI: 10.1007/s10278-025-01791-1
Fatih Gelir, Taymaz Akan, Owen T Carmichael, Md Shenuarin Bhuiyan, Steven A Conrad, John A Vanchiere, Christopher G Kevil, Mohammad Alfrad Nobel Bhuiyan

The automated analysis of medical images is crucial for early disease detection. In recent years, deep learning has become popular for medical image analysis. In this study, we employed color-based topological features with deep learning for pattern recognition. The data topology provides information about the image's shape and global features such as connectivity and holes. We used different color channels to identify changes in topological footprints by altering the image's color. We extracted topological, local binary pattern (LBP), and Gabor features and used machine learning and deep learning models for disease classification. The model's performance was tested using three open-source fundus image databases: the Asia Pacific Tele-ophthalmology Society (APTOS 2019) data, the Optic Retinal Image Database for Glaucoma Analysis (ORIGA), and the Automatic Detection Challenge on Age-Related Macular Degeneration (ICHALLENGE-AMD). We have found that topological features from different color models provide important information for disease diagnosis.

医学图像的自动分析对于疾病的早期检测至关重要。近年来,深度学习已成为医学图像分析的热门领域。在这项研究中,我们采用基于颜色的拓扑特征和深度学习进行模式识别。数据拓扑提供有关图像形状和全局特征(如连通性和孔洞)的信息。我们使用不同的颜色通道,通过改变图像的颜色来识别拓扑足迹的变化。我们提取了拓扑、局部二值模式(LBP)和Gabor特征,并使用机器学习和深度学习模型进行疾病分类。该模型的性能使用三个开源眼底图像数据库进行测试:亚太远程眼科学会(APTOS 2019)数据、青光眼分析光学视网膜图像数据库(ORIGA)和年龄相关性黄斑变性自动检测挑战(ichallengeamd)。我们发现不同颜色模型的拓扑特征为疾病诊断提供了重要信息。
{"title":"Topological Feature Extraction from Multi-color Channels for Pattern Recognition: An Application to Fundus Image Analysis.","authors":"Fatih Gelir, Taymaz Akan, Owen T Carmichael, Md Shenuarin Bhuiyan, Steven A Conrad, John A Vanchiere, Christopher G Kevil, Mohammad Alfrad Nobel Bhuiyan","doi":"10.1007/s10278-025-01791-1","DOIUrl":"https://doi.org/10.1007/s10278-025-01791-1","url":null,"abstract":"<p><p>The automated analysis of medical images is crucial for early disease detection. In recent years, deep learning has become popular for medical image analysis. In this study, we employed color-based topological features with deep learning for pattern recognition. The data topology provides information about the image's shape and global features such as connectivity and holes. We used different color channels to identify changes in topological footprints by altering the image's color. We extracted topological, local binary pattern (LBP), and Gabor features and used machine learning and deep learning models for disease classification. The model's performance was tested using three open-source fundus image databases: the Asia Pacific Tele-ophthalmology Society (APTOS 2019) data, the Optic Retinal Image Database for Glaucoma Analysis (ORIGA), and the Automatic Detection Challenge on Age-Related Macular Degeneration (ICHALLENGE-AMD). We have found that topological features from different color models provide important information for disease diagnosis.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145812597","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid YOLOv8s+Swin-T Transformer Approach for Automated Caries Detection on Periapical Radiographs. YOLOv8s+ swing - t变压器混合方法在根尖周x线片上自动检测龋病。
Pub Date : 2025-12-22 DOI: 10.1007/s10278-025-01763-5
D Meghana, A Manimaran

Early dental caries detection is essential for timely diagnosis and treatment. However, current deep learning (DL) models exhibit inconsistent accuracy across different dental X-ray datasets, revealing limitations in their robustness and adaptability. To automate caries detection in intraoral periapical radiographs, this study presents a hybrid object detector that integrates a Swin-T transformer with a YOLOv8s backbone. The model was trained on 1887 radiographs collected from the Sibar Institute of Dental Sciences, Guntur. To detect dental caries in intricate intraoral structures, this work presents an improved feature extraction through its hierarchical attention mechanism that outperforms convolutional neural network (CNN)-based models in both spatial understanding and contextual awareness. We evaluated the method against single-stage YOLOv8 variants (n, s, m, l) and a representative two-stage detector (Faster R-CNN with ResNet-50-FPNv2) under a consistent protocol. The proposed YOLOv8s+Swin-T outperformed all baselines in precision, recall, F1-score, and mAP@0.5, achieving 0.97 for precision/recall/F1 and 0.99 for mAP@0.5. These results underscore the model's clinical applicability and robustness, providing a reliable tool for accurate caries detection and supporting routine AI-assisted diagnosis.

早期发现龋齿对及时诊断和治疗至关重要。然而,目前的深度学习(DL)模型在不同的牙科x射线数据集上表现出不一致的准确性,揭示了其鲁棒性和适应性的局限性。为了在口腔内根尖周x线片中自动检测龋齿,本研究提出了一种混合物体检测器,该检测器集成了swing - t变压器和YOLOv8s骨干。该模型是根据从Guntur的Sibar牙科科学研究所收集的1887张x光片进行训练的。为了检测复杂口腔内结构中的龋齿,本研究提出了一种改进的特征提取方法,该方法通过分层注意机制在空间理解和上下文感知方面优于基于卷积神经网络(CNN)的模型。我们在一致协议下对单级YOLOv8变体(n, s, m, l)和具有代表性的两级检测器(Faster R-CNN with ResNet-50-FPNv2)进行了评估。提出的YOLOv8s+ swwin - t在精度、召回率、F1得分和mAP@0.5方面优于所有基线,精度/召回率/F1达到0.97,mAP@0.5达到0.99。这些结果强调了该模型的临床适用性和鲁棒性,为准确检测龋齿和支持常规人工智能辅助诊断提供了可靠的工具。
{"title":"A Hybrid YOLOv8s+Swin-T Transformer Approach for Automated Caries Detection on Periapical Radiographs.","authors":"D Meghana, A Manimaran","doi":"10.1007/s10278-025-01763-5","DOIUrl":"https://doi.org/10.1007/s10278-025-01763-5","url":null,"abstract":"<p><p>Early dental caries detection is essential for timely diagnosis and treatment. However, current deep learning (DL) models exhibit inconsistent accuracy across different dental X-ray datasets, revealing limitations in their robustness and adaptability. To automate caries detection in intraoral periapical radiographs, this study presents a hybrid object detector that integrates a Swin-T transformer with a YOLOv8s backbone. The model was trained on 1887 radiographs collected from the Sibar Institute of Dental Sciences, Guntur. To detect dental caries in intricate intraoral structures, this work presents an improved feature extraction through its hierarchical attention mechanism that outperforms convolutional neural network (CNN)-based models in both spatial understanding and contextual awareness. We evaluated the method against single-stage YOLOv8 variants (n, s, m, l) and a representative two-stage detector (Faster R-CNN with ResNet-50-FPNv2) under a consistent protocol. The proposed YOLOv8s+Swin-T outperformed all baselines in precision, recall, F1-score, and mAP@0.5, achieving 0.97 for precision/recall/F1 and 0.99 for mAP@0.5. These results underscore the model's clinical applicability and robustness, providing a reliable tool for accurate caries detection and supporting routine AI-assisted diagnosis.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145812599","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Harnessing Native-Resolution 2D Embeddings for Lung Cancer Classification: A Feasibility Study with the RAD-DINO Self-supervised Foundation Model. 利用原生分辨率二维嵌入进行肺癌分类:基于RAD-DINO自监督基础模型的可行性研究。
Pub Date : 2025-12-19 DOI: 10.1007/s10278-025-01748-4
Md Enamul Hoq, Lawrence Tarbox, Donald Johann, Linda Larson-Prior, Fred Prior

Low‑dose CT (LDCT) screening reduces lung cancer mortality but yields high false‑positive findings, motivating practical AI that aligns with slice‑based clinical review. We evaluate a label‑efficient 2D pipeline that uses frozen RAD‑DINO embeddings from native‑resolution axial slices and a lightweight multilayer perceptron for patient‑level risk estimation via mean aggregation and isotonic calibration. Using the NLST CT arm with outcomes defined over a 0-24‑month window, we construct a fixed patient‑level split (one CT per patient; no cross‑split leakage) and perform 25 repeated imbalanced test draws (~ 6% prevalence) to approximate screening conditions. At screening prevalence, RAD‑DINO + MLP achieves PR‑AUC = 0.705 (calibrated; raw 0.554) and ROC‑AUC = 0.817 (raw; 0.736 calibrated), with improved probability reliability following calibration; operating points are selected on validation and reported on test. For secondary ablations only, a near‑balanced cohort (N = 1984) yields accuracy 0.966, precision 0.974, recall 0.973, F1 0.973, and ROC‑AUC 0.912. Beyond classification, retrieval with triplet‑fine‑tuned embeddings attains Precision@5 = 0.853. Interpretability analyses show that cancer cases sustain higher top‑k slice scores and that directional SHAP concentrates on a small subset of high‑probability slices; label‑colored t‑SNE provides qualitative views of embedding structure. Limitations include single‑cohort evaluation, lack of Lung‑RADS labels in public NLST, and a CXR → CT pretraining shift; future work will pursue external validation and CT‑native self‑supervised continuation. Overall, frozen 2D foundation embeddings provide a strong, transparent, and computationally practical starting point for LDCT screening workflows under realistic prevalence.

低剂量CT (LDCT)筛查降低了肺癌死亡率,但产生了很高的假阳性结果,推动了与基于切片的临床评价相一致的实用人工智能。我们评估了一个标签高效的2D管道,该管道使用来自原生分辨率轴向切片的冷冻RAD - DINO嵌入和一个轻量级多层感知器,通过平均聚合和等渗校准进行患者级风险估计。使用NLST CT组,结果定义为0-24个月的窗口,我们构建了一个固定的患者水平分割(每个患者一台CT;无交叉分割泄漏),并进行了25次重复的不平衡试验(~ 6%的患病率)来近似筛选条件。在筛查流行率方面,RAD‑DINO + MLP的PR‑AUC = 0.705(校正后0.554),ROC‑AUC = 0.817(校正后0.736),校正后的概率信度有所提高;工作点在验证时选择,并在测试时报告。仅对于二次消融,近平衡队列(N = 1984)的准确度为0.966,精密度为0.974,召回率为0.973,F1为0.973,ROC - AUC为0.912。除了分类,检索与三联体微调嵌入达到Precision@5 = 0.853。可解释性分析表明,癌症病例维持较高的top - k切片分数,定向SHAP集中在高概率切片的一小部分;标签彩色t - SNE提供嵌入结构的定性视图。局限性包括单队列评估,公共NLST中缺乏Lung - RADS标签,以及CXR→CT预训练转换;未来的工作将追求外部验证和CT原生的自我监督延续。总的来说,冻结的2D基础嵌入为LDCT筛查工作流程提供了一个强大的、透明的、计算上实用的起点。
{"title":"Harnessing Native-Resolution 2D Embeddings for Lung Cancer Classification: A Feasibility Study with the RAD-DINO Self-supervised Foundation Model.","authors":"Md Enamul Hoq, Lawrence Tarbox, Donald Johann, Linda Larson-Prior, Fred Prior","doi":"10.1007/s10278-025-01748-4","DOIUrl":"https://doi.org/10.1007/s10278-025-01748-4","url":null,"abstract":"<p><p>Low‑dose CT (LDCT) screening reduces lung cancer mortality but yields high false‑positive findings, motivating practical AI that aligns with slice‑based clinical review. We evaluate a label‑efficient 2D pipeline that uses frozen RAD‑DINO embeddings from native‑resolution axial slices and a lightweight multilayer perceptron for patient‑level risk estimation via mean aggregation and isotonic calibration. Using the NLST CT arm with outcomes defined over a 0-24‑month window, we construct a fixed patient‑level split (one CT per patient; no cross‑split leakage) and perform 25 repeated imbalanced test draws (~ 6% prevalence) to approximate screening conditions. At screening prevalence, RAD‑DINO + MLP achieves PR‑AUC = 0.705 (calibrated; raw 0.554) and ROC‑AUC = 0.817 (raw; 0.736 calibrated), with improved probability reliability following calibration; operating points are selected on validation and reported on test. For secondary ablations only, a near‑balanced cohort (N = 1984) yields accuracy 0.966, precision 0.974, recall 0.973, F1 0.973, and ROC‑AUC 0.912. Beyond classification, retrieval with triplet‑fine‑tuned embeddings attains Precision@5 = 0.853. Interpretability analyses show that cancer cases sustain higher top‑k slice scores and that directional SHAP concentrates on a small subset of high‑probability slices; label‑colored t‑SNE provides qualitative views of embedding structure. Limitations include single‑cohort evaluation, lack of Lung‑RADS labels in public NLST, and a CXR → CT pretraining shift; future work will pursue external validation and CT‑native self‑supervised continuation. Overall, frozen 2D foundation embeddings provide a strong, transparent, and computationally practical starting point for LDCT screening workflows under realistic prevalence.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145795865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Transformer-based Deep Learning Models with Shape Guidance for Predicting Breast Cancer in Mammography Images. 基于变形器的深度学习模型与形状指导在乳房x线摄影图像中预测乳腺癌。
Pub Date : 2025-12-19 DOI: 10.1007/s10278-025-01773-3
Kengo Takahashi, Yuwen Zeng, Zhang Zhang, Kei Ichiji, Takuma Usuzaki, Ryusei Inamori, Haoyang Liu, Noriyasu Homma

Recent breast cancer research has investigated shape-based attention guidance in Vision Transformer (ViT) models, focusing on anatomical structures and the heterogeneity surrounding tumors. However, few studies have clarified the optimal transformer encoder layer stage for applying attention guidance. Our study aimed to evaluate the effectiveness of shape-guidance strategies by varying the combinations of encoder layers that guide attention to breast structures and by comparing the proposed models with conventional models. For the shape-guidance strategy, we applied breast masks to the attention mechanism to emphasize spatial dependencies and enhance the learning of positional relationships within breast anatomy. We then compared the representative models-Masked Transformer models that demonstrated the best performance across layer combinations-with the conventional ResNet50, ViT, and SwinT V2. In our study, a total of 2,436 publicly available mammography images from the Chinese Mammography Database via The Cancer Imaging Archive were analyzed. Three-fold cross-validation was employed, with a patient-wise split of 70% for training and 30% for validation. Model performance on differentiating breast cancer from non-cancer images was assessed by the area under the receiver-operating characteristic curve (AUROC). The results showed that applying masks at the Shallow and Deep stages gave the highest AUROC for Masked ViT. The Masked ViT achieved an AUROC of 0.885 [95% confidence interval: 0.849-0.918], a sensitivity of 0.876, and a specificity of 0.802, outperforming all other conventional models. These results indicate that incorporating mask guidance into particular Transformer encoders promotes representation learning, highlighting their potential as decision-support tools in breast cancer diagnosis.

最近的乳腺癌研究在视觉变形(ViT)模型中研究了基于形状的注意力引导,重点关注肿瘤周围的解剖结构和异质性。然而,很少有研究明确变压器编码器的最佳层阶段,以应用注意引导。我们的研究旨在通过改变编码器层的组合来评估形状引导策略的有效性,这些编码器层可以引导人们注意乳房结构,并将所提出的模型与传统模型进行比较。对于形状引导策略,我们将乳房面具应用于注意机制,以强调空间依赖性并增强乳房解剖中位置关系的学习。然后,我们比较了代表性的模型——展示跨层组合最佳性能的屏蔽变压器模型——与传统的ResNet50、ViT和SwinT V2。在我们的研究中,共有2436张公开的乳房x线摄影图像从中国乳房x线摄影数据库通过癌症影像档案进行分析。采用三重交叉验证,其中70%用于培训,30%用于验证。模型在区分乳腺癌和非癌症图像上的表现通过接受者操作特征曲线下的面积(AUROC)来评估。结果表明,在浅、深两阶段应用掩膜对掩膜ViT的AUROC最高。蒙面ViT的AUROC为0.885[95%可信区间:0.849-0.918],灵敏度为0.876,特异性为0.802,优于所有其他传统模型。这些结果表明,将掩膜指导结合到特定的Transformer编码器中可以促进表征学习,突出了它们作为乳腺癌诊断决策支持工具的潜力。
{"title":"Transformer-based Deep Learning Models with Shape Guidance for Predicting Breast Cancer in Mammography Images.","authors":"Kengo Takahashi, Yuwen Zeng, Zhang Zhang, Kei Ichiji, Takuma Usuzaki, Ryusei Inamori, Haoyang Liu, Noriyasu Homma","doi":"10.1007/s10278-025-01773-3","DOIUrl":"https://doi.org/10.1007/s10278-025-01773-3","url":null,"abstract":"<p><p>Recent breast cancer research has investigated shape-based attention guidance in Vision Transformer (ViT) models, focusing on anatomical structures and the heterogeneity surrounding tumors. However, few studies have clarified the optimal transformer encoder layer stage for applying attention guidance. Our study aimed to evaluate the effectiveness of shape-guidance strategies by varying the combinations of encoder layers that guide attention to breast structures and by comparing the proposed models with conventional models. For the shape-guidance strategy, we applied breast masks to the attention mechanism to emphasize spatial dependencies and enhance the learning of positional relationships within breast anatomy. We then compared the representative models-Masked Transformer models that demonstrated the best performance across layer combinations-with the conventional ResNet50, ViT, and SwinT V2. In our study, a total of 2,436 publicly available mammography images from the Chinese Mammography Database via The Cancer Imaging Archive were analyzed. Three-fold cross-validation was employed, with a patient-wise split of 70% for training and 30% for validation. Model performance on differentiating breast cancer from non-cancer images was assessed by the area under the receiver-operating characteristic curve (AUROC). The results showed that applying masks at the Shallow and Deep stages gave the highest AUROC for Masked ViT. The Masked ViT achieved an AUROC of 0.885 [95% confidence interval: 0.849-0.918], a sensitivity of 0.876, and a specificity of 0.802, outperforming all other conventional models. These results indicate that incorporating mask guidance into particular Transformer encoders promotes representation learning, highlighting their potential as decision-support tools in breast cancer diagnosis.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145795921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multimodal Machine Learning Integrating N-13 Ammonia PET and Clinical Variables Predicts Major Adverse Cardiac Events. 整合N-13氨PET和临床变量的多模态机器学习预测主要不良心脏事件。
Pub Date : 2025-12-19 DOI: 10.1007/s10278-025-01779-x
Ryo Mikurino, Michinobu Nagao, Masateru Kawakubo, Atsushi Yamamoto, Risako Nakao, Yuka Matsuo, Akiko Sakai, Shuji Sakai

The need for dynamic and static acquisitions under stress and rest in myocardial perfusion positron emission tomography (PET) is burdensome, and the short half-life of N-13 ammonia places additional constraints on scanning protocols. This study investigated whether combining static PET-derived images with clinical parameters via machine learning predicts major adverse cardiac events (MACE) to expand the utility of ammonia PET without dynamic scanning. The cohort comprised 386 patients, and during a mean follow-up of 345 days, MACE occurred in 35 patients. We applied stratified fivefold cross-validation based on MACE prediction with balanced and random shuffles. A logistic regression model was trained using all the explanatory variables after removing highly collinear features. Based on the cumulative importance of MACE prediction, a model was developed using the minimum number of top-ranked features accounting for > 50% of total cumulative importance. The predictive performance of a simple threshold-based classification using myocardial flow reserve (MFR) < 2.0 was also evaluated for comparison. The model was trained on 308 cases using three features: age, dyslipidemia, and resting end-diastolic volume. When tested on an independent set of 78 cases with fivefold cross-validation, it achieved an accuracy of 0.74 ± 0.06, a sensitivity of 0.74 ± 0.23, and a specificity of 0.74 ± 0.07. The accuracy, sensitivity, and specificity of simple MFR < 2.0 prediction were 0.58 ± 0.05, 0.77 ± 0.13, and 0.56 ± 0.05, respectively. A multimodal machine learning approach potentially serves as a clinically useful alternative to dynamic PET scans.

心肌灌注正电子发射断层扫描(PET)需要在压力和休息下进行动态和静态采集,这是繁重的,N-13氨的短半衰期对扫描方案施加了额外的限制。本研究探讨了通过机器学习将静态PET衍生图像与临床参数相结合是否可以预测主要不良心脏事件(MACE),以扩大氨PET在不进行动态扫描的情况下的应用。该队列包括386名患者,在平均345天的随访期间,35名患者发生了MACE。我们应用分层五重交叉验证基于MACE预测与平衡和随机洗牌。在去除高度共线性特征后,使用所有解释变量训练逻辑回归模型。基于MACE预测的累积重要度,利用占总累积重要度50%的最小排名特征数建立模型。基于心肌血流储备的简单阈值分类的预测性能
{"title":"Multimodal Machine Learning Integrating N-13 Ammonia PET and Clinical Variables Predicts Major Adverse Cardiac Events.","authors":"Ryo Mikurino, Michinobu Nagao, Masateru Kawakubo, Atsushi Yamamoto, Risako Nakao, Yuka Matsuo, Akiko Sakai, Shuji Sakai","doi":"10.1007/s10278-025-01779-x","DOIUrl":"https://doi.org/10.1007/s10278-025-01779-x","url":null,"abstract":"<p><p>The need for dynamic and static acquisitions under stress and rest in myocardial perfusion positron emission tomography (PET) is burdensome, and the short half-life of N-13 ammonia places additional constraints on scanning protocols. This study investigated whether combining static PET-derived images with clinical parameters via machine learning predicts major adverse cardiac events (MACE) to expand the utility of ammonia PET without dynamic scanning. The cohort comprised 386 patients, and during a mean follow-up of 345 days, MACE occurred in 35 patients. We applied stratified fivefold cross-validation based on MACE prediction with balanced and random shuffles. A logistic regression model was trained using all the explanatory variables after removing highly collinear features. Based on the cumulative importance of MACE prediction, a model was developed using the minimum number of top-ranked features accounting for > 50% of total cumulative importance. The predictive performance of a simple threshold-based classification using myocardial flow reserve (MFR) < 2.0 was also evaluated for comparison. The model was trained on 308 cases using three features: age, dyslipidemia, and resting end-diastolic volume. When tested on an independent set of 78 cases with fivefold cross-validation, it achieved an accuracy of 0.74 ± 0.06, a sensitivity of 0.74 ± 0.23, and a specificity of 0.74 ± 0.07. The accuracy, sensitivity, and specificity of simple MFR < 2.0 prediction were 0.58 ± 0.05, 0.77 ± 0.13, and 0.56 ± 0.05, respectively. A multimodal machine learning approach potentially serves as a clinically useful alternative to dynamic PET scans.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145795869","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Artificial Intelligence Detection and Classification of Ventriculoperitoneal Shunt Valves Utilizing Fine-Tuning of a Detection Model. 利用检测模型微调的脑室腹腔分流阀的人工智能检测和分类。
Pub Date : 2025-12-19 DOI: 10.1007/s10278-025-01794-y
Zari O'Connor, Austin Fullenkamp, Morgan P McBee

Ventriculoperitoneal (VP) shunts are a mainstay treatment for hydrocephalus and identifying the correct valve type is essential to determine its setting. The proliferation of valve models over the last few decades makes identification on radiographs more difficult and time consuming. We trained a deep learning model to detect and classify VP shunts on radiographs. We curated 2263 skull radiographs spanning 11 valve types and split the data 80/10/10 for training/validation/test. A YOLOv8-large model was fine-tuned with augmentation and hyperparameter search (300 epochs, batch 32, 640 × 640 input). The fine-tuned YOLOv8 model demonstrated robust performance on the test set, achieving a precision (P) of 0.949, recall (R) of 0.930, mean average precision (mAP50) of 0.951, mAP50-90 of 0.755, and max F1 score of 0.952 across all classes at a confidence threshold of 0.524. For individual valve types, mAP50 ranged from 0.841 to 0.995, mAP50-90 from 0.562 to 0.963, P from 0.8 to 1, and R from 0.667 to 1 demonstrating that the model overall performs well but does perform better for some valve types. The fine-tuned YOLOv8 model demonstrates high accuracy and generalization across a range of valve types, suggesting its potential for clinical application. Compared with prior purely classification approaches, detection explicitly localizes each valve, accommodates patients with multiple valves, and enables creation of downstream models to determine valve settings.

脑室-腹膜(VP)分流是脑积水的主要治疗方法,确定正确的瓣膜类型对于确定其设置至关重要。在过去的几十年里,瓣膜模型的激增使得x光片上的识别变得更加困难和耗时。我们训练了一个深度学习模型来检测和分类x光片上的VP分流器。我们收集了2263张颅骨x光片,涵盖11种瓣膜类型,并将数据分成80/10/10用于训练/验证/测试。使用增强和超参数搜索对YOLOv8-large模型进行微调(300 epoch, batch 32, 640 × 640输入)。微调后的YOLOv8模型在测试集上表现出稳健的性能,在置信阈值为0.524的情况下,所有类别的精度(P)为0.949,召回率(R)为0.930,平均平均精度(mAP50)为0.951,mAP50-90为0.755,最大F1分数为0.952。对于单个阀门类型,mAP50范围为0.841至0.995,mAP50-90范围为0.562至0.963,P范围为0.8至1,R范围为0.667至1,表明该模型总体上表现良好,但对某些阀门类型确实表现更好。经过微调的YOLOv8模型在一系列瓣膜类型中显示出较高的准确性和通用性,表明其具有临床应用潜力。与之前的纯粹分类方法相比,检测明确定位每个瓣膜,适应有多个瓣膜的患者,并能够创建下游模型来确定瓣膜设置。
{"title":"Artificial Intelligence Detection and Classification of Ventriculoperitoneal Shunt Valves Utilizing Fine-Tuning of a Detection Model.","authors":"Zari O'Connor, Austin Fullenkamp, Morgan P McBee","doi":"10.1007/s10278-025-01794-y","DOIUrl":"https://doi.org/10.1007/s10278-025-01794-y","url":null,"abstract":"<p><p>Ventriculoperitoneal (VP) shunts are a mainstay treatment for hydrocephalus and identifying the correct valve type is essential to determine its setting. The proliferation of valve models over the last few decades makes identification on radiographs more difficult and time consuming. We trained a deep learning model to detect and classify VP shunts on radiographs. We curated 2263 skull radiographs spanning 11 valve types and split the data 80/10/10 for training/validation/test. A YOLOv8-large model was fine-tuned with augmentation and hyperparameter search (300 epochs, batch 32, 640 × 640 input). The fine-tuned YOLOv8 model demonstrated robust performance on the test set, achieving a precision (P) of 0.949, recall (R) of 0.930, mean average precision (mAP50) of 0.951, mAP50-90 of 0.755, and max F1 score of 0.952 across all classes at a confidence threshold of 0.524. For individual valve types, mAP50 ranged from 0.841 to 0.995, mAP50-90 from 0.562 to 0.963, P from 0.8 to 1, and R from 0.667 to 1 demonstrating that the model overall performs well but does perform better for some valve types. The fine-tuned YOLOv8 model demonstrates high accuracy and generalization across a range of valve types, suggesting its potential for clinical application. Compared with prior purely classification approaches, detection explicitly localizes each valve, accommodates patients with multiple valves, and enables creation of downstream models to determine valve settings.</p>","PeriodicalId":516858,"journal":{"name":"Journal of imaging informatics in medicine","volume":" ","pages":""},"PeriodicalIF":0.0,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145795897","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of imaging informatics in medicine
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1