首页 > 最新文献

Journal of Imaging最新文献

英文 中文
Adjunct Automated Breast Ultrasound in Mammographic Screening: A Systematic Review and Meta-Analysis. 辅助自动乳腺超声在乳房x线摄影筛查中的应用:一项系统综述和荟萃分析。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-22 DOI: 10.3390/jimaging12010003
Ghufran Jassim, Fahad AlZayani, Suchita Dsilva

Mammographic sensitivity is reduced in women with dense breasts, leading to missed cancers and a higher burden of interval cancers. Automated breast ultrasound (ABUS) and ultrasound tomography (UST) have been introduced as supplemental breast imaging modalities, but primary studies are heterogeneous, and previous reviews have not focused on screening settings or on head-to-head comparisons with handheld ultrasound (HHUS). We systematically searched PubMed, Embase, Web of Science and the Cochrane Library for studies from 1 January 2000 to 31 May 2025 evaluating ABUS or UST as adjuncts to mammographic screening. Two reviewers independently selected studies and assessed risk of bias. When at least two clinically comparable studies were available, we pooled sensitivity and specificity using random-effects bivariate meta-analysis. Eighteen studies (just over 20,000 screening or recall episodes) met the inclusion criteria; 16 evaluated ABUS/ABVS and 2 UST. Adding ABUS to mammography increased sensitivity by 6-35 percentage points and improved cancer detection by 2.4-4.3 per 1000 women with dense breasts, with higher recall rates and modest reductions in specificity. When ABUS was compared directly with HHUS, pooled sensitivity was 0.90 and specificity 0.89, with HHUS showing slightly lower sensitivity and slightly higher specificity. Only two studies had an overall low risk of bias, and heterogeneity (particularly for specificity) was substantial. ABUS is a practical and scalable adjunct to mammography that increases cancer detection in women with dense breasts, with an expected trade-off of higher recall and modest specificity loss. Its comparative diagnostic accuracy appears broadly non-inferior to HHUS. However, the predominance of high-risk-of-bias studies and between-study heterogeneity means that high-quality population-based trials and standardised reporting are still required before widespread implementation in organised screening programmes.

乳房致密的女性乳房x光检查灵敏度降低,导致癌症漏诊,间隔期癌症的负担更高。自动乳房超声(ABUS)和超声断层扫描(UST)已被引入作为补充乳房成像方式,但主要研究是异质性的,以前的综述没有关注筛查设置或与手持式超声(HHUS)的头对头比较。我们系统地检索了PubMed、Embase、Web of Science和Cochrane图书馆2000年1月1日至2025年5月31日期间评估ABUS或UST作为乳腺x线摄影筛查辅助手段的研究。两位审稿人独立选择研究并评估偏倚风险。当至少有两项临床可比研究可用时,我们使用随机效应双变量荟萃分析合并敏感性和特异性。18项研究(超过20,000个筛查或回忆事件)符合纳入标准;16例评估ABUS/ABVS, 2例评估UST。在乳房x线摄影中加入ABUS可使敏感性提高6-35个百分点,每1000名乳腺致密的女性中可提高2.4-4.3个癌症检出率,召回率更高,特异性略有降低。当ABUS与HHUS直接比较时,合并敏感性为0.90,特异性为0.89,其中HHUS敏感性略低,特异性略高。只有两项研究总体偏倚风险较低,异质性(特别是特异性)很大。ABUS是一种实用且可扩展的乳房x线摄影辅助手段,可提高致密乳房女性的癌症检出率,预期具有更高的召回率和适度的特异性损失。其相对诊断的准确性似乎不低于hus。然而,高风险偏倚研究的优势和研究间异质性意味着,在有组织的筛查计划中广泛实施之前,仍然需要高质量的基于人群的试验和标准化报告。
{"title":"Adjunct Automated Breast Ultrasound in Mammographic Screening: A Systematic Review and Meta-Analysis.","authors":"Ghufran Jassim, Fahad AlZayani, Suchita Dsilva","doi":"10.3390/jimaging12010003","DOIUrl":"10.3390/jimaging12010003","url":null,"abstract":"<p><p>Mammographic sensitivity is reduced in women with dense breasts, leading to missed cancers and a higher burden of interval cancers. Automated breast ultrasound (ABUS) and ultrasound tomography (UST) have been introduced as supplemental breast imaging modalities, but primary studies are heterogeneous, and previous reviews have not focused on screening settings or on head-to-head comparisons with handheld ultrasound (HHUS). We systematically searched PubMed, Embase, Web of Science and the Cochrane Library for studies from 1 January 2000 to 31 May 2025 evaluating ABUS or UST as adjuncts to mammographic screening. Two reviewers independently selected studies and assessed risk of bias. When at least two clinically comparable studies were available, we pooled sensitivity and specificity using random-effects bivariate meta-analysis. Eighteen studies (just over 20,000 screening or recall episodes) met the inclusion criteria; 16 evaluated ABUS/ABVS and 2 UST. Adding ABUS to mammography increased sensitivity by 6-35 percentage points and improved cancer detection by 2.4-4.3 per 1000 women with dense breasts, with higher recall rates and modest reductions in specificity. When ABUS was compared directly with HHUS, pooled sensitivity was 0.90 and specificity 0.89, with HHUS showing slightly lower sensitivity and slightly higher specificity. Only two studies had an overall low risk of bias, and heterogeneity (particularly for specificity) was substantial. ABUS is a practical and scalable adjunct to mammography that increases cancer detection in women with dense breasts, with an expected trade-off of higher recall and modest specificity loss. Its comparative diagnostic accuracy appears broadly non-inferior to HHUS. However, the predominance of high-risk-of-bias studies and between-study heterogeneity means that high-quality population-based trials and standardised reporting are still required before widespread implementation in organised screening programmes.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843147/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Development of a Multispectral Image Database in Visible-Near-Infrared for Demosaicking and Machine Learning Applications. 用于去马赛克和机器学习应用的可见-近红外多光谱图像数据库的开发。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-20 DOI: 10.3390/jimaging12010002
Vahid Mohammadi, Sovi Guillaume Sodjinou, Pierre Gouton

The use of Multispectral (MS) imaging is growing fast across many research fields. However, one of the obstacles researchers face is the limited availability of multispectral image databases. This arises from two factors: multispectral cameras are a relatively recent technology, and they are not widely available. Hence, the development of an image database is crucial for research on multispectral images. This study takes advantage of two high-end MS cameras in visible and near-infrared based on filter array technology developed in the PImRob platform, the University of Burgundy, to provide a freely accessible database. The database includes high-resolution MS images taken from different plants and weeds, along with annotated images and masks. The original raw images and the demosaicked images have been provided. The database has been developed for research on demosaicking techniques, segmentation algorithms, or deep learning for crop/weed discrimination.

多光谱(MS)成像在许多研究领域的应用正在迅速增长。然而,研究人员面临的障碍之一是多光谱图像数据库的有限可用性。这源于两个因素:多光谱相机是一项相对较新的技术,而且还没有广泛使用。因此,图像数据库的开发对于多光谱图像的研究至关重要。本研究利用在勃艮第大学(University of Burgundy)的PImRob平台上开发的滤波阵列技术,利用两台高端可见和近红外MS相机,提供一个免费访问的数据库。该数据库包括从不同植物和杂草中拍摄的高分辨率MS图像,以及注释图像和掩模。提供了原始图像和去马赛克图像。该数据库已开发用于研究去马赛克技术,分割算法或作物/杂草识别的深度学习。
{"title":"Development of a Multispectral Image Database in Visible-Near-Infrared for Demosaicking and Machine Learning Applications.","authors":"Vahid Mohammadi, Sovi Guillaume Sodjinou, Pierre Gouton","doi":"10.3390/jimaging12010002","DOIUrl":"10.3390/jimaging12010002","url":null,"abstract":"<p><p>The use of Multispectral (MS) imaging is growing fast across many research fields. However, one of the obstacles researchers face is the limited availability of multispectral image databases. This arises from two factors: multispectral cameras are a relatively recent technology, and they are not widely available. Hence, the development of an image database is crucial for research on multispectral images. This study takes advantage of two high-end MS cameras in visible and near-infrared based on filter array technology developed in the PImRob platform, the University of Burgundy, to provide a freely accessible database. The database includes high-resolution MS images taken from different plants and weeds, along with annotated images and masks. The original raw images and the demosaicked images have been provided. The database has been developed for research on demosaicking techniques, segmentation algorithms, or deep learning for crop/weed discrimination.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842297/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Non-Destructive Mangosteen Volume Estimation via Multi-View Instance Segmentation and Hybrid Geometric Modeling. 基于多视图实例分割和混合几何建模的山竹无损体积估计。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-19 DOI: 10.3390/jimaging12010001
Wattanapong Kurdthongmee, Arsanchai Sukkuea, Md Eshrat E Alahi, Qi Zeng

In precision agriculture, accurate, non-destructive estimation of fruit volume is crucial for quality grading, yield prediction, and post-harvest management. While vision-based methods provided some usefulness, fruits with complex geometry-such as mangosteen (Garcinia mangostana L.)-are difficult due to their large calyx, which may lead to difficulties in solving using traditional form-modeling methods. Traditional geometric solutions such as ellipsoid approximations, diameter-height estimation, and shape-from-silhouette reconstruction often fail because the irregular calyx generates asymmetric protrusions that violate their basic form assumptions. We offer a novel study framework employing both multi-view instance segmentation and hybrid geometrical feature modeling to quantitatively model mangosteen volume with traditional 2D imaging. A You Only Look Once (YOLO)-based segmentation model was employed to explicitly separate the fruit body from the calyx. Calyx inclusion resulted in dense geometric noise and reduced model performance (R2<0.40). We trained eight regression models on a curated and augmented 900 image dataset (N=720, test N=180). The models used single-view and multi-view geometric regressors (V∝A1.5), polynomial hybrid configurations, ellipsoid-based approximations, as well as hybrid feature formulations. Multi-view models consistently outperformed single-view models, and the average predictive accuracy improved from R2=0.6493 to R2=0.7290. The best model is indeed a hybrid linear regression model with side- and bottom-area features-(As1.5, Ab1.5)-combined with ellipsoid-derived volume estimation-(Vellipsoid)-which resulted in R2=0.7290, a Mean Absolute Percentage Error (MAPE) of 16.04%, and a Root Mean Square Error (RMSE) of 31.9 cm3 on the test set. These results confirm the proposed model as a low-cost, interpretable, and flexible model for real-time fruit volume estimation, ready for incorporation into automated sorting and grading systems integrated in post-harvest processing pipelines.

在精准农业中,水果体积的准确、无损估计对于质量分级、产量预测和收获后管理至关重要。虽然基于视觉的方法提供了一些有用的东西,但具有复杂几何形状的水果,如山竹(Garcinia mangostana L.),由于它们的花萼很大,因此很难使用传统的形状建模方法来解决问题。传统的几何解决方案,如椭球近似、直径-高度估计和形状-轮廓重建往往失败,因为不规则的花萼产生不对称的突起,违反了它们的基本形式假设。我们提出了一种新的研究框架,采用多视图实例分割和混合几何特征建模,对传统二维成像的山竹体积进行定量建模。采用基于YOLO (You Only Look Once)的分割模型将果体与花萼明确分离。花萼夹杂导致几何噪声密集,模型性能降低(R2)
{"title":"Non-Destructive Mangosteen Volume Estimation via Multi-View Instance Segmentation and Hybrid Geometric Modeling.","authors":"Wattanapong Kurdthongmee, Arsanchai Sukkuea, Md Eshrat E Alahi, Qi Zeng","doi":"10.3390/jimaging12010001","DOIUrl":"10.3390/jimaging12010001","url":null,"abstract":"<p><p>In precision agriculture, accurate, non-destructive estimation of fruit volume is crucial for quality grading, yield prediction, and post-harvest management. While vision-based methods provided some usefulness, fruits with complex geometry-such as mangosteen (<i>Garcinia mangostana</i> L.)-are difficult due to their large calyx, which may lead to difficulties in solving using traditional form-modeling methods. Traditional geometric solutions such as ellipsoid approximations, diameter-height estimation, and shape-from-silhouette reconstruction often fail because the irregular calyx generates asymmetric protrusions that violate their basic form assumptions. We offer a novel study framework employing both multi-view instance segmentation and hybrid geometrical feature modeling to quantitatively model mangosteen volume with traditional 2D imaging. A You Only Look Once (YOLO)-based segmentation model was employed to explicitly separate the fruit body from the calyx. Calyx inclusion resulted in dense geometric noise and reduced model performance (R2<0.40). We trained eight regression models on a curated and augmented 900 image dataset (N=720, test N=180). The models used single-view and multi-view geometric regressors (V∝A1.5), polynomial hybrid configurations, ellipsoid-based approximations, as well as hybrid feature formulations. Multi-view models consistently outperformed single-view models, and the average predictive accuracy improved from R2=0.6493 to R2=0.7290. The best model is indeed a hybrid linear regression model with side- and bottom-area features-(As1.5, Ab1.5)-combined with ellipsoid-derived volume estimation-(Vellipsoid)-which resulted in R2=0.7290, a Mean Absolute Percentage Error (MAPE) of 16.04%, and a Root Mean Square Error (RMSE) of 31.9 cm3 on the test set. These results confirm the proposed model as a low-cost, interpretable, and flexible model for real-time fruit volume estimation, ready for incorporation into automated sorting and grading systems integrated in post-harvest processing pipelines.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842316/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Structured Review and Quantitative Profiling of Public Brain MRI Datasets for Foundation Model Development. 用于基础模型开发的公共脑MRI数据集的结构化回顾和定量分析。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-18 DOI: 10.3390/jimaging11120454
Minh Sao Khue Luu, Margaret V Benedichuk, Ekaterina I Roppert, Roman M Kenzhin, Bair N Tuchinov

The development of foundation models for brain MRI depends critically on the scale, diversity, and consistency of available data, yet systematic assessments of these factors remain scarce. In this study, we analyze 54 publicly accessible brain MRI datasets encompassing over 538,031 scans to provide a structured, multi-level overview tailored to foundation model development. At the dataset level, we characterize modality composition, disease coverage, and dataset scale, revealing strong imbalances between large healthy cohorts and smaller clinical populations. At the image level, we quantify voxel spacing, orientation, and intensity distributions across 14 representative datasets, demonstrating substantial heterogeneity that can influence representation learning. We then perform a quantitative evaluation of preprocessing variability, examining how intensity normalization, bias field correction, skull stripping, spatial registration, and interpolation alter voxel statistics and geometry. While these steps improve within-dataset consistency, residual differences persist between datasets. Finally, a feature-space case study using a 3D DenseNet121 shows measurable residual covariate shift after standardized preprocessing, confirming that harmonization alone cannot eliminate inter-dataset bias. Together, these analyses provide a unified characterization of variability in public brain MRI resources and emphasize the need for preprocessing-aware and domain-adaptive strategies in the design of generalizable brain MRI foundation models.

脑MRI基础模型的发展主要取决于可用数据的规模、多样性和一致性,但对这些因素的系统评估仍然很少。在本研究中,我们分析了54个可公开访问的脑MRI数据集,包括超过538,031次扫描,以提供针对基础模型开发量身定制的结构化,多层次概述。在数据集水平上,我们描述了模式组成、疾病覆盖和数据集规模,揭示了大型健康队列和较小临床人群之间的强烈不平衡。在图像层面,我们量化了14个代表性数据集的体素间距、方向和强度分布,证明了可能影响表征学习的实质性异质性。然后,我们对预处理变异性进行定量评估,检查强度归一化、偏置场校正、颅骨剥离、空间配准和插值如何改变体素统计和几何。虽然这些步骤提高了数据集内的一致性,但数据集之间的残余差异仍然存在。最后,使用3D DenseNet121的特征空间案例研究显示,经过标准化预处理后,残差协变量移位可测量,证实仅靠协调无法消除数据集间偏差。总之,这些分析为公共脑MRI资源的可变性提供了统一的表征,并强调了在设计可推广的脑MRI基础模型时需要预处理感知和领域自适应策略。
{"title":"A Structured Review and Quantitative Profiling of Public Brain MRI Datasets for Foundation Model Development.","authors":"Minh Sao Khue Luu, Margaret V Benedichuk, Ekaterina I Roppert, Roman M Kenzhin, Bair N Tuchinov","doi":"10.3390/jimaging11120454","DOIUrl":"10.3390/jimaging11120454","url":null,"abstract":"<p><p>The development of foundation models for brain MRI depends critically on the scale, diversity, and consistency of available data, yet systematic assessments of these factors remain scarce. In this study, we analyze 54 publicly accessible brain MRI datasets encompassing over 538,031 scans to provide a structured, multi-level overview tailored to foundation model development. At the dataset level, we characterize modality composition, disease coverage, and dataset scale, revealing strong imbalances between large healthy cohorts and smaller clinical populations. At the image level, we quantify voxel spacing, orientation, and intensity distributions across 14 representative datasets, demonstrating substantial heterogeneity that can influence representation learning. We then perform a quantitative evaluation of preprocessing variability, examining how intensity normalization, bias field correction, skull stripping, spatial registration, and interpolation alter voxel statistics and geometry. While these steps improve within-dataset consistency, residual differences persist between datasets. Finally, a feature-space case study using a 3D DenseNet121 shows measurable residual covariate shift after standardized preprocessing, confirming that harmonization alone cannot eliminate inter-dataset bias. Together, these analyses provide a unified characterization of variability in public brain MRI resources and emphasize the need for preprocessing-aware and domain-adaptive strategies in the design of generalizable brain MRI foundation models.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733697/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Salient Object Detection in Optical Remote Sensing Images Based on Hierarchical Semantic Interaction. 基于层次语义交互的光学遥感图像显著目标检测。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-17 DOI: 10.3390/jimaging11120453
Jingfan Xu, Qi Zhang, Jinwen Xing, Mingquan Zhou, Guohua Geng

Existing salient object detection methods for optical remote sensing images still face certain limitations due to complex background variations, significant scale discrepancies among targets, severe background interference, and diverse topological structures. On the one hand, the feature transmission process often neglects the constraints and complementary effects of high-level features on low-level features, leading to insufficient feature interaction and weakened model representation. On the other hand, decoder architectures generally rely on simple cascaded structures, which fail to adequately exploit and utilize contextual information. To address these challenges, this study proposes a Hierarchical Semantic Interaction Module to enhance salient object detection performance in optical remote sensing scenarios. The module introduces foreground content modeling and a hierarchical semantic interaction mechanism within a multi-scale feature space, reinforcing the synergy and complementarity among features at different levels. This effectively highlights multi-scale and multi-type salient regions in complex backgrounds. Extensive experiments on multiple optical remote sensing datasets demonstrate the effectiveness of the proposed method. Specifically, on the EORSSD dataset, our full model integrating both CA and PA modules improves the max F-measure from 0.8826 to 0.9100 (↑2.74%), increases maxE from 0.9603 to 0.9727 (↑1.24%), and enhances the S-measure from 0.9026 to 0.9295 (↑2.69%) compared with the baseline. These results clearly demonstrate the effectiveness of the proposed modules and verify the robustness and strong generalization capability of our method in complex remote sensing scenarios.

由于背景变化复杂、目标间尺度差异大、背景干扰严重、拓扑结构多样等原因,现有的光学遥感图像显著目标检测方法仍存在一定的局限性。一方面,特征传递过程往往忽略了高层特征对低层特征的约束和互补作用,导致特征交互不足,模型表征弱化。另一方面,解码器架构通常依赖于简单的级联结构,不能充分利用上下文信息。为了解决这些挑战,本研究提出了一个分层语义交互模块,以提高光学遥感场景下的显著目标检测性能。该模块在多尺度特征空间中引入前景内容建模和分层语义交互机制,加强了不同层次特征之间的协同和互补。这有效地突出了复杂背景下的多尺度、多类型突出区域。在多个光学遥感数据集上的大量实验证明了该方法的有效性。具体而言,在EORSSD数据集上,与基线相比,集成CA和PA模块的完整模型将最大f -测度从0.8826提高到0.9100(↑2.74%),将maxE从0.9603提高到0.9727(↑1.24%),将s -测度从0.9026提高到0.9295(↑2.69%)。这些结果清楚地证明了所提出模块的有效性,并验证了我们的方法在复杂遥感场景下的鲁棒性和强大的泛化能力。
{"title":"Salient Object Detection in Optical Remote Sensing Images Based on Hierarchical Semantic Interaction.","authors":"Jingfan Xu, Qi Zhang, Jinwen Xing, Mingquan Zhou, Guohua Geng","doi":"10.3390/jimaging11120453","DOIUrl":"10.3390/jimaging11120453","url":null,"abstract":"<p><p>Existing salient object detection methods for optical remote sensing images still face certain limitations due to complex background variations, significant scale discrepancies among targets, severe background interference, and diverse topological structures. On the one hand, the feature transmission process often neglects the constraints and complementary effects of high-level features on low-level features, leading to insufficient feature interaction and weakened model representation. On the other hand, decoder architectures generally rely on simple cascaded structures, which fail to adequately exploit and utilize contextual information. To address these challenges, this study proposes a Hierarchical Semantic Interaction Module to enhance salient object detection performance in optical remote sensing scenarios. The module introduces foreground content modeling and a hierarchical semantic interaction mechanism within a multi-scale feature space, reinforcing the synergy and complementarity among features at different levels. This effectively highlights multi-scale and multi-type salient regions in complex backgrounds. Extensive experiments on multiple optical remote sensing datasets demonstrate the effectiveness of the proposed method. Specifically, on the EORSSD dataset, our full model integrating both CA and PA modules improves the max F-measure from 0.8826 to 0.9100 (↑2.74%), increases maxE from 0.9603 to 0.9727 (↑1.24%), and enhances the S-measure from 0.9026 to 0.9295 (↑2.69%) compared with the baseline. These results clearly demonstrate the effectiveness of the proposed modules and verify the robustness and strong generalization capability of our method in complex remote sensing scenarios.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733419/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SRE-FMaps: A Sinkhorn-Regularized Elastic Functional Map Framework for Non-Isometric 3D Shape Matching. SRE-FMaps:一种用于非等距三维形状匹配的sinkhorn正则化弹性功能地图框架。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-16 DOI: 10.3390/jimaging11120452
Dan Zhang, Yue Zhang, Ning Wang, Dong Zhao

Precise 3D shape correspondence is a fundamental prerequisite for critical applications ranging from medical anatomical modeling to visual recognition. However, non-isometric 3D shape matching remains a challenging task due to the limited sensitivity of traditional Laplace-Beltrami (LB) bases to local geometric deformations such as stretching and bending. To address these limitations, this paper proposes a Sinkhorn-Regularized Elastic Functional Map framework (SRE-FMaps) that integrates entropy-regularized optimal transport with an elastic thin-shell energy basis. First, a sparse Sinkhorn transport plan is adopted to initialize a bijective correspondence with linear computational complexity. Then, a non-orthogonal elastic basis, derived from the Hessian of thin-shell deformation energy, is introduced to enhance high-frequency feature perception. Finally, correspondence stability is quantified through a cosine-based elastic distance metric, enabling retrieval and classification. Experiments on the SHREC2015, McGill, and Face datasets demonstrate that SRE-FMaps reduces the correspondence error by a maximum of 32% and achieves an average of 92.3% classification accuracy (with a peak of 94.74% on the Face dataset). Moreover, the framework exhibits superior robustness, yielding a recall of up to 91.67% and an F1-score of 0.94, effectively handling bending, stretching, and folding deformations compared with conventional LB-based functional map pipelines. The proposed framework provides a scalable solution for non-isometric shape correspondence in medical modeling, 3D reconstruction, and visual recognition.

精确的3D形状对应是从医学解剖建模到视觉识别等关键应用的基本先决条件。然而,由于传统的Laplace-Beltrami (LB)基对局部几何变形(如拉伸和弯曲)的灵敏度有限,非等距三维形状匹配仍然是一项具有挑战性的任务。为了解决这些限制,本文提出了一个sinkhorn -正则化弹性功能映射框架(SRE-FMaps),该框架将熵正则化最优传输与弹性薄壳能量基础相结合。首先,采用稀疏的Sinkhorn传输方案初始化计算复杂度为线性的双目标对应。然后,引入基于薄壳变形能Hessian的非正交弹性基来增强高频特征感知;最后,通过基于余弦的弹性距离度量来量化对应稳定性,从而实现检索和分类。在SHREC2015、McGill和Face数据集上的实验表明,SRE-FMaps最大减少了32%的对应误差,平均达到了92.3%的分类准确率(在Face数据集上达到了94.74%的峰值)。此外,与传统的基于lb的功能图管道相比,该框架具有出色的鲁棒性,召回率高达91.67%,f1得分为0.94,能够有效地处理弯曲、拉伸和折叠变形。该框架为医学建模、三维重建和视觉识别中的非等距形状对应提供了可扩展的解决方案。
{"title":"SRE-FMaps: A Sinkhorn-Regularized Elastic Functional Map Framework for Non-Isometric 3D Shape Matching.","authors":"Dan Zhang, Yue Zhang, Ning Wang, Dong Zhao","doi":"10.3390/jimaging11120452","DOIUrl":"10.3390/jimaging11120452","url":null,"abstract":"<p><p>Precise 3D shape correspondence is a fundamental prerequisite for critical applications ranging from medical anatomical modeling to visual recognition. However, non-isometric 3D shape matching remains a challenging task due to the limited sensitivity of traditional Laplace-Beltrami (LB) bases to local geometric deformations such as stretching and bending. To address these limitations, this paper proposes a Sinkhorn-Regularized Elastic Functional Map framework (SRE-FMaps) that integrates entropy-regularized optimal transport with an elastic thin-shell energy basis. First, a sparse Sinkhorn transport plan is adopted to initialize a bijective correspondence with linear computational complexity. Then, a non-orthogonal elastic basis, derived from the Hessian of thin-shell deformation energy, is introduced to enhance high-frequency feature perception. Finally, correspondence stability is quantified through a cosine-based elastic distance metric, enabling retrieval and classification. Experiments on the SHREC2015, McGill, and Face datasets demonstrate that SRE-FMaps reduces the correspondence error by a maximum of 32% and achieves an average of 92.3% classification accuracy (with a peak of 94.74% on the Face dataset). Moreover, the framework exhibits superior robustness, yielding a recall of up to 91.67% and an F1-score of 0.94, effectively handling bending, stretching, and folding deformations compared with conventional LB-based functional map pipelines. The proposed framework provides a scalable solution for non-isometric shape correspondence in medical modeling, 3D reconstruction, and visual recognition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733647/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Sensory Representation of Neural Networks Using Sound and Color for Medical Imaging Segmentation. 基于声音和颜色的神经网络感官表征在医学影像分割中的应用。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-15 DOI: 10.3390/jimaging11120449
Irenel Lopo Da Silva, Nicolas Francisco Lori, José Manuel Ferreira Machado

This paper introduces a novel framework for sensory representation of brain imaging data, combining deep learning-based segmentation with multimodal visual and auditory outputs. Structural magnetic resonance imaging (MRI) predictions are converted into color-coded maps and stereophonic/MIDI sonifications, enabling intuitive interpretation of cortical activation patterns. High-precision U-Net models efficiently generate these outputs, supporting clinical decision-making, cognitive research, and creative applications. Spatial, intensity, and anomalous features are encoded into perceivable visual and auditory cues, facilitating early detection and introducing the concept of "auditory biomarkers" for potential pathological identification. Despite current limitations, including dataset size, absence of clinical validation, and heuristic-based sonification, the pipeline demonstrates technical feasibility and robustness. Future work will focus on clinical user studies, the application of functional MRI (fMRI) time-series for dynamic sonification, and the integration of real-time emotional feedback in cinematic contexts. This multisensory approach offers a promising avenue for enhancing the interpretability of complex neuroimaging data across medical, research, and artistic domains.

本文介绍了一种新的脑成像数据感官表征框架,将基于深度学习的分割与多模态视觉和听觉输出相结合。结构磁共振成像(MRI)预测被转换成彩色编码图和立体声/MIDI超声,从而能够直观地解释皮层激活模式。高精度U-Net模型有效地生成这些输出,支持临床决策、认知研究和创造性应用。空间、强度和异常特征被编码为可感知的视觉和听觉线索,促进早期发现,并引入“听觉生物标志物”的概念,用于潜在的病理识别。尽管目前存在一些限制,包括数据集大小、缺乏临床验证和基于启发式的超声,但该管道展示了技术可行性和鲁棒性。未来的工作将集中在临床用户研究,功能磁共振成像(fMRI)时间序列动态超声的应用,以及电影背景下实时情绪反馈的整合。这种多感官方法为提高医学、研究和艺术领域复杂神经成像数据的可解释性提供了一条有前途的途径。
{"title":"Sensory Representation of Neural Networks Using Sound and Color for Medical Imaging Segmentation.","authors":"Irenel Lopo Da Silva, Nicolas Francisco Lori, José Manuel Ferreira Machado","doi":"10.3390/jimaging11120449","DOIUrl":"10.3390/jimaging11120449","url":null,"abstract":"<p><p>This paper introduces a novel framework for sensory representation of brain imaging data, combining deep learning-based segmentation with multimodal visual and auditory outputs. Structural magnetic resonance imaging (MRI) predictions are converted into color-coded maps and stereophonic/MIDI sonifications, enabling intuitive interpretation of cortical activation patterns. High-precision U-Net models efficiently generate these outputs, supporting clinical decision-making, cognitive research, and creative applications. Spatial, intensity, and anomalous features are encoded into perceivable visual and auditory cues, facilitating early detection and introducing the concept of \"auditory biomarkers\" for potential pathological identification. Despite current limitations, including dataset size, absence of clinical validation, and heuristic-based sonification, the pipeline demonstrates technical feasibility and robustness. Future work will focus on clinical user studies, the application of functional MRI (fMRI) time-series for dynamic sonification, and the integration of real-time emotional feedback in cinematic contexts. This multisensory approach offers a promising avenue for enhancing the interpretability of complex neuroimaging data across medical, research, and artistic domains.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733452/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Generative Adversarial Networks to Improve COVID-19 Classification on Ultrasound Images. 生成对抗网络在超声图像上改进COVID-19分类中的应用
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-15 DOI: 10.3390/jimaging11120451
Pedro Sérgio Tôrres Figueiredo Silva, Antonio Mauricio Ferreira Leite Miranda de Sá, Wagner Coelho de Albuquerque Pereira, Leonardo Bonato Felix, José Manoel de Seixas

COVID-19 screening is crucial for the early diagnosis and treatment of the disease, with lung ultrasound posing as a cost-effective alternative to other imaging techniques. Given the dependency on medical expertise and experience to accurately identify patterns in ultrasound exams, deep learning techniques have been explored for automatically classifying patients' conditions. However, the limited availability of public medical databases remains a significant obstacle to the development of more advanced models. To address the data scarcity problem, this study proposes a method that leverages generative adversarial networks (GANs) to generate synthetic lung ultrasound images, which are subsequently used to train frame-based classification models. Two types of GANs are considered: Wasserstein GANs (WGAN) and Pix2Pix. Specific tools are used to show that the synthetic data produced present a distribution close to the original data. The classification models trained with synthetic data achieved a peak accuracy of 96.32% ± 4.17%, significantly outperforming the maximum accuracy of 82.69% ± 10.42% obtained when training only with the original data. Furthermore, the best results are comparable to, and in some cases surpass, those reported in recent related studies.

COVID-19筛查对于疾病的早期诊断和治疗至关重要,肺部超声是其他成像技术的一种具有成本效益的替代方案。考虑到在超声检查中准确识别模式依赖于医学专业知识和经验,人们已经探索了深度学习技术来自动分类患者的病情。然而,公共医疗数据库的有限可用性仍然是开发更先进模型的重大障碍。为了解决数据稀缺问题,本研究提出了一种利用生成对抗网络(GANs)生成合成肺部超声图像的方法,该方法随后用于训练基于框架的分类模型。考虑了两种类型的gan: Wasserstein gan (WGAN)和Pix2Pix。使用特定的工具来显示生成的合成数据呈现接近原始数据的分布。使用合成数据训练的分类模型的最高准确率为96.32%±4.17%,明显优于仅使用原始数据训练时的最高准确率82.69%±10.42%。此外,最好的结果与最近的相关研究报告相当,在某些情况下甚至超过了这些研究报告。
{"title":"Application of Generative Adversarial Networks to Improve COVID-19 Classification on Ultrasound Images.","authors":"Pedro Sérgio Tôrres Figueiredo Silva, Antonio Mauricio Ferreira Leite Miranda de Sá, Wagner Coelho de Albuquerque Pereira, Leonardo Bonato Felix, José Manoel de Seixas","doi":"10.3390/jimaging11120451","DOIUrl":"10.3390/jimaging11120451","url":null,"abstract":"<p><p>COVID-19 screening is crucial for the early diagnosis and treatment of the disease, with lung ultrasound posing as a cost-effective alternative to other imaging techniques. Given the dependency on medical expertise and experience to accurately identify patterns in ultrasound exams, deep learning techniques have been explored for automatically classifying patients' conditions. However, the limited availability of public medical databases remains a significant obstacle to the development of more advanced models. To address the data scarcity problem, this study proposes a method that leverages generative adversarial networks (GANs) to generate synthetic lung ultrasound images, which are subsequently used to train frame-based classification models. Two types of GANs are considered: Wasserstein GANs (WGAN) and Pix2Pix. Specific tools are used to show that the synthetic data produced present a distribution close to the original data. The classification models trained with synthetic data achieved a peak accuracy of 96.32% ± 4.17%, significantly outperforming the maximum accuracy of 82.69% ± 10.42% obtained when training only with the original data. Furthermore, the best results are comparable to, and in some cases surpass, those reported in recent related studies.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733862/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Applying Radiomics to Predict Outcomes in Patients with High-Grade Retroperitoneal Sarcoma Treated with Preoperative Radiotherapy. 应用放射组学预测术前放疗治疗的高级别腹膜后肉瘤的预后。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-15 DOI: 10.3390/jimaging11120450
Adel Shahnam, Nicholas Hardcastle, David E Gyorki, Katrina M Ingley, Krystel Tran, Catherine Mitchell, Sarat Chander, Julie Chu, Michael Henderson, Alan Herschtal, Mathias Bressel, Jeremy Lewin

Retroperitoneal sarcomas (RPS) are rare tumours, primarily treated with surgical resection. However, recurrences are frequent. Combining clinical factors with CT-derived radiomic features could enhance treatment stratification and personalization. This study aims to assess whether radiomic features provide additional prognostic value beyond clinicopathological features in patients with high-risk RPS treated with preoperative radiotherapy. This retrospective study included patients aged 18 or older with non-recurrent and non-metastatic RPS treated with preoperative radiotherapy between 2008 and 2016. Hazard ratios (HR) were calculated using Cox proportional hazards regression to assess the impact of clinical and radiomic features on time to event outcomes. Predictive accuracy was assessed with c-statistics. Radiomic analysis was performed on the high-risk group (undifferentiated pleomorphic sarcoma, well-differentiated/de-differentiated liposarcoma or grade 2/3 leiomyosarcoma). Seventy-two patients were included, with a median follow-up of 3.7 years, the 5-year overall survival (OS) was 67%. Multivariable analysis showed older age (HR: 1.3 per 5-year increase, p = 0.04), grade 3 (HR: 180.3, p = 0.02), and larger tumours (HR: 4.0 per 10 cm increase, p = 0.02) predicted worse OS. In the higher-risk group, the c-statistic for the clinical model was 0.59 (time to distant metastasis (TDM)) and 0.56 (OS). Among 27 radiomic features, kurtosis improved OS prediction (c-statistic 0.69, p = 0.013), and Neighbourhood Gray-Tone Difference Matrix (NGTDM) busyness improved it to 0.73 (p = 0.036). Kurtosis also improved TDM prediction (c-statistic 0.72, p = 0.023). Radiomic features may complement clinicopathological factors in predicting overall survival and time to distant metastasis in high-risk retroperitoneal sarcoma. These exploratory findings warrant validation in larger, multi-institutional studies.

腹膜后肉瘤(RPS)是一种罕见的肿瘤,主要通过手术切除治疗。然而,复发是频繁的。将临床因素与ct放射学特征相结合,可以加强治疗分层和个性化。本研究旨在评估放射学特征是否在术前放疗的高危RPS患者的临床病理特征之外提供额外的预后价值。这项回顾性研究纳入了2008年至2016年期间接受术前放疗的18岁及以上非复发性和非转移性RPS患者。使用Cox比例风险回归计算风险比(HR),以评估临床和放射学特征对事件发生时间的影响。用c-statistics评估预测准确性。对高危组(未分化多形性肉瘤、高分化/去分化脂肪肉瘤或2/3级平滑肌肉瘤)进行放射组学分析。纳入72例患者,中位随访3.7年,5年总生存率(OS)为67%。多变量分析显示,年龄较大(HR:每5年增加1.3,p = 0.04)、3级(HR: 180.3, p = 0.02)和肿瘤较大(HR:每10厘米增加4.0,p = 0.02)预测较差的OS。在高危组,临床模型的c统计量为0.59 (TDM)和0.56 (OS)。在27个放射学特征中,峰度改善了OS预测(c-统计量0.69,p = 0.013),邻域灰调差分矩阵(NGTDM)繁忙度改善了OS预测(p = 0.036)。峰度也能改善TDM预测(c统计量0.72,p = 0.023)。放射学特征可以补充临床病理因素,预测高危腹膜后肉瘤的总生存期和远处转移时间。这些探索性的发现在更大的、多机构的研究中得到了验证。
{"title":"Applying Radiomics to Predict Outcomes in Patients with High-Grade Retroperitoneal Sarcoma Treated with Preoperative Radiotherapy.","authors":"Adel Shahnam, Nicholas Hardcastle, David E Gyorki, Katrina M Ingley, Krystel Tran, Catherine Mitchell, Sarat Chander, Julie Chu, Michael Henderson, Alan Herschtal, Mathias Bressel, Jeremy Lewin","doi":"10.3390/jimaging11120450","DOIUrl":"10.3390/jimaging11120450","url":null,"abstract":"<p><p>Retroperitoneal sarcomas (RPS) are rare tumours, primarily treated with surgical resection. However, recurrences are frequent. Combining clinical factors with CT-derived radiomic features could enhance treatment stratification and personalization. This study aims to assess whether radiomic features provide additional prognostic value beyond clinicopathological features in patients with high-risk RPS treated with preoperative radiotherapy. This retrospective study included patients aged 18 or older with non-recurrent and non-metastatic RPS treated with preoperative radiotherapy between 2008 and 2016. Hazard ratios (HR) were calculated using Cox proportional hazards regression to assess the impact of clinical and radiomic features on time to event outcomes. Predictive accuracy was assessed with c-statistics. Radiomic analysis was performed on the high-risk group (undifferentiated pleomorphic sarcoma, well-differentiated/de-differentiated liposarcoma or grade 2/3 leiomyosarcoma). Seventy-two patients were included, with a median follow-up of 3.7 years, the 5-year overall survival (OS) was 67%. Multivariable analysis showed older age (HR: 1.3 per 5-year increase, <i>p</i> = 0.04), grade 3 (HR: 180.3, <i>p</i> = 0.02), and larger tumours (HR: 4.0 per 10 cm increase, <i>p</i> = 0.02) predicted worse OS. In the higher-risk group, the c-statistic for the clinical model was 0.59 (time to distant metastasis (TDM)) and 0.56 (OS). Among 27 radiomic features, kurtosis improved OS prediction (c-statistic 0.69, <i>p</i> = 0.013), and Neighbourhood Gray-Tone Difference Matrix (NGTDM) busyness improved it to 0.73 (<i>p</i> = 0.036). Kurtosis also improved TDM prediction (c-statistic 0.72, <i>p</i> = 0.023). Radiomic features may complement clinicopathological factors in predicting overall survival and time to distant metastasis in high-risk retroperitoneal sarcoma. These exploratory findings warrant validation in larger, multi-institutional studies.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733741/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Dual-Path Convolutional Neural Network with Squeeze-and-Excitation Attention for Lung and Colon Histopathology Classification. 双路径卷积神经网络在肺和结肠组织病理学分类中的应用。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-14 DOI: 10.3390/jimaging11120448
Helala AlShehri

Lung and colon cancers remain among the leading causes of cancer-related mortality worldwide, underscoring the need for rapid and accurate histopathological diagnosis. Manual examination of biopsy slides is often time-consuming and prone to inter-observer variability, which highlights the importance of developing reliable and explainable automated diagnostic systems. This study presents DPCSE-Net, a lightweight dual-path convolutional neural network enhanced with a squeeze-and-excitation (SE) attention mechanism for lung and colon cancer classification. The dual-path structure captures both fine-grained cellular textures and global contextual information through multiscale feature extraction, while the SE attention module adaptively recalibrates channel responses to emphasize discriminative features. To enhance transparency and interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM), attention heatmaps, and Integrated Gradients are employed to visualize class-specific activation patterns and verify that the model's focus aligns with diagnostically relevant tissue regions. Evaluated on the publicly available LC25000 dataset, DPCSE-Net achieved state-of-the-art performance with 99.88% accuracy and F1-score, while maintaining low computational complexity. Ablation experiments confirmed the contribution of the dual-path design and SE module, and qualitative analyses demonstrated the model's strong interpretability. These results establish DPCSE-Net as an accurate, efficient, and explainable framework for computer-aided histopathological diagnosis, supporting the broader goals of explainable AI in computer vision.

肺癌和结肠癌仍然是世界范围内癌症相关死亡的主要原因,强调了快速准确的组织病理学诊断的必要性。手工检查活检切片通常很耗时,而且容易在观察者之间发生变化,这突出了开发可靠和可解释的自动诊断系统的重要性。本研究提出了DPCSE-Net,一个轻量级的双路径卷积神经网络,增强了挤压和激发(SE)注意机制,用于肺癌和结肠癌的分类。双路径结构通过多尺度特征提取捕获细粒度细胞纹理和全局上下文信息,而SE注意模块自适应地重新校准通道响应以强调判别特征。为了提高透明度和可解释性,梯度加权类激活映射(Grad-CAM)、注意力热图和集成梯度被用于可视化类特定的激活模式,并验证模型的焦点与诊断相关的组织区域一致。在公开可用的LC25000数据集上进行评估,DPCSE-Net达到了最先进的性能,准确率为99.88%,得分为f1,同时保持了较低的计算复杂度。烧蚀实验证实了双路径设计和SE模块的贡献,定性分析表明模型具有较强的可解释性。这些结果建立了DPCSE-Net作为计算机辅助组织病理学诊断的准确、高效和可解释的框架,支持计算机视觉中可解释人工智能的更广泛目标。
{"title":"Dual-Path Convolutional Neural Network with Squeeze-and-Excitation Attention for Lung and Colon Histopathology Classification.","authors":"Helala AlShehri","doi":"10.3390/jimaging11120448","DOIUrl":"10.3390/jimaging11120448","url":null,"abstract":"<p><p>Lung and colon cancers remain among the leading causes of cancer-related mortality worldwide, underscoring the need for rapid and accurate histopathological diagnosis. Manual examination of biopsy slides is often time-consuming and prone to inter-observer variability, which highlights the importance of developing reliable and explainable automated diagnostic systems. This study presents DPCSE-Net, a lightweight dual-path convolutional neural network enhanced with a squeeze-and-excitation (SE) attention mechanism for lung and colon cancer classification. The dual-path structure captures both fine-grained cellular textures and global contextual information through multiscale feature extraction, while the SE attention module adaptively recalibrates channel responses to emphasize discriminative features. To enhance transparency and interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM), attention heatmaps, and Integrated Gradients are employed to visualize class-specific activation patterns and verify that the model's focus aligns with diagnostically relevant tissue regions. Evaluated on the publicly available LC25000 dataset, DPCSE-Net achieved state-of-the-art performance with 99.88% accuracy and F1-score, while maintaining low computational complexity. Ablation experiments confirmed the contribution of the dual-path design and SE module, and qualitative analyses demonstrated the model's strong interpretability. These results establish DPCSE-Net as an accurate, efficient, and explainable framework for computer-aided histopathological diagnosis, supporting the broader goals of explainable AI in computer vision.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733646/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1