Pub Date : 2025-12-22DOI: 10.3390/jimaging12010003
Ghufran Jassim, Fahad AlZayani, Suchita Dsilva
Mammographic sensitivity is reduced in women with dense breasts, leading to missed cancers and a higher burden of interval cancers. Automated breast ultrasound (ABUS) and ultrasound tomography (UST) have been introduced as supplemental breast imaging modalities, but primary studies are heterogeneous, and previous reviews have not focused on screening settings or on head-to-head comparisons with handheld ultrasound (HHUS). We systematically searched PubMed, Embase, Web of Science and the Cochrane Library for studies from 1 January 2000 to 31 May 2025 evaluating ABUS or UST as adjuncts to mammographic screening. Two reviewers independently selected studies and assessed risk of bias. When at least two clinically comparable studies were available, we pooled sensitivity and specificity using random-effects bivariate meta-analysis. Eighteen studies (just over 20,000 screening or recall episodes) met the inclusion criteria; 16 evaluated ABUS/ABVS and 2 UST. Adding ABUS to mammography increased sensitivity by 6-35 percentage points and improved cancer detection by 2.4-4.3 per 1000 women with dense breasts, with higher recall rates and modest reductions in specificity. When ABUS was compared directly with HHUS, pooled sensitivity was 0.90 and specificity 0.89, with HHUS showing slightly lower sensitivity and slightly higher specificity. Only two studies had an overall low risk of bias, and heterogeneity (particularly for specificity) was substantial. ABUS is a practical and scalable adjunct to mammography that increases cancer detection in women with dense breasts, with an expected trade-off of higher recall and modest specificity loss. Its comparative diagnostic accuracy appears broadly non-inferior to HHUS. However, the predominance of high-risk-of-bias studies and between-study heterogeneity means that high-quality population-based trials and standardised reporting are still required before widespread implementation in organised screening programmes.
乳房致密的女性乳房x光检查灵敏度降低,导致癌症漏诊,间隔期癌症的负担更高。自动乳房超声(ABUS)和超声断层扫描(UST)已被引入作为补充乳房成像方式,但主要研究是异质性的,以前的综述没有关注筛查设置或与手持式超声(HHUS)的头对头比较。我们系统地检索了PubMed、Embase、Web of Science和Cochrane图书馆2000年1月1日至2025年5月31日期间评估ABUS或UST作为乳腺x线摄影筛查辅助手段的研究。两位审稿人独立选择研究并评估偏倚风险。当至少有两项临床可比研究可用时,我们使用随机效应双变量荟萃分析合并敏感性和特异性。18项研究(超过20,000个筛查或回忆事件)符合纳入标准;16例评估ABUS/ABVS, 2例评估UST。在乳房x线摄影中加入ABUS可使敏感性提高6-35个百分点,每1000名乳腺致密的女性中可提高2.4-4.3个癌症检出率,召回率更高,特异性略有降低。当ABUS与HHUS直接比较时,合并敏感性为0.90,特异性为0.89,其中HHUS敏感性略低,特异性略高。只有两项研究总体偏倚风险较低,异质性(特别是特异性)很大。ABUS是一种实用且可扩展的乳房x线摄影辅助手段,可提高致密乳房女性的癌症检出率,预期具有更高的召回率和适度的特异性损失。其相对诊断的准确性似乎不低于hus。然而,高风险偏倚研究的优势和研究间异质性意味着,在有组织的筛查计划中广泛实施之前,仍然需要高质量的基于人群的试验和标准化报告。
{"title":"Adjunct Automated Breast Ultrasound in Mammographic Screening: A Systematic Review and Meta-Analysis.","authors":"Ghufran Jassim, Fahad AlZayani, Suchita Dsilva","doi":"10.3390/jimaging12010003","DOIUrl":"10.3390/jimaging12010003","url":null,"abstract":"<p><p>Mammographic sensitivity is reduced in women with dense breasts, leading to missed cancers and a higher burden of interval cancers. Automated breast ultrasound (ABUS) and ultrasound tomography (UST) have been introduced as supplemental breast imaging modalities, but primary studies are heterogeneous, and previous reviews have not focused on screening settings or on head-to-head comparisons with handheld ultrasound (HHUS). We systematically searched PubMed, Embase, Web of Science and the Cochrane Library for studies from 1 January 2000 to 31 May 2025 evaluating ABUS or UST as adjuncts to mammographic screening. Two reviewers independently selected studies and assessed risk of bias. When at least two clinically comparable studies were available, we pooled sensitivity and specificity using random-effects bivariate meta-analysis. Eighteen studies (just over 20,000 screening or recall episodes) met the inclusion criteria; 16 evaluated ABUS/ABVS and 2 UST. Adding ABUS to mammography increased sensitivity by 6-35 percentage points and improved cancer detection by 2.4-4.3 per 1000 women with dense breasts, with higher recall rates and modest reductions in specificity. When ABUS was compared directly with HHUS, pooled sensitivity was 0.90 and specificity 0.89, with HHUS showing slightly lower sensitivity and slightly higher specificity. Only two studies had an overall low risk of bias, and heterogeneity (particularly for specificity) was substantial. ABUS is a practical and scalable adjunct to mammography that increases cancer detection in women with dense breasts, with an expected trade-off of higher recall and modest specificity loss. Its comparative diagnostic accuracy appears broadly non-inferior to HHUS. However, the predominance of high-risk-of-bias studies and between-study heterogeneity means that high-quality population-based trials and standardised reporting are still required before widespread implementation in organised screening programmes.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843147/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054095","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-20DOI: 10.3390/jimaging12010002
Vahid Mohammadi, Sovi Guillaume Sodjinou, Pierre Gouton
The use of Multispectral (MS) imaging is growing fast across many research fields. However, one of the obstacles researchers face is the limited availability of multispectral image databases. This arises from two factors: multispectral cameras are a relatively recent technology, and they are not widely available. Hence, the development of an image database is crucial for research on multispectral images. This study takes advantage of two high-end MS cameras in visible and near-infrared based on filter array technology developed in the PImRob platform, the University of Burgundy, to provide a freely accessible database. The database includes high-resolution MS images taken from different plants and weeds, along with annotated images and masks. The original raw images and the demosaicked images have been provided. The database has been developed for research on demosaicking techniques, segmentation algorithms, or deep learning for crop/weed discrimination.
多光谱(MS)成像在许多研究领域的应用正在迅速增长。然而,研究人员面临的障碍之一是多光谱图像数据库的有限可用性。这源于两个因素:多光谱相机是一项相对较新的技术,而且还没有广泛使用。因此,图像数据库的开发对于多光谱图像的研究至关重要。本研究利用在勃艮第大学(University of Burgundy)的PImRob平台上开发的滤波阵列技术,利用两台高端可见和近红外MS相机,提供一个免费访问的数据库。该数据库包括从不同植物和杂草中拍摄的高分辨率MS图像,以及注释图像和掩模。提供了原始图像和去马赛克图像。该数据库已开发用于研究去马赛克技术,分割算法或作物/杂草识别的深度学习。
{"title":"Development of a Multispectral Image Database in Visible-Near-Infrared for Demosaicking and Machine Learning Applications.","authors":"Vahid Mohammadi, Sovi Guillaume Sodjinou, Pierre Gouton","doi":"10.3390/jimaging12010002","DOIUrl":"10.3390/jimaging12010002","url":null,"abstract":"<p><p>The use of Multispectral (MS) imaging is growing fast across many research fields. However, one of the obstacles researchers face is the limited availability of multispectral image databases. This arises from two factors: multispectral cameras are a relatively recent technology, and they are not widely available. Hence, the development of an image database is crucial for research on multispectral images. This study takes advantage of two high-end MS cameras in visible and near-infrared based on filter array technology developed in the PImRob platform, the University of Burgundy, to provide a freely accessible database. The database includes high-resolution MS images taken from different plants and weeds, along with annotated images and masks. The original raw images and the demosaicked images have been provided. The database has been developed for research on demosaicking techniques, segmentation algorithms, or deep learning for crop/weed discrimination.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842297/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054167","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-19DOI: 10.3390/jimaging12010001
Wattanapong Kurdthongmee, Arsanchai Sukkuea, Md Eshrat E Alahi, Qi Zeng
In precision agriculture, accurate, non-destructive estimation of fruit volume is crucial for quality grading, yield prediction, and post-harvest management. While vision-based methods provided some usefulness, fruits with complex geometry-such as mangosteen (Garcinia mangostana L.)-are difficult due to their large calyx, which may lead to difficulties in solving using traditional form-modeling methods. Traditional geometric solutions such as ellipsoid approximations, diameter-height estimation, and shape-from-silhouette reconstruction often fail because the irregular calyx generates asymmetric protrusions that violate their basic form assumptions. We offer a novel study framework employing both multi-view instance segmentation and hybrid geometrical feature modeling to quantitatively model mangosteen volume with traditional 2D imaging. A You Only Look Once (YOLO)-based segmentation model was employed to explicitly separate the fruit body from the calyx. Calyx inclusion resulted in dense geometric noise and reduced model performance (R2<0.40). We trained eight regression models on a curated and augmented 900 image dataset (N=720, test N=180). The models used single-view and multi-view geometric regressors (V∝A1.5), polynomial hybrid configurations, ellipsoid-based approximations, as well as hybrid feature formulations. Multi-view models consistently outperformed single-view models, and the average predictive accuracy improved from R2=0.6493 to R2=0.7290. The best model is indeed a hybrid linear regression model with side- and bottom-area features-(As1.5, Ab1.5)-combined with ellipsoid-derived volume estimation-(Vellipsoid)-which resulted in R2=0.7290, a Mean Absolute Percentage Error (MAPE) of 16.04%, and a Root Mean Square Error (RMSE) of 31.9 cm3 on the test set. These results confirm the proposed model as a low-cost, interpretable, and flexible model for real-time fruit volume estimation, ready for incorporation into automated sorting and grading systems integrated in post-harvest processing pipelines.
在精准农业中,水果体积的准确、无损估计对于质量分级、产量预测和收获后管理至关重要。虽然基于视觉的方法提供了一些有用的东西,但具有复杂几何形状的水果,如山竹(Garcinia mangostana L.),由于它们的花萼很大,因此很难使用传统的形状建模方法来解决问题。传统的几何解决方案,如椭球近似、直径-高度估计和形状-轮廓重建往往失败,因为不规则的花萼产生不对称的突起,违反了它们的基本形式假设。我们提出了一种新的研究框架,采用多视图实例分割和混合几何特征建模,对传统二维成像的山竹体积进行定量建模。采用基于YOLO (You Only Look Once)的分割模型将果体与花萼明确分离。花萼夹杂导致几何噪声密集,模型性能降低(R2)
{"title":"Non-Destructive Mangosteen Volume Estimation via Multi-View Instance Segmentation and Hybrid Geometric Modeling.","authors":"Wattanapong Kurdthongmee, Arsanchai Sukkuea, Md Eshrat E Alahi, Qi Zeng","doi":"10.3390/jimaging12010001","DOIUrl":"10.3390/jimaging12010001","url":null,"abstract":"<p><p>In precision agriculture, accurate, non-destructive estimation of fruit volume is crucial for quality grading, yield prediction, and post-harvest management. While vision-based methods provided some usefulness, fruits with complex geometry-such as mangosteen (<i>Garcinia mangostana</i> L.)-are difficult due to their large calyx, which may lead to difficulties in solving using traditional form-modeling methods. Traditional geometric solutions such as ellipsoid approximations, diameter-height estimation, and shape-from-silhouette reconstruction often fail because the irregular calyx generates asymmetric protrusions that violate their basic form assumptions. We offer a novel study framework employing both multi-view instance segmentation and hybrid geometrical feature modeling to quantitatively model mangosteen volume with traditional 2D imaging. A You Only Look Once (YOLO)-based segmentation model was employed to explicitly separate the fruit body from the calyx. Calyx inclusion resulted in dense geometric noise and reduced model performance (R2<0.40). We trained eight regression models on a curated and augmented 900 image dataset (N=720, test N=180). The models used single-view and multi-view geometric regressors (V∝A1.5), polynomial hybrid configurations, ellipsoid-based approximations, as well as hybrid feature formulations. Multi-view models consistently outperformed single-view models, and the average predictive accuracy improved from R2=0.6493 to R2=0.7290. The best model is indeed a hybrid linear regression model with side- and bottom-area features-(As1.5, Ab1.5)-combined with ellipsoid-derived volume estimation-(Vellipsoid)-which resulted in R2=0.7290, a Mean Absolute Percentage Error (MAPE) of 16.04%, and a Root Mean Square Error (RMSE) of 31.9 cm3 on the test set. These results confirm the proposed model as a low-cost, interpretable, and flexible model for real-time fruit volume estimation, ready for incorporation into automated sorting and grading systems integrated in post-harvest processing pipelines.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842316/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-18DOI: 10.3390/jimaging11120454
Minh Sao Khue Luu, Margaret V Benedichuk, Ekaterina I Roppert, Roman M Kenzhin, Bair N Tuchinov
The development of foundation models for brain MRI depends critically on the scale, diversity, and consistency of available data, yet systematic assessments of these factors remain scarce. In this study, we analyze 54 publicly accessible brain MRI datasets encompassing over 538,031 scans to provide a structured, multi-level overview tailored to foundation model development. At the dataset level, we characterize modality composition, disease coverage, and dataset scale, revealing strong imbalances between large healthy cohorts and smaller clinical populations. At the image level, we quantify voxel spacing, orientation, and intensity distributions across 14 representative datasets, demonstrating substantial heterogeneity that can influence representation learning. We then perform a quantitative evaluation of preprocessing variability, examining how intensity normalization, bias field correction, skull stripping, spatial registration, and interpolation alter voxel statistics and geometry. While these steps improve within-dataset consistency, residual differences persist between datasets. Finally, a feature-space case study using a 3D DenseNet121 shows measurable residual covariate shift after standardized preprocessing, confirming that harmonization alone cannot eliminate inter-dataset bias. Together, these analyses provide a unified characterization of variability in public brain MRI resources and emphasize the need for preprocessing-aware and domain-adaptive strategies in the design of generalizable brain MRI foundation models.
{"title":"A Structured Review and Quantitative Profiling of Public Brain MRI Datasets for Foundation Model Development.","authors":"Minh Sao Khue Luu, Margaret V Benedichuk, Ekaterina I Roppert, Roman M Kenzhin, Bair N Tuchinov","doi":"10.3390/jimaging11120454","DOIUrl":"10.3390/jimaging11120454","url":null,"abstract":"<p><p>The development of foundation models for brain MRI depends critically on the scale, diversity, and consistency of available data, yet systematic assessments of these factors remain scarce. In this study, we analyze 54 publicly accessible brain MRI datasets encompassing over 538,031 scans to provide a structured, multi-level overview tailored to foundation model development. At the dataset level, we characterize modality composition, disease coverage, and dataset scale, revealing strong imbalances between large healthy cohorts and smaller clinical populations. At the image level, we quantify voxel spacing, orientation, and intensity distributions across 14 representative datasets, demonstrating substantial heterogeneity that can influence representation learning. We then perform a quantitative evaluation of preprocessing variability, examining how intensity normalization, bias field correction, skull stripping, spatial registration, and interpolation alter voxel statistics and geometry. While these steps improve within-dataset consistency, residual differences persist between datasets. Finally, a feature-space case study using a 3D DenseNet121 shows measurable residual covariate shift after standardized preprocessing, confirming that harmonization alone cannot eliminate inter-dataset bias. Together, these analyses provide a unified characterization of variability in public brain MRI resources and emphasize the need for preprocessing-aware and domain-adaptive strategies in the design of generalizable brain MRI foundation models.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733697/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Existing salient object detection methods for optical remote sensing images still face certain limitations due to complex background variations, significant scale discrepancies among targets, severe background interference, and diverse topological structures. On the one hand, the feature transmission process often neglects the constraints and complementary effects of high-level features on low-level features, leading to insufficient feature interaction and weakened model representation. On the other hand, decoder architectures generally rely on simple cascaded structures, which fail to adequately exploit and utilize contextual information. To address these challenges, this study proposes a Hierarchical Semantic Interaction Module to enhance salient object detection performance in optical remote sensing scenarios. The module introduces foreground content modeling and a hierarchical semantic interaction mechanism within a multi-scale feature space, reinforcing the synergy and complementarity among features at different levels. This effectively highlights multi-scale and multi-type salient regions in complex backgrounds. Extensive experiments on multiple optical remote sensing datasets demonstrate the effectiveness of the proposed method. Specifically, on the EORSSD dataset, our full model integrating both CA and PA modules improves the max F-measure from 0.8826 to 0.9100 (↑2.74%), increases maxE from 0.9603 to 0.9727 (↑1.24%), and enhances the S-measure from 0.9026 to 0.9295 (↑2.69%) compared with the baseline. These results clearly demonstrate the effectiveness of the proposed modules and verify the robustness and strong generalization capability of our method in complex remote sensing scenarios.
{"title":"Salient Object Detection in Optical Remote Sensing Images Based on Hierarchical Semantic Interaction.","authors":"Jingfan Xu, Qi Zhang, Jinwen Xing, Mingquan Zhou, Guohua Geng","doi":"10.3390/jimaging11120453","DOIUrl":"10.3390/jimaging11120453","url":null,"abstract":"<p><p>Existing salient object detection methods for optical remote sensing images still face certain limitations due to complex background variations, significant scale discrepancies among targets, severe background interference, and diverse topological structures. On the one hand, the feature transmission process often neglects the constraints and complementary effects of high-level features on low-level features, leading to insufficient feature interaction and weakened model representation. On the other hand, decoder architectures generally rely on simple cascaded structures, which fail to adequately exploit and utilize contextual information. To address these challenges, this study proposes a Hierarchical Semantic Interaction Module to enhance salient object detection performance in optical remote sensing scenarios. The module introduces foreground content modeling and a hierarchical semantic interaction mechanism within a multi-scale feature space, reinforcing the synergy and complementarity among features at different levels. This effectively highlights multi-scale and multi-type salient regions in complex backgrounds. Extensive experiments on multiple optical remote sensing datasets demonstrate the effectiveness of the proposed method. Specifically, on the EORSSD dataset, our full model integrating both CA and PA modules improves the max F-measure from 0.8826 to 0.9100 (↑2.74%), increases maxE from 0.9603 to 0.9727 (↑1.24%), and enhances the S-measure from 0.9026 to 0.9295 (↑2.69%) compared with the baseline. These results clearly demonstrate the effectiveness of the proposed modules and verify the robustness and strong generalization capability of our method in complex remote sensing scenarios.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733419/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821377","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-16DOI: 10.3390/jimaging11120452
Dan Zhang, Yue Zhang, Ning Wang, Dong Zhao
Precise 3D shape correspondence is a fundamental prerequisite for critical applications ranging from medical anatomical modeling to visual recognition. However, non-isometric 3D shape matching remains a challenging task due to the limited sensitivity of traditional Laplace-Beltrami (LB) bases to local geometric deformations such as stretching and bending. To address these limitations, this paper proposes a Sinkhorn-Regularized Elastic Functional Map framework (SRE-FMaps) that integrates entropy-regularized optimal transport with an elastic thin-shell energy basis. First, a sparse Sinkhorn transport plan is adopted to initialize a bijective correspondence with linear computational complexity. Then, a non-orthogonal elastic basis, derived from the Hessian of thin-shell deformation energy, is introduced to enhance high-frequency feature perception. Finally, correspondence stability is quantified through a cosine-based elastic distance metric, enabling retrieval and classification. Experiments on the SHREC2015, McGill, and Face datasets demonstrate that SRE-FMaps reduces the correspondence error by a maximum of 32% and achieves an average of 92.3% classification accuracy (with a peak of 94.74% on the Face dataset). Moreover, the framework exhibits superior robustness, yielding a recall of up to 91.67% and an F1-score of 0.94, effectively handling bending, stretching, and folding deformations compared with conventional LB-based functional map pipelines. The proposed framework provides a scalable solution for non-isometric shape correspondence in medical modeling, 3D reconstruction, and visual recognition.
{"title":"SRE-FMaps: A Sinkhorn-Regularized Elastic Functional Map Framework for Non-Isometric 3D Shape Matching.","authors":"Dan Zhang, Yue Zhang, Ning Wang, Dong Zhao","doi":"10.3390/jimaging11120452","DOIUrl":"10.3390/jimaging11120452","url":null,"abstract":"<p><p>Precise 3D shape correspondence is a fundamental prerequisite for critical applications ranging from medical anatomical modeling to visual recognition. However, non-isometric 3D shape matching remains a challenging task due to the limited sensitivity of traditional Laplace-Beltrami (LB) bases to local geometric deformations such as stretching and bending. To address these limitations, this paper proposes a Sinkhorn-Regularized Elastic Functional Map framework (SRE-FMaps) that integrates entropy-regularized optimal transport with an elastic thin-shell energy basis. First, a sparse Sinkhorn transport plan is adopted to initialize a bijective correspondence with linear computational complexity. Then, a non-orthogonal elastic basis, derived from the Hessian of thin-shell deformation energy, is introduced to enhance high-frequency feature perception. Finally, correspondence stability is quantified through a cosine-based elastic distance metric, enabling retrieval and classification. Experiments on the SHREC2015, McGill, and Face datasets demonstrate that SRE-FMaps reduces the correspondence error by a maximum of 32% and achieves an average of 92.3% classification accuracy (with a peak of 94.74% on the Face dataset). Moreover, the framework exhibits superior robustness, yielding a recall of up to 91.67% and an F1-score of 0.94, effectively handling bending, stretching, and folding deformations compared with conventional LB-based functional map pipelines. The proposed framework provides a scalable solution for non-isometric shape correspondence in medical modeling, 3D reconstruction, and visual recognition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733647/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821343","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-15DOI: 10.3390/jimaging11120449
Irenel Lopo Da Silva, Nicolas Francisco Lori, José Manuel Ferreira Machado
This paper introduces a novel framework for sensory representation of brain imaging data, combining deep learning-based segmentation with multimodal visual and auditory outputs. Structural magnetic resonance imaging (MRI) predictions are converted into color-coded maps and stereophonic/MIDI sonifications, enabling intuitive interpretation of cortical activation patterns. High-precision U-Net models efficiently generate these outputs, supporting clinical decision-making, cognitive research, and creative applications. Spatial, intensity, and anomalous features are encoded into perceivable visual and auditory cues, facilitating early detection and introducing the concept of "auditory biomarkers" for potential pathological identification. Despite current limitations, including dataset size, absence of clinical validation, and heuristic-based sonification, the pipeline demonstrates technical feasibility and robustness. Future work will focus on clinical user studies, the application of functional MRI (fMRI) time-series for dynamic sonification, and the integration of real-time emotional feedback in cinematic contexts. This multisensory approach offers a promising avenue for enhancing the interpretability of complex neuroimaging data across medical, research, and artistic domains.
{"title":"Sensory Representation of Neural Networks Using Sound and Color for Medical Imaging Segmentation.","authors":"Irenel Lopo Da Silva, Nicolas Francisco Lori, José Manuel Ferreira Machado","doi":"10.3390/jimaging11120449","DOIUrl":"10.3390/jimaging11120449","url":null,"abstract":"<p><p>This paper introduces a novel framework for sensory representation of brain imaging data, combining deep learning-based segmentation with multimodal visual and auditory outputs. Structural magnetic resonance imaging (MRI) predictions are converted into color-coded maps and stereophonic/MIDI sonifications, enabling intuitive interpretation of cortical activation patterns. High-precision U-Net models efficiently generate these outputs, supporting clinical decision-making, cognitive research, and creative applications. Spatial, intensity, and anomalous features are encoded into perceivable visual and auditory cues, facilitating early detection and introducing the concept of \"auditory biomarkers\" for potential pathological identification. Despite current limitations, including dataset size, absence of clinical validation, and heuristic-based sonification, the pipeline demonstrates technical feasibility and robustness. Future work will focus on clinical user studies, the application of functional MRI (fMRI) time-series for dynamic sonification, and the integration of real-time emotional feedback in cinematic contexts. This multisensory approach offers a promising avenue for enhancing the interpretability of complex neuroimaging data across medical, research, and artistic domains.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733452/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821305","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-15DOI: 10.3390/jimaging11120451
Pedro Sérgio Tôrres Figueiredo Silva, Antonio Mauricio Ferreira Leite Miranda de Sá, Wagner Coelho de Albuquerque Pereira, Leonardo Bonato Felix, José Manoel de Seixas
COVID-19 screening is crucial for the early diagnosis and treatment of the disease, with lung ultrasound posing as a cost-effective alternative to other imaging techniques. Given the dependency on medical expertise and experience to accurately identify patterns in ultrasound exams, deep learning techniques have been explored for automatically classifying patients' conditions. However, the limited availability of public medical databases remains a significant obstacle to the development of more advanced models. To address the data scarcity problem, this study proposes a method that leverages generative adversarial networks (GANs) to generate synthetic lung ultrasound images, which are subsequently used to train frame-based classification models. Two types of GANs are considered: Wasserstein GANs (WGAN) and Pix2Pix. Specific tools are used to show that the synthetic data produced present a distribution close to the original data. The classification models trained with synthetic data achieved a peak accuracy of 96.32% ± 4.17%, significantly outperforming the maximum accuracy of 82.69% ± 10.42% obtained when training only with the original data. Furthermore, the best results are comparable to, and in some cases surpass, those reported in recent related studies.
COVID-19筛查对于疾病的早期诊断和治疗至关重要,肺部超声是其他成像技术的一种具有成本效益的替代方案。考虑到在超声检查中准确识别模式依赖于医学专业知识和经验,人们已经探索了深度学习技术来自动分类患者的病情。然而,公共医疗数据库的有限可用性仍然是开发更先进模型的重大障碍。为了解决数据稀缺问题,本研究提出了一种利用生成对抗网络(GANs)生成合成肺部超声图像的方法,该方法随后用于训练基于框架的分类模型。考虑了两种类型的gan: Wasserstein gan (WGAN)和Pix2Pix。使用特定的工具来显示生成的合成数据呈现接近原始数据的分布。使用合成数据训练的分类模型的最高准确率为96.32%±4.17%,明显优于仅使用原始数据训练时的最高准确率82.69%±10.42%。此外,最好的结果与最近的相关研究报告相当,在某些情况下甚至超过了这些研究报告。
{"title":"Application of Generative Adversarial Networks to Improve COVID-19 Classification on Ultrasound Images.","authors":"Pedro Sérgio Tôrres Figueiredo Silva, Antonio Mauricio Ferreira Leite Miranda de Sá, Wagner Coelho de Albuquerque Pereira, Leonardo Bonato Felix, José Manoel de Seixas","doi":"10.3390/jimaging11120451","DOIUrl":"10.3390/jimaging11120451","url":null,"abstract":"<p><p>COVID-19 screening is crucial for the early diagnosis and treatment of the disease, with lung ultrasound posing as a cost-effective alternative to other imaging techniques. Given the dependency on medical expertise and experience to accurately identify patterns in ultrasound exams, deep learning techniques have been explored for automatically classifying patients' conditions. However, the limited availability of public medical databases remains a significant obstacle to the development of more advanced models. To address the data scarcity problem, this study proposes a method that leverages generative adversarial networks (GANs) to generate synthetic lung ultrasound images, which are subsequently used to train frame-based classification models. Two types of GANs are considered: Wasserstein GANs (WGAN) and Pix2Pix. Specific tools are used to show that the synthetic data produced present a distribution close to the original data. The classification models trained with synthetic data achieved a peak accuracy of 96.32% ± 4.17%, significantly outperforming the maximum accuracy of 82.69% ± 10.42% obtained when training only with the original data. Furthermore, the best results are comparable to, and in some cases surpass, those reported in recent related studies.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733862/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821053","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-15DOI: 10.3390/jimaging11120450
Adel Shahnam, Nicholas Hardcastle, David E Gyorki, Katrina M Ingley, Krystel Tran, Catherine Mitchell, Sarat Chander, Julie Chu, Michael Henderson, Alan Herschtal, Mathias Bressel, Jeremy Lewin
Retroperitoneal sarcomas (RPS) are rare tumours, primarily treated with surgical resection. However, recurrences are frequent. Combining clinical factors with CT-derived radiomic features could enhance treatment stratification and personalization. This study aims to assess whether radiomic features provide additional prognostic value beyond clinicopathological features in patients with high-risk RPS treated with preoperative radiotherapy. This retrospective study included patients aged 18 or older with non-recurrent and non-metastatic RPS treated with preoperative radiotherapy between 2008 and 2016. Hazard ratios (HR) were calculated using Cox proportional hazards regression to assess the impact of clinical and radiomic features on time to event outcomes. Predictive accuracy was assessed with c-statistics. Radiomic analysis was performed on the high-risk group (undifferentiated pleomorphic sarcoma, well-differentiated/de-differentiated liposarcoma or grade 2/3 leiomyosarcoma). Seventy-two patients were included, with a median follow-up of 3.7 years, the 5-year overall survival (OS) was 67%. Multivariable analysis showed older age (HR: 1.3 per 5-year increase, p = 0.04), grade 3 (HR: 180.3, p = 0.02), and larger tumours (HR: 4.0 per 10 cm increase, p = 0.02) predicted worse OS. In the higher-risk group, the c-statistic for the clinical model was 0.59 (time to distant metastasis (TDM)) and 0.56 (OS). Among 27 radiomic features, kurtosis improved OS prediction (c-statistic 0.69, p = 0.013), and Neighbourhood Gray-Tone Difference Matrix (NGTDM) busyness improved it to 0.73 (p = 0.036). Kurtosis also improved TDM prediction (c-statistic 0.72, p = 0.023). Radiomic features may complement clinicopathological factors in predicting overall survival and time to distant metastasis in high-risk retroperitoneal sarcoma. These exploratory findings warrant validation in larger, multi-institutional studies.
{"title":"Applying Radiomics to Predict Outcomes in Patients with High-Grade Retroperitoneal Sarcoma Treated with Preoperative Radiotherapy.","authors":"Adel Shahnam, Nicholas Hardcastle, David E Gyorki, Katrina M Ingley, Krystel Tran, Catherine Mitchell, Sarat Chander, Julie Chu, Michael Henderson, Alan Herschtal, Mathias Bressel, Jeremy Lewin","doi":"10.3390/jimaging11120450","DOIUrl":"10.3390/jimaging11120450","url":null,"abstract":"<p><p>Retroperitoneal sarcomas (RPS) are rare tumours, primarily treated with surgical resection. However, recurrences are frequent. Combining clinical factors with CT-derived radiomic features could enhance treatment stratification and personalization. This study aims to assess whether radiomic features provide additional prognostic value beyond clinicopathological features in patients with high-risk RPS treated with preoperative radiotherapy. This retrospective study included patients aged 18 or older with non-recurrent and non-metastatic RPS treated with preoperative radiotherapy between 2008 and 2016. Hazard ratios (HR) were calculated using Cox proportional hazards regression to assess the impact of clinical and radiomic features on time to event outcomes. Predictive accuracy was assessed with c-statistics. Radiomic analysis was performed on the high-risk group (undifferentiated pleomorphic sarcoma, well-differentiated/de-differentiated liposarcoma or grade 2/3 leiomyosarcoma). Seventy-two patients were included, with a median follow-up of 3.7 years, the 5-year overall survival (OS) was 67%. Multivariable analysis showed older age (HR: 1.3 per 5-year increase, <i>p</i> = 0.04), grade 3 (HR: 180.3, <i>p</i> = 0.02), and larger tumours (HR: 4.0 per 10 cm increase, <i>p</i> = 0.02) predicted worse OS. In the higher-risk group, the c-statistic for the clinical model was 0.59 (time to distant metastasis (TDM)) and 0.56 (OS). Among 27 radiomic features, kurtosis improved OS prediction (c-statistic 0.69, <i>p</i> = 0.013), and Neighbourhood Gray-Tone Difference Matrix (NGTDM) busyness improved it to 0.73 (<i>p</i> = 0.036). Kurtosis also improved TDM prediction (c-statistic 0.72, <i>p</i> = 0.023). Radiomic features may complement clinicopathological factors in predicting overall survival and time to distant metastasis in high-risk retroperitoneal sarcoma. These exploratory findings warrant validation in larger, multi-institutional studies.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733741/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821002","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-12-14DOI: 10.3390/jimaging11120448
Helala AlShehri
Lung and colon cancers remain among the leading causes of cancer-related mortality worldwide, underscoring the need for rapid and accurate histopathological diagnosis. Manual examination of biopsy slides is often time-consuming and prone to inter-observer variability, which highlights the importance of developing reliable and explainable automated diagnostic systems. This study presents DPCSE-Net, a lightweight dual-path convolutional neural network enhanced with a squeeze-and-excitation (SE) attention mechanism for lung and colon cancer classification. The dual-path structure captures both fine-grained cellular textures and global contextual information through multiscale feature extraction, while the SE attention module adaptively recalibrates channel responses to emphasize discriminative features. To enhance transparency and interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM), attention heatmaps, and Integrated Gradients are employed to visualize class-specific activation patterns and verify that the model's focus aligns with diagnostically relevant tissue regions. Evaluated on the publicly available LC25000 dataset, DPCSE-Net achieved state-of-the-art performance with 99.88% accuracy and F1-score, while maintaining low computational complexity. Ablation experiments confirmed the contribution of the dual-path design and SE module, and qualitative analyses demonstrated the model's strong interpretability. These results establish DPCSE-Net as an accurate, efficient, and explainable framework for computer-aided histopathological diagnosis, supporting the broader goals of explainable AI in computer vision.
{"title":"Dual-Path Convolutional Neural Network with Squeeze-and-Excitation Attention for Lung and Colon Histopathology Classification.","authors":"Helala AlShehri","doi":"10.3390/jimaging11120448","DOIUrl":"10.3390/jimaging11120448","url":null,"abstract":"<p><p>Lung and colon cancers remain among the leading causes of cancer-related mortality worldwide, underscoring the need for rapid and accurate histopathological diagnosis. Manual examination of biopsy slides is often time-consuming and prone to inter-observer variability, which highlights the importance of developing reliable and explainable automated diagnostic systems. This study presents DPCSE-Net, a lightweight dual-path convolutional neural network enhanced with a squeeze-and-excitation (SE) attention mechanism for lung and colon cancer classification. The dual-path structure captures both fine-grained cellular textures and global contextual information through multiscale feature extraction, while the SE attention module adaptively recalibrates channel responses to emphasize discriminative features. To enhance transparency and interpretability, Gradient-weighted Class Activation Mapping (Grad-CAM), attention heatmaps, and Integrated Gradients are employed to visualize class-specific activation patterns and verify that the model's focus aligns with diagnostically relevant tissue regions. Evaluated on the publicly available LC25000 dataset, DPCSE-Net achieved state-of-the-art performance with 99.88% accuracy and F1-score, while maintaining low computational complexity. Ablation experiments confirmed the contribution of the dual-path design and SE module, and qualitative analyses demonstrated the model's strong interpretability. These results establish DPCSE-Net as an accurate, efficient, and explainable framework for computer-aided histopathological diagnosis, supporting the broader goals of explainable AI in computer vision.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733646/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}