首页 > 最新文献

Journal of Imaging最新文献

英文 中文
LLM-Based Pose Normalization and Multimodal Fusion for Facial Expression Recognition in Extreme Poses. 基于llm的姿态归一化和多模态融合的极端姿态面部表情识别。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-04 DOI: 10.3390/jimaging12010024
Bohan Chen, Bowen Qu, Yu Zhou, Han Huang, Jianing Guo, Yanning Xian, Longxiang Ma, Jinxuan Yu, Jingyu Chen

Facial expression recognition (FER) technology has progressively matured over time. However, existing FER methods are primarily optimized for frontal face images, and their recognition accuracy significantly degrades when processing profile or large-angle rotated facial images. Consequently, this limitation hinders the practical deployment of FER systems. To mitigate the interference caused by large pose variations and improve recognition accuracy, we propose a FER method based on profile-to-frontal transformation and multimodal learning. Specifically, we first leverage the visual understanding and generation capabilities of Qwen-Image-Edit that transform profile images to frontal viewpoints, preserving key expression features while standardizing facial poses. Second, we introduce the CLIP model to enhance the semantic representation capability of expression features through vision-language joint learning. The qualitative and quantitative experiments on the RAF (89.39%), EXPW (67.17%), and AffectNet-7 (62.66%) datasets demonstrate that our method outperforms the existing approaches.

面部表情识别(FER)技术随着时间的推移逐渐成熟。然而,现有的FER方法主要针对正面人脸图像进行优化,在处理轮廓或大角度旋转人脸图像时,识别精度明显下降。因此,这一限制阻碍了FER系统的实际部署。为了减轻姿态变化带来的干扰,提高识别精度,提出了一种基于面-面变换和多模态学习的人脸识别方法。具体来说,我们首先利用Qwen-Image-Edit的视觉理解和生成能力,将侧面图像转换为正面视点,在标准化面部姿势的同时保留关键的表情特征。其次,引入CLIP模型,通过视觉语言联合学习增强表达特征的语义表示能力。在RAF(89.39%)、EXPW(67.17%)和AffectNet-7(62.66%)数据集上进行的定性和定量实验表明,本文方法优于现有方法。
{"title":"LLM-Based Pose Normalization and Multimodal Fusion for Facial Expression Recognition in Extreme Poses.","authors":"Bohan Chen, Bowen Qu, Yu Zhou, Han Huang, Jianing Guo, Yanning Xian, Longxiang Ma, Jinxuan Yu, Jingyu Chen","doi":"10.3390/jimaging12010024","DOIUrl":"10.3390/jimaging12010024","url":null,"abstract":"<p><p>Facial expression recognition (FER) technology has progressively matured over time. However, existing FER methods are primarily optimized for frontal face images, and their recognition accuracy significantly degrades when processing profile or large-angle rotated facial images. Consequently, this limitation hinders the practical deployment of FER systems. To mitigate the interference caused by large pose variations and improve recognition accuracy, we propose a FER method based on profile-to-frontal transformation and multimodal learning. Specifically, we first leverage the visual understanding and generation capabilities of Qwen-Image-Edit that transform profile images to frontal viewpoints, preserving key expression features while standardizing facial poses. Second, we introduce the CLIP model to enhance the semantic representation capability of expression features through vision-language joint learning. The qualitative and quantitative experiments on the RAF (89.39%), EXPW (67.17%), and AffectNet-7 (62.66%) datasets demonstrate that our method outperforms the existing approaches.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842521/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054251","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Comparative Evaluation of Vision-Language Models for Detecting and Localizing Dental Lesions from Intraoral Images. 从口腔内图像检测和定位牙齿病变的视觉语言模型的比较评价。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-03 DOI: 10.3390/jimaging12010022
Maria Jahan, Al Ibne Siam, Lamim Zakir Pronay, Saif Ahmed, Nabeel Mohammed, James Dudley, Taseef Hasan Farook

To assess the efficiency of vision-language models in detecting and classifying carious and non-carious lesions from intraoral photo imaging. A dataset of 172 annotated images were classified for microcavitation, cavitated lesions, staining, calculus, and non-carious lesions. Florence-2, PaLI-Gemma, and YOLOv8 models were trained on the dataset and model performance. The dataset was divided into 80:10:10 split, and the model performance was evaluated using mean average precision (mAP), mAP50-95, class-specific precision and recall. YOLOv8 outperformed the vision-language models, achieving a mean average precision (mAP) of 37% with a precision of 42.3% (with 100% for cavitation detection) and 31.3% recall. PaLI-Gemma produced a recall of 13% and 21%. Florence-2 yielded a mean average precision of 10% with a precision and recall was 51% and 35%. YOLOv8 achieved the strongest overall performance. Florence-2 and PaLI-Gemma models underperformed relative to YOLOv8 despite the potential for multimodal contextual understanding, highlighting the need for larger, more diverse datasets and hybrid architectures to achieve improved performance.

目的评估视觉语言模型在口腔内照片中检测和分类龋齿和非龋齿病变的效率。172张带注释的图像数据集被分类为微空化、空化病变、染色、结石和非龋齿病变。在数据集和模型性能上训练了Florence-2、pal - gemma和YOLOv8模型。将数据集分为80:10:10分割,并使用平均平均精度(mAP)、mAP50-95、类别特定精度和召回率对模型性能进行评估。YOLOv8优于视觉语言模型,平均精度(mAP)为37%,精度为42.3%(空化检测为100%),召回率为31.3%。PaLI-Gemma的召回率分别为13%和21%。Florence-2的平均准确率为10%,准确率和召回率分别为51%和35%。YOLOv8取得了最强的整体性能。尽管具有多模态上下文理解的潜力,但Florence-2和pal - gemma模型相对于YOLOv8表现不佳,这突出了对更大、更多样化的数据集和混合架构的需求,以实现更高的性能。
{"title":"Comparative Evaluation of Vision-Language Models for Detecting and Localizing Dental Lesions from Intraoral Images.","authors":"Maria Jahan, Al Ibne Siam, Lamim Zakir Pronay, Saif Ahmed, Nabeel Mohammed, James Dudley, Taseef Hasan Farook","doi":"10.3390/jimaging12010022","DOIUrl":"10.3390/jimaging12010022","url":null,"abstract":"<p><p>To assess the efficiency of vision-language models in detecting and classifying carious and non-carious lesions from intraoral photo imaging. A dataset of 172 annotated images were classified for microcavitation, cavitated lesions, staining, calculus, and non-carious lesions. Florence-2, PaLI-Gemma, and YOLOv8 models were trained on the dataset and model performance. The dataset was divided into 80:10:10 split, and the model performance was evaluated using mean average precision (mAP), mAP50-95, class-specific precision and recall. YOLOv8 outperformed the vision-language models, achieving a mean average precision (mAP) of 37% with a precision of 42.3% (with 100% for cavitation detection) and 31.3% recall. PaLI-Gemma produced a recall of 13% and 21%. Florence-2 yielded a mean average precision of 10% with a precision and recall was 51% and 35%. YOLOv8 achieved the strongest overall performance. Florence-2 and PaLI-Gemma models underperformed relative to YOLOv8 despite the potential for multimodal contextual understanding, highlighting the need for larger, more diverse datasets and hybrid architectures to achieve improved performance.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842643/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054239","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multi-Temporal Shoreline Monitoring and Analysis in Bangkok Bay, Thailand, Using Remote Sensing and GIS Techniques. 基于遥感和GIS技术的泰国曼谷湾岸线多时相监测与分析
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-01 DOI: 10.3390/jimaging12010021
Yan Wang, Adisorn Sirikham, Jessada Konpang, Chunguang Li

Drastic alterations have been observed in the coastline of Bangkok Bay, Thailand, over the past three decades. Understanding how coastlines change plays a key role in developing strategies for coastal protection and sustainable resource utilization. This study investigates the temporal and spatial changes in the Bangkok Bay coastline, Thailand, using remote sensing and GIS techniques from 1989 to 2024. The historical rate of coastline change for a typical segment was analyzed using the EPR method, and the underlying causes of these changes were discussed. Finally, the variation trend of the total shoreline length and the characteristics of erosion and sedimentation for a typical shoreline in Bangkok Bay, Thailand, over the past 35 years were obtained. An overall increase in coastline length was observed in Bangkok Bay, Thailand, over the 35-year period from 1989 to 2024, with a net gain from 507.23 km to 571.38 km. The rate of growth has transitioned from rapid to slow, with the most significant changes occurring during the period 1989-1994. Additionally, the average and maximum erosion rates for the typical shoreline segment were notably high during 1989-1994, with values of -21.61 m/a and -55.49 m/a, respectively. The maximum sedimentation rate along the coastline was relatively high from 2014 to 2024, reaching 10.57 m/a. Overall, the entire coastline of the Samut Sakhon-Bangkok-Samut Prakan Provinces underwent net erosion from 1989 to 2024, driven by a confluence of natural and anthropogenic factors.

在过去的三十年里,泰国曼谷湾的海岸线发生了巨大的变化。了解海岸线如何变化在制定海岸保护和可持续资源利用战略方面起着关键作用。利用遥感和GIS技术,对1989 - 2024年泰国曼谷湾海岸线的时空变化进行了研究。利用EPR方法分析了典型段海岸线变化的历史速率,并探讨了这些变化的根本原因。最后,分析了泰国曼谷湾典型岸线近35年的总岸线长度变化趋势和侵蚀沉积特征。1989年至2024年的35年间,泰国曼谷湾的海岸线长度总体增加,净增长从507.23公里增加到571.38公里。增长率已由快速转为缓慢,最显著的变化发生在1989-1994年期间。典型岸线段的平均侵蚀速率和最大侵蚀速率在1989—1994年显著较高,分别为-21.61 m/a和-55.49 m/a。2014 - 2024年,岸线最大沉积速率较高,达10.57 m/a。总体而言,1989年至2024年,在自然和人为因素的共同作用下,Samut - sakhong - bangkok -Samut Prakan三省的整个海岸线经历了净侵蚀。
{"title":"Multi-Temporal Shoreline Monitoring and Analysis in Bangkok Bay, Thailand, Using Remote Sensing and GIS Techniques.","authors":"Yan Wang, Adisorn Sirikham, Jessada Konpang, Chunguang Li","doi":"10.3390/jimaging12010021","DOIUrl":"10.3390/jimaging12010021","url":null,"abstract":"<p><p>Drastic alterations have been observed in the coastline of Bangkok Bay, Thailand, over the past three decades. Understanding how coastlines change plays a key role in developing strategies for coastal protection and sustainable resource utilization. This study investigates the temporal and spatial changes in the Bangkok Bay coastline, Thailand, using remote sensing and GIS techniques from 1989 to 2024. The historical rate of coastline change for a typical segment was analyzed using the EPR method, and the underlying causes of these changes were discussed. Finally, the variation trend of the total shoreline length and the characteristics of erosion and sedimentation for a typical shoreline in Bangkok Bay, Thailand, over the past 35 years were obtained. An overall increase in coastline length was observed in Bangkok Bay, Thailand, over the 35-year period from 1989 to 2024, with a net gain from 507.23 km to 571.38 km. The rate of growth has transitioned from rapid to slow, with the most significant changes occurring during the period 1989-1994. Additionally, the average and maximum erosion rates for the typical shoreline segment were notably high during 1989-1994, with values of -21.61 m/a and -55.49 m/a, respectively. The maximum sedimentation rate along the coastline was relatively high from 2014 to 2024, reaching 10.57 m/a. Overall, the entire coastline of the Samut Sakhon-Bangkok-Samut Prakan Provinces underwent net erosion from 1989 to 2024, driven by a confluence of natural and anthropogenic factors.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842628/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054176","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Object Detection on Road: Vehicle's Detection Based on Re-Training Models on NVIDIA-Jetson Platform. 道路上的目标检测:基于NVIDIA-Jetson平台上再训练模型的车辆检测。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-01 DOI: 10.3390/jimaging12010020
Sleiter Ramos-Sanchez, Jinmi Lezama, Ricardo Yauri, Joyce Zevallos

The increasing use of artificial intelligence (AI) and deep learning (DL) techniques has driven advances in vehicle classification and detection applications for embedded devices with deployment constraints due to computational cost and response time. In the case of urban environments with high traffic congestion, such as the city of Lima, it is important to determine the trade-off between model accuracy, type of embedded system, and the dataset used. This study was developed using a methodology adapted from the CRISP-DM approach, which included the acquisition of traffic videos in the city of Lima, their segmentation, and manual labeling. Subsequently, three SSD-based detection models (MobileNetV1-SSD, MobileNetV2-SSD-Lite, and VGG16-SSD) were trained on the NVIDIA Jetson Orin NX 16 GB platform. The results show that the VGG16-SSD model achieved the highest average precision (mAP ≈90.7%), with a longer training time, while the MobileNetV1-SSD (512×512) model achieved comparable performance (mAP ≈90.4%) with a shorter time. Additionally, data augmentation through contrast adjustment improved the detection of minority classes such as Tuk-tuk and Motorcycle. The results indicate that, among the evaluated models, MobileNetV1-SSD (512×512) achieved the best balance between accuracy and computational load for its implementation in ADAS embedded systems in congested urban environments.

人工智能(AI)和深度学习(DL)技术的使用越来越多,推动了嵌入式设备在车辆分类和检测应用方面的进步,这些设备由于计算成本和响应时间而受到部署限制。在交通拥堵严重的城市环境中,比如利马市,确定模型精度、嵌入式系统类型和使用的数据集之间的权衡是很重要的。本研究采用了一种改编自CRISP-DM方法的方法,其中包括获取利马市的交通视频,对其进行分割和手动标记。随后,在NVIDIA Jetson Orin NX 16gb平台上训练了三个基于ssd的检测模型(MobileNetV1-SSD、MobileNetV2-SSD-Lite和VGG16-SSD)。结果表明,VGG16-SSD模型的平均精度最高(mAP≈90.7%),训练时间较长,而MobileNetV1-SSD (512×512)模型在较短的训练时间内取得了相当的性能(mAP≈90.4%)。此外,通过对比度调整的数据增强提高了Tuk-tuk和Motorcycle等少数类别的检测。结果表明,在所评估的模型中,MobileNetV1-SSD (512×512)在城市拥挤环境下的ADAS嵌入式系统中实现了精度和计算负荷之间的最佳平衡。
{"title":"Object Detection on Road: Vehicle's Detection Based on Re-Training Models on NVIDIA-Jetson Platform.","authors":"Sleiter Ramos-Sanchez, Jinmi Lezama, Ricardo Yauri, Joyce Zevallos","doi":"10.3390/jimaging12010020","DOIUrl":"10.3390/jimaging12010020","url":null,"abstract":"<p><p>The increasing use of artificial intelligence (AI) and deep learning (DL) techniques has driven advances in vehicle classification and detection applications for embedded devices with deployment constraints due to computational cost and response time. In the case of urban environments with high traffic congestion, such as the city of Lima, it is important to determine the trade-off between model accuracy, type of embedded system, and the dataset used. This study was developed using a methodology adapted from the CRISP-DM approach, which included the acquisition of traffic videos in the city of Lima, their segmentation, and manual labeling. Subsequently, three SSD-based detection models (MobileNetV1-SSD, MobileNetV2-SSD-Lite, and VGG16-SSD) were trained on the NVIDIA Jetson Orin NX 16 GB platform. The results show that the VGG16-SSD model achieved the highest average precision (mAP ≈90.7%), with a longer training time, while the MobileNetV1-SSD (512×512) model achieved comparable performance (mAP ≈90.4%) with a shorter time. Additionally, data augmentation through contrast adjustment improved the detection of minority classes such as Tuk-tuk and Motorcycle. The results indicate that, among the evaluated models, MobileNetV1-SSD (512×512) achieved the best balance between accuracy and computational load for its implementation in ADAS embedded systems in congested urban environments.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842717/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Double-Gated Mamba Multi-Scale Adaptive Feature Learning Network for Unsupervised Single RGB Image Hyperspectral Image Reconstruction. 基于双门门曼巴多尺度自适应特征学习网络的无监督单幅RGB高光谱图像重建。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-31 DOI: 10.3390/jimaging12010019
Zhongmin Jiang, Zhen Wang, Wenju Wang, Jifan Zhu

Existing methods for reconstructing hyperspectral images from single RGB images struggle to obtain a large number of labeled RGB-HSI paired images. These methods face issues such as detail loss, insufficient robustness, low reconstruction accuracy, and the difficulty of balancing the spatial-spectral trade-off. To address these challenges, a Double-Gated Mamba Multi-Scale Adaptive Feature (DMMAF) learning network model is proposed. DMMAF designs a reflection dot-product adaptive dual-noise-aware feature extraction method, which is used to supplement edge detail information in spectral images and improve robustness. DMMAF also constructs a deformable attention-based global feature extraction method and a double-gated Mamba local feature extraction approach, enhancing the interaction between local and global information during the reconstruction process, thereby improving image accuracy. Meanwhile, DMMAF introduces a structure-aware smooth loss function, which, by combining smoothing, curvature, and attention supervision losses, effectively resolves the spatial-spectral resolution balance problem. This network model is applied to three datasets-NTIRE 2020, Harvard, and CAVE-achieving state-of-the-art unsupervised reconstruction performance compared to existing advanced algorithms. Experiments on the NTIRE 2020, Harvard, and CAVE datasets demonstrate that this model achieves state-of-the-art unsupervised reconstruction performance. On the NTIRE 2020 dataset, our method attains MRAE, RMSE, and PSNR values of 0.133, 0.040, and 31.314, respectively. On the Harvard dataset, it achieves RMSE and PSNR values of 0.025 and 34.955, respectively, while on the CAVE dataset, it achieves RMSE and PSNR values of 0.041 and 30.983, respectively.

现有的从单个RGB图像重建高光谱图像的方法难以获得大量标记的RGB- hsi配对图像。这些方法面临着细节丢失、鲁棒性不足、重建精度低以及难以平衡空间-频谱权衡等问题。为了解决这些问题,提出了一种双门门曼巴多尺度自适应特征(DMMAF)学习网络模型。DMMAF设计了一种反射点积自适应双噪声感知特征提取方法,用于补充光谱图像中的边缘细节信息,提高鲁棒性。DMMAF还构建了一种基于可变形注意力的全局特征提取方法和一种双门门的Mamba局部特征提取方法,增强了重建过程中局部信息与全局信息的交互作用,从而提高了图像精度。同时,DMMAF引入了结构感知平滑损失函数,该函数结合平滑、曲率和注意监督损失,有效地解决了空间-光谱分辨率平衡问题。该网络模型应用于三个数据集——ntire 2020、Harvard和cave——与现有的先进算法相比,实现了最先进的无监督重建性能。在NTIRE 2020、Harvard和CAVE数据集上的实验表明,该模型实现了最先进的无监督重建性能。在整个2020数据集上,我们的方法获得的MRAE、RMSE和PSNR值分别为0.133、0.040和31.314。在Harvard数据集上,RMSE和PSNR分别为0.025和34.955,而在CAVE数据集上,RMSE和PSNR分别为0.041和30.983。
{"title":"Double-Gated Mamba Multi-Scale Adaptive Feature Learning Network for Unsupervised Single RGB Image Hyperspectral Image Reconstruction.","authors":"Zhongmin Jiang, Zhen Wang, Wenju Wang, Jifan Zhu","doi":"10.3390/jimaging12010019","DOIUrl":"10.3390/jimaging12010019","url":null,"abstract":"<p><p>Existing methods for reconstructing hyperspectral images from single RGB images struggle to obtain a large number of labeled RGB-HSI paired images. These methods face issues such as detail loss, insufficient robustness, low reconstruction accuracy, and the difficulty of balancing the spatial-spectral trade-off. To address these challenges, a Double-Gated Mamba Multi-Scale Adaptive Feature (DMMAF) learning network model is proposed. DMMAF designs a reflection dot-product adaptive dual-noise-aware feature extraction method, which is used to supplement edge detail information in spectral images and improve robustness. DMMAF also constructs a deformable attention-based global feature extraction method and a double-gated Mamba local feature extraction approach, enhancing the interaction between local and global information during the reconstruction process, thereby improving image accuracy. Meanwhile, DMMAF introduces a structure-aware smooth loss function, which, by combining smoothing, curvature, and attention supervision losses, effectively resolves the spatial-spectral resolution balance problem. This network model is applied to three datasets-NTIRE 2020, Harvard, and CAVE-achieving state-of-the-art unsupervised reconstruction performance compared to existing advanced algorithms. Experiments on the NTIRE 2020, Harvard, and CAVE datasets demonstrate that this model achieves state-of-the-art unsupervised reconstruction performance. On the NTIRE 2020 dataset, our method attains MRAE, RMSE, and PSNR values of 0.133, 0.040, and 31.314, respectively. On the Harvard dataset, it achieves RMSE and PSNR values of 0.025 and 34.955, respectively, while on the CAVE dataset, it achieves RMSE and PSNR values of 0.041 and 30.983, respectively.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843217/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054122","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Advancing Medical Decision-Making with AI: A Comprehensive Exploration of the Evolution from Convolutional Neural Networks to Capsule Networks. 用人工智能推进医疗决策:从卷积神经网络到胶囊网络进化的全面探索。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-30 DOI: 10.3390/jimaging12010017
Ichrak Khoulqi, Zakariae El Ouazzani

In this paper, we propose a literature review regarding two deep learning architectures, namely Convolutional Neural Networks (CNNs) and Capsule Networks (CapsNets), applied to medical images, in order to analyze them to help in medical decision support. CNNs demonstrate their capacity in the medical diagnostic field; however, their reliability decreases when there is slight spatial variability, which can affect diagnosis, especially since the anatomical structure of the human body can differ from one patient to another. In contrast, CapsNets encode not only feature activation but also spatial relationships, hence improving the reliability and stability of model generalization. This paper proposes a structured comparison by reviewing studies published from 2018 to 2025 across major databases, including IEEE Xplore, ScienceDirect, SpringerLink, and MDPI. The applications in the reviewed papers are based on the benchmark datasets BraTS, INbreast, ISIC, and COVIDx. This paper review compares the core architectural principles, performance, and interpretability of both architectures. To conclude the paper, we underline the complementary roles of these two architectures in medical decision-making and propose future directions toward hybrid, explainable, and computationally efficient deep learning systems for real clinical environments, thereby increasing survival rates by helping prevent diseases at an early stage.

在本文中,我们提出了两种深度学习架构,即卷积神经网络(cnn)和胶囊网络(CapsNets),应用于医学图像的文献综述,以分析它们以帮助医疗决策支持。cnn展示了他们在医疗诊断领域的能力;然而,当存在轻微的空间变异性时,它们的可靠性会降低,这会影响诊断,特别是因为人体的解剖结构可能因患者而异。相比之下,capnet不仅对特征激活进行编码,还对空间关系进行编码,从而提高了模型泛化的可靠性和稳定性。本文通过回顾2018年至2025年在主要数据库(包括IEEE explore、ScienceDirect、SpringerLink和MDPI)上发表的研究,提出了一个结构化的比较。所审查论文中的应用基于基准数据集BraTS、INbreast、ISIC和covid。本文比较了这两种体系结构的核心体系结构原则、性能和可解释性。最后,我们强调了这两种架构在医疗决策中的互补作用,并提出了面向真实临床环境的混合、可解释和计算效率高的深度学习系统的未来方向,从而通过帮助在早期阶段预防疾病来提高生存率。
{"title":"Advancing Medical Decision-Making with AI: A Comprehensive Exploration of the Evolution from Convolutional Neural Networks to Capsule Networks.","authors":"Ichrak Khoulqi, Zakariae El Ouazzani","doi":"10.3390/jimaging12010017","DOIUrl":"10.3390/jimaging12010017","url":null,"abstract":"<p><p>In this paper, we propose a literature review regarding two deep learning architectures, namely Convolutional Neural Networks (CNNs) and Capsule Networks (CapsNets), applied to medical images, in order to analyze them to help in medical decision support. CNNs demonstrate their capacity in the medical diagnostic field; however, their reliability decreases when there is slight spatial variability, which can affect diagnosis, especially since the anatomical structure of the human body can differ from one patient to another. In contrast, CapsNets encode not only feature activation but also spatial relationships, hence improving the reliability and stability of model generalization. This paper proposes a structured comparison by reviewing studies published from 2018 to 2025 across major databases, including IEEE Xplore, ScienceDirect, SpringerLink, and MDPI. The applications in the reviewed papers are based on the benchmark datasets BraTS, INbreast, ISIC, and COVIDx. This paper review compares the core architectural principles, performance, and interpretability of both architectures. To conclude the paper, we underline the complementary roles of these two architectures in medical decision-making and propose future directions toward hybrid, explainable, and computationally efficient deep learning systems for real clinical environments, thereby increasing survival rates by helping prevent diseases at an early stage.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842556/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Revisiting Underwater Image Enhancement for Object Detection: A Unified Quality-Detection Evaluation Framework. 重述水下图像增强用于目标检测:一个统一的质量检测评估框架。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-30 DOI: 10.3390/jimaging12010018
Ali Awad, Ashraf Saleem, Sidike Paheding, Evan Lucas, Serein Al-Ratrout, Timothy C Havens

Underwater images often suffer from severe color distortion, low contrast, and reduced visibility, motivating the widespread use of image enhancement as a preprocessing step for downstream computer vision tasks. However, recent studies have questioned whether enhancement actually improves object detection performance. In this work, we conduct a comprehensive and rigorous evaluation of nine state-of-the-art enhancement methods and their interactions with modern object detectors. We propose a unified evaluation framework that integrates (1) a distribution-level quality assessment using a composite quality index (Q-index), (2) a fine-grained per-image detection protocol based on COCO-style mAP, and (3) a mixed-set upper-bound analysis that quantifies the theoretical performance achievable through ideal selective enhancement. Our findings reveal that traditional image quality metrics do not reliably predict detection performance, and that dataset-level conclusions often overlook substantial image-level variability. Through per-image evaluation, we identify numerous cases in which enhancement significantly improves detection accuracy-primarily for low-quality inputs-while also demonstrating conditions under which enhancement degrades performance. The mixed-set analysis shows that selective enhancement can yield substantial gains over both original and fully enhanced datasets, establishing a new direction for designing enhancement models optimized for downstream vision tasks. This study provides the most comprehensive evidence to date that underwater image enhancement can be beneficial for object detection when evaluated at the appropriate granularity and guided by informed selection strategies. The data generated and code developed are publicly available.

水下图像经常遭受严重的色彩失真,对比度低,能见度降低,激励图像增强作为下游计算机视觉任务的预处理步骤的广泛使用。然而,最近的研究质疑增强是否真的提高了目标检测性能。在这项工作中,我们对九种最先进的增强方法及其与现代目标探测器的相互作用进行了全面而严格的评估。我们提出了一个统一的评估框架,该框架集成了(1)使用复合质量指数(Q-index)的分布级质量评估,(2)基于coco风格mAP的细粒度每幅图像检测协议,以及(3)混合集上限分析,该分析量化了通过理想选择性增强可以实现的理论性能。我们的研究结果表明,传统的图像质量指标不能可靠地预测检测性能,并且数据集级别的结论往往忽略了大量的图像级别的可变性。通过每幅图像评估,我们确定了许多增强显著提高检测精度的情况(主要是针对低质量输入),同时也证明了增强会降低性能的情况。混合集分析表明,选择性增强可以在原始和完全增强的数据集上获得可观的收益,为设计针对下游视觉任务优化的增强模型开辟了新的方向。该研究提供了迄今为止最全面的证据,表明水下图像增强可以在适当的粒度下进行评估,并在知情的选择策略指导下有益于目标检测。生成的数据和开发的代码是公开的。
{"title":"Revisiting Underwater Image Enhancement for Object Detection: A Unified Quality-Detection Evaluation Framework.","authors":"Ali Awad, Ashraf Saleem, Sidike Paheding, Evan Lucas, Serein Al-Ratrout, Timothy C Havens","doi":"10.3390/jimaging12010018","DOIUrl":"10.3390/jimaging12010018","url":null,"abstract":"<p><p>Underwater images often suffer from severe color distortion, low contrast, and reduced visibility, motivating the widespread use of image enhancement as a preprocessing step for downstream computer vision tasks. However, recent studies have questioned whether enhancement actually improves object detection performance. In this work, we conduct a comprehensive and rigorous evaluation of nine state-of-the-art enhancement methods and their interactions with modern object detectors. We propose a unified evaluation framework that integrates (1) a distribution-level quality assessment using a composite quality index (Q-index), (2) a fine-grained per-image detection protocol based on COCO-style mAP, and (3) a mixed-set upper-bound analysis that quantifies the theoretical performance achievable through ideal selective enhancement. Our findings reveal that traditional image quality metrics do not reliably predict detection performance, and that dataset-level conclusions often overlook substantial image-level variability. Through per-image evaluation, we identify numerous cases in which enhancement significantly improves detection accuracy-primarily for low-quality inputs-while also demonstrating conditions under which enhancement degrades performance. The mixed-set analysis shows that selective enhancement can yield substantial gains over both original and fully enhanced datasets, establishing a new direction for designing enhancement models optimized for downstream vision tasks. This study provides the most comprehensive evidence to date that underwater image enhancement can be beneficial for object detection when evaluated at the appropriate granularity and guided by informed selection strategies. The data generated and code developed are publicly available.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842599/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FluoNeRF: Fluorescent Novel-View Synthesis Under Novel Light Source Colors and Spectra. FluoNeRF:在新光源颜色和光谱下的荧光新视野合成。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-29 DOI: 10.3390/jimaging12010016
Lin Shi, Kengo Matsufuji, Michitaka Yoshida, Ryo Kawahara, Takahiro Okabe

Synthesizing photo-realistic images of a scene from arbitrary viewpoints and under arbitrary lighting environments is one of the important research topics in computer vision and graphics. In this paper, we propose a method for synthesizing photo-realistic images of a scene with fluorescent objects from novel viewpoints and under novel lighting colors and spectra. In general, fluorescent materials absorb light with certain wavelengths and then emit light with longer wavelengths than the absorbed ones, in contrast to reflective materials, which preserve wavelengths of light. Therefore, we cannot reproduce the colors of fluorescent objects under arbitrary lighting colors by combining conventional view synthesis techniques with the white balance adjustment of the RGB channels. Accordingly, we extend the novel-view synthesis based on the neural radiance fields by incorporating the superposition principle of light; our proposed method captures a sparse set of images of a scene from varying viewpoints and under varying lighting colors or spectra with active lighting systems such as a color display or a multi-spectral light stage and then synthesizes photo-realistic images of the scene without explicitly modeling its geometric and photometric models. We conducted a number of experiments using real images captured with an LCD and confirmed that our method works better than the existing methods. Moreover, we showed that the extension of our method using more than three primary colors with a light stage enables us to reproduce the colors of fluorescent objects under common light sources.

在任意视点和任意光照环境下合成逼真的场景图像是计算机视觉和图形学领域的重要研究课题之一。在本文中,我们提出了一种从新的视点和新的光照颜色和光谱下合成具有荧光物体的场景真实感图像的方法。一般来说,荧光材料吸收一定波长的光,然后发出比被吸收的光波长更长的光,而反射材料则保留光的波长。因此,我们无法将传统的视图合成技术与RGB通道的白平衡调整相结合,在任意照明颜色下再现荧光物体的颜色。在此基础上,结合光的叠加原理,扩展了基于神经辐射场的新视点合成;我们提出的方法通过主动照明系统(如彩色显示器或多光谱光台)从不同视点和不同照明颜色或光谱下捕获场景的稀疏图像集,然后在不明确建模其几何和光度模型的情况下合成场景的照片真实感图像。我们使用LCD捕捉的真实图像进行了大量实验,并证实了我们的方法比现有方法更好。此外,我们表明,我们的方法的扩展使用超过三种原色与一个光阶段,使我们能够再现荧光物体的颜色在常见的光源。
{"title":"FluoNeRF: Fluorescent Novel-View Synthesis Under Novel Light Source Colors and Spectra.","authors":"Lin Shi, Kengo Matsufuji, Michitaka Yoshida, Ryo Kawahara, Takahiro Okabe","doi":"10.3390/jimaging12010016","DOIUrl":"10.3390/jimaging12010016","url":null,"abstract":"<p><p>Synthesizing photo-realistic images of a scene from arbitrary viewpoints and under arbitrary lighting environments is one of the important research topics in computer vision and graphics. In this paper, we propose a method for synthesizing photo-realistic images of a scene with fluorescent objects from novel viewpoints and under novel lighting colors and spectra. In general, fluorescent materials absorb light with certain wavelengths and then emit light with longer wavelengths than the absorbed ones, in contrast to reflective materials, which preserve wavelengths of light. Therefore, we cannot reproduce the colors of fluorescent objects under arbitrary lighting colors by combining conventional view synthesis techniques with the white balance adjustment of the RGB channels. Accordingly, we extend the novel-view synthesis based on the neural radiance fields by incorporating the superposition principle of light; our proposed method captures a sparse set of images of a scene from varying viewpoints and under varying lighting colors or spectra with active lighting systems such as a color display or a multi-spectral light stage and then synthesizes photo-realistic images of the scene without explicitly modeling its geometric and photometric models. We conducted a number of experiments using real images captured with an LCD and confirmed that our method works better than the existing methods. Moreover, we showed that the extension of our method using more than three primary colors with a light stage enables us to reproduce the colors of fluorescent objects under common light sources.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843175/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
M3-TransUNet: Medical Image Segmentation Based on Spatial Prior Attention and Multi-Scale Gating. 基于空间优先注意和多尺度门控的医学图像分割。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-29 DOI: 10.3390/jimaging12010015
Zhigao Zeng, Jiale Xiao, Shengqiu Yi, Qiang Liu, Yanhui Zhu

Medical image segmentation presents substantial challenges arising from the diverse scales and morphological complexities of target anatomical structures. Although existing Transformer-based models excel at capturing global dependencies, they encounter critical bottlenecks in multi-scale feature representation, spatial relationship modeling, and cross-layer feature fusion. To address these limitations, we propose the M3-TransUNet architecture, which incorporates three key innovations: (1) MSGA (Multi-Scale Gate Attention) and MSSA (Multi-Scale Selective Attention) modules to enhance multi-scale feature representation; (2) ME-MSA (Manhattan Enhanced Multi-Head Self-Attention) to integrate spatial priors into self-attention computations, thereby overcoming spatial modeling deficiencies; and (3) MKGAG (Multi-kernel Gated Attention Gate) to optimize skip connections by precisely filtering noise and preserving boundary details. Extensive experiments on public datasets-including Synapse, CVC-ClinicDB, and ISIC-demonstrate that M3-TransUNet achieves state-of-the-art performance. Specifically, on the Synapse dataset, our model outperforms recent TransUNet variants such as J-CAPA, improving the average DSC to 82.79% (compared to 82.29%) and significantly reducing the average HD95 from 19.74 mm to 10.21 mm.

由于目标解剖结构的不同尺度和形态复杂性,医学图像分割面临着巨大的挑战。尽管现有的基于transformer的模型在捕获全局依赖关系方面表现出色,但它们在多尺度特征表示、空间关系建模和跨层特征融合方面遇到了关键瓶颈。为了解决这些限制,我们提出了M3-TransUNet架构,该架构包含三个关键创新:(1)MSGA(多尺度门注意)和MSSA(多尺度选择性注意)模块,以增强多尺度特征表示;(2)曼哈顿增强多头自注意模型(ME-MSA)将空间先验整合到自注意计算中,克服了空间建模的不足;(3) MKGAG (Multi-kernel门控注意门),通过精确滤波噪声和保留边界细节来优化跳跃连接。在公共数据集(包括Synapse、CVC-ClinicDB和isic)上进行的大量实验表明,M3-TransUNet达到了最先进的性能。具体来说,在Synapse数据集上,我们的模型优于最近的TransUNet变体,如J-CAPA,将平均DSC提高到82.79%(与82.29%相比),并显着将平均HD95从19.74 mm降低到10.21 mm。
{"title":"M<sup>3</sup>-TransUNet: Medical Image Segmentation Based on Spatial Prior Attention and Multi-Scale Gating.","authors":"Zhigao Zeng, Jiale Xiao, Shengqiu Yi, Qiang Liu, Yanhui Zhu","doi":"10.3390/jimaging12010015","DOIUrl":"10.3390/jimaging12010015","url":null,"abstract":"<p><p>Medical image segmentation presents substantial challenges arising from the diverse scales and morphological complexities of target anatomical structures. Although existing Transformer-based models excel at capturing global dependencies, they encounter critical bottlenecks in multi-scale feature representation, spatial relationship modeling, and cross-layer feature fusion. To address these limitations, we propose the M<sup>3</sup>-TransUNet architecture, which incorporates three key innovations: (1) MSGA (Multi-Scale Gate Attention) and MSSA (Multi-Scale Selective Attention) modules to enhance multi-scale feature representation; (2) ME-MSA (Manhattan Enhanced Multi-Head Self-Attention) to integrate spatial priors into self-attention computations, thereby overcoming spatial modeling deficiencies; and (3) MKGAG (Multi-kernel Gated Attention Gate) to optimize skip connections by precisely filtering noise and preserving boundary details. Extensive experiments on public datasets-including Synapse, CVC-ClinicDB, and ISIC-demonstrate that M<sup>3</sup>-TransUNet achieves state-of-the-art performance. Specifically, on the Synapse dataset, our model outperforms recent TransUNet variants such as J-CAPA, improving the average DSC to 82.79% (compared to 82.29%) and significantly reducing the average HD95 from 19.74 mm to 10.21 mm.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843401/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification. 自适应归一化增强了深度学习模型在胸部x射线分类中的泛化。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-28 DOI: 10.3390/jimaging12010014
Jatsada Singthongchai, Tanachapong Wangkhamhan

This study presents a controlled benchmarking analysis of min-max scaling, Z-score normalization, and an adaptive preprocessing pipeline that combines percentile-based ROI cropping with histogram standardization. The evaluation was conducted across four public chest X-ray (CXR) datasets and three convolutional neural network architectures under controlled experimental settings. The adaptive pipeline generally improved accuracy, F1-score, and training stability on datasets with relatively stable contrast characteristics while yielding limited gains on MIMIC-CXR due to strong acquisition heterogeneity. Ablation experiments showed that histogram standardization provided the primary performance contribution, with ROI cropping offering complementary benefits, and the full pipeline achieving the best overall performance. The computational overhead of the adaptive preprocessing was minimal (+6.3% training-time cost; 5.2 ms per batch). Friedman-Nemenyi and Wilcoxon signed-rank tests confirmed that the observed improvements were statistically significant across most dataset-model configurations. Overall, adaptive normalization is positioned not as a novel algorithmic contribution, but as a practical preprocessing design choice that can enhance cross-dataset robustness and reliability in chest X-ray classification workflows.

本研究提出了对最小-最大缩放、z分数归一化和自适应预处理管道的控制基准分析,该管道将基于百分位数的ROI裁剪与直方图标准化相结合。在可控的实验设置下,对四个公共胸部x射线(CXR)数据集和三个卷积神经网络架构进行了评估。自适应管道在相对稳定的对比特征数据集上通常提高了准确性、f1分数和训练稳定性,但由于采集异质性强,MIMIC-CXR的增益有限。消融实验表明,直方图标准化提供了主要的性能贡献,ROI裁剪提供了互补的优势,全管道实现了最佳的整体性能。自适应预处理的计算开销最小(+6.3%的训练时间成本;每批5.2 ms)。Friedman-Nemenyi和Wilcoxon签名秩检验证实,在大多数数据集模型配置中,观察到的改进在统计上是显著的。总的来说,自适应归一化并不是一种新的算法贡献,而是一种实用的预处理设计选择,可以增强胸部x射线分类工作流程中的跨数据集鲁棒性和可靠性。
{"title":"Adaptive Normalization Enhances the Generalization of Deep Learning Model in Chest X-Ray Classification.","authors":"Jatsada Singthongchai, Tanachapong Wangkhamhan","doi":"10.3390/jimaging12010014","DOIUrl":"10.3390/jimaging12010014","url":null,"abstract":"<p><p>This study presents a controlled benchmarking analysis of min-max scaling, Z-score normalization, and an adaptive preprocessing pipeline that combines percentile-based ROI cropping with histogram standardization. The evaluation was conducted across four public chest X-ray (CXR) datasets and three convolutional neural network architectures under controlled experimental settings. The adaptive pipeline generally improved accuracy, F1-score, and training stability on datasets with relatively stable contrast characteristics while yielding limited gains on MIMIC-CXR due to strong acquisition heterogeneity. Ablation experiments showed that histogram standardization provided the primary performance contribution, with ROI cropping offering complementary benefits, and the full pipeline achieving the best overall performance. The computational overhead of the adaptive preprocessing was minimal (+6.3% training-time cost; 5.2 ms per batch). Friedman-Nemenyi and Wilcoxon signed-rank tests confirmed that the observed improvements were statistically significant across most dataset-model configurations. Overall, adaptive normalization is positioned not as a novel algorithmic contribution, but as a practical preprocessing design choice that can enhance cross-dataset robustness and reliability in chest X-ray classification workflows.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842669/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1