首页 > 最新文献

Journal of Imaging最新文献

英文 中文
Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks. 基于生成对抗网络的木材显微图像数据集增强研究。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-12 DOI: 10.3390/jimaging11120445
Shuo Xu, Hang Su, Lei Zhao

Microscopic wood images are vital in wood analysis and classification research. However, the high cost of acquiring microscopic images and the limitations of experimental conditions have led to a severe problem of insufficient sample data, which significantly restricts the training performance and generalization ability of deep learning models. This study first used basic image processing techniques to perform preliminary augmentation of the original dataset. The augmented data were then input into five GAN models, BGAN, DCGAN, WGAN-GP, LSGAN, and StyleGAN2, for training. The quality and model performance of the generated images were assessed by analyzing the degree of fidelity of cellular structure (e.g., earlywood, latewood, and wood rays), image clarity, and diversity of the images for each model-generated image, as well as by using KID, IS, and SSIM. The results showed that images generated by BGAN and WGAN-GP exhibited high quality, with lower KID values and higher IS values, and the generated images were visually close to real images. In contrast, the DCGAN, LSGAN, and StyleGAN2 models experienced mode collapse during training, resulting in lower image clarity and diversity compared to the other models. Through a comparative analysis of different GAN models, this study demonstrates the feasibility and effectiveness of Generative Adversarial Networks in the domain of small-sample image data augmentation, providing an important reference for further research in the field of wood identification.

木材显微图像在木材分析和分类研究中至关重要。然而,显微图像的高获取成本和实验条件的限制导致了严重的样本数据不足问题,严重制约了深度学习模型的训练性能和泛化能力。本研究首先使用基本的图像处理技术对原始数据集进行初步增强。然后将增强后的数据输入到5个GAN模型,BGAN、DCGAN、WGAN-GP、LSGAN和StyleGAN2中进行训练。通过分析细胞结构(如早木、晚木和木射线)的保真度、图像清晰度和每个模型生成图像的多样性,以及使用KID、IS和SSIM来评估生成图像的质量和模型性能。结果表明,BGAN和WGAN-GP生成的图像质量较好,KID值较低,IS值较高,生成的图像视觉上接近真实图像。相比之下,DCGAN、LSGAN和StyleGAN2模型在训练过程中经历了模式崩溃,导致图像清晰度和多样性低于其他模型。本研究通过对不同GAN模型的对比分析,论证了生成对抗网络在小样本图像数据增强领域的可行性和有效性,为木材识别领域的进一步研究提供了重要参考。
{"title":"Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks.","authors":"Shuo Xu, Hang Su, Lei Zhao","doi":"10.3390/jimaging11120445","DOIUrl":"10.3390/jimaging11120445","url":null,"abstract":"<p><p>Microscopic wood images are vital in wood analysis and classification research. However, the high cost of acquiring microscopic images and the limitations of experimental conditions have led to a severe problem of insufficient sample data, which significantly restricts the training performance and generalization ability of deep learning models. This study first used basic image processing techniques to perform preliminary augmentation of the original dataset. The augmented data were then input into five GAN models, BGAN, DCGAN, WGAN-GP, LSGAN, and StyleGAN2, for training. The quality and model performance of the generated images were assessed by analyzing the degree of fidelity of cellular structure (e.g., earlywood, latewood, and wood rays), image clarity, and diversity of the images for each model-generated image, as well as by using KID, IS, and SSIM. The results showed that images generated by BGAN and WGAN-GP exhibited high quality, with lower KID values and higher IS values, and the generated images were visually close to real images. In contrast, the DCGAN, LSGAN, and StyleGAN2 models experienced mode collapse during training, resulting in lower image clarity and diversity compared to the other models. Through a comparative analysis of different GAN models, this study demonstrates the feasibility and effectiveness of Generative Adversarial Networks in the domain of small-sample image data augmentation, providing an important reference for further research in the field of wood identification.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733676/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821362","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AI-Driven Clinical Decision Support System for Automated Ventriculomegaly Classification from Fetal Brain MRI. 人工智能驱动的胎儿脑MRI脑室肿大自动分类临床决策支持系统。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-12 DOI: 10.3390/jimaging11120444
Mannam Subbarao, Simi Surendran, Seena Thomas, Hemanth Lakshman, Vinjanampati Goutham, Keshagani Goud, Suhas Udayakumaran

Fetal ventriculomegaly (VM) is a condition characterized by abnormal enlargement of the cerebral ventricles of the fetus brain that often causes developmental disorders in children. Manual segmentation and classification of ventricular structures from brain MRI scans are time-consuming and require clinical expertise. To address this challenge, we develop an automated pipeline for ventricle segmentation, ventricular width estimation, and VM severity classification using a publicly available dataset. An adaptive slice selection strategy converts 3D MRI volumes into the most informative 2D slices, which are then segmented to isolate the lateral ventricles and deep gray matter. Ventricular width is automatically estimated to assign severity levels based on clinical thresholds, generating labeled data for training a deep learning classifier. Finally, an explainability module using a large language model integrates the MRI slices, segmentation masks, and predicted severity to provide interpretable clinical reasoning. Experimental results demonstrate that the proposed decision support system delivers robust performance, achieving dice scores of 89% and 87.5% for the 2D and 3D segmentation models, respectively. Also, the classification network attains an accuracy of 86% and an F1-score of 0.84 in VM analysis.

胎儿脑室肿大(VM)是一种以胎儿脑室异常增大为特征的疾病,常引起儿童发育障碍。从脑MRI扫描中手动分割和分类心室结构是耗时的,并且需要临床专业知识。为了解决这一挑战,我们开发了一个自动化的管道,用于心室分割、心室宽度估计和VM严重性分类,使用公开可用的数据集。自适应切片选择策略将3D MRI体积转换为最具信息量的2D切片,然后对其进行分割以分离侧脑室和深部灰质。根据临床阈值自动估计心室宽度以分配严重程度,生成标记数据用于训练深度学习分类器。最后,使用大型语言模型的可解释性模块集成了MRI切片,分割掩模和预测的严重程度,以提供可解释的临床推理。实验结果表明,所提出的决策支持系统具有良好的性能,在2D和3D分割模型上分别达到89%和87.5%的骰子分数。在VM分析中,该分类网络的准确率为86%,f1得分为0.84。
{"title":"AI-Driven Clinical Decision Support System for Automated Ventriculomegaly Classification from Fetal Brain MRI.","authors":"Mannam Subbarao, Simi Surendran, Seena Thomas, Hemanth Lakshman, Vinjanampati Goutham, Keshagani Goud, Suhas Udayakumaran","doi":"10.3390/jimaging11120444","DOIUrl":"10.3390/jimaging11120444","url":null,"abstract":"<p><p>Fetal ventriculomegaly (VM) is a condition characterized by abnormal enlargement of the cerebral ventricles of the fetus brain that often causes developmental disorders in children. Manual segmentation and classification of ventricular structures from brain MRI scans are time-consuming and require clinical expertise. To address this challenge, we develop an automated pipeline for ventricle segmentation, ventricular width estimation, and VM severity classification using a publicly available dataset. An adaptive slice selection strategy converts 3D MRI volumes into the most informative 2D slices, which are then segmented to isolate the lateral ventricles and deep gray matter. Ventricular width is automatically estimated to assign severity levels based on clinical thresholds, generating labeled data for training a deep learning classifier. Finally, an explainability module using a large language model integrates the MRI slices, segmentation masks, and predicted severity to provide interpretable clinical reasoning. Experimental results demonstrate that the proposed decision support system delivers robust performance, achieving dice scores of 89% and 87.5% for the 2D and 3D segmentation models, respectively. Also, the classification network attains an accuracy of 86% and an F1-score of 0.84 in VM analysis.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12734337/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pixel-Wise Sky-Obstacle Segmentation in Fisheye Imagery Using Deep Learning and Gradient Boosting. 基于深度学习和梯度增强的鱼眼图像像素级天空障碍物分割。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-12 DOI: 10.3390/jimaging11120446
Némo Bouillon, Vincent Boitier

Accurate sky-obstacle segmentation in hemispherical fisheye imagery is essential for solar irradiance forecasting, photovoltaic system design, and environmental monitoring. However, existing methods often rely on expensive all-sky imagers and region-specific training data, produce coarse sky-obstacle boundaries, and ignore the optical properties of fisheye lenses. We propose a low-cost segmentation framework designed for fisheye imagery that combines synthetic data generation, lens-aware augmentation, and a hybrid deep-learning pipeline. Synthetic fisheye training images are created from publicly available street-view panoramas to cover diverse environments without dedicated hardware, and lens-aware augmentations model fisheye projection and photometric effects to improve robustness across devices. On this dataset, we train a convolutional neural network (CNN) and refine its output with gradient-boosted decision trees (GBDT) to sharpen sky-obstacle boundaries. The method is evaluated on real fisheye images captured with smartphones and low-cost clip-on lenses across multiple sites, achieving an Intersection over Union (IoU) of 96.63% and an F1 score of 98.29%, along with high boundary accuracy. An additional evaluation on an external panoramic baseline dataset confirms strong cross-dataset generalization. Together, these results show that the proposed framework enables accurate, low-cost, and widely deployable hemispherical sky segmentation for practical solar and environmental imaging applications.

在半球形鱼眼图像中精确的天空障碍物分割对于太阳辐照度预测、光伏系统设计和环境监测至关重要。然而,现有的方法往往依赖于昂贵的全天成像仪和特定区域的训练数据,产生粗糙的天空障碍边界,并且忽略了鱼眼镜头的光学特性。我们提出了一种针对鱼眼图像设计的低成本分割框架,该框架结合了合成数据生成、镜头感知增强和混合深度学习管道。合成鱼眼训练图像是从公开可用的街景全景图创建的,以覆盖不同的环境,而无需专用硬件,镜头感知增强模型鱼眼投影和光度效果,以提高跨设备的鲁棒性。在这个数据集上,我们训练了一个卷积神经网络(CNN),并用梯度增强决策树(GBDT)改进其输出,以锐化天空障碍边界。用智能手机和低成本夹片镜头在多个地点拍摄的真实鱼眼图像对该方法进行了评估,实现了96.63%的交叉点(IoU)和98.29%的F1分数,并具有较高的边界精度。对外部全景基线数据集的额外评估证实了强大的跨数据集泛化。总之,这些结果表明,所提出的框架能够实现精确、低成本和广泛部署的半球天空分割,用于实际的太阳和环境成像应用。
{"title":"Pixel-Wise Sky-Obstacle Segmentation in Fisheye Imagery Using Deep Learning and Gradient Boosting.","authors":"Némo Bouillon, Vincent Boitier","doi":"10.3390/jimaging11120446","DOIUrl":"10.3390/jimaging11120446","url":null,"abstract":"<p><p>Accurate sky-obstacle segmentation in hemispherical fisheye imagery is essential for solar irradiance forecasting, photovoltaic system design, and environmental monitoring. However, existing methods often rely on expensive all-sky imagers and region-specific training data, produce coarse sky-obstacle boundaries, and ignore the optical properties of fisheye lenses. We propose a low-cost segmentation framework designed for fisheye imagery that combines synthetic data generation, lens-aware augmentation, and a hybrid deep-learning pipeline. Synthetic fisheye training images are created from publicly available street-view panoramas to cover diverse environments without dedicated hardware, and lens-aware augmentations model fisheye projection and photometric effects to improve robustness across devices. On this dataset, we train a convolutional neural network (CNN) and refine its output with gradient-boosted decision trees (GBDT) to sharpen sky-obstacle boundaries. The method is evaluated on real fisheye images captured with smartphones and low-cost clip-on lenses across multiple sites, achieving an Intersection over Union (IoU) of 96.63% and an F1 score of 98.29%, along with high boundary accuracy. An additional evaluation on an external panoramic baseline dataset confirms strong cross-dataset generalization. Together, these results show that the proposed framework enables accurate, low-cost, and widely deployable hemispherical sky segmentation for practical solar and environmental imaging applications.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733846/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced Object Detection Algorithms in Complex Environments via Improved CycleGAN Data Augmentation and AS-YOLO Framework. 基于改进的CycleGAN数据增强和AS-YOLO框架的复杂环境下增强目标检测算法。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-12 DOI: 10.3390/jimaging11120447
Zhen Li, Yuxuan Wang, Lingzhong Meng, Wenjuan Chu, Guang Yang

Object detection in complex environments, such as challenging lighting conditions, adverse weather, and target occlusions, poses significant difficulties for existing algorithms. To address these challenges, this study introduces a collaborative solution integrating improved CycleGAN-based data augmentation and an enhanced object detection framework, AS-YOLO. The improved CycleGAN incorporates a dual self-attention mechanism and spectral normalization to enhance feature capture and training stability. The AS-YOLO framework integrates a channel-spatial parallel attention mechanism, an AFPN structure for improved feature fusion, and the Inner_IoU loss function for better generalization. The experimental results show that compared with YOLOv8n, mAP@0.5 and mAP@0.95 of the AS-YOLO algorithm have increased by 1.5% and 0.6%, respectively. After data augmentation and style transfer, mAP@0.5 and mAP@0.95 have increased by 14.6% and 17.8%, respectively, demonstrating the effectiveness of the proposed method in improving the performance of the model in complex scenarios.

复杂环境下的目标检测,如具有挑战性的光照条件、恶劣的天气和目标遮挡,对现有算法构成了重大困难。为了应对这些挑战,本研究引入了一种协作解决方案,该解决方案集成了改进的基于cyclegan的数据增强和增强的目标检测框架AS-YOLO。改进的CycleGAN结合了双重自注意机制和谱归一化,以增强特征捕获和训练稳定性。AS-YOLO框架集成了信道-空间并行注意机制、用于改进特征融合的AFPN结构和用于更好泛化的Inner_IoU损失函数。实验结果表明,与YOLOv8n相比,AS-YOLO算法的mAP@0.5和mAP@0.95分别提高了1.5%和0.6%。经过数据增强和风格迁移后,mAP@0.5和mAP@0.95分别提高了14.6%和17.8%,表明本文方法在复杂场景下提高模型性能的有效性。
{"title":"Enhanced Object Detection Algorithms in Complex Environments via Improved CycleGAN Data Augmentation and AS-YOLO Framework.","authors":"Zhen Li, Yuxuan Wang, Lingzhong Meng, Wenjuan Chu, Guang Yang","doi":"10.3390/jimaging11120447","DOIUrl":"10.3390/jimaging11120447","url":null,"abstract":"<p><p>Object detection in complex environments, such as challenging lighting conditions, adverse weather, and target occlusions, poses significant difficulties for existing algorithms. To address these challenges, this study introduces a collaborative solution integrating improved CycleGAN-based data augmentation and an enhanced object detection framework, AS-YOLO. The improved CycleGAN incorporates a dual self-attention mechanism and spectral normalization to enhance feature capture and training stability. The AS-YOLO framework integrates a channel-spatial parallel attention mechanism, an AFPN structure for improved feature fusion, and the Inner_IoU loss function for better generalization. The experimental results show that compared with YOLOv8n, mAP@0.5 and mAP@0.95 of the AS-YOLO algorithm have increased by 1.5% and 0.6%, respectively. After data augmentation and style transfer, mAP@0.5 and mAP@0.95 have increased by 14.6% and 17.8%, respectively, demonstrating the effectiveness of the proposed method in improving the performance of the model in complex scenarios.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12734041/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
VMPANet: Vision Mamba Skin Lesion Image Segmentation Model Based on Prompt and Attention Mechanism Fusion. 基于提示机制和注意机制融合的视觉曼巴皮肤病变图像分割模型。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-11 DOI: 10.3390/jimaging11120443
Zinuo Peng, Shuxian Liu, Chenhao Li

In the realm of medical image processing, the segmentation of dermatological lesions is a pivotal technique for the early detection of skin cancer. However, existing methods for segmenting images of skin lesions often encounter limitations when dealing with intricate boundaries and diverse lesion shapes. To address these challenges, we propose VMPANet, designed to accurately localize critical targets and capture edge structures. VMPANet employs an inverted pyramid convolution to extract multi-scale features while utilizing the visual Mamba module to capture long-range dependencies among image features. Additionally, we leverage previously extracted masks as cues to facilitate efficient feature propagation. Furthermore, VMPANet integrates parallel depthwise separable convolutions to enhance feature extraction and introduces innovative mechanisms for edge enhancement, spatial attention, and channel attention to adaptively extract edge information and complex spatial relationships. Notably, VMPANet refines a novel cross-attention mechanism, which effectively facilitates the interaction between deep semantic cues and shallow texture details, thereby generating comprehensive feature representations while reducing computational load and redundancy. We conducted comparative and ablation experiments on two public skin lesion datasets (ISIC2017 and ISIC2018). The results demonstrate that VMPANet outperforms existing mainstream methods. On the ISIC2017 dataset, its mIoU and DSC metrics are 1.38% and 0.83% higher than those of VM-Unet respectively; on the ISIC2018 dataset, these metrics are 1.10% and 0.67% higher than those of EMCAD, respectively. Moreover, VMPANet boasts a parameter count of only 0.383 M and a computational load of 1.159 GFLOPs.

在医学图像处理领域,皮肤病变的分割是早期发现皮肤癌的关键技术。然而,现有的皮肤病变图像分割方法在处理复杂的边界和不同的病变形状时经常遇到局限性。为了解决这些挑战,我们提出了VMPANet,旨在准确定位关键目标并捕获边缘结构。VMPANet采用倒金字塔卷积提取多尺度特征,同时利用视觉曼巴模块捕获图像特征之间的远程依赖关系。此外,我们利用之前提取的蒙版作为线索,以促进有效的特征传播。此外,VMPANet集成了并行深度可分卷积来增强特征提取,并引入了边缘增强、空间注意和通道注意的创新机制,以自适应地提取边缘信息和复杂的空间关系。值得注意的是,VMPANet改进了一种新的交叉注意机制,有效地促进了深层语义线索和浅层纹理细节之间的交互,从而在减少计算负荷和冗余的同时生成全面的特征表示。我们在两个公共皮肤病变数据集(ISIC2017和ISIC2018)上进行了对比和消融实验。结果表明,VMPANet优于现有的主流方法。在ISIC2017数据集上,其mIoU和DSC指标分别比VM-Unet高1.38%和0.83%;在ISIC2018数据集上,这些指标分别比EMCAD高1.10%和0.67%。vppanet的参数数仅为0.383 M,计算负载为1.159 GFLOPs。
{"title":"VMPANet: Vision Mamba Skin Lesion Image Segmentation Model Based on Prompt and Attention Mechanism Fusion.","authors":"Zinuo Peng, Shuxian Liu, Chenhao Li","doi":"10.3390/jimaging11120443","DOIUrl":"10.3390/jimaging11120443","url":null,"abstract":"<p><p>In the realm of medical image processing, the segmentation of dermatological lesions is a pivotal technique for the early detection of skin cancer. However, existing methods for segmenting images of skin lesions often encounter limitations when dealing with intricate boundaries and diverse lesion shapes. To address these challenges, we propose VMPANet, designed to accurately localize critical targets and capture edge structures. VMPANet employs an inverted pyramid convolution to extract multi-scale features while utilizing the visual Mamba module to capture long-range dependencies among image features. Additionally, we leverage previously extracted masks as cues to facilitate efficient feature propagation. Furthermore, VMPANet integrates parallel depthwise separable convolutions to enhance feature extraction and introduces innovative mechanisms for edge enhancement, spatial attention, and channel attention to adaptively extract edge information and complex spatial relationships. Notably, VMPANet refines a novel cross-attention mechanism, which effectively facilitates the interaction between deep semantic cues and shallow texture details, thereby generating comprehensive feature representations while reducing computational load and redundancy. We conducted comparative and ablation experiments on two public skin lesion datasets (ISIC2017 and ISIC2018). The results demonstrate that VMPANet outperforms existing mainstream methods. On the ISIC2017 dataset, its mIoU and DSC metrics are 1.38% and 0.83% higher than those of VM-Unet respectively; on the ISIC2018 dataset, these metrics are 1.10% and 0.67% higher than those of EMCAD, respectively. Moreover, VMPANet boasts a parameter count of only 0.383 M and a computational load of 1.159 GFLOPs.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733779/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
HDR Merging of RAW Exposure Series for All-Sky Cameras: A Comparative Study for Circumsolar Radiometry. 全天相机RAW曝光序列的HDR合并:太阳周围辐射测量的比较研究。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-11 DOI: 10.3390/jimaging11120442
Paul Matteschk, Max Aragón, Jose Gomez, Jacob K Thorning, Stefanie Meilinger, Sebastian Houben

All-sky imagers (ASIs) used in solar energy meteorology face an extreme intra-image dynamic range, with the circumsolar neighborhood orders of magnitude brighter than the diffuse dome. Many operational ASI pipelines address this gap with high-dynamic-range (HDR) bracketing inside the camera's image signal processor (ISP), i.e., after demosaicing and color processing in a nonlinear 8-bit RGB domain. Near the Sun, such ISP-domain HDR can down-weight the shortest exposure, retain clipped or near-clipped samples from longer frames, and compress highlight contrast, thereby increasing circumsolar saturation and flattening aureole gradients. A radiance-linear HDR fusion in the sensor/RAW domain (RAW-HDR) is therefore contrasted with the vendor ISP-based HDR mode (ISP-HDR). Solar-based geometric calibration enables Sun-centered analysis. Paired, interleaved acquisitions under clear-sky and broken-cloud conditions are evaluated using two circumsolar performance criteria per RGB channel: (i) saturated-area fraction in concentric rings and (ii) a median-based radial gradient in defined arcs. All quantitative analyses operate on the radiance-linear HDR result; post-merge tone mapping is only used for visualization. Across conditions, ISP-HDR exhibits roughly double the near-saturation within 0-4° of the Sun and about a three- to fourfold weaker circumsolar radial gradient within 0-6° relative to RAW-HDR. These findings indicate that radiance-linear fusion in the RAW domain better preserves circumsolar structure than the examined ISP-domain HDR mode and thus provides more suitable input for downstream tasks such as cloud-edge detection, aerosol retrieval, and irradiance estimation.

用于太阳能气象的全天成像仪(ASIs)面临着极端的图像内动态范围,其太阳周边比漫射圆顶亮几个数量级。许多可操作的ASI管道通过相机图像信号处理器(ISP)内部的高动态范围(HDR)覆盖来解决这一差距,即在非线性8位RGB域中进行去马赛克和色彩处理后。在太阳附近,这样的isp域HDR可以降低最短曝光的权重,保留较长帧中剪切或接近剪切的样本,并压缩高光对比度,从而增加太阳周围的饱和度并使光晕梯度变平。因此,传感器/RAW域的辐射线性HDR融合(RAW-HDR)与基于供应商isp的HDR模式(ISP-HDR)进行了对比。基于太阳的几何校准使太阳为中心的分析。在晴朗的天空和破碎的云条件下,使用每个RGB通道的两个周太阳性能标准来评估成对的、交错的采集:(i)同心圆中的饱和面积分数和(ii)在确定的弧中基于中值的径向梯度。所有定量分析都基于辐射线性HDR结果;合并后音调映射仅用于可视化。在各种条件下,相对于RAW-HDR, ISP-HDR在太阳0-4°范围内表现出大约两倍的近饱和,在0-6°范围内表现出大约三到四倍的弱太阳径向梯度。这些发现表明,与isp域的HDR模式相比,RAW域的辐射-线性融合模式更好地保留了太阳周围的结构,从而为云边缘检测、气溶胶检索和辐照度估计等下游任务提供了更合适的输入。
{"title":"HDR Merging of RAW Exposure Series for All-Sky Cameras: A Comparative Study for Circumsolar Radiometry.","authors":"Paul Matteschk, Max Aragón, Jose Gomez, Jacob K Thorning, Stefanie Meilinger, Sebastian Houben","doi":"10.3390/jimaging11120442","DOIUrl":"10.3390/jimaging11120442","url":null,"abstract":"<p><p>All-sky imagers (ASIs) used in solar energy meteorology face an extreme intra-image dynamic range, with the circumsolar neighborhood orders of magnitude brighter than the diffuse dome. Many operational ASI pipelines address this gap with high-dynamic-range (HDR) bracketing inside the camera's image signal processor (ISP), i.e., after demosaicing and color processing in a nonlinear 8-bit RGB domain. Near the Sun, such ISP-domain HDR can down-weight the shortest exposure, retain clipped or near-clipped samples from longer frames, and compress highlight contrast, thereby increasing circumsolar saturation and flattening aureole gradients. A radiance-linear HDR fusion in the sensor/RAW domain (RAW-HDR) is therefore contrasted with the vendor ISP-based HDR mode (ISP-HDR). Solar-based geometric calibration enables Sun-centered analysis. Paired, interleaved acquisitions under clear-sky and broken-cloud conditions are evaluated using two circumsolar performance criteria per RGB channel: (i) saturated-area fraction in concentric rings and (ii) a median-based radial gradient in defined arcs. All quantitative analyses operate on the radiance-linear HDR result; post-merge tone mapping is only used for visualization. Across conditions, ISP-HDR exhibits roughly double the near-saturation within 0-4° of the Sun and about a three- to fourfold weaker circumsolar radial gradient within 0-6° relative to RAW-HDR. These findings indicate that radiance-linear fusion in the RAW domain better preserves circumsolar structure than the examined ISP-domain HDR mode and thus provides more suitable input for downstream tasks such as cloud-edge detection, aerosol retrieval, and irradiance estimation.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733980/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821364","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Multi-Scale Neural Network with Attention-Based Fusion for Fruit Crop Disease Identification. 基于注意力融合的混合多尺度神经网络在水果作物病害识别中的应用。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-10 DOI: 10.3390/jimaging11120440
Shakhmaran Seilov, Akniyet Nurzhaubayev, Marat Baideldinov, Bibinur Zhursinbek, Medet Ashimgaliyev, Ainur Zhumadillayeva

Unobserved fruit crop illnesses are a major threat to agricultural productivity worldwide and frequently cause farmers to suffer large financial losses. Manual field inspection-based disease detection techniques are time-consuming, unreliable, and unsuitable for extensive monitoring. Deep learning approaches, in particular convolutional neural networks, have shown promise for automated plant disease identification, although they still face significant obstacles. These include poor generalization across complicated visual backdrops, limited resilience to different illness sizes, and high processing needs that make deployment on resource-constrained edge devices difficult. We suggest a Hybrid Multi-Scale Neural Network (HMCT-AF with GSAF) architecture for precise and effective fruit crop disease identification in order to overcome these drawbacks. In order to extract long-range dependencies, HMCT-AF with GSAF combines a Vision Transformer-based structural branch with multi-scale convolutional branches to capture both high-level contextual patterns and fine-grained local information. These disparate features are adaptively combined using a novel HMCT-AF with a GSAF module, which enhances model interpretability and classification performance. We conduct evaluations on both PlantVillage (controlled environment) and CLD (real-world in-field conditions), observing consistent performance gains that indicate strong resilience to natural lighting variations and background complexity. With an accuracy of up to 93.79%, HMCT-AF with GSAF outperforms vanilla Transformer models, EfficientNet, and traditional CNNs. These findings demonstrate how well the model captures scale-variant disease symptoms and how it may be used in real-time agricultural applications using hardware that is compatible with the edge. According to our research, HMCT-AF with GSAF presents a viable basis for intelligent, scalable plant disease monitoring systems in contemporary precision farming.

未被发现的水果作物病害是全球农业生产力的主要威胁,并经常使农民遭受巨大的经济损失。基于现场检查的人工疾病检测技术耗时、不可靠、不适合广泛监测。深度学习方法,特别是卷积神经网络,已经显示出自动化植物病害识别的希望,尽管它们仍然面临着重大障碍。这些问题包括在复杂的视觉背景下泛化能力差,对不同疾病大小的弹性有限,以及在资源受限的边缘设备上部署困难的高处理需求。为了克服这些缺点,我们提出了一种混合多尺度神经网络(HMCT-AF与GSAF)结构来精确有效地识别水果作物病害。为了提取远程依赖关系,HMCT-AF与GSAF结合了基于Vision transformer的结构分支和多尺度卷积分支,以捕获高级上下文模式和细粒度本地信息。使用新的HMCT-AF和GSAF模块自适应地组合了这些不同的特征,从而增强了模型的可解释性和分类性能。我们对PlantVillage(受控环境)和CLD(真实的野外条件)进行了评估,观察到一致的性能增益表明对自然光照变化和背景复杂性的强弹性。具有GSAF的HMCT-AF的准确率高达93.79%,优于普通Transformer模型、EfficientNet和传统cnn。这些发现证明了该模型捕捉尺度变异疾病症状的效果如何,以及如何使用与边缘兼容的硬件将其用于实时农业应用。根据我们的研究,具有GSAF的HMCT-AF为现代精准农业中智能、可扩展的植物病害监测系统提供了可行的基础。
{"title":"Hybrid Multi-Scale Neural Network with Attention-Based Fusion for Fruit Crop Disease Identification.","authors":"Shakhmaran Seilov, Akniyet Nurzhaubayev, Marat Baideldinov, Bibinur Zhursinbek, Medet Ashimgaliyev, Ainur Zhumadillayeva","doi":"10.3390/jimaging11120440","DOIUrl":"10.3390/jimaging11120440","url":null,"abstract":"<p><p>Unobserved fruit crop illnesses are a major threat to agricultural productivity worldwide and frequently cause farmers to suffer large financial losses. Manual field inspection-based disease detection techniques are time-consuming, unreliable, and unsuitable for extensive monitoring. Deep learning approaches, in particular convolutional neural networks, have shown promise for automated plant disease identification, although they still face significant obstacles. These include poor generalization across complicated visual backdrops, limited resilience to different illness sizes, and high processing needs that make deployment on resource-constrained edge devices difficult. We suggest a Hybrid Multi-Scale Neural Network (HMCT-AF with GSAF) architecture for precise and effective fruit crop disease identification in order to overcome these drawbacks. In order to extract long-range dependencies, HMCT-AF with GSAF combines a Vision Transformer-based structural branch with multi-scale convolutional branches to capture both high-level contextual patterns and fine-grained local information. These disparate features are adaptively combined using a novel HMCT-AF with a GSAF module, which enhances model interpretability and classification performance. We conduct evaluations on both PlantVillage (controlled environment) and CLD (real-world in-field conditions), observing consistent performance gains that indicate strong resilience to natural lighting variations and background complexity. With an accuracy of up to 93.79%, HMCT-AF with GSAF outperforms vanilla Transformer models, EfficientNet, and traditional CNNs. These findings demonstrate how well the model captures scale-variant disease symptoms and how it may be used in real-time agricultural applications using hardware that is compatible with the edge. According to our research, HMCT-AF with GSAF presents a viable basis for intelligent, scalable plant disease monitoring systems in contemporary precision farming.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12734175/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821313","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Application of Artificial Intelligence and Computer Vision for Measuring and Counting Oysters. 人工智能与计算机视觉在牡蛎计量与计数中的应用。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-10 DOI: 10.3390/jimaging11120439
Julio Antonio Laria Pino, Jesús David Terán Villanueva, Julio Laria Menchaca, Leobardo Garcia Solorio, Salvador Ibarra Martínez, Mirna Patricia Ponce Flores, Aurelio Alejandro Santiago Pineda

One of the most important activities in any oyster farm is the measurement of oyster size; this activity is time-consuming and conducted manually, generally using a caliper, which leads to high measurement variability. This paper proposes a methodology to count and obtain the length and width averages of a sample of oysters from an image, relying on artificial intelligence (AI), which refers to systems capable of learning and decision-making, and computer vision (CV), which enables the extraction of information from digital images. The proposed approach employs the DBScan clustering algorithm, an artificial neural network (ANN), and a random forest classifier to enable automatic oyster classification, counting, and size estimation from images. As a result of the proposed methodology, the speed in measuring the length and width of the oysters was 86.7 times faster than manual measurement. Regarding the counting, the process missed the total count of oysters in two of the ten images. These results demonstrate the feasibility of using the proposed methodology to measure oyster size and count in oyster farms.

牡蛎养殖场最重要的活动之一是测量牡蛎的大小;这个活动是耗时的,并且是手动进行的,通常使用卡尺,这导致了高度的测量可变性。本文提出了一种方法,依靠人工智能(AI)和计算机视觉(CV),从图像中计算和获得牡蛎样本的长度和宽度平均值。人工智能(AI)是指能够学习和决策的系统,计算机视觉(CV)能够从数字图像中提取信息。该方法采用DBScan聚类算法、人工神经网络(ANN)和随机森林分类器实现牡蛎的自动分类、计数和大小估计。采用该方法,测量牡蛎的长度和宽度的速度比人工测量快86.7倍。在计数方面,该过程在10张图像中有两张遗漏了牡蛎的总数。这些结果证明了在牡蛎养殖场中使用所提出的方法测量牡蛎大小和数量的可行性。
{"title":"Application of Artificial Intelligence and Computer Vision for Measuring and Counting Oysters.","authors":"Julio Antonio Laria Pino, Jesús David Terán Villanueva, Julio Laria Menchaca, Leobardo Garcia Solorio, Salvador Ibarra Martínez, Mirna Patricia Ponce Flores, Aurelio Alejandro Santiago Pineda","doi":"10.3390/jimaging11120439","DOIUrl":"10.3390/jimaging11120439","url":null,"abstract":"<p><p>One of the most important activities in any oyster farm is the measurement of oyster size; this activity is time-consuming and conducted manually, generally using a caliper, which leads to high measurement variability. This paper proposes a methodology to count and obtain the length and width averages of a sample of oysters from an image, relying on artificial intelligence (AI), which refers to systems capable of learning and decision-making, and computer vision (CV), which enables the extraction of information from digital images. The proposed approach employs the DBScan clustering algorithm, an artificial neural network (ANN), and a random forest classifier to enable automatic oyster classification, counting, and size estimation from images. As a result of the proposed methodology, the speed in measuring the length and width of the oysters was 86.7 times faster than manual measurement. Regarding the counting, the process missed the total count of oysters in two of the ten images. These results demonstrate the feasibility of using the proposed methodology to measure oyster size and count in oyster farms.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733815/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion. 基于子带特征提取和融合的压缩小波系数直接分类。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-10 DOI: 10.3390/jimaging11120441
Xin Li, Baile Sun

A major computational bottleneck in classifying large-scale hyperspectral images (HSI) is the mandatory data decompression prior to processing. Compressed-domain computing offers a solution by enabling deep learning on partially compressed data. However, existing compressed-domain methods are predominantly tailored for the Discrete Cosine Transform (DCT) used in natural images, while HSIs are typically compressed using the Discrete Wavelet Transform (DWT). The fundamental structural mismatch between the block-based DCT and the hierarchical DWT sub-bands presents two core challenges: how to extract features from multiple wavelet sub-bands, and how to fuse these features effectively? To address these issues, we propose a novel framework that extracts and fuses features from different DWT sub-bands directly. We design a multi-branch feature extractor with sub-band feature alignment loss that processes functionally different sub-bands in parallel, preserving the independence of each frequency feature. We then employ a sub-band cross-attention mechanism that inverts the typical attention paradigm by using the sparse, high-frequency detail sub-bands as queries to adaptively select and enhance salient features from the dense, information-rich low-frequency sub-bands. This enables a targeted fusion of global context and fine-grained structural information without data reconstruction. Experiments on three benchmark datasets demonstrate that our method achieves classification accuracy comparable to state-of-the-art spatial-domain approaches while eliminating at least 56% of the decompression overhead.

分类大规模高光谱图像(HSI)的主要计算瓶颈是处理前的强制数据解压缩。压缩域计算通过在部分压缩数据上进行深度学习提供了一种解决方案。然而,现有的压缩域方法主要针对自然图像中使用的离散余弦变换(DCT),而hsi通常使用离散小波变换(DWT)进行压缩。基于块的DCT与分层小波变换子带之间的基本结构不匹配提出了两个核心挑战:如何从多个小波子带中提取特征,以及如何有效地融合这些特征。为了解决这些问题,我们提出了一种新的框架,直接从不同的DWT子带提取和融合特征。我们设计了一种具有子带特征对齐损失的多分支特征提取器,可以并行处理不同的子带,同时保持每个频率特征的独立性。然后,我们采用了一种子带交叉注意机制,该机制通过使用稀疏的高频细节子带作为查询,从密集的、信息丰富的低频子带中自适应地选择和增强显著特征,从而颠覆了典型的注意范式。这使得全局上下文和细粒度结构信息的目标融合无需数据重建。在三个基准数据集上的实验表明,我们的方法达到了与最先进的空域方法相当的分类精度,同时消除了至少56%的解压开销。
{"title":"WaveletHSI: Direct HSI Classification from Compressed Wavelet Coefficients via Sub-Band Feature Extraction and Fusion.","authors":"Xin Li, Baile Sun","doi":"10.3390/jimaging11120441","DOIUrl":"10.3390/jimaging11120441","url":null,"abstract":"<p><p>A major computational bottleneck in classifying large-scale hyperspectral images (HSI) is the mandatory data decompression prior to processing. Compressed-domain computing offers a solution by enabling deep learning on partially compressed data. However, existing compressed-domain methods are predominantly tailored for the Discrete Cosine Transform (DCT) used in natural images, while HSIs are typically compressed using the Discrete Wavelet Transform (DWT). The fundamental structural mismatch between the block-based DCT and the hierarchical DWT sub-bands presents two core challenges: how to extract features from multiple wavelet sub-bands, and how to fuse these features effectively? To address these issues, we propose a novel framework that extracts and fuses features from different DWT sub-bands directly. We design a multi-branch feature extractor with sub-band feature alignment loss that processes functionally different sub-bands in parallel, preserving the independence of each frequency feature. We then employ a sub-band cross-attention mechanism that inverts the typical attention paradigm by using the sparse, high-frequency detail sub-bands as queries to adaptively select and enhance salient features from the dense, information-rich low-frequency sub-bands. This enables a targeted fusion of global context and fine-grained structural information without data reconstruction. Experiments on three benchmark datasets demonstrate that our method achieves classification accuracy comparable to state-of-the-art spatial-domain approaches while eliminating at least 56% of the decompression overhead.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12733817/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821374","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Texture-Based Preprocessing Framework with nnU-Net Model for Accurate Intracranial Artery Segmentation. 基于纹理的nnU-Net模型预处理框架颅内动脉精确分割。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-09 DOI: 10.3390/jimaging11120438
Kyuseok Kim, Ji-Youn Kim

Accurate intracranial artery segmentation from digital subtraction angiography (DSA) is critical for neurovascular diagnosis and intervention planning. Vascular extraction, which combines preprocessing methods and deep learning models, yields a high level of results, but limited preprocessing results constrain the improvement of results. We propose a texture-based contrast enhancement preprocessing framework integrated with the nnU-Net model to improve vessel segmentation in time-sequential DSA images. The method generates a combined feature mask by fusing local contrast, local entropy, and brightness threshold maps, which is then used as input for deep learning-based segmentation. Segmentation performance was evaluated using the DIAS dataset with various standard quantitative metrics. The proposed preprocessing significantly improved segmentation across all metrics compared to both the baseline and contrast-limited adaptive histogram equalization (CLAHE). Using nnU-Net, the method achieved a Dice Similarity Coefficient (DICE) of 0.83 ± 0.20 and an Intersection over Union (IoU) of 0.72 ± 0.14, outperforming CLAHE (DICE 0.79 ± 0.41, IoU 0.70 ± 0.23) and the baseline (DICE 0.65 ± 0.15, IoU 0.47 ± 0.20). Most notably, vessel connectivity (VC) dropped by over 65% relative to unprocessed images, indicating marked improvements in VC and topological accuracy. This study demonstrates that combining texture-based preprocessing with nnU-Net delivers robust, noise-tolerant, and clinically interpretable segmentation of intracranial arteries from DSA.

数字减影血管造影(DSA)对颅内动脉的准确分割是神经血管诊断和干预计划的关键。血管提取结合了预处理方法和深度学习模型,得到了高水平的结果,但预处理结果有限,制约了结果的提高。我们提出了一种基于纹理的对比度增强预处理框架,结合nnU-Net模型来改善时间序列DSA图像中的血管分割。该方法通过融合局部对比度、局部熵和亮度阈值图生成组合特征掩码,然后将其作为基于深度学习的分割的输入。使用DIAS数据集和各种标准定量指标评估分割性能。与基线和对比度有限的自适应直方图均衡化(CLAHE)相比,所提出的预处理显著改善了所有指标的分割。使用nnU-Net,该方法获得了骰子相似系数(Dice)为0.83±0.20,交集比(IoU)为0.72±0.14,优于CLAHE (Dice 0.79±0.41,IoU 0.70±0.23)和基线(Dice 0.65±0.15,IoU 0.47±0.20)。最值得注意的是,与未经处理的图像相比,血管连通性(VC)下降了65%以上,这表明VC和拓扑精度有了显著提高。该研究表明,将基于纹理的预处理与nnU-Net相结合,可以从DSA中获得鲁棒性、耐噪性和临床可解释的颅内动脉分割。
{"title":"Texture-Based Preprocessing Framework with nnU-Net Model for Accurate Intracranial Artery Segmentation.","authors":"Kyuseok Kim, Ji-Youn Kim","doi":"10.3390/jimaging11120438","DOIUrl":"10.3390/jimaging11120438","url":null,"abstract":"<p><p>Accurate intracranial artery segmentation from digital subtraction angiography (DSA) is critical for neurovascular diagnosis and intervention planning. Vascular extraction, which combines preprocessing methods and deep learning models, yields a high level of results, but limited preprocessing results constrain the improvement of results. We propose a texture-based contrast enhancement preprocessing framework integrated with the nnU-Net model to improve vessel segmentation in time-sequential DSA images. The method generates a combined feature mask by fusing local contrast, local entropy, and brightness threshold maps, which is then used as input for deep learning-based segmentation. Segmentation performance was evaluated using the DIAS dataset with various standard quantitative metrics. The proposed preprocessing significantly improved segmentation across all metrics compared to both the baseline and contrast-limited adaptive histogram equalization (CLAHE). Using nnU-Net, the method achieved a Dice Similarity Coefficient (DICE) of 0.83 ± 0.20 and an Intersection over Union (IoU) of 0.72 ± 0.14, outperforming CLAHE (DICE 0.79 ± 0.41, IoU 0.70 ± 0.23) and the baseline (DICE 0.65 ± 0.15, IoU 0.47 ± 0.20). Most notably, vessel connectivity (VC) dropped by over 65% relative to unprocessed images, indicating marked improvements in VC and topological accuracy. This study demonstrates that combining texture-based preprocessing with nnU-Net delivers robust, noise-tolerant, and clinically interpretable segmentation of intracranial arteries from DSA.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 12","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12734170/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145821370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1