首页 > 最新文献

Journal of Imaging最新文献

英文 中文
Assessing Change in Stone Burden on Baseline and Follow-Up CT: Radiologist and Radiomics Evaluations. 评估基线和随访CT上结石负荷的变化:放射科医生和放射组学评估。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-27 DOI: 10.3390/jimaging12010013
Parisa Kaviani, Matthias F Froelich, Bernardo Bizzo, Andrew Primak, Giridhar Dasegowda, Emiliano Garza-Frias, Lina Karout, Anushree Burade, Seyedehelaheh Hosseini, Javier Eduardo Contreras Yametti, Keith Dreyer, Sanjay Saini, Mannudeep Kalra

This retrospective diagnostic accuracy study compared radiologist-based qualitative assessments and radiomics-based analyses with an automated artificial intelligence (AI)-based volumetric approach for evaluating changes in kidney stone burden on follow-up CT examinations. With institutional review board approval, 157 patients (mean age, 61 ± 13 years; 99 men, 58 women) who underwent baseline and follow-up non-contrast abdomen-pelvis CT for kidney stone evaluation were included. The index test was an automated AI-based whole-kidney and stone segmentation radiomics prototype (Frontier, Siemens Healthineers), which segmented both kidneys and isolated stone volumes using a fixed threshold of 130 Hounsfield units, providing stone volume and maximum diameter per kidney. The reference standard was a threshold-defined volumetric assessment of stone burden change between baseline and follow-up CTs. The radiologist's performance was assessed using (1) interpretations from clinical radiology reports and (2) an independent radiologist's assessment of stone burden change (stable, increased, or decreased). Diagnostic accuracy was evaluated using multivariable logistic regression and receiver operating characteristic (ROC) analysis. Automated volumetric assessment identified stable (n = 44), increased (n = 109), and decreased (n = 108) stone burden across the evaluated kidneys. Qualitative assessments from radiology reports demonstrated weak diagnostic performance (AUC range, 0.55-0.62), similar to the independent radiologist (AUC range, 0.41-0.72) for differentiating changes in stone burden. A model incorporating higher-order radiomics features achieved an AUC of 0.71 for distinguishing increased versus decreased stone burdens compared with the baseline CT (p < 0.001), but did not outperform threshold-based volumetric assessment. The automated threshold-based volumetric quantification of kidney stone burdens provides higher diagnostic accuracy than qualitative radiologist assessments and radiomics-based analyses for identifying a stable, increased, or decreased stone burden on follow-up CT examinations.

本回顾性诊断准确性研究将基于放射科医生的定性评估和基于放射组学的分析与基于人工智能(AI)的自动体积方法进行比较,以评估随访CT检查中肾结石负担的变化。经机构审查委员会批准,157例患者(平均年龄61±13岁;男性99例,女性58例)接受了基线和随访的非对比腹部-骨盆CT肾结石评估。指数测试是一种自动化的基于人工智能的全肾和结石分割放射组学原型(Frontier, Siemens Healthineers),它使用130 Hounsfield单位的固定阈值分割肾脏和分离结石体积,提供每个肾脏的结石体积和最大直径。参考标准是基线和随访ct之间结石负荷变化的阈值定义的体积评估。评估放射科医生的表现使用(1)临床放射学报告的解释和(2)独立放射科医生对结石负荷变化(稳定、增加或减少)的评估。采用多变量logistic回归和受试者工作特征(ROC)分析评估诊断准确性。自动容量评估确定了稳定(n = 44)、增加(n = 109)和减少(n = 108)的肾脏结石负担。放射学报告的定性评估显示诊断能力较弱(AUC范围,0.55-0.62),与独立放射科医生(AUC范围,0.41-0.72)在鉴别结石负荷变化方面相似。与基线CT相比,结合高阶放射组学特征的模型在区分结石负担增加和减少方面的AUC为0.71 (p < 0.001),但其表现并不优于基于阈值的体积评估。在后续CT检查中,基于阈值的肾结石负荷自动体积量化比定性放射科医生评估和基于放射组学的分析在确定稳定、增加或减少的结石负荷方面提供了更高的诊断准确性。
{"title":"Assessing Change in Stone Burden on Baseline and Follow-Up CT: Radiologist and Radiomics Evaluations.","authors":"Parisa Kaviani, Matthias F Froelich, Bernardo Bizzo, Andrew Primak, Giridhar Dasegowda, Emiliano Garza-Frias, Lina Karout, Anushree Burade, Seyedehelaheh Hosseini, Javier Eduardo Contreras Yametti, Keith Dreyer, Sanjay Saini, Mannudeep Kalra","doi":"10.3390/jimaging12010013","DOIUrl":"10.3390/jimaging12010013","url":null,"abstract":"<p><p>This retrospective diagnostic accuracy study compared radiologist-based qualitative assessments and radiomics-based analyses with an automated artificial intelligence (AI)-based volumetric approach for evaluating changes in kidney stone burden on follow-up CT examinations. With institutional review board approval, 157 patients (mean age, 61 ± 13 years; 99 men, 58 women) who underwent baseline and follow-up non-contrast abdomen-pelvis CT for kidney stone evaluation were included. The index test was an automated AI-based whole-kidney and stone segmentation radiomics prototype (Frontier, Siemens Healthineers), which segmented both kidneys and isolated stone volumes using a fixed threshold of 130 Hounsfield units, providing stone volume and maximum diameter per kidney. The reference standard was a threshold-defined volumetric assessment of stone burden change between baseline and follow-up CTs. The radiologist's performance was assessed using (1) interpretations from clinical radiology reports and (2) an independent radiologist's assessment of stone burden change (stable, increased, or decreased). Diagnostic accuracy was evaluated using multivariable logistic regression and receiver operating characteristic (ROC) analysis. Automated volumetric assessment identified stable (<i>n</i> = 44), increased (<i>n</i> = 109), and decreased (<i>n</i> = 108) stone burden across the evaluated kidneys. Qualitative assessments from radiology reports demonstrated weak diagnostic performance (AUC range, 0.55-0.62), similar to the independent radiologist (AUC range, 0.41-0.72) for differentiating changes in stone burden. A model incorporating higher-order radiomics features achieved an AUC of 0.71 for distinguishing increased versus decreased stone burdens compared with the baseline CT (<i>p</i> < 0.001), but did not outperform threshold-based volumetric assessment. The automated threshold-based volumetric quantification of kidney stone burdens provides higher diagnostic accuracy than qualitative radiologist assessments and radiomics-based analyses for identifying a stable, increased, or decreased stone burden on follow-up CT examinations.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842500/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hybrid Vision Transformer-BiRNN Architecture for Direct k-Space to Image Reconstruction in Accelerated MRI. 一种用于加速MRI直接k空间到图像重建的混合视觉变压器- birnn结构。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-26 DOI: 10.3390/jimaging12010011
Changheun Oh

Long scan times remain a fundamental challenge in Magnetic Resonance Imaging (MRI). Accelerated MRI, which undersamples k-space, requires robust reconstruction methods to solve the ill-posed inverse problem. Recent methods have shown promise by processing image-domain features to capture global spatial context. However, these approaches are often limited, as they fail to fully leverage the unique, sequential characteristics of the k-space data themselves, which are critical for disentangling aliasing artifacts. This study introduces a novel, hybrid, dual-domain deep learning architecture that combines a ViT-based autoencoder with Bidirectional Recurrent Neural Networks (BiRNNs). The proposed architecture is designed to synergistically process information from both domains: it uses the ViT to learn features from image patches and the BiRNNs to model sequential dependencies directly from k-space data. We conducted a comprehensive comparative analysis against a standard ViT with only an MLP head (Model 1), a ViT autoencoder operating solely in the image domain (Model 2), and a competitive UNet baseline. Evaluations were performed on retrospectively undersampled neuro-MRI data using R = 4 and R = 8 acceleration factors with both regular and random sampling patterns. The proposed architecture demonstrated superior performance and robustness, significantly outperforming all other models in challenging high-acceleration and random-sampling scenarios. The results confirm that integrating sequential k-space processing via BiRNNs is critical for superior artifact suppression, offering a robust solution for accelerated MRI.

长扫描时间仍然是磁共振成像(MRI)的一个基本挑战。加速MRI的k空间采样不足,需要鲁棒的重建方法来解决不适定逆问题。最近的方法通过处理图像域特征来捕获全球空间背景显示出前景。然而,这些方法通常是有限的,因为它们不能充分利用k空间数据本身的独特的、连续的特征,而这些特征对于解耦混叠工件是至关重要的。本研究介绍了一种新型的混合双域深度学习架构,该架构结合了基于vit的自编码器和双向循环神经网络(birnn)。所提出的体系结构旨在协同处理来自两个领域的信息:它使用ViT从图像补丁中学习特征,使用birnn直接从k空间数据中建模顺序依赖关系。我们对只有MLP头部的标准ViT(模型1)、仅在图像域中工作的ViT自动编码器(模型2)和竞争性UNet基线进行了全面的比较分析。在常规和随机抽样模式下,采用R = 4和R = 8加速因子对回顾性欠采样神经mri数据进行评估。所提出的架构表现出卓越的性能和鲁棒性,在具有挑战性的高加速和随机抽样场景中显著优于所有其他模型。结果证实,通过birnn集成顺序k空间处理对于抑制伪影至关重要,为加速MRI提供了强大的解决方案。
{"title":"A Hybrid Vision Transformer-BiRNN Architecture for Direct k-Space to Image Reconstruction in Accelerated MRI.","authors":"Changheun Oh","doi":"10.3390/jimaging12010011","DOIUrl":"10.3390/jimaging12010011","url":null,"abstract":"<p><p>Long scan times remain a fundamental challenge in Magnetic Resonance Imaging (MRI). Accelerated MRI, which undersamples k-space, requires robust reconstruction methods to solve the ill-posed inverse problem. Recent methods have shown promise by processing image-domain features to capture global spatial context. However, these approaches are often limited, as they fail to fully leverage the unique, sequential characteristics of the k-space data themselves, which are critical for disentangling aliasing artifacts. This study introduces a novel, hybrid, dual-domain deep learning architecture that combines a ViT-based autoencoder with Bidirectional Recurrent Neural Networks (BiRNNs). The proposed architecture is designed to synergistically process information from both domains: it uses the ViT to learn features from image patches and the BiRNNs to model sequential dependencies directly from k-space data. We conducted a comprehensive comparative analysis against a standard ViT with only an MLP head (Model 1), a ViT autoencoder operating solely in the image domain (Model 2), and a competitive UNet baseline. Evaluations were performed on retrospectively undersampled neuro-MRI data using R = 4 and R = 8 acceleration factors with both regular and random sampling patterns. The proposed architecture demonstrated superior performance and robustness, significantly outperforming all other models in challenging high-acceleration and random-sampling scenarios. The results confirm that integrating sequential k-space processing via BiRNNs is critical for superior artifact suppression, offering a robust solution for accelerated MRI.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842503/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Patched-Based Swin Transformer Hyperprior for Learned Image Compression. 基于补丁的Swin变压器学习图像压缩超先验算法。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-26 DOI: 10.3390/jimaging12010012
Sibusiso B Buthelezi, Jules R Tapamo

We present a hybrid end-to-end learned image compression framework that combines a CNN-based variational autoencoder (VAE) with an efficient hierarchical Swin Transformer to address the limitations of existing entropy models in capturing global dependencies under computational constraints. Traditional VAE-based codecs typically rely on CNN-based priors with localized receptive fields, which are insufficient for modelling the complex, high-dimensional dependencies of the latent space, thereby limiting compression efficiency. While fully global transformer-based models can capture long-range dependencies, their high computational complexity makes them impractical for high-resolution image compression. To overcome this trade-off, our approach couples a CNN-based VAE with a patch-based hierarchical Swin Transformer hyperprior that employs shifted window self-attention to effectively model both local and global contextual information while maintaining computational efficiency. The proposed framework tightly integrates this expressive entropy model with an end-to-end differentiable quantization module, enabling joint optimization of the complete rate-distortion objective. By learning a more accurate probability distribution of the latent representation, the model achieves improved bitrate estimation and a more compact latent representation, resulting in enhanced compression performance. We validate our approach on the widely used Kodak, JPEG AI, and CLIC datasets, demonstrating that the proposed hybrid architecture achieves superior rate-distortion performance, delivering higher visual quality at lower bitrates compared to methods relying on simpler CNN-based entropy priors. This work demonstrates the effectiveness of integrating efficient transformer architectures into learned image compression and highlights their potential for advancing entropy modelling beyond conventional CNN-based designs.

我们提出了一种混合端到端学习图像压缩框架,该框架结合了基于cnn的变分自编码器(VAE)和高效的分层Swin变压器,以解决现有熵模型在计算约束下捕获全局依赖关系的局限性。传统的基于vae的编解码器通常依赖于具有局部接受域的基于cnn的先验,这不足以对潜在空间的复杂、高维依赖性进行建模,从而限制了压缩效率。虽然完全基于全局变压器的模型可以捕获远程依赖关系,但它们的高计算复杂性使它们对于高分辨率图像压缩不切实际。为了克服这种权衡,我们的方法将基于cnn的VAE与基于补丁的分层Swin Transformer超先验结合起来,该超先验利用转移窗口自关注来有效地建模局部和全局上下文信息,同时保持计算效率。该框架将该表达性熵模型与端到端可微量化模块紧密集成,实现了完整率失真目标的联合优化。通过学习更精确的潜在表示的概率分布,该模型实现了改进的比特率估计和更紧凑的潜在表示,从而提高了压缩性能。我们在广泛使用的柯达、JPEG AI和CLIC数据集上验证了我们的方法,证明了所提出的混合架构实现了卓越的率失真性能,与依赖于更简单的基于cnn的熵先验的方法相比,可以在更低的比特率下提供更高的视觉质量。这项工作证明了将高效变压器架构集成到学习图像压缩中的有效性,并强调了它们在超越传统的基于cnn的设计的熵建模方面的潜力。
{"title":"Patched-Based Swin Transformer Hyperprior for Learned Image Compression.","authors":"Sibusiso B Buthelezi, Jules R Tapamo","doi":"10.3390/jimaging12010012","DOIUrl":"10.3390/jimaging12010012","url":null,"abstract":"<p><p>We present a hybrid end-to-end learned image compression framework that combines a CNN-based variational autoencoder (VAE) with an efficient hierarchical Swin Transformer to address the limitations of existing entropy models in capturing global dependencies under computational constraints. Traditional VAE-based codecs typically rely on CNN-based priors with localized receptive fields, which are insufficient for modelling the complex, high-dimensional dependencies of the latent space, thereby limiting compression efficiency. While fully global transformer-based models can capture long-range dependencies, their high computational complexity makes them impractical for high-resolution image compression. To overcome this trade-off, our approach couples a CNN-based VAE with a patch-based hierarchical Swin Transformer hyperprior that employs shifted window self-attention to effectively model both local and global contextual information while maintaining computational efficiency. The proposed framework tightly integrates this expressive entropy model with an end-to-end differentiable quantization module, enabling joint optimization of the complete rate-distortion objective. By learning a more accurate probability distribution of the latent representation, the model achieves improved bitrate estimation and a more compact latent representation, resulting in enhanced compression performance. We validate our approach on the widely used Kodak, JPEG AI, and CLIC datasets, demonstrating that the proposed hybrid architecture achieves superior rate-distortion performance, delivering higher visual quality at lower bitrates compared to methods relying on simpler CNN-based entropy priors. This work demonstrates the effectiveness of integrating efficient transformer architectures into learned image compression and highlights their potential for advancing entropy modelling beyond conventional CNN-based designs.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842631/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054199","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Terrain-Constrained TIN Approach for High-Precision DEM Reconstruction Using UAV Point Clouds. 基于无人机点云的高精度DEM重建方法
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-25 DOI: 10.3390/jimaging12010008
Ziye He, Shu Gan, Xiping Yuan

To address the decline in self-consistency and limited spatial adaptability of traditional interpolation methods in complex terrain, this study proposes a terrain-constrained Triangulated Irregular Network (TIN) interpolation method based on UAV point clouds. The method was tested in the southern margin of the Lufeng Dinosaur National Geopark, Yunnan Province, using ground points at different sampling densities (90%, 70%, 50%, 30%, and 10%), and compared with Spline, Kriging, ANUDEM, and IDW methods. Results show that the proposed method maintains the lowest RMSE and MAE across all densities, demonstrating higher stability and self-consistency and better preserving terrain undulations. This provides technical support for high-precision DEM reconstruction from UAV point clouds in complex terrain.

针对传统插值方法在复杂地形中自一致性下降和空间适应性有限的问题,提出了一种基于无人机点云的地形约束不规则三角网(TIN)插值方法。在云南陆丰恐龙国家地质公园南缘采用不同采样密度(90%、70%、50%、30%、10%)的地点对该方法进行了试验,并与样条法、克里格法、ANUDEM法和IDW法进行了比较。结果表明,该方法在所有密度下均能保持最低的RMSE和MAE,具有较高的稳定性和自一致性,并能较好地保留地形起伏。这为复杂地形下无人机点云的高精度DEM重建提供了技术支持。
{"title":"A Terrain-Constrained TIN Approach for High-Precision DEM Reconstruction Using UAV Point Clouds.","authors":"Ziye He, Shu Gan, Xiping Yuan","doi":"10.3390/jimaging12010008","DOIUrl":"10.3390/jimaging12010008","url":null,"abstract":"<p><p>To address the decline in self-consistency and limited spatial adaptability of traditional interpolation methods in complex terrain, this study proposes a terrain-constrained Triangulated Irregular Network (TIN) interpolation method based on UAV point clouds. The method was tested in the southern margin of the Lufeng Dinosaur National Geopark, Yunnan Province, using ground points at different sampling densities (90%, 70%, 50%, 30%, and 10%), and compared with Spline, Kriging, ANUDEM, and IDW methods. Results show that the proposed method maintains the lowest RMSE and MAE across all densities, demonstrating higher stability and self-consistency and better preserving terrain undulations. This provides technical support for high-precision DEM reconstruction from UAV point clouds in complex terrain.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842391/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Accurate Segmentation of Vegetation in UAV Desert Imagery Using HSV-GLCM Features and SVM Classification. 基于HSV-GLCM特征和SVM分类的无人机沙漠影像植被精确分割
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-25 DOI: 10.3390/jimaging12010009
Thani Jintasuttisak, Patompong Chabplan, Sasitorn Issaro, Orawan Saeung, Thamasan Suwanroj

Segmentation of vegetation from images is an important task in precision agriculture applications, particularly in challenging desert environments where sparse vegetation, varying soil colors, and strong shadows pose significant difficulties. In this paper, we present a machine learning approach to robust green-vegetation segmentation in drone imagery captured over desert farmlands. The proposed method combines HSV color-space representation with Gray-Level Co-occurrence Matrix (GLCM) texture features and employs Support Vector Machine (SVM) as the learning algorithm. To enhance robustness, we incorporate comprehensive preprocessing, including Gaussian filtering, illumination normalization, and bilateral filtering, followed by morphological post-processing to improve segmentation quality. The method is evaluated against both traditional spectral index methods (ExG and CIVE) and a modern deep learning baseline using comprehensive metrics including accuracy, precision, recall, F1-score, and Intersection over Union (IoU). Experimental results on 120 high-resolution drone images from UAE desert farmlands demonstrate that the proposed method achieves superior performance with an accuracy of 0.91, F1-score of 0.88, and IoU of 0.82, showing significant improvement over baseline methods in handling challenging desert conditions, including shadows, varying soil colors, and sparse vegetation patterns. The method provides practical computational performance with a processing time of 25 s per image and a training time of 28 min, making it suitable for agricultural applications where accuracy is prioritized over processing speed.

在精准农业应用中,从图像中分割植被是一项重要的任务,特别是在具有挑战性的沙漠环境中,在这些环境中,植被稀疏、土壤颜色变化和强烈的阴影构成了重大困难。在本文中,我们提出了一种机器学习方法,用于在沙漠农田上捕获的无人机图像中进行鲁棒绿色植被分割。该方法将HSV色彩空间表示与灰度共生矩阵(GLCM)纹理特征相结合,采用支持向量机(SVM)作为学习算法。为了增强鲁棒性,我们结合了全面的预处理,包括高斯滤波、照明归一化和双边滤波,然后进行形态学后处理以提高分割质量。该方法是根据传统的光谱指数方法(ExG和CIVE)和现代深度学习基线进行评估的,使用综合指标包括准确性、精密度、召回率、F1-score和Union交集(IoU)。来自阿联酋沙漠农田的120幅高分辨率无人机图像的实验结果表明,该方法的精度为0.91,f1得分为0.88,IoU为0.82,在处理具有挑战性的沙漠条件(包括阴影、不同土壤颜色和稀疏植被模式)方面比基线方法有了显着改善。该方法提供了实用的计算性能,每张图像的处理时间为25秒,训练时间为28分钟,使其适用于精度优先于处理速度的农业应用。
{"title":"Accurate Segmentation of Vegetation in UAV Desert Imagery Using HSV-GLCM Features and SVM Classification.","authors":"Thani Jintasuttisak, Patompong Chabplan, Sasitorn Issaro, Orawan Saeung, Thamasan Suwanroj","doi":"10.3390/jimaging12010009","DOIUrl":"10.3390/jimaging12010009","url":null,"abstract":"<p><p>Segmentation of vegetation from images is an important task in precision agriculture applications, particularly in challenging desert environments where sparse vegetation, varying soil colors, and strong shadows pose significant difficulties. In this paper, we present a machine learning approach to robust green-vegetation segmentation in drone imagery captured over desert farmlands. The proposed method combines HSV color-space representation with Gray-Level Co-occurrence Matrix (GLCM) texture features and employs Support Vector Machine (SVM) as the learning algorithm. To enhance robustness, we incorporate comprehensive preprocessing, including Gaussian filtering, illumination normalization, and bilateral filtering, followed by morphological post-processing to improve segmentation quality. The method is evaluated against both traditional spectral index methods (ExG and CIVE) and a modern deep learning baseline using comprehensive metrics including accuracy, precision, recall, F1-score, and Intersection over Union (IoU). Experimental results on 120 high-resolution drone images from UAE desert farmlands demonstrate that the proposed method achieves superior performance with an accuracy of 0.91, F1-score of 0.88, and IoU of 0.82, showing significant improvement over baseline methods in handling challenging desert conditions, including shadows, varying soil colors, and sparse vegetation patterns. The method provides practical computational performance with a processing time of 25 s per image and a training time of 28 min, making it suitable for agricultural applications where accuracy is prioritized over processing speed.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843393/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054076","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Long-Term Prognostic Value in Nuclear Cardiology: Expert Scoring Combined with Automated Measurements vs. Angiographic Score. 核心学的长期预后价值:专家评分结合自动测量vs血管造影评分。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-25 DOI: 10.3390/jimaging12010006
George Angelidis, Stavroula Giannakou, Varvara Valotassiou, Emmanouil Panagiotidis, Ioannis Tsougos, Chara Tzavara, Dimitrios Psimadas, Evdoxia Theodorou, Charalampos Ziangas, John Skoularigis, Filippos Triposkiadis, Panagiotis Georgoulias

The evaluation of myocardial perfusion imaging (MPI) studies is based on the visual interpretation of the reconstructed images, while the measurements obtained through software packages may contribute to the investigation, mainly in cases of ambiguous scintigraphic findings. We aimed to investigate the long-term prognostic value of expert reading of Summed Stress Score (SSS), Summed Rest Score (SRS), and Summed Difference Score (SDS), combined with the automated measurements of these parameters, in comparison to the prognostic ability of the angiographic score for soft and hard cardiac events. The study was conducted at the Nuclear Medicine Laboratory of the University of Thessaly, in Larissa, Greece. Overall, 378 consecutive patients with known or suspected coronary artery disease (CAD) were enrolled. Automated measurements of SSS, SRS, and SDS were obtained using the Emory Cardiac Toolbox, Myovation, and Quantitative Perfusion SPECT software packages. Coronary angiographies were scored according to a four-point scoring system (angiographic score). Follow-up data were recorded after phone contact, as well as through review of hospital records. All participants were followed up for at least 36 months. Soft and hard cardiac events were recorded in 31.7% and 11.6% of the sample, respectively, while any cardiac event was recorded in 36.5%. For hard cardiac events, the prognostic value of expert scoring, combined with the prognostic value of the automated measurements, was significantly greater compared to the prognostic ability of the angiographic score (p < 0.001). As far as any cardiac event, the prognostic value of expert scoring, combined with the prognostic value of the automated analyses, was significantly greater compared to the prognostic ability of the angiographic score (p < 0.001). According to our results, in patients with known or suspected CAD, the combination of expert reading and automated measurements of SSS, SRS, and SDS shows a superior prognostic ability in comparison to the angiographic score.

心肌灌注成像(MPI)研究的评估是基于对重建图像的视觉解释,而通过软件包获得的测量可能有助于调查,主要是在扫描结果不明确的情况下。我们的目的是研究总结应激评分(SSS)、总结休息评分(SRS)和总结差异评分(SDS)专家读数的长期预后价值,并结合这些参数的自动测量,与血管造影评分对软性和硬性心脏事件的预后能力进行比较。这项研究是在希腊拉里萨的色萨利大学核医学实验室进行的。总的来说,378名已知或疑似冠状动脉疾病(CAD)的连续患者被纳入研究。使用Emory Cardiac Toolbox、Myovation和定量灌注SPECT软件包自动测量SSS、SRS和SDS。冠状动脉造影按照四分制评分(血管造影评分)。在电话联系后以及通过查阅医院记录记录后续数据。所有参与者都被随访了至少36个月。软性和硬性心脏事件分别占31.7%和11.6%,而任何心脏事件均有记录的占36.5%。对于硬心事件,专家评分的预后价值,结合自动测量的预后价值,与血管造影评分的预后能力相比,显著更高(p < 0.001)。就任何心脏事件而言,专家评分和自动分析的预后价值明显高于血管造影评分的预后能力(p < 0.001)。根据我们的结果,在已知或疑似CAD的患者中,与血管造影评分相比,专家读数和自动测量SSS、SRS和SDS的组合显示出更好的预后能力。
{"title":"Long-Term Prognostic Value in Nuclear Cardiology: Expert Scoring Combined with Automated Measurements vs. Angiographic Score.","authors":"George Angelidis, Stavroula Giannakou, Varvara Valotassiou, Emmanouil Panagiotidis, Ioannis Tsougos, Chara Tzavara, Dimitrios Psimadas, Evdoxia Theodorou, Charalampos Ziangas, John Skoularigis, Filippos Triposkiadis, Panagiotis Georgoulias","doi":"10.3390/jimaging12010006","DOIUrl":"10.3390/jimaging12010006","url":null,"abstract":"<p><p>The evaluation of myocardial perfusion imaging (MPI) studies is based on the visual interpretation of the reconstructed images, while the measurements obtained through software packages may contribute to the investigation, mainly in cases of ambiguous scintigraphic findings. We aimed to investigate the long-term prognostic value of expert reading of Summed Stress Score (SSS), Summed Rest Score (SRS), and Summed Difference Score (SDS), combined with the automated measurements of these parameters, in comparison to the prognostic ability of the angiographic score for soft and hard cardiac events. The study was conducted at the Nuclear Medicine Laboratory of the University of Thessaly, in Larissa, Greece. Overall, 378 consecutive patients with known or suspected coronary artery disease (CAD) were enrolled. Automated measurements of SSS, SRS, and SDS were obtained using the Emory Cardiac Toolbox, Myovation, and Quantitative Perfusion SPECT software packages. Coronary angiographies were scored according to a four-point scoring system (angiographic score). Follow-up data were recorded after phone contact, as well as through review of hospital records. All participants were followed up for at least 36 months. Soft and hard cardiac events were recorded in 31.7% and 11.6% of the sample, respectively, while any cardiac event was recorded in 36.5%. For hard cardiac events, the prognostic value of expert scoring, combined with the prognostic value of the automated measurements, was significantly greater compared to the prognostic ability of the angiographic score (<i>p</i> < 0.001). As far as any cardiac event, the prognostic value of expert scoring, combined with the prognostic value of the automated analyses, was significantly greater compared to the prognostic ability of the angiographic score (<i>p</i> < 0.001). According to our results, in patients with known or suspected CAD, the combination of expert reading and automated measurements of SSS, SRS, and SDS shows a superior prognostic ability in comparison to the angiographic score.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843285/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054266","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AKAZE-GMS-PROSAC: A New Progressive Framework for Matching Dynamic Characteristics of Flotation Foam. AKAZE-GMS-PROSAC:一种新的匹配浮选泡沫动态特性的渐进框架。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-25 DOI: 10.3390/jimaging12010007
Zhen Peng, Zhihong Jiang, Pengcheng Zhu, Gaipin Cai, Xiaoyan Luo

The dynamic characteristics of flotation foam, such as velocity and breakage rate, are critical factors that influence mineral separation efficiency. However, challenges inherent in foam images, including weak textures, severe deformations, and motion blur, present significant technical hurdles for dynamic monitoring. These issues lead to a fundamental conflict between the efficiency and accuracy of traditional feature matching algorithms. This paper introduces a novel progressive framework for dynamic feature matching in flotation foam images, termed "stable extraction, efficient coarse screening, and precise matching." This framework first employs the Accelerated-KAZE (AKAZE) algorithm to extract robust, scale- and rotation-invariant feature points from a non-linear scale-space, effectively addressing the challenge of weak textures. Subsequently, it innovatively incorporates the Grid-based Motion Statistics (GMS) algorithm to perform efficient coarse screening based on motion consistency, rapidly filtering out a large number of obvious mismatches. Finally, the Progressive Sample and Consensus (PROSAC) algorithm is used for precise matching, eliminating the remaining subtle mismatches through progressive sampling and geometric constraints. This framework enables the precise analysis of dynamic foam characteristics, including displacement, velocity, and breakage rate (enhanced by a robust "foam lifetime" mechanism). Comparative experimental results demonstrate that, compared to ORB-GMS-RANSAC (with a Mean Absolute Error, MAE of 1.20 pixels and a Mean Relative Error, MRE of 9.10%) and ORB-RANSAC (MAE: 3.53 pixels, MRE: 27.36%), the proposed framework achieves significantly lower error rates (MAE: 0.23 pixels, MRE: 2.13%). It exhibits exceptional stability and accuracy, particularly in complex scenarios involving low texture and minor displacements. This research provides a high-precision, high-robustness technical solution for the dynamic monitoring and intelligent control of the flotation process.

浮选泡沫的速度和破碎率等动态特性是影响矿物分离效率的关键因素。然而,泡沫图像固有的挑战,包括弱纹理,严重变形和运动模糊,为动态监测提出了重大的技术障碍。这些问题导致了传统特征匹配算法的效率和准确性之间的根本冲突。本文介绍了一种新的渐进式浮选泡沫图像动态特征匹配框架,称为“稳定提取、高效粗筛选和精确匹配”。该框架首先采用加速kaze (AKAZE)算法从非线性尺度空间中提取鲁棒、尺度和旋转不变的特征点,有效地解决了弱纹理的挑战。随后,创新地结合基于网格的运动统计(Grid-based Motion Statistics, GMS)算法,基于运动一致性进行高效粗筛选,快速滤除大量明显的不匹配。最后,采用渐进采样和一致性(PROSAC)算法进行精确匹配,通过渐进采样和几何约束消除剩余的细微不匹配。该框架能够精确分析动态泡沫特性,包括位移、速度和破碎率(通过强大的“泡沫寿命”机制增强)。对比实验结果表明,与ORB-GMS-RANSAC(平均绝对误差,MAE为1.20像素,平均相对误差,MRE为9.10%)和ORB-RANSAC (MAE: 3.53像素,MRE: 27.36%)相比,该框架的错误率显著降低(MAE: 0.23像素,MRE: 2.13%)。它表现出卓越的稳定性和准确性,特别是在复杂的情况下,涉及低纹理和小位移。该研究为浮选过程的动态监测和智能控制提供了高精度、高鲁棒性的技术解决方案。
{"title":"AKAZE-GMS-PROSAC: A New Progressive Framework for Matching Dynamic Characteristics of Flotation Foam.","authors":"Zhen Peng, Zhihong Jiang, Pengcheng Zhu, Gaipin Cai, Xiaoyan Luo","doi":"10.3390/jimaging12010007","DOIUrl":"10.3390/jimaging12010007","url":null,"abstract":"<p><p>The dynamic characteristics of flotation foam, such as velocity and breakage rate, are critical factors that influence mineral separation efficiency. However, challenges inherent in foam images, including weak textures, severe deformations, and motion blur, present significant technical hurdles for dynamic monitoring. These issues lead to a fundamental conflict between the efficiency and accuracy of traditional feature matching algorithms. This paper introduces a novel progressive framework for dynamic feature matching in flotation foam images, termed \"stable extraction, efficient coarse screening, and precise matching.\" This framework first employs the Accelerated-KAZE (AKAZE) algorithm to extract robust, scale- and rotation-invariant feature points from a non-linear scale-space, effectively addressing the challenge of weak textures. Subsequently, it innovatively incorporates the Grid-based Motion Statistics (GMS) algorithm to perform efficient coarse screening based on motion consistency, rapidly filtering out a large number of obvious mismatches. Finally, the Progressive Sample and Consensus (PROSAC) algorithm is used for precise matching, eliminating the remaining subtle mismatches through progressive sampling and geometric constraints. This framework enables the precise analysis of dynamic foam characteristics, including displacement, velocity, and breakage rate (enhanced by a robust \"foam lifetime\" mechanism). Comparative experimental results demonstrate that, compared to ORB-GMS-RANSAC (with a Mean Absolute Error, MAE of 1.20 pixels and a Mean Relative Error, MRE of 9.10%) and ORB-RANSAC (MAE: 3.53 pixels, MRE: 27.36%), the proposed framework achieves significantly lower error rates (MAE: 0.23 pixels, MRE: 2.13%). It exhibits exceptional stability and accuracy, particularly in complex scenarios involving low texture and minor displacements. This research provides a high-precision, high-robustness technical solution for the dynamic monitoring and intelligent control of the flotation process.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843094/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054097","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Render-Rank-Refine: Accurate 6D Indoor Localization via Circular Rendering. Render-Rank-Refine:通过循环渲染实现6D室内精确定位。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-25 DOI: 10.3390/jimaging12010010
Haya Monawwar, Guoliang Fan

Accurate six-degree-of-freedom (6-DoF) camera pose estimation is essential for augmented reality, robotics navigation, and indoor mapping. Existing pipelines often depend on detailed floorplans, strict Manhattan-world priors, and dense structural annotations, which lead to failures in ambiguous room layouts where multiple rooms appear in a query image and their boundaries may overlap or be partially occluded. We present Render-Rank-Refine, a two-stage framework operating on coarse semantic meshes without requiring textured models or per-scene fine-tuning. First, panoramas rendered from the mesh enable global retrieval of coarse pose hypotheses. Then, perspective views from the top-k candidates are compared to the query via rotation-invariant circular descriptors, which re-ranks the matches before final translation and rotation refinement. Our method increases camera localization accuracy compared to the state-of-the-art SPVLoc baseline by reducing the translation error by 40.4% and the rotation error by 29.7% in ambiguous layouts, as evaluated on the Zillow Indoor Dataset. In terms of inference throughput, our method achieves 25.8-26.4 QPS, (Queries Per Second) which is significantly faster than other recent comparable methods, while maintaining accuracy comparable to or better than the SPVLoc baseline. These results demonstrate robust, near-real-time indoor localization that overcomes structural ambiguities and heavy geometric assumptions.

准确的六自由度(6-DoF)相机姿态估计对于增强现实,机器人导航和室内测绘至关重要。现有的管道通常依赖于详细的平面图,严格的曼哈顿世界先验和密集的结构注释,这导致模糊的房间布局失败,其中多个房间出现在查询图像中,并且它们的边界可能重叠或部分遮挡。我们提出了Render-Rank-Refine,这是一个在粗糙语义网格上运行的两阶段框架,不需要纹理模型或逐场景微调。首先,从网格中渲染的全景图可以实现粗姿态假设的全局检索。然后,通过旋转不变的循环描述符将来自前k个候选对象的透视图与查询进行比较,这在最终的转换和旋转细化之前重新排列匹配项。根据Zillow室内数据集的评估,与最先进的SPVLoc基线相比,我们的方法通过减少模糊布局中的平移误差40.4%和旋转误差29.7%,提高了相机定位精度。在推理吞吐量方面,我们的方法达到了25.8-26.4 QPS(每秒查询数),这比其他最近的可比方法要快得多,同时保持了与SPVLoc基线相当或更好的准确性。这些结果证明了强大的、近实时的室内定位,克服了结构模糊性和繁重的几何假设。
{"title":"Render-Rank-Refine: Accurate 6D Indoor Localization via Circular Rendering.","authors":"Haya Monawwar, Guoliang Fan","doi":"10.3390/jimaging12010010","DOIUrl":"10.3390/jimaging12010010","url":null,"abstract":"<p><p>Accurate six-degree-of-freedom (6-DoF) camera pose estimation is essential for augmented reality, robotics navigation, and indoor mapping. Existing pipelines often depend on detailed floorplans, strict Manhattan-world priors, and dense structural annotations, which lead to failures in ambiguous room layouts where multiple rooms appear in a query image and their boundaries may overlap or be partially occluded. We present <i>Render-Rank-Refine</i>, a two-stage framework operating on coarse semantic meshes without requiring textured models or per-scene fine-tuning. First, panoramas rendered from the mesh enable global retrieval of coarse pose hypotheses. Then, perspective views from the top-<i>k</i> candidates are compared to the query via rotation-invariant circular descriptors, which re-ranks the matches before final translation and rotation refinement. Our method increases camera localization accuracy compared to the state-of-the-art SPVLoc baseline by reducing the translation error by 40.4% and the rotation error by 29.7% in ambiguous layouts, as evaluated on the Zillow Indoor Dataset. In terms of inference throughput, our method achieves 25.8-26.4 QPS, (Queries Per Second) which is significantly faster than other recent comparable methods, while maintaining accuracy comparable to or better than the SPVLoc baseline. These results demonstrate robust, near-real-time indoor localization that overcomes structural ambiguities and heavy geometric assumptions.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842481/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054236","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Bone Changes in Mandibular Condyle of Temporomandibular Dysfunction Patients Recognized on Magnetic Resonance Imaging. 磁共振成像识别颞下颌功能障碍患者下颌髁骨变化。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-24 DOI: 10.3390/jimaging12010005
Fumi Mizuhashi, Ichiro Ogura, Ryo Mizuhashi, Yuko Watarai, Tatsuhiro Suzuki, Momoka Kawana, Kotono Nagata, Tomonori Niitsuma, Makoto Oohashi

We aimed to investigate the type of bone changes in temporomandibular disorder patients with disc displacement. The subjects were 117 temporomandibular joints that were diagnosed with anterior disc displacement using magnetic resonance imaging (MRI). Temporomandibular joint (TMJ) pain and opening dysfunction were examined. Disc displacement with and without reduction, joint effusion, and bone changes in the mandibular condyle were assessed on MRI. The types of bone changes were classified into erosion, flattening, osteophyte, and atrophy on the MR images. Fisher's exact test and χ2 test were performed for analyses. Bone changes were found on 30.8% of subjects with erosion, flattening, osteophyte, and atrophy types (p < 0.001). The occurrence of joint effusion appearance (p < 0.001), TMJ pain (p = 0.027), and opening dysfunction (p = 0.002) differed among the types of bone changes. Gender differences were also found among the types of bone changes (p < 0.001). The rate of disc displacement with reduction was significantly smaller than that of disc displacement without reduction on flattening and osteophyte (p < 0.001). The results made it clear that the symptoms, gender, and presence or absence of disc reduction differed among the types of bone changes.

我们的目的是研究伴有椎间盘移位的颞下颌关节紊乱患者的骨改变类型。研究对象为117个经磁共振成像诊断为前盘移位的颞下颌关节。观察颞下颌关节(TMJ)疼痛及开口功能障碍。在MRI上评估有复位和没有复位的椎间盘移位、关节积液和下颌髁的骨变化。骨改变的类型在MR图像上分为侵蚀、扁平、骨赘和萎缩。采用Fisher精确检验和χ2检验进行分析。30.8%的受试者出现骨改变,表现为糜烂、扁平、骨赘和萎缩(p < 0.001)。关节积液样貌(p < 0.001)、颞下颌关节疼痛(p = 0.027)、开口功能障碍(p = 0.002)的发生在不同骨改变类型之间存在差异。骨变化类型也存在性别差异(p < 0.001)。有复位的椎间盘移位率明显小于无复位的扁平和骨赘椎间盘移位率(p < 0.001)。结果清楚地表明,不同类型的骨改变在症状、性别和有无椎间盘复位方面存在差异。
{"title":"Bone Changes in Mandibular Condyle of Temporomandibular Dysfunction Patients Recognized on Magnetic Resonance Imaging.","authors":"Fumi Mizuhashi, Ichiro Ogura, Ryo Mizuhashi, Yuko Watarai, Tatsuhiro Suzuki, Momoka Kawana, Kotono Nagata, Tomonori Niitsuma, Makoto Oohashi","doi":"10.3390/jimaging12010005","DOIUrl":"10.3390/jimaging12010005","url":null,"abstract":"<p><p>We aimed to investigate the type of bone changes in temporomandibular disorder patients with disc displacement. The subjects were 117 temporomandibular joints that were diagnosed with anterior disc displacement using magnetic resonance imaging (MRI). Temporomandibular joint (TMJ) pain and opening dysfunction were examined. Disc displacement with and without reduction, joint effusion, and bone changes in the mandibular condyle were assessed on MRI. The types of bone changes were classified into erosion, flattening, osteophyte, and atrophy on the MR images. Fisher's exact test and χ<sup>2</sup> test were performed for analyses. Bone changes were found on 30.8% of subjects with erosion, flattening, osteophyte, and atrophy types (<i>p</i> < 0.001). The occurrence of joint effusion appearance (<i>p</i> < 0.001), TMJ pain (<i>p</i> = 0.027), and opening dysfunction (<i>p</i> = 0.002) differed among the types of bone changes. Gender differences were also found among the types of bone changes (<i>p</i> < 0.001). The rate of disc displacement with reduction was significantly smaller than that of disc displacement without reduction on flattening and osteophyte (<i>p</i> < 0.001). The results made it clear that the symptoms, gender, and presence or absence of disc reduction differed among the types of bone changes.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842661/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054100","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Empirical Mode Decomposition-Based Deep Learning Model Development for Medical Imaging: Feasibility Study for Gastrointestinal Endoscopic Image Classification. 基于经验模式分解的医学影像深度学习模型开发:胃肠内镜图像分类的可行性研究。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2025-12-22 DOI: 10.3390/jimaging12010004
Mou Deb, Mrinal Kanti Dhar, Poonguzhali Elangovan, Keerthy Gopalakrishnan, Divyanshi Sood, Aaftab Sethi, Sabah Afroze, Sourav Bansal, Aastha Goudel, Charmy Parikh, Avneet Kaur, Swetha Rapolu, Gianeshwaree Alias Rachna Panjwani, Rabiah Aslam Ansari, Naghmeh Asadimanesh, Shiva Sankari Karuppiah, Scott A Helgeson, Venkata S Akshintala, Shivaram P Arunachalam

This study proposes a novel two-dimensional Empirical Mode Decomposition (2D EMD)-based deep learning framework to enhance model performance in multi-class image classification tasks and potential early detection of diseases in healthcare using medical imaging. To validate this approach, we apply it to gastrointestinal (GI) endoscopic image classification using the publicly available Kvasir dataset, which contains eight GI image classes with 1000 images each. The proposed 2D EMD-based design procedure decomposes images into a full set of intrinsic mode functions (IMFs) to enhance image features beneficial for AI model development. Integrating 2D EMD into a deep learning pipeline, we evaluate its impact on four popular models (ResNet152, VGG19bn, MobileNetV3L, and SwinTransformerV2S). The results demonstrate that subtracting IMFs from the original image consistently improves accuracy, F1-score, and AUC for all models. The study reveals a notable enhancement in model performance, with an approximately 9% increase in accuracy compared to counterparts without EMD integration for ResNet152. Similarly, there is an increase of around 18% for VGG19L, 3% for MobileNetV3L, and 8% for SwinTransformerV2. Additionally, explainable AI (XAI) techniques, such as Grad-CAM, illustrate that the model focuses on GI regions for predictions. This study highlights the efficacy of 2D EMD in enhancing deep learning model performance for GI image classification, with potential applications in other domains.

本研究提出了一种新的基于二维经验模式分解(2D EMD)的深度学习框架,以提高模型在多类图像分类任务中的性能,并利用医学成像潜在地早期检测医疗保健中的疾病。为了验证这种方法,我们使用公开可用的Kvasir数据集将其应用于胃肠道(GI)内窥镜图像分类,该数据集包含8个胃肠道图像类别,每个类别有1000张图像。提出的基于二维emd的设计过程将图像分解为一整套内在模式函数(IMFs),以增强有利于AI模型开发的图像特征。将2D EMD集成到深度学习管道中,我们评估了其对四种流行模型(ResNet152, vgg190 bn, MobileNetV3L和SwinTransformerV2S)的影响。结果表明,从原始图像中减去imf可以持续提高所有模型的精度、f1分数和AUC。该研究揭示了模型性能的显著增强,与未集成EMD的ResNet152相比,准确率提高了约9%。类似地,VGG19L增加了18%,MobileNetV3L增加了3%,SwinTransformerV2增加了8%。此外,可解释的AI (XAI)技术,如Grad-CAM,说明该模型将重点放在GI区域进行预测。本研究强调了2D EMD在增强GI图像分类的深度学习模型性能方面的有效性,在其他领域具有潜在的应用前景。
{"title":"Empirical Mode Decomposition-Based Deep Learning Model Development for Medical Imaging: Feasibility Study for Gastrointestinal Endoscopic Image Classification.","authors":"Mou Deb, Mrinal Kanti Dhar, Poonguzhali Elangovan, Keerthy Gopalakrishnan, Divyanshi Sood, Aaftab Sethi, Sabah Afroze, Sourav Bansal, Aastha Goudel, Charmy Parikh, Avneet Kaur, Swetha Rapolu, Gianeshwaree Alias Rachna Panjwani, Rabiah Aslam Ansari, Naghmeh Asadimanesh, Shiva Sankari Karuppiah, Scott A Helgeson, Venkata S Akshintala, Shivaram P Arunachalam","doi":"10.3390/jimaging12010004","DOIUrl":"10.3390/jimaging12010004","url":null,"abstract":"<p><p>This study proposes a novel two-dimensional Empirical Mode Decomposition (2D EMD)-based deep learning framework to enhance model performance in multi-class image classification tasks and potential early detection of diseases in healthcare using medical imaging. To validate this approach, we apply it to gastrointestinal (GI) endoscopic image classification using the publicly available Kvasir dataset, which contains eight GI image classes with 1000 images each. The proposed 2D EMD-based design procedure decomposes images into a full set of intrinsic mode functions (IMFs) to enhance image features beneficial for AI model development. Integrating 2D EMD into a deep learning pipeline, we evaluate its impact on four popular models (ResNet152, VGG19bn, MobileNetV3L, and SwinTransformerV2S). The results demonstrate that subtracting IMFs from the original image consistently improves accuracy, F1-score, and AUC for all models. The study reveals a notable enhancement in model performance, with an approximately 9% increase in accuracy compared to counterparts without EMD integration for ResNet152. Similarly, there is an increase of around 18% for VGG19L, 3% for MobileNetV3L, and 8% for SwinTransformerV2. Additionally, explainable AI (XAI) techniques, such as Grad-CAM, illustrate that the model focuses on GI regions for predictions. This study highlights the efficacy of 2D EMD in enhancing deep learning model performance for GI image classification, with potential applications in other domains.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2025-12-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842276/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1