首页 > 最新文献

Journal of Imaging最新文献

英文 中文
From Visual to Multimodal: Systematic Ablation of Encoders and Fusion Strategies in Animal Identification. 从视觉到多模态:动物识别中编码器的系统消融和融合策略。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-07 DOI: 10.3390/jimaging12010030
Vasiliy Kudryavtsev, Kirill Borodin, German Berezin, Kirill Bubenchikov, Grach Mkrtchian, Alexander Ryzhkov

Automated animal identification is a practical task for reuniting lost pets with their owners, yet current systems often struggle due to limited dataset scale and reliance on unimodal visual cues. This study introduces a multimodal verification framework that enhances visual features with semantic identity priors derived from synthetic textual descriptions. We constructed a massive training corpus of 1.9 million photographs covering 695,091 unique animals to support this investigation. Through systematic ablation studies, we identified SigLIP2-Giant and E5-Small-v2 as the optimal vision and text backbones. We further evaluated fusion strategies ranging from simple concatenation to adaptive gating to determine the best method for integrating these modalities. Our proposed approach utilizes a gated fusion mechanism and achieved a Top-1 accuracy of 84.28% and an Equal Error Rate of 0.0422 on a comprehensive test protocol. These results represent an 11% improvement over leading unimodal baselines and demonstrate that integrating synthesized semantic descriptions significantly refines decision boundaries in large-scale pet re-identification.

自动动物识别是让走失的宠物与主人团聚的一项实际任务,但由于数据集规模有限和依赖单模视觉线索,目前的系统经常遇到困难。本研究引入了一个多模态验证框架,该框架利用合成文本描述衍生的语义身份先验来增强视觉特征。我们构建了一个庞大的训练语料库,包含190万张照片,涵盖695,091种独特的动物,以支持这项调查。通过系统消融研究,我们确定了SigLIP2-Giant和E5-Small-v2为最佳的视觉和文本主干。我们进一步评估了从简单连接到自适应门控的融合策略,以确定整合这些模式的最佳方法。我们提出的方法利用门控融合机制,在综合测试协议上实现了84.28%的Top-1精度和0.0422的等错误率。这些结果比领先的单峰基线提高了11%,并表明集成综合语义描述显着改善了大规模宠物再识别的决策边界。
{"title":"From Visual to Multimodal: Systematic Ablation of Encoders and Fusion Strategies in Animal Identification.","authors":"Vasiliy Kudryavtsev, Kirill Borodin, German Berezin, Kirill Bubenchikov, Grach Mkrtchian, Alexander Ryzhkov","doi":"10.3390/jimaging12010030","DOIUrl":"10.3390/jimaging12010030","url":null,"abstract":"<p><p>Automated animal identification is a practical task for reuniting lost pets with their owners, yet current systems often struggle due to limited dataset scale and reliance on unimodal visual cues. This study introduces a multimodal verification framework that enhances visual features with semantic identity priors derived from synthetic textual descriptions. We constructed a massive training corpus of 1.9 million photographs covering 695,091 unique animals to support this investigation. Through systematic ablation studies, we identified SigLIP2-Giant and E5-Small-v2 as the optimal vision and text backbones. We further evaluated fusion strategies ranging from simple concatenation to adaptive gating to determine the best method for integrating these modalities. Our proposed approach utilizes a gated fusion mechanism and achieved a Top-1 accuracy of 84.28% and an Equal Error Rate of 0.0422 on a comprehensive test protocol. These results represent an 11% improvement over leading unimodal baselines and demonstrate that integrating synthesized semantic descriptions significantly refines decision boundaries in large-scale pet re-identification.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843040/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054204","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hybrid Skeleton-Based Motion Templates for Cross-View and Appearance-Robust Gait Recognition. 交叉视角和外观鲁棒步态识别的混合骨架运动模板。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-07 DOI: 10.3390/jimaging12010032
João Ferreira Nunes, Pedro Miguel Moreira, João Manuel R S Tavares

Gait recognition methods based on silhouette templates, such as the Gait Energy Image (GEI), achieve high accuracy under controlled conditions but often degrade when appearance varies due to viewpoint, clothing, or carried objects. In contrast, skeleton-based approaches provide interpretable motion cues but remain sensitive to pose-estimation noise. This work proposes two compact 2D skeletal descriptors-Gait Skeleton Images (GSIs)-that encode 3D joint trajectories into line-based and joint-based static templates compatible with standard 2D CNN architectures. A unified processing pipeline is introduced, including skeletal topology normalization, rigid view alignment, orthographic projection, and pixel-level rendering. Core design factors are analyzed on the GRIDDS dataset, where depth-based 3D coordinates provide stable ground truth for evaluating structural choices and rendering parameters. An extensive evaluation is then conducted on the widely used CASIA-B dataset, using 3D coordinates estimated via human pose estimation, to assess robustness under viewpoint, clothing, and carrying covariates. Results show that although GEIs achieve the highest same-view accuracy, GSI variants exhibit reduced degradation under appearance changes and demonstrate greater stability under severe cross-view conditions. These findings indicate that compact skeletal templates can complement appearance-based descriptors and may benefit further from continued advances in 3D human pose estimation.

基于轮廓模板的步态识别方法,如步态能量图像(GEI),在受控条件下实现高精度,但当外观因视点、服装或携带的物体而变化时,往往会降低精度。相比之下,基于骨架的方法提供了可解释的运动线索,但仍然对姿态估计噪声敏感。这项工作提出了两个紧凑的2D骨骼描述符-步态骨骼图像(GSIs)-将3D关节轨迹编码为与标准2D CNN架构兼容的基于线和基于关节的静态模板。引入了一个统一的处理流程,包括骨架拓扑归一化、刚性视图对齐、正交投影和像素级渲染。在GRIDDS数据集上分析了核心设计因素,其中基于深度的三维坐标为评估结构选择和渲染参数提供了稳定的地面真相。然后对广泛使用的CASIA-B数据集进行了广泛的评估,使用通过人体姿态估计估计的3D坐标来评估视点、服装和携带协变量下的鲁棒性。结果表明,尽管gei实现了最高的同视图精度,但GSI变体在外观变化下表现出较少的退化,并在严重的交叉视图条件下表现出更大的稳定性。这些发现表明,紧凑的骨骼模板可以补充基于外观的描述符,并可能从3D人体姿势估计的持续进步中进一步受益。
{"title":"Hybrid Skeleton-Based Motion Templates for Cross-View and Appearance-Robust Gait Recognition.","authors":"João Ferreira Nunes, Pedro Miguel Moreira, João Manuel R S Tavares","doi":"10.3390/jimaging12010032","DOIUrl":"10.3390/jimaging12010032","url":null,"abstract":"<p><p>Gait recognition methods based on silhouette templates, such as the Gait Energy Image (GEI), achieve high accuracy under controlled conditions but often degrade when appearance varies due to viewpoint, clothing, or carried objects. In contrast, skeleton-based approaches provide interpretable motion cues but remain sensitive to pose-estimation noise. This work proposes two compact 2D skeletal descriptors-Gait Skeleton Images (GSIs)-that encode 3D joint trajectories into line-based and joint-based static templates compatible with standard 2D CNN architectures. A unified processing pipeline is introduced, including skeletal topology normalization, rigid view alignment, orthographic projection, and pixel-level rendering. Core design factors are analyzed on the GRIDDS dataset, where depth-based 3D coordinates provide stable ground truth for evaluating structural choices and rendering parameters. An extensive evaluation is then conducted on the widely used CASIA-B dataset, using 3D coordinates estimated via human pose estimation, to assess robustness under viewpoint, clothing, and carrying covariates. Results show that although GEIs achieve the highest same-view accuracy, GSI variants exhibit reduced degradation under appearance changes and demonstrate greater stability under severe cross-view conditions. These findings indicate that compact skeletal templates can complement appearance-based descriptors and may benefit further from continued advances in 3D human pose estimation.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843325/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054210","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Unified Complex-Fresnel Model for Physically Based Long-Wave Infrared Imaging and Simulation. 基于物理的长波红外成像与仿真的统一复菲涅耳模型。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-07 DOI: 10.3390/jimaging12010033
Peter Ter Heerdt, William Keustermans, Ivan De Boi, Steve Vanlanduit

Accurate modelling of reflection, transmission, absorption, and emission at material interfaces is essential for infrared imaging, rendering, and the simulation of optical and sensing systems. This need is particularly pronounced across the short-wave to long-wave infrared (SWIR-LWIR) spectrum, where many materials exhibit dispersion- and wavelength-dependent attenuation described by complex refractive indices. In this work, we introduce a unified formulation of the full Fresnel equations that directly incorporates wavelength-dependent complex refractive-index data and provides physically consistent interface behaviour for both dielectrics and conductors. The approach reformulates the classical Fresnel expressions to eliminate sign ambiguities and numerical instabilities, resulting in a stable evaluation across incidence angles and for strongly absorbing materials. We demonstrate the model through spectral-rendering simulations that illustrate realistic reflectance and transmittance behaviour for materials with different infrared optical properties. To assess its suitability for thermal-infrared applications, we also compare the simulated long-wave emission of a heated glass sphere with measurements from a LWIR camera. The agreement between measured and simulated radiometric trends indicates that the proposed formulation offers a practical and physically grounded tool for wavelength-parametric interface modelling in infrared imaging, supporting applications in spectral rendering, synthetic data generation, and infrared system analysis.

材料界面上反射、透射、吸收和发射的精确建模对于红外成像、渲染和光学和传感系统的模拟至关重要。这种需求在短波到长波红外(SWIR-LWIR)光谱中尤其明显,其中许多材料表现出由复折射率描述的色散和波长相关衰减。在这项工作中,我们引入了完整菲涅耳方程的统一公式,该公式直接包含波长相关的复杂折射率数据,并为介电体和导体提供物理一致的界面行为。该方法重新制定了经典菲涅耳表达式,以消除符号歧义和数值不稳定性,从而在入射角和强吸收材料之间产生稳定的评估。我们通过光谱渲染模拟来展示具有不同红外光学特性的材料的真实反射率和透射率行为。为了评估其对热红外应用的适用性,我们还将加热玻璃球的模拟长波发射与LWIR相机的测量结果进行了比较。测量和模拟的辐射趋势之间的一致性表明,所提出的公式为红外成像中的波长参数界面建模提供了实用和物理基础的工具,支持光谱绘制,合成数据生成和红外系统分析的应用。
{"title":"A Unified Complex-Fresnel Model for Physically Based Long-Wave Infrared Imaging and Simulation.","authors":"Peter Ter Heerdt, William Keustermans, Ivan De Boi, Steve Vanlanduit","doi":"10.3390/jimaging12010033","DOIUrl":"10.3390/jimaging12010033","url":null,"abstract":"<p><p>Accurate modelling of reflection, transmission, absorption, and emission at material interfaces is essential for infrared imaging, rendering, and the simulation of optical and sensing systems. This need is particularly pronounced across the short-wave to long-wave infrared (SWIR-LWIR) spectrum, where many materials exhibit dispersion- and wavelength-dependent attenuation described by complex refractive indices. In this work, we introduce a unified formulation of the full Fresnel equations that directly incorporates wavelength-dependent complex refractive-index data and provides physically consistent interface behaviour for both dielectrics and conductors. The approach reformulates the classical Fresnel expressions to eliminate sign ambiguities and numerical instabilities, resulting in a stable evaluation across incidence angles and for strongly absorbing materials. We demonstrate the model through spectral-rendering simulations that illustrate realistic reflectance and transmittance behaviour for materials with different infrared optical properties. To assess its suitability for thermal-infrared applications, we also compare the simulated long-wave emission of a heated glass sphere with measurements from a LWIR camera. The agreement between measured and simulated radiometric trends indicates that the proposed formulation offers a practical and physically grounded tool for wavelength-parametric interface modelling in infrared imaging, supporting applications in spectral rendering, synthetic data generation, and infrared system analysis.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842791/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deep Learning-Assisted Autofocus for Aerial Cameras in Maritime Photography. 海洋摄影中航拍相机的深度学习辅助自动对焦。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-07 DOI: 10.3390/jimaging12010031
Haiying Liu, Yingchao Li, Shilong Xu, Haoyu Wang, Qiang Fu, Huilin Jiang

To address the unreliable autofocus problem of drone-mounted visible-light aerial cameras in low-contrast maritime environments, this paper proposes an autofocus system that combines deep-learning-based coarse focusing with traditional search-based fine adjustment. The system uses a built-in high-contrast resolution test chart as the signal source. Images captured by the imaging sensor are fed into a lightweight convolutional neural network to regress the defocus distance, enabling fast focus positioning. This avoids the weak signal and inaccurate focusing often encountered when adjusting focus directly on low-contrast sea surfaces. In the fine-focusing stage, a hybrid strategy integrating hill-climbing search and inverse correction is adopted. By evaluating the image sharpness function, the system accurately locks onto the optimal focal plane, forming intelligent closed-loop control. Experiments show that this method, which combines imaging of the built-in calibration target with deep-learning-based coarse focusing, significantly improves focusing efficiency. Compared with traditional full-range search strategies, the focusing speed is increased by approximately 60%. While ensuring high accuracy and strong adaptability, the proposed approach effectively enhances the overall imaging performance of aerial cameras in low-contrast maritime conditions.

针对无人机机载可见光航拍相机在低对比度海洋环境下自动对焦不可靠的问题,提出了一种基于深度学习的粗对焦与传统的基于搜索的微调相结合的自动对焦系统。系统采用内置的高对比度分辨率测试图作为信号源。成像传感器捕获的图像被输入到一个轻量级的卷积神经网络中,以回归离焦距离,从而实现快速对焦定位。这避免了在低对比度海面上直接调整焦距时经常遇到的信号弱和聚焦不准确的问题。在精细聚焦阶段,采用爬坡搜索和逆校正相结合的混合策略。通过评估图像清晰度函数,系统精确锁定最佳焦平面,形成智能闭环控制。实验表明,该方法将内置标定目标成像与基于深度学习的粗聚焦相结合,显著提高了聚焦效率。与传统的全范围搜索策略相比,聚焦速度提高了约60%。该方法在保证高精度和强适应性的同时,有效提高了航空相机在低对比度海上条件下的整体成像性能。
{"title":"Deep Learning-Assisted Autofocus for Aerial Cameras in Maritime Photography.","authors":"Haiying Liu, Yingchao Li, Shilong Xu, Haoyu Wang, Qiang Fu, Huilin Jiang","doi":"10.3390/jimaging12010031","DOIUrl":"10.3390/jimaging12010031","url":null,"abstract":"<p><p>To address the unreliable autofocus problem of drone-mounted visible-light aerial cameras in low-contrast maritime environments, this paper proposes an autofocus system that combines deep-learning-based coarse focusing with traditional search-based fine adjustment. The system uses a built-in high-contrast resolution test chart as the signal source. Images captured by the imaging sensor are fed into a lightweight convolutional neural network to regress the defocus distance, enabling fast focus positioning. This avoids the weak signal and inaccurate focusing often encountered when adjusting focus directly on low-contrast sea surfaces. In the fine-focusing stage, a hybrid strategy integrating hill-climbing search and inverse correction is adopted. By evaluating the image sharpness function, the system accurately locks onto the optimal focal plane, forming intelligent closed-loop control. Experiments show that this method, which combines imaging of the built-in calibration target with deep-learning-based coarse focusing, significantly improves focusing efficiency. Compared with traditional full-range search strategies, the focusing speed is increased by approximately 60%. While ensuring high accuracy and strong adaptability, the proposed approach effectively enhances the overall imaging performance of aerial cameras in low-contrast maritime conditions.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843086/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054231","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
DynMultiDep: A Dynamic Multimodal Fusion and Multi-Scale Time Series Modeling Approach for Depression Detection. 基于动态多模态融合和多尺度时间序列建模的抑郁症检测方法。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-06 DOI: 10.3390/jimaging12010029
Jincheng Li, Menglin Zheng, Jiongyi Yang, Yihui Zhan, Xing Xie

Depression is a prevalent mental disorder that imposes a significant public health burden worldwide. Although multimodal detection methods have shown potential, existing techniques still face two critical bottlenecks: (i) insufficient integration of global patterns and local fluctuations in long-sequence modeling and (ii) static fusion strategies that fail to dynamically adapt to the complementarity and redundancy among modalities. To address these challenges, this paper proposes a dynamic multimodal depression detection framework, DynMultiDep, which combines multi-scale temporal modeling with an adaptive fusion mechanism. The core innovations of DynMultiDep lie in its Multi-scale Temporal Experts Module (MTEM) and Dynamic Multimodal Fusion module (DynMM). On one hand, MTEM employs Mamba experts to extract long-term trend features and utilizes local-window Transformers to capture short-term dynamic fluctuations, achieving adaptive fusion through a long-short routing mechanism. On the other hand, DynMM introduces modality-level and fusion-level dynamic decision-making, selecting critical modality paths and optimizing cross-modal interaction strategies based on input characteristics. The experimental results demonstrate that DynMultiDep outperforms existing state-of-the-art methods in detection performance on two widely used large-scale depression datasets.

抑郁症是一种普遍存在的精神障碍,在全世界造成了重大的公共卫生负担。虽然多模态检测方法显示出潜力,但现有技术仍然面临两个关键瓶颈:(i)长序列建模中全球模式和局部波动的整合不足;(ii)静态融合策略未能动态适应模式之间的互补性和冗余性。为了解决这些问题,本文提出了一种动态多模态凹陷检测框架DynMultiDep,该框架将多尺度时间建模与自适应融合机制相结合。dynmultideep的核心创新在于其多尺度时间专家模块(MTEM)和动态多模态融合模块(DynMM)。MTEM一方面利用曼巴专家提取长期趋势特征,利用局部窗口变压器捕捉短期动态波动,通过长-短路由机制实现自适应融合。另一方面,DynMM引入了模态级和融合级动态决策,选择关键模态路径,并基于输入特性优化跨模态交互策略。实验结果表明,DynMultiDep在两个广泛使用的大型洼地数据集上的检测性能优于现有的最先进的方法。
{"title":"DynMultiDep: A Dynamic Multimodal Fusion and Multi-Scale Time Series Modeling Approach for Depression Detection.","authors":"Jincheng Li, Menglin Zheng, Jiongyi Yang, Yihui Zhan, Xing Xie","doi":"10.3390/jimaging12010029","DOIUrl":"10.3390/jimaging12010029","url":null,"abstract":"<p><p>Depression is a prevalent mental disorder that imposes a significant public health burden worldwide. Although multimodal detection methods have shown potential, existing techniques still face two critical bottlenecks: (i) insufficient integration of global patterns and local fluctuations in long-sequence modeling and (ii) static fusion strategies that fail to dynamically adapt to the complementarity and redundancy among modalities. To address these challenges, this paper proposes a dynamic multimodal depression detection framework, DynMultiDep, which combines multi-scale temporal modeling with an adaptive fusion mechanism. The core innovations of DynMultiDep lie in its Multi-scale Temporal Experts Module (MTEM) and Dynamic Multimodal Fusion module (DynMM). On one hand, MTEM employs Mamba experts to extract long-term trend features and utilizes local-window Transformers to capture short-term dynamic fluctuations, achieving adaptive fusion through a long-short routing mechanism. On the other hand, DynMM introduces modality-level and fusion-level dynamic decision-making, selecting critical modality paths and optimizing cross-modal interaction strategies based on input characteristics. The experimental results demonstrate that DynMultiDep outperforms existing state-of-the-art methods in detection performance on two widely used large-scale depression datasets.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12842938/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054173","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ultrashort Echo Time Quantitative Susceptibility Source Separation in Musculoskeletal System: A Feasibility Study. 肌肉骨骼系统中超短回波时间定量敏感源分离的可行性研究。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-06 DOI: 10.3390/jimaging12010028
Sam Sedaghat, Jin Il Park, Eddie Fu, Annette von Drygalski, Yajun Ma, Eric Y Chang, Jiang Du, Lorenzo Nardo, Hyungseok Jang

This study aims to demonstrate the feasibility of ultrashort echo time (UTE)-based susceptibility source separation for musculoskeletal (MSK) imaging, enabling discrimination between diamagnetic and paramagnetic tissue components, with a particular focus on hemophilic arthropathy (HA). Three key techniques were integrated to achieve UTE-based susceptibility source separation: Iterative decomposition of water and fat with echo asymmetry and least-squares estimation for B0 field estimation, projection onto dipole fields for local field mapping, and χ-separation for quantitative susceptibility mapping (QSM) with source decomposition. A phantom containing varying concentrations of diamagnetic (CaCO3) and paramagnetic (Fe3O4) materials was used to validate the method. In addition, in vivo UTE-QSM scans of the knees and ankles were performed on five HA patients using a 3T clinical MRI scanner. In the phantom, conventional QSM underestimated susceptibility values due to the mixed-source cancelling the effect. In contrast, source-separated maps provided distinct diamagnetic and paramagnetic susceptibility values that correlated strongly with CaCO3 and Fe3O4 concentrations (r = -0.99 and 0.95, p < 0.05). In vivo, paramagnetic maps enabled improved visualization of hemosiderin deposits in joints of HA patients, which were poorly visualized or obscured in conventional QSM due to susceptibility cancellation by surrounding diamagnetic tissues such as bone. This study demonstrates, for the first time, the feasibility of UTE-based quantitative susceptibility source separation for MSK applications. The approach enhances the detection of paramagnetic substances like hemosiderin in HA and offers potential for improved assessment of bone and joint tissue composition.

本研究旨在证明基于超短回波时间(UTE)的敏感性源分离用于肌肉骨骼(MSK)成像的可行性,实现抗磁性和顺磁性组织成分的区分,特别关注血友病关节病(HA)。结合3个关键技术实现基于ute的敏感性源分离:基于回波不对称的水和脂肪迭代分解和最小二乘估计的B0场估计,投影到偶极子场的局部场映射,以及基于源分解的χ-分离的定量敏感性图(QSM)。使用含有不同浓度的抗磁性(CaCO3)和顺磁性(Fe3O4)材料的模型来验证该方法。此外,使用3T临床MRI扫描仪对5例HA患者的膝关节和脚踝进行了体内UTE-QSM扫描。在模体中,由于混合源抵消了效应,传统的QSM低估了磁化率值。相比之下,源分离图提供了明显的抗磁性和顺磁性值,与CaCO3和Fe3O4浓度密切相关(r = -0.99和0.95,p < 0.05)。在体内,顺磁图可以改善HA患者关节中含铁血黄素沉积的可视化,而在传统的QSM中,由于周围的抗磁性组织(如骨)的易感性抵消,这些沉积的可视化效果很差或模糊不清。该研究首次证明了基于ute的定量磁化率源分离在MSK应用中的可行性。该方法增强了血凝素中含铁血黄素等顺磁性物质的检测,并为改进骨和关节组织成分的评估提供了潜力。
{"title":"Ultrashort Echo Time Quantitative Susceptibility Source Separation in Musculoskeletal System: A Feasibility Study.","authors":"Sam Sedaghat, Jin Il Park, Eddie Fu, Annette von Drygalski, Yajun Ma, Eric Y Chang, Jiang Du, Lorenzo Nardo, Hyungseok Jang","doi":"10.3390/jimaging12010028","DOIUrl":"10.3390/jimaging12010028","url":null,"abstract":"<p><p>This study aims to demonstrate the feasibility of ultrashort echo time (UTE)-based susceptibility source separation for musculoskeletal (MSK) imaging, enabling discrimination between diamagnetic and paramagnetic tissue components, with a particular focus on hemophilic arthropathy (HA). Three key techniques were integrated to achieve UTE-based susceptibility source separation: Iterative decomposition of water and fat with echo asymmetry and least-squares estimation for B0 field estimation, projection onto dipole fields for local field mapping, and χ-separation for quantitative susceptibility mapping (QSM) with source decomposition. A phantom containing varying concentrations of diamagnetic (CaCO<sub>3</sub>) and paramagnetic (Fe<sub>3</sub>O<sub>4</sub>) materials was used to validate the method. In addition, in vivo UTE-QSM scans of the knees and ankles were performed on five HA patients using a 3T clinical MRI scanner. In the phantom, conventional QSM underestimated susceptibility values due to the mixed-source cancelling the effect. In contrast, source-separated maps provided distinct diamagnetic and paramagnetic susceptibility values that correlated strongly with CaCO<sub>3</sub> and Fe<sub>3</sub>O<sub>4</sub> concentrations (r = -0.99 and 0.95, <i>p</i> < 0.05). In vivo, paramagnetic maps enabled improved visualization of hemosiderin deposits in joints of HA patients, which were poorly visualized or obscured in conventional QSM due to susceptibility cancellation by surrounding diamagnetic tissues such as bone. This study demonstrates, for the first time, the feasibility of UTE-based quantitative susceptibility source separation for MSK applications. The approach enhances the detection of paramagnetic substances like hemosiderin in HA and offers potential for improved assessment of bone and joint tissue composition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843048/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling. 使用可学习的gababoraware纹理建模的高保真三维人脸重建的分层多分辨率自监督框架。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-05 DOI: 10.3390/jimaging12010026
Pichet Mareo, Rerkchai Fooprateepsiri

High-fidelity 3D face reconstruction from a single image is challenging, owing to the inherently ambiguous depth cues and the strong entanglement of multi-scale facial textures. In this regard, we propose a hierarchical multi-resolution self-supervised framework (HMR-Framework), which reconstructs coarse-, medium-, and fine-scale facial geometry progressively through a unified pipeline. A coarse geometric prior is first estimated via 3D morphable model regression, followed by medium-scale refinement using a vertex deformation map constrained by a global-local Markov random field loss to preserve structural coherence. In order to improve fine-scale fidelity, a learnable Gabor-aware texture enhancement module has been proposed to decouple spatial-frequency information and thus improve sensitivity for high-frequency facial attributes. Additionally, we employ a wavelet-based detail perception loss to preserve the edge-aware texture features while mitigating noise commonly observed in in-the-wild images. Extensive qualitative and quantitative evaluation of benchmark datasets indicate that the proposed framework provides better fine-detail reconstruction than existing state-of-the-art methods, while maintaining robustness over pose variations. Notably, the hierarchical design increases semantic consistency across multiple geometric scales, providing a functional solution for high-fidelity 3D face reconstruction from monocular images.

单幅图像的高保真三维人脸重建具有一定的挑战性,这主要是由于图像本身具有模糊的深度线索和多尺度面部纹理的强烈纠缠。在这方面,我们提出了一种分层多分辨率自监督框架(HMR-Framework),该框架通过统一的管道逐步重建粗、中、细尺度的面部几何。首先通过3D变形模型回归估计粗糙的几何先验,然后使用全局局部马尔可夫随机场损失约束的顶点变形映射进行中等尺度细化,以保持结构一致性。为了提高精细尺度的保真度,提出了一种可学习的gabor感知纹理增强模块来解耦空间频率信息,从而提高对高频面部属性的灵敏度。此外,我们采用基于小波的细节感知损失来保留边缘感知纹理特征,同时减轻在野外图像中常见的噪声。对基准数据集的广泛定性和定量评估表明,所提出的框架比现有的最先进的方法提供更好的细节重建,同时保持对姿态变化的鲁棒性。值得注意的是,分层设计增加了跨多个几何尺度的语义一致性,为单眼图像的高保真3D人脸重建提供了功能解决方案。
{"title":"A Hierarchical Multi-Resolution Self-Supervised Framework for High-Fidelity 3D Face Reconstruction Using Learnable Gabor-Aware Texture Modeling.","authors":"Pichet Mareo, Rerkchai Fooprateepsiri","doi":"10.3390/jimaging12010026","DOIUrl":"10.3390/jimaging12010026","url":null,"abstract":"<p><p>High-fidelity 3D face reconstruction from a single image is challenging, owing to the inherently ambiguous depth cues and the strong entanglement of multi-scale facial textures. In this regard, we propose a hierarchical multi-resolution self-supervised framework (HMR-Framework), which reconstructs coarse-, medium-, and fine-scale facial geometry progressively through a unified pipeline. A coarse geometric prior is first estimated via 3D morphable model regression, followed by medium-scale refinement using a vertex deformation map constrained by a global-local Markov random field loss to preserve structural coherence. In order to improve fine-scale fidelity, a learnable Gabor-aware texture enhancement module has been proposed to decouple spatial-frequency information and thus improve sensitivity for high-frequency facial attributes. Additionally, we employ a wavelet-based detail perception loss to preserve the edge-aware texture features while mitigating noise commonly observed in in-the-wild images. Extensive qualitative and quantitative evaluation of benchmark datasets indicate that the proposed framework provides better fine-detail reconstruction than existing state-of-the-art methods, while maintaining robustness over pose variations. Notably, the hierarchical design increases semantic consistency across multiple geometric scales, providing a functional solution for high-fidelity 3D face reconstruction from monocular images.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843383/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053813","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Vision-Based People Counting and Tracking for Urban Environments. 基于视觉的城市环境人口统计与跟踪。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-05 DOI: 10.3390/jimaging12010027
Daniyar Nurseitov, Kairat Bostanbekov, Nazgul Toiganbayeva, Aidana Zhalgas, Didar Yedilkhan, Beibut Amirgaliyev

Population growth and expansion of urban areas increase the need for the introduction of intelligent passenger traffic monitoring systems. Accurate estimation of the number of passengers is an important condition for improving the efficiency, safety and quality of transport services. This paper proposes an approach to the automatic detection and counting of people using computer vision and deep learning methods. While YOLOv8 and DeepSORT have been widely explored individually, our contribution lies in a task-specific modification of the DeepSORT tracking pipeline, optimized for dense passenger environments, strong occlusions, and dynamic lighting, as well as in a unified architecture that integrates detection, tracking, and automatic event-log generation. Our new proprietary dataset of 4047 images and 8918 labeled objects has achieved 92% detection accuracy and 85% counting accuracy, which confirms the effectiveness of the solution. Compared to Mask R-CNN and DETR, the YOLOv8 model demonstrates an optimal balance between speed, accuracy, and computational efficiency. The results confirm that computer vision can become an efficient and scalable replacement for traditional sensory passenger counting systems. The developed architecture (YOLO + Tracking) combines recognition, tracking and counting of people into a single system that automatically generates annotated video streams and event logs. In the future, it is planned to expand the dataset, introduce support for multicamera integration, and adapt the model for embedded devices to improve the accuracy and energy efficiency of the solution in real-world conditions.

人口增长和城市面积的扩大增加了引入智能客运交通监控系统的需求。准确估计客运量是提高运输服务效率、安全性和质量的重要条件。本文提出了一种利用计算机视觉和深度学习方法对人进行自动检测和计数的方法。虽然YOLOv8和DeepSORT已经被广泛地单独探索,但我们的贡献在于对DeepSORT跟踪管道的特定任务修改,针对密集的乘客环境,强遮挡和动态照明进行了优化,以及集成了检测,跟踪和自动事件日志生成的统一架构。我们的新专有数据集包含4047张图像和8918个标记对象,检测准确率达到92%,计数准确率达到85%,证实了该解决方案的有效性。与Mask R-CNN和DETR相比,YOLOv8模型在速度、精度和计算效率之间取得了最佳平衡。结果证实,计算机视觉可以成为传统感官乘客计数系统的有效和可扩展的替代品。开发的体系结构(YOLO + Tracking)将人员识别、跟踪和计数结合到一个系统中,该系统自动生成带注释的视频流和事件日志。未来,计划扩展数据集,引入对多摄像机集成的支持,并使模型适应嵌入式设备,以提高解决方案在现实条件下的精度和能效。
{"title":"Vision-Based People Counting and Tracking for Urban Environments.","authors":"Daniyar Nurseitov, Kairat Bostanbekov, Nazgul Toiganbayeva, Aidana Zhalgas, Didar Yedilkhan, Beibut Amirgaliyev","doi":"10.3390/jimaging12010027","DOIUrl":"10.3390/jimaging12010027","url":null,"abstract":"<p><p>Population growth and expansion of urban areas increase the need for the introduction of intelligent passenger traffic monitoring systems. Accurate estimation of the number of passengers is an important condition for improving the efficiency, safety and quality of transport services. This paper proposes an approach to the automatic detection and counting of people using computer vision and deep learning methods. While YOLOv8 and DeepSORT have been widely explored individually, our contribution lies in a task-specific modification of the DeepSORT tracking pipeline, optimized for dense passenger environments, strong occlusions, and dynamic lighting, as well as in a unified architecture that integrates detection, tracking, and automatic event-log generation. Our new proprietary dataset of 4047 images and 8918 labeled objects has achieved 92% detection accuracy and 85% counting accuracy, which confirms the effectiveness of the solution. Compared to Mask R-CNN and DETR, the YOLOv8 model demonstrates an optimal balance between speed, accuracy, and computational efficiency. The results confirm that computer vision can become an efficient and scalable replacement for traditional sensory passenger counting systems. The developed architecture (YOLO + Tracking) combines recognition, tracking and counting of people into a single system that automatically generates annotated video streams and event logs. In the future, it is planned to expand the dataset, introduce support for multicamera integration, and adapt the model for embedded devices to improve the accuracy and energy efficiency of the solution in real-world conditions.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843365/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054264","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Slicer-Independent Framework for Measuring G-Code Accuracy in Medical 3D Printing. 用于测量医疗3D打印中g代码精度的切片机独立框架。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-04 DOI: 10.3390/jimaging12010025
Michel Beyer, Alexandru Burde, Andreas E Roser, Maximiliane Beyer, Sead Abazi, Florian M Thieringer

In medical 3D printing, accuracy is critical for fabricating patient-specific implants and anatomical models. Although printer performance has been widely examined, the influence of slicing software on geometric fidelity is less frequently quantified. The slicing step, which converts STL files into printer-readable G-code, may introduce deviations that affect the final printed object. To quantify slicer-induced G-code deviations by comparing G-code-derived geometries with their reference STL modelsTwenty mandibular models were processed using five slicers (PrusaSlicer (version 2.9.1.), Cura (version 5.2.2.), Simplify3D (version 4.1.2.), Slic3r (version 1.3.0.) and Fusion 360 (version 2.0.19725)). A custom Python workflow converted the G-code into point clouds and reconstructed STL meshes through XY and Z corrections, marching cubes surface extraction, and volumetric extrusion. A calibration object enabled coordinate normalization across slicers. Accuracy was assessed using Mean Surface Distance (MSD), Root Mean Square (RMS) deviation, and Volume Difference. MSD ranged from 0.071 to 0.095 mm, and RMS deviation from 0.084 to 0.113 mm, depending on the slicer. Volumetric differences were slicer-dependent. PrusaSlicer yielded the highest surface accuracy; Simplify3D and Slic3r showed best repeatability. Fusion 360 produced the largest deviations. The slicers introduced geometric deviations below 0.1 mm that represent a substantial proportion of the overall error in the FDM workflow.

在医疗3D打印中,准确性对于制造患者特定植入物和解剖模型至关重要。虽然打印机的性能已经被广泛研究,切片软件对几何保真度的影响很少被量化。切片步骤将STL文件转换为打印机可读的g代码,可能会引入影响最终打印对象的偏差。为了通过比较g码导出的几何形状与参考的STL模型来量化切片机诱导的g码偏差,使用5种切片机(PrusaSlicer(2.9.1.)、Cura(5.2.2.)、Simplify3D(4.1.2.)、Slic3r(1.3.0.)和Fusion 360(2.0.19725 .))对20个下颌模型进行处理。自定义Python工作流将g代码转换为点云,并通过XY和Z校正、移动立方体表面提取和体积挤压来重建STL网格。校准对象支持跨切片器的坐标规范化。使用平均表面距离(MSD)、均方根(RMS)偏差和体积差来评估准确性。MSD范围为0.071至0.095 mm, RMS偏差范围为0.084至0.113 mm,具体取决于切片机。体积差异与切片机有关。普鲁士激光产生了最高的表面精度;Simplify3D和Slic3r的重复性最好。Fusion 360产生的偏差最大。切片机引入了小于0.1 mm的几何偏差,这代表了FDM工作流程中总体误差的很大一部分。
{"title":"A Slicer-Independent Framework for Measuring G-Code Accuracy in Medical 3D Printing.","authors":"Michel Beyer, Alexandru Burde, Andreas E Roser, Maximiliane Beyer, Sead Abazi, Florian M Thieringer","doi":"10.3390/jimaging12010025","DOIUrl":"10.3390/jimaging12010025","url":null,"abstract":"<p><p>In medical 3D printing, accuracy is critical for fabricating patient-specific implants and anatomical models. Although printer performance has been widely examined, the influence of slicing software on geometric fidelity is less frequently quantified. The slicing step, which converts STL files into printer-readable G-code, may introduce deviations that affect the final printed object. To quantify slicer-induced G-code deviations by comparing G-code-derived geometries with their reference STL modelsTwenty mandibular models were processed using five slicers (PrusaSlicer (version 2.9.1.), Cura (version 5.2.2.), Simplify3D (version 4.1.2.), Slic3r (version 1.3.0.) and Fusion 360 (version 2.0.19725)). A custom Python workflow converted the G-code into point clouds and reconstructed STL meshes through XY and Z corrections, marching cubes surface extraction, and volumetric extrusion. A calibration object enabled coordinate normalization across slicers. Accuracy was assessed using Mean Surface Distance (MSD), Root Mean Square (RMS) deviation, and Volume Difference. MSD ranged from 0.071 to 0.095 mm, and RMS deviation from 0.084 to 0.113 mm, depending on the slicer. Volumetric differences were slicer-dependent. PrusaSlicer yielded the highest surface accuracy; Simplify3D and Slic3r showed best repeatability. Fusion 360 produced the largest deviations. The slicers introduced geometric deviations below 0.1 mm that represent a substantial proportion of the overall error in the FDM workflow.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843157/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146053902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
State of the Art of Remote Sensing Data: Gradient Pattern in Pseudocolor Composite Images. 遥感数据的研究现状:伪彩色合成图像中的梯度模式。
IF 2.7 Q3 IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY Pub Date : 2026-01-04 DOI: 10.3390/jimaging12010023
Alexey Terekhov, Ravil I Mukhamediev, Igor Savin

The thematic processing of pseudocolor composite images, especially those created from remote sensing data, is of considerable interest. The set of spectral classes comprising such images is typically described by a nominal scale, meaning the absence of any predetermined relationships between the classes. However, in many cases, images of this type may contain elements of a regular spatial order, one variant of which is a gradient structure. Gradient structures are characterized by a certain regular spatial ordering of spectral classes. Recognizing gradient patterns in the structure of pseudocolor composite images opens up new possibilities for deeper thematic images processing. This article describes an algorithm for analyzing the spatial structure of a pseudocolor composite image to identify gradient patterns. In this process, the initial nominal scale of spectral classes is transformed into a rank scale of the gradient legend. The algorithm is based on the analysis of Moore neighborhoods for each image pixel. This creates an array of the prevalence of all types of local binary patterns (the pixel's nearest neighbors). All possible variants of the spectral class rank scale composition are then considered. The rank scale variant that describes the largest proportion of image pixels within its gradient order is used as a final result. The user can independently define the criteria for the significance of the gradient order in the analyzed image, focusing either on the overall statistics of the proportion of pixels consistent with the spatial structure of the selected gradient or on the statistics of a selected key image region. The proposed algorithm is illustrated using analysis of test examples.

伪彩色合成图像的专题处理,特别是从遥感数据产生的伪彩色合成图像的专题处理,引起了相当大的兴趣。包含这些图像的光谱类别集合通常用标称尺度来描述,这意味着类别之间没有任何预定的关系。然而,在许多情况下,这种类型的图像可能包含规则空间秩序的元素,其中一种变体是梯度结构。梯度结构的特征是谱类具有一定的规则空间顺序。识别伪彩色合成图像结构中的梯度模式为更深层次的主题图像处理开辟了新的可能性。本文描述了一种用于分析伪彩色合成图像的空间结构以识别渐变模式的算法。在此过程中,光谱类的初始标称尺度被转换为梯度图例的等级尺度。该算法基于对每个图像像素的摩尔邻域分析。这将创建一个包含所有类型的本地二进制模式(像素的最近邻居)的流行度的数组。然后考虑光谱类等级尺度组成的所有可能变体。描述其梯度顺序内图像像素的最大比例的等级尺度变量被用作最终结果。用户可以独立定义被分析图像中梯度顺序显著性的标准,关注与所选梯度空间结构一致的像素比例的总体统计,或者关注所选关键图像区域的统计。通过对测试实例的分析,说明了该算法的有效性。
{"title":"State of the Art of Remote Sensing Data: Gradient Pattern in Pseudocolor Composite Images.","authors":"Alexey Terekhov, Ravil I Mukhamediev, Igor Savin","doi":"10.3390/jimaging12010023","DOIUrl":"10.3390/jimaging12010023","url":null,"abstract":"<p><p>The thematic processing of pseudocolor composite images, especially those created from remote sensing data, is of considerable interest. The set of spectral classes comprising such images is typically described by a nominal scale, meaning the absence of any predetermined relationships between the classes. However, in many cases, images of this type may contain elements of a regular spatial order, one variant of which is a gradient structure. Gradient structures are characterized by a certain regular spatial ordering of spectral classes. Recognizing gradient patterns in the structure of pseudocolor composite images opens up new possibilities for deeper thematic images processing. This article describes an algorithm for analyzing the spatial structure of a pseudocolor composite image to identify gradient patterns. In this process, the initial nominal scale of spectral classes is transformed into a rank scale of the gradient legend. The algorithm is based on the analysis of Moore neighborhoods for each image pixel. This creates an array of the prevalence of all types of local binary patterns (the pixel's nearest neighbors). All possible variants of the spectral class rank scale composition are then considered. The rank scale variant that describes the largest proportion of image pixels within its gradient order is used as a final result. The user can independently define the criteria for the significance of the gradient order in the analyzed image, focusing either on the overall statistics of the proportion of pixels consistent with the spatial structure of the selected gradient or on the statistics of a selected key image region. The proposed algorithm is illustrated using analysis of test examples.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"12 1","pages":""},"PeriodicalIF":2.7,"publicationDate":"2026-01-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12843327/pdf/","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146054256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Journal of Imaging
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1