Pub Date : 2025-08-28DOI: 10.1109/TCI.2025.3603689
Edward P. Chandler;Shirin Shoushtari;Brendt Wohlberg;Ulugbek S. Kamilov
Total variation (TV) is a widely used function for regularizing imaging inverse problems that is particularly appropriate for images whose underlying structure is piecewise constant. TV regularized optimization problems are typically solved using proximal methods, but the way in which they are applied is constrained by the absence of a closed-form expression for the proximal operator of the TV function. A closed-form approximation of the TV proximal operator has previously been proposed, but its accuracy was not theoretically explored in detail. We address this gap by making several new theoretical contributions, proving that the approximation leads to a proximal operator of some convex function, it is equivalent to a gradient descent step on a smoothed version of TV, and that its error can be fully characterized and controlled with its scaling parameter. We experimentally validate our theoretical results on image denoising and sparse-view computed tomography (CT) image reconstruction.
{"title":"Closed-Form Approximation of the Total Variation Proximal Operator","authors":"Edward P. Chandler;Shirin Shoushtari;Brendt Wohlberg;Ulugbek S. Kamilov","doi":"10.1109/TCI.2025.3603689","DOIUrl":"https://doi.org/10.1109/TCI.2025.3603689","url":null,"abstract":"Total variation (TV) is a widely used function for regularizing imaging inverse problems that is particularly appropriate for images whose underlying structure is piecewise constant. TV regularized optimization problems are typically solved using proximal methods, but the way in which they are applied is constrained by the absence of a closed-form expression for the proximal operator of the TV function. A closed-form approximation of the TV proximal operator has previously been proposed, but its accuracy was not theoretically explored in detail. We address this gap by making several new theoretical contributions, proving that the approximation leads to a proximal operator of some convex function, it is equivalent to a gradient descent step on a smoothed version of TV, and that its error can be fully characterized and controlled with its scaling parameter. We experimentally validate our theoretical results on image denoising and sparse-view computed tomography (CT) image reconstruction.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1217-1228"},"PeriodicalIF":4.8,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090148","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-28DOI: 10.1109/TCI.2025.3603741
Emmanuel Soubies;Wolfgang Bacsa
As opposed to popular far-field and near-field optical microscopy techniques, Interference Scanning Optical probe Microscopy (ISOM) operates in the intermediate-field region, where the probing distance is typically of the order of the wavelength of incident light. Specifically, ISOM enables the imaging of nanostructures through numerical inverse scattering of standing waves generated by the interference between the incident (or reflected) and scattered waves. In this work, we shed new light on this microscopy modality through an in-depth Fourier analysis. Our analysis reveals insights on the required acquisition sampling step as well as on the resolution limit of the system. Moreover, we propose two novel methods to address the associated inverse scattering problem, leveraging the intrinsic structure of the image formation model to reduce computational complexity and sensitivity to errors in model parameters. Finally, we illustrate our theoretical findings with numerical experiments.
{"title":"Fourier Analysis of Interference Scanning Optical Probe Microscopy","authors":"Emmanuel Soubies;Wolfgang Bacsa","doi":"10.1109/TCI.2025.3603741","DOIUrl":"https://doi.org/10.1109/TCI.2025.3603741","url":null,"abstract":"As opposed to popular far-field and near-field optical microscopy techniques, Interference Scanning Optical probe Microscopy (ISOM) operates in the intermediate-field region, where the probing distance is typically of the order of the wavelength of incident light. Specifically, ISOM enables the imaging of nanostructures through numerical inverse scattering of standing waves generated by the interference between the incident (or reflected) and scattered waves. In this work, we shed new light on this microscopy modality through an in-depth Fourier analysis. Our analysis reveals insights on the required acquisition sampling step as well as on the resolution limit of the system. Moreover, we propose two novel methods to address the associated inverse scattering problem, leveraging the intrinsic structure of the image formation model to reduce computational complexity and sensitivity to errors in model parameters. Finally, we illustrate our theoretical findings with numerical experiments.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1206-1216"},"PeriodicalIF":4.8,"publicationDate":"2025-08-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145036203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-25DOI: 10.1109/TCI.2025.3602315
Yunfei Li;Qian Liu;Fuhong Cai
Biological tissue optics has garnered significant attention in biomedical research for its non-destructive, high-sensitivity nature. However, the scattering and absorption properties of biological tissues fundamentally limit the penetration depth of optical imaging. Fluorescence molecular tomography (FMT) offers a solution balancing imaging depth and resolution, yet tissue scattering and absorption continue to challenge depth-resolved reconstruction accuracy. This study develops a sensitive near-infrared II (NIR-II) hyperspectral imaging system to investigate the relationship between fluorescence penetration depth and tissue absorption/scattering coefficients. By leveraging the strong water absorption peak around 1450 nm, we strategically divide the reconstruction object into layers within the FMT model, significantly improving the ill-posed inverse problem. We then utilize hyperspectral data to select wavelengths with progressively decreasing absorption coefficients relative to the 1450 nm peak. This enables layer-by-layer 3D reconstruction of deep biological tissues, overcoming the limitations of conventional FMT. Our method demonstrates single-perspective FMT reconstruction capable of resolving heterogeneous targets at 10 mm depth with a 0.74 Dice coefficient in depth discrimination. This spectraldimension-enhanced FMT method enables accurate 3D reconstruction from single-view measurements. By exploiting the depth-dependent light-tissue interactions at selected NIR-II wavelengths, our approach achieves imaging quality comparable to multi-angle systems while simplifying the experimental setup. Both simulation and phantom experiments demonstrate precise target localization and shape recovery, suggesting promising potential for small animal imaging applications where system complexity and acquisition speed are critical.
{"title":"Single-View Fluorescence Molecular Tomography Based on Hyperspectral NIR-II Imaging","authors":"Yunfei Li;Qian Liu;Fuhong Cai","doi":"10.1109/TCI.2025.3602315","DOIUrl":"https://doi.org/10.1109/TCI.2025.3602315","url":null,"abstract":"Biological tissue optics has garnered significant attention in biomedical research for its non-destructive, high-sensitivity nature. However, the scattering and absorption properties of biological tissues fundamentally limit the penetration depth of optical imaging. Fluorescence molecular tomography (FMT) offers a solution balancing imaging depth and resolution, yet tissue scattering and absorption continue to challenge depth-resolved reconstruction accuracy. This study develops a sensitive near-infrared II (NIR-II) hyperspectral imaging system to investigate the relationship between fluorescence penetration depth and tissue absorption/scattering coefficients. By leveraging the strong water absorption peak around 1450 nm, we strategically divide the reconstruction object into layers within the FMT model, significantly improving the ill-posed inverse problem. We then utilize hyperspectral data to select wavelengths with progressively decreasing absorption coefficients relative to the 1450 nm peak. This enables layer-by-layer 3D reconstruction of deep biological tissues, overcoming the limitations of conventional FMT. Our method demonstrates single-perspective FMT reconstruction capable of resolving heterogeneous targets at 10 mm depth with a 0.74 Dice coefficient in depth discrimination. This spectraldimension-enhanced FMT method enables accurate 3D reconstruction from single-view measurements. By exploiting the depth-dependent light-tissue interactions at selected NIR-II wavelengths, our approach achieves imaging quality comparable to multi-angle systems while simplifying the experimental setup. Both simulation and phantom experiments demonstrate precise target localization and shape recovery, suggesting promising potential for small animal imaging applications where system complexity and acquisition speed are critical.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1161-1173"},"PeriodicalIF":4.8,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144998033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-20DOI: 10.1109/TCI.2025.3598440
Xinhui Xue;Hai-Miao Hu;Zhuang He;Haowen Zheng
Color constancy seeks to keep the perceived color of objects consistent under varying illumination conditions. However, existing methods often rely on restrictive prior assumptions or suffer from limited generalization capability, posing significant challenges in complex scenes with multiple light sources. In this paper, we propose a neural network-enhanced, physics-based approach to multi-illuminant color constancy that leverages spectral imaging—highly sensitive to illumination variation. First, we analyze the physical image-formation process under mixed lighting and introduce a master–subordinate illumination model, extending conventional correlated-color-temperature re-illumination techniques. Our neural network framework explicitly models the correlation between narrow-band spectral reflectance and the spectral power distribution (SPD) of the illumination, enabling accurate recovery of the scene light’s full SPD. Using this model, we fuse RGB images with the estimated illumination spectra to predict illuminant chromaticity precisely, then correct image colors to a standard reference light. Extensive experiments on synthetic multi–color-temperature datasets and real-world spectral datasets demonstrate that our neural network-based method achieves state-of-the-art accuracy in spectral estimation and color-constancy correction.
{"title":"Towards Multi-Source Illumination Color Constancy Through Physics-Based Rendering and Spectral Power Distribution Embedding","authors":"Xinhui Xue;Hai-Miao Hu;Zhuang He;Haowen Zheng","doi":"10.1109/TCI.2025.3598440","DOIUrl":"https://doi.org/10.1109/TCI.2025.3598440","url":null,"abstract":"Color constancy seeks to keep the perceived color of objects consistent under varying illumination conditions. However, existing methods often rely on restrictive prior assumptions or suffer from limited generalization capability, posing significant challenges in complex scenes with multiple light sources. In this paper, we propose a neural network-enhanced, physics-based approach to multi-illuminant color constancy that leverages spectral imaging—highly sensitive to illumination variation. First, we analyze the physical image-formation process under mixed lighting and introduce a master–subordinate illumination model, extending conventional correlated-color-temperature re-illumination techniques. Our neural network framework explicitly models the correlation between narrow-band spectral reflectance and the spectral power distribution (SPD) of the illumination, enabling accurate recovery of the scene light’s full SPD. Using this model, we fuse RGB images with the estimated illumination spectra to predict illuminant chromaticity precisely, then correct image colors to a standard reference light. Extensive experiments on synthetic multi–color-temperature datasets and real-world spectral datasets demonstrate that our neural network-based method achieves state-of-the-art accuracy in spectral estimation and color-constancy correction.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1349-1360"},"PeriodicalIF":4.8,"publicationDate":"2025-08-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145255921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
A sparse camera array captures multiple images of a scene within the same spatial plane, enabling super-resolution reconstruction. However, existing methods often fail to fully exploit time as an additional dimension for enhanced information acquisition. Even when temporal and spatial observations are collected simultaneously, their individual contributions are often conflated. Analysis of the system’s imaging model reveals that the spatiotemporal camera system, integrating a camera array with video sequences, holds greater potential for degradation recovery. Based on these insights, we propose a novel multi-video super-resolution network for spatiotemporal information fusion. Guided by explicit physical dimensional orientation, the network effectively integrates spatial information and propagates it along the temporal dimension. By utilizing diverse and informative spatiotemporal sampling, our method more readily addresses challenges arising from ill-posed mapping matrices during reconstruction. Experimental results on both synthetic and real-world datasets show that the components of our network, with information fully propagated and spatiotemporally fused, work synergistically to enhance super-resolution performance, providing substantial improvements over state-of-the-art methods. We believe our study can inspire innovations for future super-resolution tasks by optimizing information acquisition and utilization.
{"title":"Multi-Video Super-Resolution: Spatiotemporal Fusion for Sparse Camera Array","authors":"Xudong Liu;Tianren Li;Yu Zhang;Yufu Qu;Zhenzhong Wei","doi":"10.1109/TCI.2025.3599774","DOIUrl":"https://doi.org/10.1109/TCI.2025.3599774","url":null,"abstract":"A sparse camera array captures multiple images of a scene within the same spatial plane, enabling super-resolution reconstruction. However, existing methods often fail to fully exploit time as an additional dimension for enhanced information acquisition. Even when temporal and spatial observations are collected simultaneously, their individual contributions are often conflated. Analysis of the system’s imaging model reveals that the spatiotemporal camera system, integrating a camera array with video sequences, holds greater potential for degradation recovery. Based on these insights, we propose a novel multi-video super-resolution network for spatiotemporal information fusion. Guided by explicit physical dimensional orientation, the network effectively integrates spatial information and propagates it along the temporal dimension. By utilizing diverse and informative spatiotemporal sampling, our method more readily addresses challenges arising from ill-posed mapping matrices during reconstruction. Experimental results on both synthetic and real-world datasets show that the components of our network, with information fully propagated and spatiotemporally fused, work synergistically to enhance super-resolution performance, providing substantial improvements over state-of-the-art methods. We believe our study can inspire innovations for future super-resolution tasks by optimizing information acquisition and utilization.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1087-1098"},"PeriodicalIF":4.8,"publicationDate":"2025-08-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144909295","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-13DOI: 10.1109/TCI.2025.3598421
Zi Wang;Jiahao Huang;Mingkai Huang;Chengyan Wang;Guang Yang;Xiaobo Qu
Accelerated dynamic magnetic resonance imaging (MRI) is highly expected in clinical applications. However, its reconstruction remains challenging due to the inherently high dimensionality and spatiotemporal complexity. While diffusion models have demonstrated robust performance in spatial imaging, their application to spatiotemporal data has been underexplored. To address this gap, we propose a novel spatiotemporal diffusion model (STDM) specifically designed for robust dynamic MRI reconstruction. Our approach decomposes the complex 3D diffusion process into manageable sub-problems by focusing on 2D spatiotemporal images, thereby reducing dimensionality and enhancing computational efficiency. Each 2D image is treated independently, allowing for a parallel reverse diffusion process guided by data consistency to ensure measurement alignment. To further improve the image quality, we introduce a dual-directional diffusion framework (dSTDM), which simultaneously performs reverse diffusion along two orthogonal directions, effectively capturing the full 3D data distribution. Comprehensive experiments on cardiac cine MRI datasets demonstrate that our approach achieves state-of-the-art performance in highly accelerated reconstruction. Additionally, it exhibits preliminary robustness across various undersampling scenarios and unseen datasets, including patient data, non-Cartesian radial sampling, and different anatomies.
{"title":"Robust Cardiac Cine MRI Reconstruction With Spatiotemporal Diffusion Model","authors":"Zi Wang;Jiahao Huang;Mingkai Huang;Chengyan Wang;Guang Yang;Xiaobo Qu","doi":"10.1109/TCI.2025.3598421","DOIUrl":"https://doi.org/10.1109/TCI.2025.3598421","url":null,"abstract":"Accelerated dynamic magnetic resonance imaging (MRI) is highly expected in clinical applications. However, its reconstruction remains challenging due to the inherently high dimensionality and spatiotemporal complexity. While diffusion models have demonstrated robust performance in spatial imaging, their application to spatiotemporal data has been underexplored. To address this gap, we propose a novel spatiotemporal diffusion model (STDM) specifically designed for robust dynamic MRI reconstruction. Our approach decomposes the complex 3D diffusion process into manageable sub-problems by focusing on 2D spatiotemporal images, thereby reducing dimensionality and enhancing computational efficiency. Each 2D image is treated independently, allowing for a parallel reverse diffusion process guided by data consistency to ensure measurement alignment. To further improve the image quality, we introduce a dual-directional diffusion framework (dSTDM), which simultaneously performs reverse diffusion along two orthogonal directions, effectively capturing the full 3D data distribution. Comprehensive experiments on cardiac cine MRI datasets demonstrate that our approach achieves state-of-the-art performance in highly accelerated reconstruction. Additionally, it exhibits preliminary robustness across various undersampling scenarios and unseen datasets, including patient data, non-Cartesian radial sampling, and different anatomies.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1258-1270"},"PeriodicalIF":4.8,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145090149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Objective: Magnetic particle imaging (MPI) is a novel non-destructive medical imaging method that visualizes the spatial distribution of superparamagnetic iron oxide nanoparticles. However, due to the non-uniformity of the selection and drive field, the unsatisfactory of the receive coil and the different components of the magnetization signal (induced electromotive force) detected by the orthogonal coil, processing the voltage signals measured by the receiving coils in different directions without discrimination will affect the reconstruction quality. Methods: This study introduces the Double-Branched and Multi-Magnetic Directions Feature Fusion Network (DB&MDF2-Net) to address these challenges. The dual-branch(DB) strategy processes X and Y-directional magnetic field components independently, reducing information confusion. Each branch has a dual-sampling feature(DSF) layer that captures multi-scale spatial information and preserves spatial structure, enhancing the extraction of particle distribution and edge details. Additionally, a multi-head self-attention transformer(MSA-T) layer efficiently integrates features from different modules, allowing the network to learn complex inter-feature relationships. Results: The effectiveness of the DB strategy, DSF and MSA-T layers in our proposed method were validated through ablation experiments. Simulate and phantom experiments further demonstrate significant improvements in detail capture and anti-noise capability of DB&MDF2-Net without any hardware modifications, enabling more precise restoration of real particle distribution characteristics. Conclusion: These findings suggest that DB&MDF2-Net can significantly improve the imaging accuracy of MPI. Significance: This research is expected to enhance the practicality of MPI in biomedical applications and contribute to the future development of MPI technology.
{"title":"Double-Branched and Multi-Magnetic Directions Feature Fusion Network (DB&MDF2-Net) for the Accurate Reconstruction of Magnetic Particle Imaging","authors":"Jintao Li;Lizhi Zhang;Shuangchen Li;Huanlong Gao;Shuaishuai He;Yizhe Zhao;Xiaowei He;Yuqing Hou;Hongbo Guo","doi":"10.1109/TCI.2025.3598455","DOIUrl":"https://doi.org/10.1109/TCI.2025.3598455","url":null,"abstract":"<italic>Objective:</i> Magnetic particle imaging (MPI) is a novel non-destructive medical imaging method that visualizes the spatial distribution of superparamagnetic iron oxide nanoparticles. However, due to the non-uniformity of the selection and drive field, the unsatisfactory of the receive coil and the different components of the magnetization signal (induced electromotive force) detected by the orthogonal coil, processing the voltage signals measured by the receiving coils in different directions without discrimination will affect the reconstruction quality. <italic>Methods:</i> This study introduces the Double-Branched and Multi-Magnetic Directions Feature Fusion Network (DB&MDF2-Net) to address these challenges. The dual-branch(DB) strategy processes X and Y-directional magnetic field components independently, reducing information confusion. Each branch has a dual-sampling feature(DSF) layer that captures multi-scale spatial information and preserves spatial structure, enhancing the extraction of particle distribution and edge details. Additionally, a multi-head self-attention transformer(MSA-T) layer efficiently integrates features from different modules, allowing the network to learn complex inter-feature relationships. <italic>Results:</i> The effectiveness of the DB strategy, DSF and MSA-T layers in our proposed method were validated through ablation experiments. Simulate and phantom experiments further demonstrate significant improvements in detail capture and anti-noise capability of DB&MDF2-Net without any hardware modifications, enabling more precise restoration of real particle distribution characteristics. <italic>Conclusion:</i> These findings suggest that DB&MDF2-Net can significantly improve the imaging accuracy of MPI. <italic>Significance:</i> This research is expected to enhance the practicality of MPI in biomedical applications and contribute to the future development of MPI technology.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1074-1086"},"PeriodicalIF":4.8,"publicationDate":"2025-08-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144918326","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Photoacoustic imaging combines the high contrast of optical imaging with the deep penetration depth of ultrasonic imaging, showing great potential in cerebrovascular disease detection. However, the ultrasonic wave suffers strong attenuation and multi-scattering when it passes through the skull tissue, resulting in the distortion of the collected photoacoustic signal. In this paper, inspired by the principles of deep learning and non-line-of-sight imaging, we propose an image reconstruction framework named HDN (Hybrid Deep-learning and Non-line-of-sight), which consists of the signal extraction part and difference utilization part. The signal extraction part is used to correct the distorted signal and reconstruct an initial image. The difference utilization part is used to make further use of the signal difference between the distorted signal and corrected signal, reconstructing the residual image between the initial image and the target image. The test results on a photoacoustic digital brain simulation dataset show that compared with the traditional method (delay-and-sum) and deep-learning-based method (UNet), the HDN achieved superior performance in both signal correction and image reconstruction. Specifically for the structural similarity index, the HDN reached 0.661 in imaging results, compared to 0.157 for the delay-and-sum method and 0.305 for the deep-learning-based method.
光声成像结合了光学成像的高对比度和超声成像的深穿透深度,在脑血管疾病检测中显示出巨大的潜力。然而,超声波在穿过颅骨组织时受到强烈的衰减和多重散射,导致采集到的光声信号失真。本文受深度学习和非视距成像原理的启发,提出了一种图像重建框架HDN (Hybrid deep -learning and non-line-of-sight),该框架由信号提取部分和差分利用部分组成。信号提取部分用于校正失真信号并重建初始图像。差值利用部分用于进一步利用失真信号与校正信号之间的信号差,重构初始图像与目标图像之间的残差图像。在光声数字脑模拟数据集上的测试结果表明,与传统的延迟求和方法和基于深度学习的UNet方法相比,HDN在信号校正和图像重建方面都取得了更好的效果。具体到结构相似指数,成像结果HDN达到0.661,而延迟和方法的HDN为0.157,基于深度学习的HDN为0.305。
{"title":"HDN: Hybrid Deep-Learning and Non-Line-of-Sight Reconstruction Framework for Transcranial Photoacoustic Imaging of Human Brain","authors":"Pengcheng Wan;Fan Zhang;Yuting Shen;Hulin Zhao;Xiran Cai;Xiaohua Feng;Fei Gao","doi":"10.1109/TCI.2025.3594073","DOIUrl":"https://doi.org/10.1109/TCI.2025.3594073","url":null,"abstract":"Photoacoustic imaging combines the high contrast of optical imaging with the deep penetration depth of ultrasonic imaging, showing great potential in cerebrovascular disease detection. However, the ultrasonic wave suffers strong attenuation and multi-scattering when it passes through the skull tissue, resulting in the distortion of the collected photoacoustic signal. In this paper, inspired by the principles of deep learning and non-line-of-sight imaging, we propose an image reconstruction framework named HDN (Hybrid Deep-learning and Non-line-of-sight), which consists of the signal extraction part and difference utilization part. The signal extraction part is used to correct the distorted signal and reconstruct an initial image. The difference utilization part is used to make further use of the signal difference between the distorted signal and corrected signal, reconstructing the residual image between the initial image and the target image. The test results on a photoacoustic digital brain simulation dataset show that compared with the traditional method (delay-and-sum) and deep-learning-based method (UNet), the HDN achieved superior performance in both signal correction and image reconstruction. Specifically for the structural similarity index, the HDN reached 0.661 in imaging results, compared to 0.157 for the delay-and-sum method and 0.305 for the deep-learning-based method.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1142-1149"},"PeriodicalIF":4.8,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"144918327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-11DOI: 10.1109/TCI.2025.3597449
Chunliang Ma;Kaiwen Tan;Yunxiang Li;Shouhua Luo
Nanovision static CT is an innovative CT scanning technique that features the arrangement of the X-ray source array and detector array on two parallel planes with a consistent offset. This configuration significantly enhances temporal resolution compared to conventional CT, providing particular advantages for dynamic organ imaging and low-dose imaging applications. However, it also introduces cone angle and sparse angle artifacts during helical scanning. To address this, this paper proposes a novel theoretical analysis framework to systematically analyze the artifact generation mechanism of the traditional FDK algorithm in this scenario. Through numerical solutions and data superposition, we are able to attribute the causes of artifacts for the first time to two types of data incompleteness issues arising from the lack of cone angle data and insufficient sparse angular sampling. Building on these insights, we propose an innovative dual-module collaborative reconstruction network. First, we introduce the Helical Bi-directional xFDK algorithm (HbixFDK), which employs a limited-angle weighted compensation strategy to mitigate data incompleteness in the cone angle region. Next, we develop the attention-based Helical FISTA network (HFISTA-Net), which utilizes the output from HbixFDK as the initial reconstruction to effectively suppress sparse sampling artifacts. Extensive experiments conducted on the TCIA dataset and clinical static CT scans demonstrate that our proposed method significantly reduces both cone angle and sparse angle artifacts in static CT helical scanning. The approach achieves rapid and high-precision helical reconstruction, showcasing superior accuracy and computational efficiency.
{"title":"A Helical Reconstruction Network for Multi-Source Static CT","authors":"Chunliang Ma;Kaiwen Tan;Yunxiang Li;Shouhua Luo","doi":"10.1109/TCI.2025.3597449","DOIUrl":"https://doi.org/10.1109/TCI.2025.3597449","url":null,"abstract":"Nanovision static CT is an innovative CT scanning technique that features the arrangement of the X-ray source array and detector array on two parallel planes with a consistent offset. This configuration significantly enhances temporal resolution compared to conventional CT, providing particular advantages for dynamic organ imaging and low-dose imaging applications. However, it also introduces cone angle and sparse angle artifacts during helical scanning. To address this, this paper proposes a novel theoretical analysis framework to systematically analyze the artifact generation mechanism of the traditional FDK algorithm in this scenario. Through numerical solutions and data superposition, we are able to attribute the causes of artifacts for the first time to two types of data incompleteness issues arising from the lack of cone angle data and insufficient sparse angular sampling. Building on these insights, we propose an innovative dual-module collaborative reconstruction network. First, we introduce the Helical Bi-directional xFDK algorithm (HbixFDK), which employs a limited-angle weighted compensation strategy to mitigate data incompleteness in the cone angle region. Next, we develop the attention-based Helical FISTA network (HFISTA-Net), which utilizes the output from HbixFDK as the initial reconstruction to effectively suppress sparse sampling artifacts. Extensive experiments conducted on the TCIA dataset and clinical static CT scans demonstrate that our proposed method significantly reduces both cone angle and sparse angle artifacts in static CT helical scanning. The approach achieves rapid and high-precision helical reconstruction, showcasing superior accuracy and computational efficiency.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1174-1189"},"PeriodicalIF":4.8,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145011319","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2025-08-11DOI: 10.1109/TCI.2025.3597462
Xiang Cai;Shunjun Wei;Mou Wang;Hao Zhang;Kun Chen;Xinyuan Liu;Jun Shi;Guolong Cui
High-precision 2-D/3-D Synthetic Aperture Radar (SAR) image reconstruction from the indirect scattered echoes of hidden targets represents a core technical challenge in millimeter-wave (mmW) Non-Line-of-Sight (NLOS) environmental perception. Deep learning approaches have demonstrated exceptional performance in SAR imaging. However, existing methods are predominantly designed for Line-of-Sight (LOS) scenarios, where clean LOS simulation signals can be acquired for training purposes, a condition often difficult or impossible to meet in NLOS imaging due to complex multipath environments and noise. To tackle this issue within specific NLOS configurations, particularly those involving strong specular reflections from discrete, isolated hidden objects, we propose an Equivariant Imaging (EI) framework tailored for mmW SAR. The EI framework is a fully self-supervised learning approach that leverages the group invariance present in signal distributions, enabling robust image reconstruction from partial NLOS measurements contaminated with noise and multipath artifacts. In our method, the reconstruction function is based on a deep unfolding network with Total Variation (TV) constraints, mapping the NLOS scattered echoes to the target image. Moreover, we introduce an Adaptive Peak Convolution Network (APConv) into the reconstruction process to dynamically adjust thresholds, replacing traditional fixed-threshold methods. This enhances imaging flexibility and quality under these defined NLOS conditions. Finally, we validate the proposed method using various NLOS echo data collected through an experimental mmW system. Numerical and visual results both demonstrate the effectiveness of our approach for NLOS mmW SAR imaging tasks. The proposed EI framework thus offers a promising approach for advancing NLOS mmW SAR perception capabilities, particularly for environments and target configurations aligning with those investigated and supported by our current experiments.
{"title":"Non-Line-of-Sight mmW SAR Imaging With Equivariant Adaptive Threshold Learning","authors":"Xiang Cai;Shunjun Wei;Mou Wang;Hao Zhang;Kun Chen;Xinyuan Liu;Jun Shi;Guolong Cui","doi":"10.1109/TCI.2025.3597462","DOIUrl":"https://doi.org/10.1109/TCI.2025.3597462","url":null,"abstract":"High-precision 2-D/3-D Synthetic Aperture Radar (SAR) image reconstruction from the indirect scattered echoes of hidden targets represents a core technical challenge in millimeter-wave (mmW) Non-Line-of-Sight (NLOS) environmental perception. Deep learning approaches have demonstrated exceptional performance in SAR imaging. However, existing methods are predominantly designed for Line-of-Sight (LOS) scenarios, where clean LOS simulation signals can be acquired for training purposes, a condition often difficult or impossible to meet in NLOS imaging due to complex multipath environments and noise. To tackle this issue within specific NLOS configurations, particularly those involving strong specular reflections from discrete, isolated hidden objects, we propose an Equivariant Imaging (EI) framework tailored for mmW SAR. The EI framework is a fully self-supervised learning approach that leverages the group invariance present in signal distributions, enabling robust image reconstruction from partial NLOS measurements contaminated with noise and multipath artifacts. In our method, the reconstruction function is based on a deep unfolding network with Total Variation (TV) constraints, mapping the NLOS scattered echoes to the target image. Moreover, we introduce an Adaptive Peak Convolution Network (APConv) into the reconstruction process to dynamically adjust thresholds, replacing traditional fixed-threshold methods. This enhances imaging flexibility and quality under these defined NLOS conditions. Finally, we validate the proposed method using various NLOS echo data collected through an experimental mmW system. Numerical and visual results both demonstrate the effectiveness of our approach for NLOS mmW SAR imaging tasks. The proposed EI framework thus offers a promising approach for advancing NLOS mmW SAR perception capabilities, particularly for environments and target configurations aligning with those investigated and supported by our current experiments.","PeriodicalId":56022,"journal":{"name":"IEEE Transactions on Computational Imaging","volume":"11 ","pages":"1190-1205"},"PeriodicalIF":4.8,"publicationDate":"2025-08-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145036849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}