首页 > 最新文献

IEEE Signal Processing Letters最新文献

英文 中文
A Fractional-Order Cauchy Penalty With Enhanced Adaptability for Signal Recovery 具有增强信号恢复适应性的分数阶柯西惩罚
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-26 DOI: 10.1109/LSP.2025.3648640
Zhen Gao;Yiping Jiang;Rong Yang;Xingqun Zhan
Recovering signals from noisy observations remains challenging due to the ill-posedness of inverse problems. While non-convex regularization methods like the standard Cauchy penalty improve estimation accuracy, it lacks adaptability across diverse scenarios. In response, this letter proposes a fractional-order Cauchy (q-Cauchy) penalty inspired by the Lq maximum likelihood estimation. By introducing the parameter $q$, the q-Cauchy penalty achieves greater adaptability in diverse scenarios. Specifically, we also derive sufficient convexity conditions for its proximal operator and propose a forward-backward solver. Simulation results demonstrate that the q-Cauchy with the appropriate $q$ outperforms the baseline methods in both 1D signal denoising and 2D image deblurring tasks.
由于逆问题的病态性,从噪声观测中恢复信号仍然具有挑战性。虽然像标准柯西惩罚这样的非凸正则化方法提高了估计精度,但它缺乏对不同场景的适应性。作为回应,这封信提出了一个分数阶柯西(q-Cauchy)惩罚,灵感来自Lq最大似然估计。通过引入参数$q$, q- cauchy penalty在不同的场景下具有更强的适应性。具体地说,我们还得到了它的近算子的充分凸性条件,并提出了一个正向向后求解器。仿真结果表明,适当的q- cauchy方法在一维信号去噪和二维图像去模糊任务中都优于基线方法。
{"title":"A Fractional-Order Cauchy Penalty With Enhanced Adaptability for Signal Recovery","authors":"Zhen Gao;Yiping Jiang;Rong Yang;Xingqun Zhan","doi":"10.1109/LSP.2025.3648640","DOIUrl":"https://doi.org/10.1109/LSP.2025.3648640","url":null,"abstract":"Recovering signals from noisy observations remains challenging due to the ill-posedness of inverse problems. While non-convex regularization methods like the standard Cauchy penalty improve estimation accuracy, it lacks adaptability across diverse scenarios. In response, this letter proposes a fractional-order Cauchy (q-Cauchy) penalty inspired by the Lq maximum likelihood estimation. By introducing the parameter <inline-formula><tex-math>$q$</tex-math></inline-formula>, the q-Cauchy penalty achieves greater adaptability in diverse scenarios. Specifically, we also derive sufficient convexity conditions for its proximal operator and propose a forward-backward solver. Simulation results demonstrate that the q-Cauchy with the appropriate <inline-formula><tex-math>$q$</tex-math></inline-formula> outperforms the baseline methods in both 1D signal denoising and 2D image deblurring tasks.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"456-460"},"PeriodicalIF":3.9,"publicationDate":"2025-12-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929583","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Bayesian Hybrid Attention Module for Underwater Acoustic Target Recognition 一种用于水声目标识别的贝叶斯混合注意模块
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-25 DOI: 10.1109/LSP.2025.3648249
Menghui Lei;Xiangyang Zeng;Mingmin Zeng;Anqi Jin
The attention mechanism improves underwater acoustic target recognition (UATR) by suppressing irrelevant features. However, due to the uncertainty and scarcity of underwater acoustic target (UWAT) signals, complicated deterministic attention modules increase the risk of model overfitting, resulting in limited improvement or even degradation in the performance of UATR. This letter proposes a Bayesian Hybrid Attention Module (BHAM) that enhances UATR based on time–frequency (T–F) features. BHAM models attention weights as random variables following Beta and Dirichlet distributions to capture uncertainty of UWAT signals and mitigate overfitting, while strengthening T–F feature representation via Bayesian channel attention and Bayesian T–F attention. By learning attention distributions in a Bayesian manner, BHAM effectively models complex dependencies in UWAT signals. Experiments on the DeepShip dataset demonstrate that BHAM alleviates overfitting and generalizes well across different network backbones.
注意机制通过抑制无关特征来提高水声目标识别能力。然而,由于水声目标(UWAT)信号的不确定性和稀缺性,复杂的确定性注意模块增加了模型过拟合的风险,导致UATR性能的提高有限甚至下降。这封信提出了一种基于时频(T-F)特征增强UATR的贝叶斯混合注意模块(BHAM)。BHAM将注意力权重建模为遵循Beta和Dirichlet分布的随机变量,以捕获UWAT信号的不确定性并减轻过拟合,同时通过贝叶斯信道注意和贝叶斯T-F注意加强T-F特征表示。通过以贝叶斯方式学习注意力分布,BHAM有效地模拟了UWAT信号中的复杂依赖关系。在DeepShip数据集上进行的实验表明,该方法可以很好地缓解过拟合问题,并能很好地泛化不同网络骨干网。
{"title":"A Bayesian Hybrid Attention Module for Underwater Acoustic Target Recognition","authors":"Menghui Lei;Xiangyang Zeng;Mingmin Zeng;Anqi Jin","doi":"10.1109/LSP.2025.3648249","DOIUrl":"https://doi.org/10.1109/LSP.2025.3648249","url":null,"abstract":"The attention mechanism improves underwater acoustic target recognition (UATR) by suppressing irrelevant features. However, due to the uncertainty and scarcity of underwater acoustic target (UWAT) signals, complicated deterministic attention modules increase the risk of model overfitting, resulting in limited improvement or even degradation in the performance of UATR. This letter proposes a Bayesian Hybrid Attention Module (BHAM) that enhances UATR based on time–frequency (T–F) features. BHAM models attention weights as random variables following Beta and Dirichlet distributions to capture uncertainty of UWAT signals and mitigate overfitting, while strengthening T–F feature representation via Bayesian channel attention and Bayesian T–F attention. By learning attention distributions in a Bayesian manner, BHAM effectively models complex dependencies in UWAT signals. Experiments on the DeepShip dataset demonstrate that BHAM alleviates overfitting and generalizes well across different network backbones.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"441-445"},"PeriodicalIF":3.9,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929434","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient CRB Estimation for Linear Models via Expectation Propagation and Monte Carlo Sampling 基于期望传播和蒙特卡罗抽样的线性模型有效CRB估计
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-25 DOI: 10.1109/LSP.2025.3648638
Fangqing Xiao;Dirk Slock
The Cramér–Rao bound (CRB) quantifies the variance lower bound for unbiased estimators, but it is intractable to evaluate in linear hierarchical Bayesian models with non-Gaussian priors due to the intractable marginal likelihood. Existing methods, including variational Bayes and Markov chain Monte Carlo (MCMC)-based approaches, often have high computational cost and slow convergence. We propose an efficient framework to approximate the Fisher information matrix (FIM) and the CRB by expressing the gradient of the log marginal likelihood as a posterior expectation. Expectation propagation (EP) is used to approximate the posterior as a Gaussian, enabling accurate moment estimation compared to pure sampling-based methods. Numerical experiments on small-scale sparse models show that the EP-based CRB approximation achieves lower average normalized mean squared error (NMSE) and faster convergence than classical baselines in non-Gaussian settings.
cram - rao界(CRB)量化了无偏估计量的方差下界,但在具有非高斯先验的线性层次贝叶斯模型中由于难以处理的边际似然而难以估计。现有的方法,包括变分贝叶斯和基于马尔可夫链蒙特卡罗(MCMC)的方法,往往计算成本高,收敛速度慢。通过将对数边际似然的梯度表示为后验期望,我们提出了一个有效的框架来近似Fisher信息矩阵(FIM)和CRB。期望传播(EP)用于将后验近似为高斯,与纯基于抽样的方法相比,能够实现准确的矩估计。在小尺度稀疏模型上的数值实验表明,在非高斯环境下,基于ep的CRB近似比经典基线具有更低的平均归一化均方误差(NMSE)和更快的收敛速度。
{"title":"Efficient CRB Estimation for Linear Models via Expectation Propagation and Monte Carlo Sampling","authors":"Fangqing Xiao;Dirk Slock","doi":"10.1109/LSP.2025.3648638","DOIUrl":"https://doi.org/10.1109/LSP.2025.3648638","url":null,"abstract":"The Cramér–Rao bound (CRB) quantifies the variance lower bound for unbiased estimators, but it is intractable to evaluate in linear hierarchical Bayesian models with non-Gaussian priors due to the intractable marginal likelihood. Existing methods, including variational Bayes and Markov chain Monte Carlo (MCMC)-based approaches, often have high computational cost and slow convergence. We propose an efficient framework to approximate the Fisher information matrix (FIM) and the CRB by expressing the gradient of the log marginal likelihood as a posterior expectation. Expectation propagation (EP) is used to approximate the posterior as a Gaussian, enabling accurate moment estimation compared to pure sampling-based methods. Numerical experiments on small-scale sparse models show that the EP-based CRB approximation achieves lower average normalized mean squared error (NMSE) and faster convergence than classical baselines in non-Gaussian settings.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"451-455"},"PeriodicalIF":3.9,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Video Face Super-Resolution With High-Precision Identity Preservation 具有高精度身份保存的视频人脸超分辨率
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-25 DOI: 10.1109/LSP.2025.3648639
Chaoliang Wu;Ting Zhang;Xianbin Zhang;Nian He;Yiwen Xu
As an emerging technology, Video Face Super-Resolution (VFSR) aims to reconstruct high-resolution facial images from low-quality video sequences while maintaining identity consistency, which makes it applicable to scenarios such as surveillance, video conferencing, and film restoration. Compared with image-based face restoration and general video super-resolution, VFSR is more challenging because it requires accurate facial detail reconstruction, strict identity preservation, and computational efficiency under varying poses and expressions. To address these challenges, we propose a High-Precision Identity Preserving VFSR framework (HPIP), which integrates a Multi-Scale Prediction Module (MPM) and an Identity Preservation Module (IPM). The MPM focuses on identity-critical facial regions (e.g., eyes, nose, and mouth) and leverages multi-scale feature prediction to improve reconstruction accuracy and robustness while maintaining computational efficiency. The IPM further projects features into a latent representation space, generating temporally consistent dictionary features and enhancing temporal coherence. Extensive experiments demonstrate that HPIP achieves superior performance in both qualitative and quantitative evaluations, producing visually pleasing facial details while maintaining an efficient and lightweight design.
视频人脸超分辨率(Video Face Super-Resolution, VFSR)是一项新兴技术,旨在从低质量的视频序列中重建高分辨率的人脸图像,同时保持身份一致性,适用于监控、视频会议和电影修复等场景。与基于图像的人脸恢复和一般的视频超分辨率相比,VFSR需要精确的人脸细节重建,严格的身份保持,以及不同姿态和表情下的计算效率,更具挑战性。为了解决这些问题,我们提出了一个高精度身份保持VFSR框架(HPIP),该框架集成了一个多尺度预测模块(MPM)和一个身份保持模块(IPM)。MPM专注于身份关键面部区域(例如,眼睛,鼻子和嘴巴),并利用多尺度特征预测来提高重建精度和鲁棒性,同时保持计算效率。IPM进一步将特征投射到潜在的表示空间中,生成时间一致的字典特征并增强时间一致性。大量的实验表明,HPIP在定性和定量评估中都取得了卓越的性能,在保持高效和轻量化设计的同时,产生了视觉上令人愉悦的面部细节。
{"title":"Video Face Super-Resolution With High-Precision Identity Preservation","authors":"Chaoliang Wu;Ting Zhang;Xianbin Zhang;Nian He;Yiwen Xu","doi":"10.1109/LSP.2025.3648639","DOIUrl":"https://doi.org/10.1109/LSP.2025.3648639","url":null,"abstract":"As an emerging technology, Video Face Super-Resolution (VFSR) aims to reconstruct high-resolution facial images from low-quality video sequences while maintaining identity consistency, which makes it applicable to scenarios such as surveillance, video conferencing, and film restoration. Compared with image-based face restoration and general video super-resolution, VFSR is more challenging because it requires accurate facial detail reconstruction, strict identity preservation, and computational efficiency under varying poses and expressions. To address these challenges, we propose a High-Precision Identity Preserving VFSR framework (HPIP), which integrates a Multi-Scale Prediction Module (MPM) and an Identity Preservation Module (IPM). The MPM focuses on identity-critical facial regions (<italic>e.g.</i>, eyes, nose, and mouth) and leverages multi-scale feature prediction to improve reconstruction accuracy and robustness while maintaining computational efficiency. The IPM further projects features into a latent representation space, generating temporally consistent dictionary features and enhancing temporal coherence. Extensive experiments demonstrate that HPIP achieves superior performance in both qualitative and quantitative evaluations, producing visually pleasing facial details while maintaining an efficient and lightweight design.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"406-410"},"PeriodicalIF":3.9,"publicationDate":"2025-12-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929369","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High-Capacity Image Steganography via Latent Diffusion Models 基于潜在扩散模型的大容量图像隐写
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-24 DOI: 10.1109/LSP.2025.3647567
Ruijie Du;Na Wang;Cheng Xiong;Chuan Qin;Xinpeng Zhang
Generative steganography has recently attracted considerable attention due to its superior security properties. However, most existing approaches suffer from limited hiding capacity. To address this issue, this paper proposes a high-capacity image steganography framework that integrates an encoder–decoder architecture with a latent diffusion model. Specifically, a message encoder is designed to transform binary secret messages into latent-space representations through a series of ResDense modules, enabling efficient hiding of large-scale information. The encoded latent features are then guided by the latent diffusion model to synthesize visually realistic stego images. During message extraction, the stego image undergoes iterative noise addition within the diffusion process to reconstruct the latent representation, from which a message decoder accurately recovers the hidden message. Extensive experimental results demonstrate that the proposed method achieves a high hiding capacity of over 30,000 bits, outperforming state-of-the-art methods while ensuring reliable message recovery under common image storage formats such as JPEG and PNG.
生成隐写术由于其优越的安全性能,近年来引起了人们的广泛关注。然而,大多数现有的方法都存在隐藏能力有限的问题。为了解决这个问题,本文提出了一个高容量的图像隐写框架,该框架将编码器-解码器架构与潜在扩散模型集成在一起。具体来说,设计了一个消息编码器,通过一系列ResDense模块将二进制秘密消息转换为潜在空间表示,从而实现大规模信息的有效隐藏。然后利用潜扩散模型对编码后的潜特征进行引导,合成视觉逼真的隐写图像。在信息提取过程中,隐写图像在扩散过程中进行迭代噪声相加,重建潜在表示,信息解码器从中准确地恢复隐藏信息。大量的实验结果表明,该方法实现了超过30,000位的高隐藏容量,在保证常见图像存储格式(如JPEG和PNG)下可靠的消息恢复的同时,优于目前最先进的方法。
{"title":"High-Capacity Image Steganography via Latent Diffusion Models","authors":"Ruijie Du;Na Wang;Cheng Xiong;Chuan Qin;Xinpeng Zhang","doi":"10.1109/LSP.2025.3647567","DOIUrl":"https://doi.org/10.1109/LSP.2025.3647567","url":null,"abstract":"Generative steganography has recently attracted considerable attention due to its superior security properties. However, most existing approaches suffer from limited hiding capacity. To address this issue, this paper proposes a high-capacity image steganography framework that integrates an encoder–decoder architecture with a latent diffusion model. Specifically, a message encoder is designed to transform binary secret messages into latent-space representations through a series of ResDense modules, enabling efficient hiding of large-scale information. The encoded latent features are then guided by the latent diffusion model to synthesize visually realistic stego images. During message extraction, the stego image undergoes iterative noise addition within the diffusion process to reconstruct the latent representation, from which a message decoder accurately recovers the hidden message. Extensive experimental results demonstrate that the proposed method achieves a high hiding capacity of over 30,000 bits, outperforming state-of-the-art methods while ensuring reliable message recovery under common image storage formats such as JPEG and PNG.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"401-405"},"PeriodicalIF":3.9,"publicationDate":"2025-12-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hierarchical Structure Dependency Whitening for Single-Domain Generalized Infrared Small Target Detection 单域广义红外小目标检测的层次结构相关性白化
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-23 DOI: 10.1109/LSP.2025.3647431
Lizhuo Liu;Songbo Wang;Yimin Fu
Existing infrared small target detection (IRSTD) methods typically assume that training and testing data share the same distribution. However, this assumption often fails in real-world applications due to environmental and sensor-induced variations, resulting in significant performance degradation caused by domain shifts. Besides, the inherently low signal-to-clutter ratio of targets in infrared images further impedes the extraction of underlying target information, increasing the risk of overfitting to domain-specific patterns. This severely constrains the generalizability of knowledge learned from source domains, particularly when training is confined to a single source domain due to the high cost of data annotation. To solve this problem, we propose hierarchical structure dependency whitening (HSDW) for single-domain generalized IRSTD. Specifically, we characterize domain discrepancies in infrared images as differences in structural information. Building upon this point, we employ feature whitening to mitigate the dependency on domain-specific structure information, whose distribution is diversely simulated by a dual-branch nonlinear transformation module. Moreover, we adopt a hierarchical suppression mechanism to alleviate the structural biases across multiple decoding stages, thereby facilitating more generalized target understanding across domains. Extensive experiments on three public IRSTD datasets demonstrate that our method achieves state-of-the-art performance.
现有的红外小目标检测方法通常假设训练数据和测试数据具有相同的分布。然而,由于环境和传感器引起的变化,这种假设在实际应用中经常失败,导致由域移位引起的显著性能下降。此外,红外图像中目标固有的低信杂比进一步阻碍了底层目标信息的提取,增加了对特定领域模式的过拟合风险。这严重限制了从源领域学习到的知识的泛化性,特别是当由于数据注释的高成本而将训练限制在单个源领域时。为了解决这一问题,我们提出了单域广义IRSTD的分层结构依赖白化(HSDW)方法。具体来说,我们将红外图像中的域差异表征为结构信息的差异。在此基础上,我们采用特征白化来减轻对特定领域结构信息的依赖,并通过双分支非线性转换模块对其分布进行多样化模拟。此外,我们采用了分层抑制机制来减轻多个解码阶段的结构偏差,从而促进跨领域更广义的目标理解。在三个公共IRSTD数据集上的大量实验表明,我们的方法达到了最先进的性能。
{"title":"Hierarchical Structure Dependency Whitening for Single-Domain Generalized Infrared Small Target Detection","authors":"Lizhuo Liu;Songbo Wang;Yimin Fu","doi":"10.1109/LSP.2025.3647431","DOIUrl":"https://doi.org/10.1109/LSP.2025.3647431","url":null,"abstract":"Existing infrared small target detection (IRSTD) methods typically assume that training and testing data share the same distribution. However, this assumption often fails in real-world applications due to environmental and sensor-induced variations, resulting in significant performance degradation caused by domain shifts. Besides, the inherently low signal-to-clutter ratio of targets in infrared images further impedes the extraction of underlying target information, increasing the risk of overfitting to domain-specific patterns. This severely constrains the generalizability of knowledge learned from source domains, particularly when training is confined to a single source domain due to the high cost of data annotation. To solve this problem, we propose hierarchical structure dependency whitening (HSDW) for single-domain generalized IRSTD. Specifically, we characterize domain discrepancies in infrared images as differences in structural information. Building upon this point, we employ feature whitening to mitigate the dependency on domain-specific structure information, whose distribution is diversely simulated by a dual-branch nonlinear transformation module. Moreover, we adopt a hierarchical suppression mechanism to alleviate the structural biases across multiple decoding stages, thereby facilitating more generalized target understanding across domains. Extensive experiments on three public IRSTD datasets demonstrate that our method achieves state-of-the-art performance.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"396-400"},"PeriodicalIF":3.9,"publicationDate":"2025-12-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
2D DOA Estimation of Coherent Signals Exploiting Moving Uniform Rectangular Array 利用运动均匀矩形阵列的相干信号二维DOA估计
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-19 DOI: 10.1109/LSP.2025.3646141
Saidur R. Pavel;Yimin D. Zhang
This letter considers two-dimensional direction of arrival (DOA) estimation of coherent signals exploiting a moving uniform rectangular array. The motion of the array induces phase variations in the received signals across spatial positions, enabling the construction of decorrelated covariance matrices through forward-backward spatial smoothing. We analyze the achievable degrees of freedom (DOFs) in terms of movement steps and examine the impact of the motion support on effective decorrelation. Notably, we show that the maximum number of DOFs can be achieved if each movement step is at least half the signal wavelength and the number of movement steps is no less than half the number of array elements. Furthermore, it is demonstrated that distributing motion across both array axes yields better decorrelation and estimation performance than restricting movement to a single dimension.
本文考虑了利用运动均匀矩形阵列的相干信号的二维到达方向(DOA)估计。阵列的运动引起接收信号在空间位置上的相位变化,使得通过前后向空间平滑构建去相关协方差矩阵成为可能。我们从运动步骤的角度分析了可实现的自由度(DOFs),并研究了运动支持对有效去相关的影响。值得注意的是,我们表明,如果每个移动步长至少为信号波长的一半,并且移动步长数量不少于阵列元素数量的一半,则可以实现最大自由度数。此外,研究表明,在两个阵列轴上分布运动比将运动限制在单个维度上产生更好的去相关和估计性能。
{"title":"2D DOA Estimation of Coherent Signals Exploiting Moving Uniform Rectangular Array","authors":"Saidur R. Pavel;Yimin D. Zhang","doi":"10.1109/LSP.2025.3646141","DOIUrl":"https://doi.org/10.1109/LSP.2025.3646141","url":null,"abstract":"This letter considers two-dimensional direction of arrival (DOA) estimation of coherent signals exploiting a moving uniform rectangular array. The motion of the array induces phase variations in the received signals across spatial positions, enabling the construction of decorrelated covariance matrices through forward-backward spatial smoothing. We analyze the achievable degrees of freedom (DOFs) in terms of movement steps and examine the impact of the motion support on effective decorrelation. Notably, we show that the maximum number of DOFs can be achieved if each movement step is at least half the signal wavelength and the number of movement steps is no less than half the number of array elements. Furthermore, it is demonstrated that distributing motion across both array axes yields better decorrelation and estimation performance than restricting movement to a single dimension.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"346-350"},"PeriodicalIF":3.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145879962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ASM-DiffConvNet: Physics-Guided Difference Convolution Network for Single-Image Restoration ASM-DiffConvNet:用于单图像恢复的物理引导差分卷积网络
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-19 DOI: 10.1109/LSP.2025.3646138
Hemkant Nehete;Amit Monga;Tharun Kumar Reddy Bollu;Balasubramanian Raman
This work proposes a physics-guided unified deep learning architecture for single image restoration targeting dehazing, deraining, and low-light enhancement. The architecture first estimates the transmission map and airlight under an atmospheric scattering model, and then refines the result with a grayscale prior. A DiffConv feature extractor is proposed which combines vanilla and difference convolutions with a Laplacian branch (to capture high-frequency features). During inference, its branches are re-parameterized into a single kernel for reducing computational complexity. The grayscale prior replaces the Y channel in the YCbCr space to suppress noise and color artifacts, while a refinement stage uses Spatial Feature Transform (SFT) to inject structural features from this grayscale prior into the RGB domain. Experiments on standard benchmarks show consistent improvements in PSNR and SSIM at lower computational cost.
这项工作提出了一种物理指导的统一深度学习架构,用于针对去雾、去训练和弱光增强的单个图像恢复。该体系结构首先估计大气散射模型下的透射图和气流,然后使用灰度先验对结果进行细化。提出了一种将差分卷积与拉普拉斯分支相结合的DiffConv特征提取器(用于捕获高频特征)。在推理过程中,为了降低计算复杂度,将其分支重新参数化为单个核。灰度先验取代YCbCr空间中的Y通道以抑制噪声和颜色伪像,而细化阶段使用空间特征变换(SFT)将该灰度先验的结构特征注入RGB域。在标准基准测试上的实验表明,在较低的计算成本下,PSNR和SSIM得到了一致的改进。
{"title":"ASM-DiffConvNet: Physics-Guided Difference Convolution Network for Single-Image Restoration","authors":"Hemkant Nehete;Amit Monga;Tharun Kumar Reddy Bollu;Balasubramanian Raman","doi":"10.1109/LSP.2025.3646138","DOIUrl":"https://doi.org/10.1109/LSP.2025.3646138","url":null,"abstract":"This work proposes a physics-guided unified deep learning architecture for single image restoration targeting dehazing, deraining, and low-light enhancement. The architecture first estimates the transmission map and airlight under an atmospheric scattering model, and then refines the result with a grayscale prior. A DiffConv feature extractor is proposed which combines vanilla and difference convolutions with a Laplacian branch (to capture high-frequency features). During inference, its branches are re-parameterized into a single kernel for reducing computational complexity. The grayscale prior replaces the Y channel in the YCbCr space to suppress noise and color artifacts, while a refinement stage uses Spatial Feature Transform (SFT) to inject structural features from this grayscale prior into the RGB domain. Experiments on standard benchmarks show consistent improvements in PSNR and SSIM at lower computational cost.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"336-340"},"PeriodicalIF":3.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wavelet-Based Denoising Transformer With Fourier Adjustment for UAV Nighttime Tracking 基于傅立叶调整小波去噪变压器的无人机夜间跟踪
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-19 DOI: 10.1109/LSP.2025.3646140
Haijun Wang;Wei Hao;Lihua Qi;Haoyu Qu;Zihao Su
Visual object tracking methods utilizing onboard cameras have significantly advanced the widespread application of uncrewed aerial vehicles (UAVs). However, the stochastic and intricate noise inherent in camera systems has critically impeded the performance of UAV trackers, particularly under low-light conditions. To solve this problem, this letter presents an efficient wavelet-based denoising transformer (WTM) integrated with a fast Fourier adjustment module (FFAM) to reduce random real noise, thereby improving UAV nighttime tracking performance. Specifically, an encoder-latent-decoder structure is designed for efficient end-to-end transformation. Additionally, the WTM in both the encoder and the decoder block introduces channel-wise transformer to extract low-frequency information. The FFAM is utilized in the latent block to adjust local texture details. Finally, a novel residual feedforward network is designed to enhance the processing of high-frequency information. Extensive experimental results validate the effectiveness of our proposed method, demonstrating significant improvements in UAV nighttime tracking capabilities by adapting to diverse enhancers and tracking algorithms.
利用机载相机的视觉目标跟踪方法极大地促进了无人机的广泛应用。然而,相机系统中固有的随机和复杂的噪声严重阻碍了无人机跟踪器的性能,特别是在低光条件下。为了解决这一问题,本文提出了一种高效的基于小波的去噪变压器(WTM),该变压器集成了快速傅立叶调整模块(FFAM),以降低随机真实噪声,从而提高无人机的夜间跟踪性能。具体来说,设计了一个编码器-潜在-解码器结构,以实现高效的端到端转换。此外,编码器和解码器块中的WTM都引入了信道方向的变压器来提取低频信息。在隐块中利用FFAM对局部纹理细节进行调整。最后,设计了一种新的残差前馈网络,增强了对高频信息的处理能力。大量的实验结果验证了我们提出的方法的有效性,表明通过适应不同的增强器和跟踪算法,无人机夜间跟踪能力有了显著的提高。
{"title":"Wavelet-Based Denoising Transformer With Fourier Adjustment for UAV Nighttime Tracking","authors":"Haijun Wang;Wei Hao;Lihua Qi;Haoyu Qu;Zihao Su","doi":"10.1109/LSP.2025.3646140","DOIUrl":"https://doi.org/10.1109/LSP.2025.3646140","url":null,"abstract":"Visual object tracking methods utilizing onboard cameras have significantly advanced the widespread application of uncrewed aerial vehicles (UAVs). However, the stochastic and intricate noise inherent in camera systems has critically impeded the performance of UAV trackers, particularly under low-light conditions. To solve this problem, this letter presents an efficient wavelet-based denoising transformer (WTM) integrated with a fast Fourier adjustment module (FFAM) to reduce random real noise, thereby improving UAV nighttime tracking performance. Specifically, an encoder-latent-decoder structure is designed for efficient end-to-end transformation. Additionally, the WTM in both the encoder and the decoder block introduces channel-wise transformer to extract low-frequency information. The FFAM is utilized in the latent block to adjust local texture details. Finally, a novel residual feedforward network is designed to enhance the processing of high-frequency information. Extensive experimental results validate the effectiveness of our proposed method, demonstrating significant improvements in UAV nighttime tracking capabilities by adapting to diverse enhancers and tracking algorithms.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"341-345"},"PeriodicalIF":3.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145886624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
No-Reference Stitched Wide Field of View Light Field Image Quality Assessment via Structured Representation and Progressive Learning 基于结构化表示和渐进式学习的无参考拼接宽视场光场图像质量评估
IF 3.9 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2025-12-19 DOI: 10.1109/LSP.2025.3646139
Yu Sun;Rui Zhou;Yueli Cui;Ming Jin;Gangyi Jiang
The limited field of view (FoV) of commercial light field cameras has driven development of various stitching techniques to generate wide-FoV light field images (WLFIs). However, these techniques often introduce local distortions and angular inconsistencies, posing significant challenges for WLFI quality assessment. In this letter, a no-reference WLFI quality assessment (WLFIQA) method based on structured representation and progressive learning is proposed. Specifically, considering the high-dimensional characteristics and distortion properties of WLFIs, a novel joint spatial–angular representation strategy is first designed. For the angular domain, horizontal and vertical sub-aperture images stacks are employed to characterize angular features; for the spatial domain each sub-aperture image is divided into four quadrants, and image blocks containing complementary cues from corresponding positions across different quadrants are used as input for subsequent feature extraction. Furthermore, a global–local feature extraction network is employed to further model multi-scale distortion characteristics. Finally, a progressive learning strategy is designed to enhance the performance of overall perceptual evaluation. Experimental results on a benchmark WLFI dataset show that the proposed method outperforms existing quality methods.
商用光场相机有限的视场,推动了各种拼接技术的发展,以产生宽视场光场图像。然而,这些技术往往会引入局部扭曲和角度不一致,给WLFI质量评估带来重大挑战。本文提出了一种基于结构化表示和渐进式学习的无参考WLFI质量评估(WLFIQA)方法。具体而言,考虑到wlfi的高维特征和畸变特性,首先设计了一种新的空间-角度联合表示策略。在角域,采用水平和垂直子孔径图像叠加来表征角特征;在空间域中,将每个子孔径图像划分为四个象限,并将来自不同象限对应位置的包含互补线索的图像块作为后续特征提取的输入。此外,采用全局-局部特征提取网络对多尺度变形特征进行进一步建模。最后,设计渐进式学习策略以提高整体知觉评估的表现。在WLFI基准数据集上的实验结果表明,该方法优于现有的质量方法。
{"title":"No-Reference Stitched Wide Field of View Light Field Image Quality Assessment via Structured Representation and Progressive Learning","authors":"Yu Sun;Rui Zhou;Yueli Cui;Ming Jin;Gangyi Jiang","doi":"10.1109/LSP.2025.3646139","DOIUrl":"https://doi.org/10.1109/LSP.2025.3646139","url":null,"abstract":"The limited field of view (FoV) of commercial light field cameras has driven development of various stitching techniques to generate wide-FoV light field images (WLFIs). However, these techniques often introduce local distortions and angular inconsistencies, posing significant challenges for WLFI quality assessment. In this letter, a no-reference WLFI quality assessment (WLFIQA) method based on structured representation and progressive learning is proposed. Specifically, considering the high-dimensional characteristics and distortion properties of WLFIs, a novel joint spatial–angular representation strategy is first designed. For the angular domain, horizontal and vertical sub-aperture images stacks are employed to characterize angular features; for the spatial domain each sub-aperture image is divided into four quadrants, and image blocks containing complementary cues from corresponding positions across different quadrants are used as input for subsequent feature extraction. Furthermore, a global–local feature extraction network is employed to further model multi-scale distortion characteristics. Finally, a progressive learning strategy is designed to enhance the performance of overall perceptual evaluation. Experimental results on a benchmark WLFI dataset show that the proposed method outperforms existing quality methods.","PeriodicalId":13154,"journal":{"name":"IEEE Signal Processing Letters","volume":"33 ","pages":"391-395"},"PeriodicalIF":3.9,"publicationDate":"2025-12-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145929527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE Signal Processing Letters
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1