首页 > 最新文献

Digital Signal Processing最新文献

英文 中文
A Review of self-interference cancellation technologies for simultaneous transmit-receive arrays 同步收发阵列自干扰消除技术综述
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-29 DOI: 10.1016/j.dsp.2026.105967
Changqing Song , Dian Xiao , Wanbing Hao , Wanzhi Ma , Hongzhi Zhao , Shihai Shao
Driven by dual demands of spectrum-intensive military electronic warfare systems and high-spectral-efficiency civilian communications, simultaneous transmit-receive (STAR) array technology has gained significant attention due to its potential for efficient spectrum reuse. However, strong self-interference (SI) between transmit and receive channels degrades the receiver sensitivity, posing a critical technical barrier to its practical implementation. This study systematically reviews the research progress in STAR array SI cancellation technologies, covering five key aspects: SI coupling channels, spatial-domain cancellation, analog-domain cancellation, digital-domain cancellation, and experimental verification. Current state-of-the-art systems demonstrate up to 137.3 dB of isolation for 256  ×  256 STAR arrays and 140.5 dB for 4  ×  4 arrays, approaching engineering feasibility. Nevertheless, the large-scale deployment of multi-antenna arrays in civil and military applications will expose STAR arrays to more severe challenges from strong near-field SI. Future research should focus on clarifying near-field coupling mechanisms, optimizing spatial degrees of freedom, reducing the complexity of SI reconstruction, and refining compensation strategies for non-ideal factors to advance the deployment of STAR technology.
在频谱密集型军事电子战系统和高频谱效率民用通信的双重需求的驱动下,同步发射接收(STAR)阵列技术由于其高效频谱复用的潜力而受到了极大的关注。然而,发射和接收信道之间的强自干扰(SI)降低了接收机的灵敏度,对其实际实施构成了关键的技术障碍。本文系统综述了星阵信号对消技术的研究进展,涵盖了信号耦合通道、空域对消、模拟域对消、数字域对消和实验验证五个关键方面。目前最先进的系统显示,256个  ×  256个STAR阵列的隔离度可达137.3 dB, 4个  ×  4阵列的隔离度可达140.5 dB,接近工程可行性。然而,多天线阵列在民用和军事应用中的大规模部署将使STAR阵列面临来自强近场SI的更严峻挑战。未来的研究应集中在明确近场耦合机制、优化空间自由度、降低SI重建复杂性、完善非理想因素补偿策略等方面,以推进STAR技术的部署。
{"title":"A Review of self-interference cancellation technologies for simultaneous transmit-receive arrays","authors":"Changqing Song ,&nbsp;Dian Xiao ,&nbsp;Wanbing Hao ,&nbsp;Wanzhi Ma ,&nbsp;Hongzhi Zhao ,&nbsp;Shihai Shao","doi":"10.1016/j.dsp.2026.105967","DOIUrl":"10.1016/j.dsp.2026.105967","url":null,"abstract":"<div><div>Driven by dual demands of spectrum-intensive military electronic warfare systems and high-spectral-efficiency civilian communications, simultaneous transmit-receive (STAR) array technology has gained significant attention due to its potential for efficient spectrum reuse. However, strong self-interference (SI) between transmit and receive channels degrades the receiver sensitivity, posing a critical technical barrier to its practical implementation. This study systematically reviews the research progress in STAR array SI cancellation technologies, covering five key aspects: SI coupling channels, spatial-domain cancellation, analog-domain cancellation, digital-domain cancellation, and experimental verification. Current state-of-the-art systems demonstrate up to 137.3 dB of isolation for 256  ×  256 STAR arrays and 140.5 dB for 4  ×  4 arrays, approaching engineering feasibility. Nevertheless, the large-scale deployment of multi-antenna arrays in civil and military applications will expose STAR arrays to more severe challenges from strong near-field SI. Future research should focus on clarifying near-field coupling mechanisms, optimizing spatial degrees of freedom, reducing the complexity of SI reconstruction, and refining compensation strategies for non-ideal factors to advance the deployment of STAR technology.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105967"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
EMFNet: An efficient multi-scale fusion network for UAV small object detection EMFNet:用于无人机小目标检测的高效多尺度融合网络
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-27 DOI: 10.1016/j.dsp.2026.105952
Mingquan Wang , Huiying Xu , Yiming Sun , Hongbo Li , Zeyu Wang , Yi Li , Ruidong Wang , Xinzhong Zhu
Object detection in UAV aerial images holds significant application value in traffic monitoring, precision agriculture, and other fields. However, this task faces numerous challenges, including significant variations in object sizes, complex background interference, high object density, and class imbalance. Additionally, processing high-resolution aerial images involves disturbances such as uneven lighting and weather variations. To address these challenges, we propose an EMFNet model. This model effectively solves the problems in object detection in drone aerial images by enhancing the response to object areas under different lighting and weather conditions, suppressing interference from complex backgrounds, and improving adaptability to changes in image object size. Specifically, firstly, the lightweight vision transformer architecture RepViT is innovatively used as the backbone of EMFNet, combined with Dual Cross-Stage Partial Attention (DCPA) to optimize multi-scale feature fusion and background suppression, thereby enhancing small object feature extraction under varying lighting and weather conditions. Second, we propose the Context Guided Downsample Block (CGDB) to improve the downsampling process and mitigate feature information loss. Finally, the DyHead detection head utilizing the three-level attention mechanism receives three appropriately located prediction heads for classification and localization, thus improving the detection accuracy of dense and rare objects. Experiments on the VisDrone and UAVDT datasets demonstrate that EMFNet, with 6.76M parameters, achieves AP improvements of 7.5% and 15.2% over the baseline models, respectively.
无人机航拍图像中的目标检测在交通监控、精准农业等领域具有重要的应用价值。然而,这项任务面临着许多挑战,包括对象大小的显著变化、复杂的背景干扰、高对象密度和类不平衡。此外,处理高分辨率航拍图像还涉及光照不均匀和天气变化等干扰。为了应对这些挑战,我们提出了一个EMFNet模型。该模型通过增强对不同光照和天气条件下目标区域的响应,抑制复杂背景的干扰,提高对图像目标尺寸变化的适应性,有效解决了无人机航拍图像中目标检测问题。具体而言,首先,创新性地将轻量级视觉转换架构RepViT作为EMFNet的主干,结合双跨阶段局部注意(Dual Cross-Stage Partial Attention, DCPA)优化多尺度特征融合和背景抑制,从而增强不同光照和天气条件下的小目标特征提取;其次,我们提出了上下文引导下采样块(Context Guided Downsample Block, CGDB)来改进下采样过程,减少特征信息的丢失。最后,利用三级注意机制的DyHead检测头接收三个位置合适的预测头进行分类和定位,从而提高了密集和稀有物体的检测精度。在VisDrone和UAVDT数据集上的实验表明,EMFNet具有676万个参数,比基线模型分别提高了7.5%和15.2%的AP。
{"title":"EMFNet: An efficient multi-scale fusion network for UAV small object detection","authors":"Mingquan Wang ,&nbsp;Huiying Xu ,&nbsp;Yiming Sun ,&nbsp;Hongbo Li ,&nbsp;Zeyu Wang ,&nbsp;Yi Li ,&nbsp;Ruidong Wang ,&nbsp;Xinzhong Zhu","doi":"10.1016/j.dsp.2026.105952","DOIUrl":"10.1016/j.dsp.2026.105952","url":null,"abstract":"<div><div>Object detection in UAV aerial images holds significant application value in traffic monitoring, precision agriculture, and other fields. However, this task faces numerous challenges, including significant variations in object sizes, complex background interference, high object density, and class imbalance. Additionally, processing high-resolution aerial images involves disturbances such as uneven lighting and weather variations. To address these challenges, we propose an EMFNet model. This model effectively solves the problems in object detection in drone aerial images by enhancing the response to object areas under different lighting and weather conditions, suppressing interference from complex backgrounds, and improving adaptability to changes in image object size. Specifically, firstly, the lightweight vision transformer architecture RepViT is innovatively used as the backbone of EMFNet, combined with <strong>D</strong>ual <strong>C</strong>ross-Stage <strong>P</strong>artial <strong>A</strong>ttention (<strong>DCPA</strong>) to optimize multi-scale feature fusion and background suppression, thereby enhancing small object feature extraction under varying lighting and weather conditions. Second, we propose the <strong>C</strong>ontext <strong>G</strong>uided <strong>D</strong>ownsample <strong>B</strong>lock (<strong>CGDB</strong>) to improve the downsampling process and mitigate feature information loss. Finally, the DyHead detection head utilizing the three-level attention mechanism receives three appropriately located prediction heads for classification and localization, thus improving the detection accuracy of dense and rare objects. Experiments on the VisDrone and UAVDT datasets demonstrate that EMFNet, with 6.76M parameters, achieves <em>AP</em> improvements of 7.5% and 15.2% over the baseline models, respectively.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105952"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146070886","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Deformable convolution and transformer hybrid network for hyperspectral image classification 高光谱图像分类的可变形卷积和变压器混合网络
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-29 DOI: 10.1016/j.dsp.2026.105962
Xiang Chen, Shuzhen Zhang, Hailong Song, Qi Yan
Recently, deformable convolutions based on convolutional neural networks have been widely used in hyperspectral image (HSI) classification due to their flexible geometric adaptability and superior local feature extraction capabilities. However, they still face significant challenges in establishing long-range dependencies and capturing global contextual information among pixel sequences. To address these challenges, a novel deformable convolution and Transformer hybrid network (DTHNet) is proposed for HSI classification. Specifically, PCA is firstly employed to reduce the dimensionality of the original HSI and a group depth joint convolution block (GDJCB) is utilized to capture the spectral-spatial features of the reduced HSI patches, avoiding the neglect of certain spectral bands. Secondly, a parallel architecture composed of a designed deformable convolution and a Transformer is utilized to jointly extract local-global spectral-spatial features and long-range dependencies in HSI. In the deformable convolution branch, a simple parameter-free attention (SimAM) enhanced spectral-spatial convolution block (SSCB) is designed to effectively prevent the loss of key information and the generation of redundant features during the convolution. In the Transformer branch, the deep integration of convolutional operation and self-attention mechanism further promotes more effective extraction of HSI features. Finally, fusion features from the two branches to obtain the more accurate HSI classification. Experimental results on three widely used HSI datasets demonstrate that the proposed DTHNet outperforms several state-of-the-art HSI classification networks.
近年来,基于卷积神经网络的可变形卷积以其灵活的几何适应性和优越的局部特征提取能力在高光谱图像分类中得到了广泛的应用。然而,它们在建立长期依赖关系和捕获像素序列之间的全局上下文信息方面仍然面临重大挑战。为了解决这些挑战,提出了一种新的可变形卷积和变压器混合网络(DTHNet)用于HSI分类。具体而言,首先利用PCA对原始HSI进行降维,并利用组深度联合卷积块(group depth joint convolution block, GDJCB)捕捉降维后HSI斑块的光谱空间特征,避免了某些光谱波段的忽略。其次,利用设计的可变形卷积和变压器组成的并行结构,联合提取局部-全局频谱空间特征和远程依赖关系。在可变形卷积分支中,设计了一种简单的无参数注意(SimAM)增强频谱空间卷积块(SSCB),有效防止了卷积过程中关键信息的丢失和冗余特征的产生。在Transformer分支中,卷积运算与自关注机制的深度融合进一步促进了HSI特征的更有效提取。最后,融合两个分支的特征,得到更准确的HSI分类。在三个广泛使用的恒指指数数据集上的实验结果表明,所提出的DTHNet优于几种最先进的恒指指数分类网络。
{"title":"Deformable convolution and transformer hybrid network for hyperspectral image classification","authors":"Xiang Chen,&nbsp;Shuzhen Zhang,&nbsp;Hailong Song,&nbsp;Qi Yan","doi":"10.1016/j.dsp.2026.105962","DOIUrl":"10.1016/j.dsp.2026.105962","url":null,"abstract":"<div><div>Recently, deformable convolutions based on convolutional neural networks have been widely used in hyperspectral image (HSI) classification due to their flexible geometric adaptability and superior local feature extraction capabilities. However, they still face significant challenges in establishing long-range dependencies and capturing global contextual information among pixel sequences. To address these challenges, a novel deformable convolution and Transformer hybrid network (DTHNet) is proposed for HSI classification. Specifically, PCA is firstly employed to reduce the dimensionality of the original HSI and a group depth joint convolution block (GDJCB) is utilized to capture the spectral-spatial features of the reduced HSI patches, avoiding the neglect of certain spectral bands. Secondly, a parallel architecture composed of a designed deformable convolution and a Transformer is utilized to jointly extract local-global spectral-spatial features and long-range dependencies in HSI. In the deformable convolution branch, a simple parameter-free attention (SimAM) enhanced spectral-spatial convolution block (SSCB) is designed to effectively prevent the loss of key information and the generation of redundant features during the convolution. In the Transformer branch, the deep integration of convolutional operation and self-attention mechanism further promotes more effective extraction of HSI features. Finally, fusion features from the two branches to obtain the more accurate HSI classification. Experimental results on three widely used HSI datasets demonstrate that the proposed DTHNet outperforms several state-of-the-art HSI classification networks.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105962"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enhanced feature fusion and detail-Preserving network for small object detection in medical microscopic images 基于增强特征融合和细节保留网络的医学显微图像小目标检测
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-23 DOI: 10.1016/j.dsp.2026.105938
Runtian Zheng, Congpeng Zhang, Ying Liu
Accurately detecting tiny targets in microscopic images is critical for tuberculosis screening yet remains difficult due to large shape variation, dense instances with weak semantics, and cluttered backgrounds. We curate a Mycobacterium tuberculosis dataset of 5,842 microscopic images and present EFDNet, an Enhanced Feature Fusion and Detail-Preserving detector. EFDNet combines an Adaptive Feature Enhancement module that dynamically shifts convolutional sampling to capture irregular, fine-grained patterns, a Cross-Stage Enhanced Feature Pyramid Network that fuses semantic and localization cues across scales to withstand crowding and background clutter, and a lightweight shared Detail-Enhanced detection head that preserves high-frequency structure through differential convolutions and shared parameters, together with a Normalized Wasserstein Distance loss that reduces localization sensitivity for small boxes. On our dataset, the Tuberculosis-Phonecamera dataset, and the cross-domain BBBC041 blood-cell benchmark, EFDNet achieves AP50 of 81.9%, 87.6%, and 95.2%, outperforming a strong baseline by +5.7, +3.2, and +3.9 points, respectively, while maintaining low computational cost. These results indicate robust small-object detection under varied microscopy conditions and support the practical utility of EFDNet for automated screening.
准确检测显微图像中的微小目标对于结核病筛查至关重要,但由于形状变化大,语义弱的密集实例和杂乱的背景,仍然很困难。我们整理了一个包含5842张显微图像的结核分枝杆菌数据集,并提出了EFDNet,一种增强的特征融合和细节保留检测器。EFDNet结合了一个自适应特征增强模块,该模块动态移动卷积采样以捕获不规则、细粒度的模式;一个跨阶段增强特征金字塔网络,融合跨尺度的语义和定位线索,以抵御拥挤和背景混乱;一个轻量级共享细节增强检测头,通过微分卷积和共享参数保留高频结构。加上标准化的Wasserstein距离损失,降低了小盒子的定位灵敏度。在我们的数据集(Tuberculosis-Phonecamera数据集)和跨域BBBC041血细胞基准上,EFDNet的AP50分别达到81.9%、87.6%和95.2%,分别比强基线高出+5.7、+3.2和+3.9点,同时保持较低的计算成本。这些结果表明,在不同的显微镜条件下,小物体检测是可靠的,并支持EFDNet在自动筛选中的实际应用。
{"title":"Enhanced feature fusion and detail-Preserving network for small object detection in medical microscopic images","authors":"Runtian Zheng,&nbsp;Congpeng Zhang,&nbsp;Ying Liu","doi":"10.1016/j.dsp.2026.105938","DOIUrl":"10.1016/j.dsp.2026.105938","url":null,"abstract":"<div><div>Accurately detecting tiny targets in microscopic images is critical for tuberculosis screening yet remains difficult due to large shape variation, dense instances with weak semantics, and cluttered backgrounds. We curate a Mycobacterium tuberculosis dataset of 5,842 microscopic images and present EFDNet, an Enhanced Feature Fusion and Detail-Preserving detector. EFDNet combines an Adaptive Feature Enhancement module that dynamically shifts convolutional sampling to capture irregular, fine-grained patterns, a Cross-Stage Enhanced Feature Pyramid Network that fuses semantic and localization cues across scales to withstand crowding and background clutter, and a lightweight shared Detail-Enhanced detection head that preserves high-frequency structure through differential convolutions and shared parameters, together with a Normalized Wasserstein Distance loss that reduces localization sensitivity for small boxes. On our dataset, the Tuberculosis-Phonecamera dataset, and the cross-domain BBBC041 blood-cell benchmark, EFDNet achieves <em>AP</em><sub>50</sub> of 81.9%, 87.6%, and 95.2%, outperforming a strong baseline by <span><math><mrow><mo>+</mo><mn>5.7</mn></mrow></math></span>, <span><math><mrow><mo>+</mo><mn>3.2</mn></mrow></math></span>, and <span><math><mrow><mo>+</mo><mn>3.9</mn></mrow></math></span> points, respectively, while maintaining low computational cost. These results indicate robust small-object detection under varied microscopy conditions and support the practical utility of EFDNet for automated screening.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105938"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146081705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A visual security image encryption algorithm based on 1D-CHCCM and super-resolution reconstruction 基于1D-CHCCM和超分辨率重构的视觉安全图像加密算法
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-02-05 DOI: 10.1016/j.dsp.2026.105981
Wei Li , Yanxue Zhou , Lvchen Cao , Wenjiao Li , Xiuli Chai
In visual security image encryption algorithms, simultaneously ensuring security and high-quality restored images remains challenging. In this paper, we propose a method that incorporates deep learning-based super-resolution reconstruction as a key post-processing step after the encryption-decryption steps. This approach aims to achieve high-quality reconstructed images while maintaining decryption security. Specifically, a 1D Chebyshev-hyperbolic composite chaotic map (1D-CHCCM) is firstly proposed. Its superior chaotic behavior and stability are validated through multidimensional analysis, including Lyapunov exponent, sample entropy, and permutation entropy. To address the traditional channel independent processing, the symmetric cross-channel circular scrambling (SC3S) and odd-even alternating diffusion (OEAD) are proposed. These mechanisms treat color images as unified entities to enhance resistance to attacks. Furthermore, for visual concealment during transmission, a texture-based adaptive data hiding (ATADH) scheme is utilized to guarantee steganographic images (STIs) are visually indistinguishable. After decryption, the decrypted image is fed into a Transformer-based super-resolution reconstruction network to obtain the final high-quality image. Quantitative analysis reveals that the proposed algorithm achieves a correlation coefficient below 0.003, an information entropy of 7.9973, and NPCR/UACI scores of 99.61% and 33.42%. In terms of visual quality, the STIs maintain excellent imperceptibility with a PSNR of 48.8 dB, and these reconstructed images have reached a PSNR of 41 dB. These results confirm that the goal of balancing security and high-quality image restoration is achieved.
在视觉安全图像加密算法中,如何同时保证图像的安全性和高质量的恢复图像仍然是一个挑战。在本文中,我们提出了一种将基于深度学习的超分辨率重建作为加密-解密步骤之后的关键后处理步骤的方法。该方法的目的是在保证解密安全性的同时实现高质量的重构图像。具体而言,首先提出了一维切比雪夫-双曲复合混沌映射(1D- chccm)。通过李雅普诺夫指数、样本熵和排列熵的多维分析,验证了其优越的混沌行为和稳定性。针对传统的信道独立处理方法,提出了对称跨信道圆形置乱(SC3S)和奇偶交替扩散(OEAD)方法。这些机制将彩色图像视为统一的实体,以增强对攻击的抵抗力。此外,为了实现传输过程中的视觉隐藏,采用了基于纹理的自适应数据隐藏(ATADH)方案,保证隐写图像在视觉上不可区分。解密后的图像被送入基于transformer的超分辨率重建网络,得到最终的高质量图像。定量分析表明,该算法的相关系数小于0.003,信息熵为7.9973,NPCR/UACI得分分别为99.61%和33.42%。在视觉质量方面,sti保持了良好的不可感知性,PSNR为48.8 dB,重建图像的PSNR达到41 dB。这些结果证实了平衡安全性和高质量图像恢复的目标是实现的。
{"title":"A visual security image encryption algorithm based on 1D-CHCCM and super-resolution reconstruction","authors":"Wei Li ,&nbsp;Yanxue Zhou ,&nbsp;Lvchen Cao ,&nbsp;Wenjiao Li ,&nbsp;Xiuli Chai","doi":"10.1016/j.dsp.2026.105981","DOIUrl":"10.1016/j.dsp.2026.105981","url":null,"abstract":"<div><div>In visual security image encryption algorithms, simultaneously ensuring security and high-quality restored images remains challenging. In this paper, we propose a method that incorporates deep learning-based super-resolution reconstruction as a key post-processing step after the encryption-decryption steps. This approach aims to achieve high-quality reconstructed images while maintaining decryption security. Specifically, a 1D Chebyshev-hyperbolic composite chaotic map (1D-CHCCM) is firstly proposed. Its superior chaotic behavior and stability are validated through multidimensional analysis, including Lyapunov exponent, sample entropy, and permutation entropy. To address the traditional channel independent processing, the symmetric cross-channel circular scrambling (SC<sup>3</sup>S) and odd-even alternating diffusion (OEAD) are proposed. These mechanisms treat color images as unified entities to enhance resistance to attacks. Furthermore, for visual concealment during transmission, a texture-based adaptive data hiding (ATADH) scheme is utilized to guarantee steganographic images (STIs) are visually indistinguishable. After decryption, the decrypted image is fed into a Transformer-based super-resolution reconstruction network to obtain the final high-quality image. Quantitative analysis reveals that the proposed algorithm achieves a correlation coefficient below 0.003, an information entropy of 7.9973, and NPCR/UACI scores of 99.61% and 33.42%. In terms of visual quality, the STIs maintain excellent imperceptibility with a PSNR of 48.8 dB, and these reconstructed images have reached a PSNR of 41 dB. These results confirm that the goal of balancing security and high-quality image restoration is achieved.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105981"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Low Complexity estimation of fractional delay-Doppler-Angle parameters in MIMO-OTFS ISAC system MIMO-OTFS ISAC系统分数阶延迟-多普勒角参数的低复杂度估计
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-02-04 DOI: 10.1016/j.dsp.2026.105960
Olivia Zacharia, Vani Devi M․
This paper presents an orthogonal time-frequency space (OTFS)-based integrated sensing and communication (ISAC) transceiver architecture designed for vehicular platforms, enabling simultaneous environment sensing and data exchange with roadside units. A novel multiple-input multiple-output (MIMO) channel matrix model is introduced to account for fractional delays and Doppler shifts misaligned with the discrete delay-Doppler resolution of the OTFS grid. We derive a sparse time-domain input-output relationship for the MIMO-OTFS system and propose a two-stage delay-Doppler-angular fractional refinement (DDAFR) algorithm for joint estimation of delay, Doppler, and angle parameters. Compared to orthogonal matching pursuit (OMP), the proposed method offers lower complexity by avoiding the use of large dictionary matrices. To further mitigate the processing overhead, we propose a modified DDAFR (MDDAFR) algorithm that first determines the angle of arrival (AoA), followed by the remaining parameters. Simulation results confirm that the proposed ISAC algorithms achieve robust estimation performance while maintaining computational efficiency.
本文提出了一种基于正交时频空间(OTFS)的集成传感和通信(ISAC)收发器架构,该架构设计用于车载平台,能够同时与路边单元进行环境传感和数据交换。提出了一种新的多输入多输出(MIMO)信道矩阵模型,用于解释与OTFS网格离散延迟-多普勒分辨率不一致的分数阶延迟和多普勒频移。我们推导了MIMO-OTFS系统的稀疏时域输入输出关系,并提出了一种用于延迟、多普勒和角度参数联合估计的两级延迟-多普勒角分数细化(DDAFR)算法。与正交匹配追踪(OMP)方法相比,该方法避免了使用大的字典矩阵,降低了复杂度。为了进一步减少处理开销,我们提出了一种改进的DDAFR (MDDAFR)算法,该算法首先确定到达角(AoA),然后确定其余参数。仿真结果表明,所提出的ISAC算法在保持计算效率的前提下实现了鲁棒估计性能。
{"title":"Low Complexity estimation of fractional delay-Doppler-Angle parameters in MIMO-OTFS ISAC system","authors":"Olivia Zacharia,&nbsp;Vani Devi M․","doi":"10.1016/j.dsp.2026.105960","DOIUrl":"10.1016/j.dsp.2026.105960","url":null,"abstract":"<div><div>This paper presents an orthogonal time-frequency space (OTFS)-based integrated sensing and communication (ISAC) transceiver architecture designed for vehicular platforms, enabling simultaneous environment sensing and data exchange with roadside units. A novel multiple-input multiple-output (MIMO) channel matrix model is introduced to account for fractional delays and Doppler shifts misaligned with the discrete delay-Doppler resolution of the OTFS grid. We derive a sparse time-domain input-output relationship for the MIMO-OTFS system and propose a two-stage delay-Doppler-angular fractional refinement (DDAFR) algorithm for joint estimation of delay, Doppler, and angle parameters. Compared to orthogonal matching pursuit (OMP), the proposed method offers lower complexity by avoiding the use of large dictionary matrices. To further mitigate the processing overhead, we propose a modified DDAFR (MDDAFR) algorithm that first determines the angle of arrival (AoA), followed by the remaining parameters. Simulation results confirm that the proposed ISAC algorithms achieve robust estimation performance while maintaining computational efficiency.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105960"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174820","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Semi-supervised radar signal sorting with multiview subspace representations and graph learning 基于多视图子空间表示和图学习的半监督雷达信号排序
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-28 DOI: 10.1016/j.dsp.2026.105963
Shuai Huang , Qiang Guo , Yuhang Tian , Hao Feng , Sergey Shulga
In complex electromagnetic environments, radar pulse signals are strongly affected by noise, and limitations of reconnaissance receivers enlarge measurement errors, causing severe pulse missing and the inclusion of numerous spurious pulses. Consequently, pulse sorting faces two key difficulties: mining pulse association relations under missing information, and maintaining inter-class separability under serious parameter feature overlap. We propose a semi-supervised radar signal sorting method based on multiview subspace representation and graph learning (MvSR-GCN-RSS). First, encoders map multiple views into the latent space, where the view-specific and universal self-representation matrices are solved, and pulse sequence adjacency relations are constructed from intrapulse and interpulse information. Then, multiview information complementarity is achieved through a consistency loss and a diversity loss. In contrast to the two-stage process of first graph construction and then spectral clustering, we couple adjacency matrix solving with a graph convolutional network (GCN) in a single end-to-end framework, jointly optimizing it with the parameters of the multiview encoders and decoders to improve sorting efficiency. Finally, we design a multiview joint loss that simultaneously optimizes view reconstruction, GCN-based classification, self-representation solving, and cross-view complementarity for radar signal sorting. Simulation results show that the sorting accuracy reaches 99.99% in ideal scenarios; under scenarios with large measurement errors, pulse missing, and numerous spurious pulses, the proposed method performs far better than the comparison algorithms.
在复杂的电磁环境中,雷达脉冲信号受噪声的影响较大,侦察接收机的局限性扩大了测量误差,造成了严重的脉冲缺失和大量的杂散脉冲。因此,脉冲分类面临两个关键难题:在信息缺失的情况下挖掘脉冲关联关系,在参数特征严重重叠的情况下保持类间可分性。提出了一种基于多视图子空间表示和图学习的半监督雷达信号分选方法(MvSR-GCN-RSS)。首先,编码器将多个视图映射到隐空间中,在隐空间中求解特定于视图和通用的自表示矩阵,并从脉冲内和脉冲间信息构建脉冲序列邻接关系。然后,通过一致性损失和多样性损失实现多视图信息互补。与先图构建再谱聚类的两阶段过程不同,我们将邻接矩阵求解与单个端到端框架中的图卷积网络(GCN)耦合起来,并与多视图编码器和解码器的参数共同优化,以提高排序效率。最后,我们设计了一个多视图联合损失,同时优化了雷达信号分类的视图重建、基于gcn的分类、自表示求解和跨视图互补。仿真结果表明,在理想情况下,分选精度达到99.99%;在测量误差大、脉冲缺失、杂散脉冲多的情况下,该方法的性能远远优于比较算法。
{"title":"Semi-supervised radar signal sorting with multiview subspace representations and graph learning","authors":"Shuai Huang ,&nbsp;Qiang Guo ,&nbsp;Yuhang Tian ,&nbsp;Hao Feng ,&nbsp;Sergey Shulga","doi":"10.1016/j.dsp.2026.105963","DOIUrl":"10.1016/j.dsp.2026.105963","url":null,"abstract":"<div><div>In complex electromagnetic environments, radar pulse signals are strongly affected by noise, and limitations of reconnaissance receivers enlarge measurement errors, causing severe pulse missing and the inclusion of numerous spurious pulses. Consequently, pulse sorting faces two key difficulties: mining pulse association relations under missing information, and maintaining inter-class separability under serious parameter feature overlap. We propose a semi-supervised radar signal sorting method based on multiview subspace representation and graph learning (MvSR-GCN-RSS). First, encoders map multiple views into the latent space, where the view-specific and universal self-representation matrices are solved, and pulse sequence adjacency relations are constructed from intrapulse and interpulse information. Then, multiview information complementarity is achieved through a consistency loss and a diversity loss. In contrast to the two-stage process of first graph construction and then spectral clustering, we couple adjacency matrix solving with a graph convolutional network (GCN) in a single end-to-end framework, jointly optimizing it with the parameters of the multiview encoders and decoders to improve sorting efficiency. Finally, we design a multiview joint loss that simultaneously optimizes view reconstruction, GCN-based classification, self-representation solving, and cross-view complementarity for radar signal sorting. Simulation results show that the sorting accuracy reaches 99.99% in ideal scenarios; under scenarios with large measurement errors, pulse missing, and numerous spurious pulses, the proposed method performs far better than the comparison algorithms.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105963"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174904","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
YOLO-MBL: An infrared small target detection algorithm based on YOLOv11 YOLO-MBL:基于YOLOv11的红外小目标检测算法
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-29 DOI: 10.1016/j.dsp.2026.105968
Yixuan Shen, Mei Da, Lin Jiang
To address the deficiencies of existing infrared image detection models in terms of detection accuracy, computational complexity, detection speed, as well as missed detections and false detections in complex backgrounds, this paper proposes a lightweight infrared small target detection algorithm: YOLO - MBL. Firstly, we design a Dynamic Convolution Multi - Path Fusion Module (DCMP) to replace the original C3k2 module to enhance the feature extraction capability of the network. Secondly, we design the SDI - BiFPN as a feature fusion module in the neck network to capture more comprehensive feature information, thereby effectively avoiding the loss of information during the transmission process. Furthermore, a Lightweight Shared Convolutional Detection Head (LSCD) is introduced to reduce the number of model parameters. Finally, the Wise - MPDIoU loss function is adopted to accelerate the model convergence process and enhance its detection accuracy. To validate the effectiveness of the YOLO - MBL algorithm, we conducted comparative experiments on the FLIR dataset and the HIT - UAV dataset. The experimental results demonstrate that the YOLO - MBL model achieves a 4.6% improvement in detection accuracy ([email protected]) on the FLIR dataset, with a parameter reduction of 0.2 M, and reaches an FPS of 81.1. On the HIT - UAV dataset, the model's detection accuracy ([email protected]) is enhanced by 3.7%, accompanied by a parameter reduction of 0.2 M, and the FPS attains 84.1. Compared with traditional algorithms and current mainstream one - stage detection algorithms, the YOLO - MBL algorithm demonstrates significant advantages in terms of detection accuracy. The code repository is available at: https://github.com/yixixi12/YOLO-MBL.git.
针对现有红外图像检测模型在检测精度、计算复杂度、检测速度以及复杂背景下的漏检和误检等方面的不足,本文提出了一种轻量级红外小目标检测算法:YOLO - MBL。首先,我们设计了一个动态卷积多路径融合模块(DCMP)来取代原有的C3k2模块,以增强网络的特征提取能力。其次,我们将SDI - BiFPN设计为颈部网络中的特征融合模块,以捕获更全面的特征信息,从而有效避免信息在传输过程中的丢失。此外,引入轻量级共享卷积检测头(LSCD)来减少模型参数的数量。最后,采用Wise - MPDIoU损失函数加速了模型的收敛过程,提高了模型的检测精度。为了验证YOLO - MBL算法的有效性,我们在FLIR数据集和HIT - UAV数据集上进行了对比实验。实验结果表明,YOLO - MBL模型在FLIR数据集上的检测精度([email protected])提高了4.6%,参数减少了0.2 M, FPS达到81.1。在HIT - UAV数据集上,该模型的检测精度([email protected])提高了3.7%,参数降低了0.2 M, FPS达到84.1。与传统算法和当前主流的一级检测算法相比,YOLO - MBL算法在检测精度上具有显著的优势。代码存储库可从https://github.com/yixixi12/YOLO-MBL.git获得。
{"title":"YOLO-MBL: An infrared small target detection algorithm based on YOLOv11","authors":"Yixuan Shen,&nbsp;Mei Da,&nbsp;Lin Jiang","doi":"10.1016/j.dsp.2026.105968","DOIUrl":"10.1016/j.dsp.2026.105968","url":null,"abstract":"<div><div>To address the deficiencies of existing infrared image detection models in terms of detection accuracy, computational complexity, detection speed, as well as missed detections and false detections in complex backgrounds, this paper proposes a lightweight infrared small target detection algorithm: YOLO - MBL. Firstly, we design a Dynamic Convolution Multi - Path Fusion Module (DCMP) to replace the original C3k2 module to enhance the feature extraction capability of the network. Secondly, we design the SDI - BiFPN as a feature fusion module in the neck network to capture more comprehensive feature information, thereby effectively avoiding the loss of information during the transmission process. Furthermore, a Lightweight Shared Convolutional Detection Head (LSCD) is introduced to reduce the number of model parameters. Finally, the Wise - MPDIoU loss function is adopted to accelerate the model convergence process and enhance its detection accuracy. To validate the effectiveness of the YOLO - MBL algorithm, we conducted comparative experiments on the FLIR dataset and the HIT - UAV dataset. The experimental results demonstrate that the YOLO - MBL model achieves a 4.6% improvement in detection accuracy ([email protected]) on the FLIR dataset, with a parameter reduction of 0.2 M, and reaches an FPS of 81.1. On the HIT - UAV dataset, the model's detection accuracy ([email protected]) is enhanced by 3.7%, accompanied by a parameter reduction of 0.2 M, and the FPS attains 84.1. Compared with traditional algorithms and current mainstream one - stage detection algorithms, the YOLO - MBL algorithm demonstrates significant advantages in terms of detection accuracy. The code repository is available at: <span><span>https://github.com/yixixi12/YOLO-MBL.git</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105968"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
WCC-Net : Lightweight automatic modulation recognition of integrated underwater acoustic signals WCC-Net:集成水声信号的轻量级自动调制识别
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-27 DOI: 10.1016/j.dsp.2026.105961
Xuerong Cui , Kai Zheng , Juan Li , Lei Li , Bin Jiang
To support the detection and communication requirements of offshore devices operating under stringent resource constraints, it is essential to overcome the challenges posed by complex underwater acoustic channels and intense ocean noise. Consequently, designing a lightweight automatic modulation recognition (AMR) algorithm for integrated underwater acoustic detection and communication signals is particularly challenging. Despite recent advances, current AMR algorithms still exhibit limitations in computational speed and resource usage. Moreover, to date, no AMR method has been specifically designed for integrated acoustic detection and communication (IADC) signal frameworks. To address these issues, this paper proposes a Wavelet Complex Convolution Network (WCC-Net) that directly uses in-phase/quadrature (I/Q) signals as input. First, the in-phase and quadrature components of the signal are each fed into two independent wavelet convolution modules, which simultaneously enlarge the receptive field and suppress noise. Then, a complex convolution module preserves the phase coupling information while efficiently mixing the feature information. Finally, an efficient feature mixing module combines and refines the high-dimensional features to produce the classification result, reducing redundant information and enhancing feature interaction. Experimental results indicate that, at about 89% recognition accuracy, WCC-Net reduces the computational complexity by 84.76% and the number of parameters by 88.82%; under the same model complexity, WCC-Net accuracy is improved by at least 6.91%. Even under real-world ocean noise conditions, WCC-Net attains competitive recognition accuracy with minimal model complexity.
为了支持在严格的资源限制下运行的海上设备的检测和通信需求,必须克服复杂的水声通道和强烈的海洋噪声带来的挑战。因此,设计一种轻量级的自动调制识别(AMR)算法来集成水声探测和通信信号是非常具有挑战性的。尽管最近取得了进展,但目前的AMR算法在计算速度和资源使用方面仍然存在局限性。此外,到目前为止,还没有专门为集成声学探测和通信(IADC)信号框架设计的AMR方法。为了解决这些问题,本文提出了一种直接使用同相/正交(I/Q)信号作为输入的小波复卷积网络(WCC-Net)。首先,将信号的同相分量和正交分量分别送入两个独立的小波卷积模块,同时放大接收野和抑制噪声。复卷积模块在有效混合特征信息的同时,保留了相位耦合信息。最后,利用高效的特征混合模块对高维特征进行组合和提炼,生成分类结果,减少冗余信息,增强特征交互性。实验结果表明,在89%左右的识别准确率下,WCC-Net将计算复杂度降低了84.76%,将参数数量减少了88.82%;在相同的模型复杂度下,WCC-Net的精度至少提高了6.91%。即使在现实世界的海洋噪声条件下,WCC-Net也能以最小的模型复杂性获得具有竞争力的识别精度。
{"title":"WCC-Net : Lightweight automatic modulation recognition of integrated underwater acoustic signals","authors":"Xuerong Cui ,&nbsp;Kai Zheng ,&nbsp;Juan Li ,&nbsp;Lei Li ,&nbsp;Bin Jiang","doi":"10.1016/j.dsp.2026.105961","DOIUrl":"10.1016/j.dsp.2026.105961","url":null,"abstract":"<div><div>To support the detection and communication requirements of offshore devices operating under stringent resource constraints, it is essential to overcome the challenges posed by complex underwater acoustic channels and intense ocean noise. Consequently, designing a lightweight automatic modulation recognition (AMR) algorithm for integrated underwater acoustic detection and communication signals is particularly challenging. Despite recent advances, current AMR algorithms still exhibit limitations in computational speed and resource usage. Moreover, to date, no AMR method has been specifically designed for integrated acoustic detection and communication (IADC) signal frameworks. To address these issues, this paper proposes a Wavelet Complex Convolution Network (WCC-Net) that directly uses in-phase/quadrature (I/Q) signals as input. First, the in-phase and quadrature components of the signal are each fed into two independent wavelet convolution modules, which simultaneously enlarge the receptive field and suppress noise. Then, a complex convolution module preserves the phase coupling information while efficiently mixing the feature information. Finally, an efficient feature mixing module combines and refines the high-dimensional features to produce the classification result, reducing redundant information and enhancing feature interaction. Experimental results indicate that, at about 89% recognition accuracy, WCC-Net reduces the computational complexity by 84.76% and the number of parameters by 88.82%; under the same model complexity, WCC-Net accuracy is improved by at least 6.91%. Even under real-world ocean noise conditions, WCC-Net attains competitive recognition accuracy with minimal model complexity.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105961"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174909","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MA-YOLO: Enhanced multi-scale attentional remote sensing detector 增强型多尺度关注遥感探测器
IF 3 3区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC Pub Date : 2026-04-15 Epub Date: 2026-01-21 DOI: 10.1016/j.dsp.2026.105948
Zikai Chen , Degang Yang , Tingting Song , Yichen Ye , Yongli Liu , Xin Zhang
With the continuous development of deep learning technology, object detection tasks in remote sensing images have received increasing attention. However, due to the diversity of object scales and the complexity of background environments, current detectors often find it difficult to control computational costs while ensuring high performance. To address these challenges, we design a remote sensing image object detector called MA-YOLO, which integrates multi-scale features and attention mechanisms. We design the mixed receptive field attention convolution (MRFAConv) module to strengthen the backbone network, which is a non-parametric shared convolution that takes into account both spatial and channel attention. Moreover, a multi-scale receptive field downsampling module (MRFD) is proposed, which can extract rich feature information from different receptive fields while effectively reducing information loss. Ultimately, a lightweight multi-scale attention module (LMSA) is designed and integrated into the neck network to further optimize the feature fusion effect. Extensive experiments conducted on the DIOR and TGRS-HRRSD datasets reveal that MA-YOLO enhances the mAP by 2.1% and 5.3%, respectively, compared to the baseline model YOLOv8n, while slightly reducing computational overhead and decreasing the number of parameters by 6.7%. These experimental results fully demonstrate the remarkable effectiveness of our proposed method in enhancing the detection accuracy of remote sensing images. The code will be available at https://github.com/Zikai-Chen/MA-YOLO.
随着深度学习技术的不断发展,遥感图像中的目标检测任务越来越受到重视。然而,由于目标尺度的多样性和背景环境的复杂性,当前检测器往往难以在保证高性能的同时控制计算成本。为了解决这些挑战,我们设计了一种名为MA-YOLO的遥感图像目标探测器,该探测器集成了多尺度特征和注意机制。为了增强骨干网,我们设计了混合感受野注意卷积(MRFAConv)模块,这是一种同时考虑空间和通道注意的非参数共享卷积。此外,提出了一种多尺度感受野降采样模块(MRFD),可以从不同的感受野中提取丰富的特征信息,同时有效降低信息损失。最后,设计轻量级多尺度注意力模块(LMSA)并集成到颈部网络中,进一步优化特征融合效果。在DIOR和TGRS-HRRSD数据集上进行的大量实验表明,与基线模型YOLOv8n相比,MA-YOLO的mAP分别提高了2.1%和5.3%,同时略微减少了计算开销,参数数量减少了6.7%。这些实验结果充分证明了本文方法在提高遥感图像检测精度方面的显著有效性。代码可在https://github.com/Zikai-Chen/MA-YOLO上获得。
{"title":"MA-YOLO: Enhanced multi-scale attentional remote sensing detector","authors":"Zikai Chen ,&nbsp;Degang Yang ,&nbsp;Tingting Song ,&nbsp;Yichen Ye ,&nbsp;Yongli Liu ,&nbsp;Xin Zhang","doi":"10.1016/j.dsp.2026.105948","DOIUrl":"10.1016/j.dsp.2026.105948","url":null,"abstract":"<div><div>With the continuous development of deep learning technology, object detection tasks in remote sensing images have received increasing attention. However, due to the diversity of object scales and the complexity of background environments, current detectors often find it difficult to control computational costs while ensuring high performance. To address these challenges, we design a remote sensing image object detector called MA-YOLO, which integrates multi-scale features and attention mechanisms. We design the mixed receptive field attention convolution (MRFAConv) module to strengthen the backbone network, which is a non-parametric shared convolution that takes into account both spatial and channel attention. Moreover, a multi-scale receptive field downsampling module (MRFD) is proposed, which can extract rich feature information from different receptive fields while effectively reducing information loss. Ultimately, a lightweight multi-scale attention module (LMSA) is designed and integrated into the neck network to further optimize the feature fusion effect. Extensive experiments conducted on the DIOR and TGRS-HRRSD datasets reveal that MA-YOLO enhances the mAP by 2.1% and 5.3%, respectively, compared to the baseline model YOLOv8n, while slightly reducing computational overhead and decreasing the number of parameters by 6.7%. These experimental results fully demonstrate the remarkable effectiveness of our proposed method in enhancing the detection accuracy of remote sensing images. The code will be available at <span><span>https://github.com/Zikai-Chen/MA-YOLO</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":51011,"journal":{"name":"Digital Signal Processing","volume":"174 ","pages":"Article 105948"},"PeriodicalIF":3.0,"publicationDate":"2026-04-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"146174910","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Digital Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1