Interactive and Supervised Dual-Mode Attention Network for Remote Sensing Image Change Detection

IF 8.6 1区 地球科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC IEEE Transactions on Geoscience and Remote Sensing Pub Date : 2025-02-11 DOI:10.1109/TGRS.2025.3540864
Hongjin Ren;Min Xia;Liguo Weng;Haifeng Lin;Junqing Huang;Kai Hu
{"title":"Interactive and Supervised Dual-Mode Attention Network for Remote Sensing Image Change Detection","authors":"Hongjin Ren;Min Xia;Liguo Weng;Haifeng Lin;Junqing Huang;Kai Hu","doi":"10.1109/TGRS.2025.3540864","DOIUrl":null,"url":null,"abstract":"With the rapid advancement of remote sensing technology, change detection using bitemporal remote sensing images has significant applications in land use planning and environmental monitoring. The emergence of convolutional neural networks (CNNs) has accelerated the development of deep learning-based change detection. However, existing deep learning algorithms exhibit limitations in understanding bitemporal feature relationships and accurately identifying change region boundaries. Moreover, they inadequately explore feature interactions between bitemporal images before extracting differential features. To address these issues, this article proposes a novel interactive and supervised dual-mode attention network (ISDANet). In the feature encoding stage, we employ the lightweight MobileNetV2 as the backbone to extract bitemporal features. Additionally, we design the neighbor feature aggregation module (NFAM) to aggregate semantic features from adjacent scales within the dual-branch backbone, enhancing the representation of temporal features. We further introduce the interactive attention enhancement module (IAEM), which effectively integrates self-attention and cross-attention mechanisms. This establishes deep interactions between bitemporal features, suppresses irrelevant noise, and ensures precise focus on true change regions. In the feature decoding stage, the supervised attention module (SAM) reweights differential features and leverages supervisory signals to guide the learning of attention mechanisms, significantly improving boundary detection accuracy. SAM dynamically aggregates multilevel features, balancing high-level semantics and low-level details to capture subtle changes in complex scenes. The proposed model achieves F1 scores that are 0.28%, 1.6%, and 0.76% higher than the best comparative method, spatiotemporal enhancement and interlevel fusion network (SEIFNet), on three CD datasets [LEVIR-CD, Guangzhou dataset (GZ-CD), and Sun Yat-sen University dataset (SYSU-CD)], respectively, while maintaining a lightweight design with only 6.93 M parameters and 3.46G floating-point operations (FLOPs). The code is available at <uri>https://github.com/RenHongjin6/ISDANet</uri>.","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-18"},"PeriodicalIF":8.6000,"publicationDate":"2025-02-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10879780/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

With the rapid advancement of remote sensing technology, change detection using bitemporal remote sensing images has significant applications in land use planning and environmental monitoring. The emergence of convolutional neural networks (CNNs) has accelerated the development of deep learning-based change detection. However, existing deep learning algorithms exhibit limitations in understanding bitemporal feature relationships and accurately identifying change region boundaries. Moreover, they inadequately explore feature interactions between bitemporal images before extracting differential features. To address these issues, this article proposes a novel interactive and supervised dual-mode attention network (ISDANet). In the feature encoding stage, we employ the lightweight MobileNetV2 as the backbone to extract bitemporal features. Additionally, we design the neighbor feature aggregation module (NFAM) to aggregate semantic features from adjacent scales within the dual-branch backbone, enhancing the representation of temporal features. We further introduce the interactive attention enhancement module (IAEM), which effectively integrates self-attention and cross-attention mechanisms. This establishes deep interactions between bitemporal features, suppresses irrelevant noise, and ensures precise focus on true change regions. In the feature decoding stage, the supervised attention module (SAM) reweights differential features and leverages supervisory signals to guide the learning of attention mechanisms, significantly improving boundary detection accuracy. SAM dynamically aggregates multilevel features, balancing high-level semantics and low-level details to capture subtle changes in complex scenes. The proposed model achieves F1 scores that are 0.28%, 1.6%, and 0.76% higher than the best comparative method, spatiotemporal enhancement and interlevel fusion network (SEIFNet), on three CD datasets [LEVIR-CD, Guangzhou dataset (GZ-CD), and Sun Yat-sen University dataset (SYSU-CD)], respectively, while maintaining a lightweight design with only 6.93 M parameters and 3.46G floating-point operations (FLOPs). The code is available at https://github.com/RenHongjin6/ISDANet.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用于遥感图像变化检测的交互式和监督式双模式注意力网络
随着遥感技术的飞速发展,利用双时相遥感影像进行变化检测在土地利用规划和环境监测等领域具有重要的应用价值。卷积神经网络(cnn)的出现加速了基于深度学习的变化检测的发展。然而,现有的深度学习算法在理解双时特征关系和准确识别变化区域边界方面存在局限性。此外,在提取差分特征之前,他们没有充分探索双时图像之间的特征交互。为了解决这些问题,本文提出了一种新的交互式和监督双模注意力网络(ISDANet)。在特征编码阶段,我们采用轻量级的MobileNetV2作为主干提取双时相特征。此外,我们设计了相邻特征聚合模块(NFAM)来聚合双分支主干内相邻尺度的语义特征,增强了时间特征的表征。我们进一步介绍了交互式注意增强模块(IAEM),它有效地集成了自注意和交叉注意机制。这在双时间特征之间建立了深层的相互作用,抑制了不相关的噪声,并确保了对真实变化区域的精确关注。在特征解码阶段,有监督注意模块(SAM)对差分特征进行加权,利用监督信号指导注意机制的学习,显著提高了边界检测的准确率。SAM动态聚合多层特征,平衡高级语义和低级细节,以捕获复杂场景中的细微变化。在三个CD数据集[levirc -CD、广州数据集(GZ-CD)和中山大学数据集(SYSU-CD)]上,该模型的F1得分分别比最佳对比方法时空增强和层间融合网络(SEIFNet)高0.28%、1.6%和0.76%,同时保持了轻量级设计,只有6.93 M个参数和3.46G浮点运算(FLOPs)。代码可在https://github.com/RenHongjin6/ISDANet上获得。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Transactions on Geoscience and Remote Sensing
IEEE Transactions on Geoscience and Remote Sensing 工程技术-地球化学与地球物理
CiteScore
11.50
自引率
28.00%
发文量
1912
审稿时长
4.0 months
期刊介绍: IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.
期刊最新文献
Distribution-Aware Infrared Small Target Detection Based on Multi-Scale Convolutional Decoder and Hypergraph Attention Detecting Weak Underwater Targets in Hyperspectral Imagery via Physics-aware Residual Reasoning Faint Bottom Echo Detection for Airborne LiDAR Bathymetry Based on a Constrained Waveform Stacking Model First Cooperative Formaldehyde Monitoring with Chinese Morning and Afternoon Satellites: Revealing Global Multi-Temporal Concentration Dynamics Fast Anchor Graph Regularized Relaxation Linear Regression for Classification of Hyperspectral Images
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1