Zhiyong Huang;Hongjiang Qiu;Mingyang Hou;Zhi Yu;Shiwei Wang;Xiaoyu Li;Jiahong Wang;Yan Yan;Yushi Liu
{"title":"MCECF: A Multiscale Complementary Enhanced Context Fusion Network for Remote Sensing Change Detection","authors":"Zhiyong Huang;Hongjiang Qiu;Mingyang Hou;Zhi Yu;Shiwei Wang;Xiaoyu Li;Jiahong Wang;Yan Yan;Yushi Liu","doi":"10.1109/TGRS.2025.3556237","DOIUrl":null,"url":null,"abstract":"Remote sensing change detection (RSCD) holds significant research value in remote sensing (RS) image processing. In recent years, many researchers have achieved remarkable results in RSCD tasks using methods based on convolutional neural networks (CNNs) or Transformers. Considering the limited receptive field of CNN models and the high computational cost of Transformers, many researchers have combined the two approaches, yielding promising results. However, most current RSCD-based models focus solely on change and temporal information, overlooking their complementary relationship. Additionally, some multiscale feature fusion methods emphasize enhancing individual scales while neglecting the correlations between different scales. To address the above issues, we propose a multiscale complementary enhanced context fusion (MCECF) network. The network first introduces a global-local context aggregation module (GLCAM) to capture global-local context information while extracting multilevel feature maps. Subsequently, a complementary enhancement difference module (CEDM) is employed to complementarily aggregate the captured change and temporal information of bi-temporal RS image features. To fully leverage the correlations between multiscale features, a progressive decoder comprising a supervised spatial attention (SSA) mechanism and a multiscale complementary enhanced fusion module (MCEFM) was developed. Moreover, to tackle the disparity between changed and unchanged regions, a dual-branch dynamic attention fusion module (DAFM) was designed to enhance the model’s adaptability to diverse scenarios. We conducted comparative experiments on five RSCD datasets against nine state-of-the-art (SOTA) methods, and the results confirmed the effectiveness of the proposed MCECF in RSCD tasks. Our code will be made available at <uri>https://github.com/kakuqikaduo/MCECF</uri>","PeriodicalId":13213,"journal":{"name":"IEEE Transactions on Geoscience and Remote Sensing","volume":"63 ","pages":"1-14"},"PeriodicalIF":8.6000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Geoscience and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10945975/","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
Remote sensing change detection (RSCD) holds significant research value in remote sensing (RS) image processing. In recent years, many researchers have achieved remarkable results in RSCD tasks using methods based on convolutional neural networks (CNNs) or Transformers. Considering the limited receptive field of CNN models and the high computational cost of Transformers, many researchers have combined the two approaches, yielding promising results. However, most current RSCD-based models focus solely on change and temporal information, overlooking their complementary relationship. Additionally, some multiscale feature fusion methods emphasize enhancing individual scales while neglecting the correlations between different scales. To address the above issues, we propose a multiscale complementary enhanced context fusion (MCECF) network. The network first introduces a global-local context aggregation module (GLCAM) to capture global-local context information while extracting multilevel feature maps. Subsequently, a complementary enhancement difference module (CEDM) is employed to complementarily aggregate the captured change and temporal information of bi-temporal RS image features. To fully leverage the correlations between multiscale features, a progressive decoder comprising a supervised spatial attention (SSA) mechanism and a multiscale complementary enhanced fusion module (MCEFM) was developed. Moreover, to tackle the disparity between changed and unchanged regions, a dual-branch dynamic attention fusion module (DAFM) was designed to enhance the model’s adaptability to diverse scenarios. We conducted comparative experiments on five RSCD datasets against nine state-of-the-art (SOTA) methods, and the results confirmed the effectiveness of the proposed MCECF in RSCD tasks. Our code will be made available at https://github.com/kakuqikaduo/MCECF
期刊介绍:
IEEE Transactions on Geoscience and Remote Sensing (TGRS) is a monthly publication that focuses on the theory, concepts, and techniques of science and engineering as applied to sensing the land, oceans, atmosphere, and space; and the processing, interpretation, and dissemination of this information.