{"title":"DAM-Net: Flood detection from SAR imagery using differential attention metric-based vision transformers","authors":"Tamer Saleh , Xingxing Weng , Shimaa Holail , Chen Hao , Gui-Song Xia","doi":"10.1016/j.isprsjprs.2024.05.018","DOIUrl":null,"url":null,"abstract":"<div><p>Flood detection from synthetic aperture radar (SAR) imagery plays an important role in crisis and disaster management. Based on pre- and post-flood SAR images, flooded areas can be extracted by detecting changes of water bodies. Existing state-of-the-art change detection methods primarily target optical image pairs. The nature of SAR images, such as scarce visual information, similar backscatter signals, and ubiquitous speckle noise, pose great challenges to identifying water bodies and mining change features, thus resulting in unsatisfactory performance. Besides, the lack of large-scale annotated datasets hinders the development of accurate flood detection methods. In this paper, we focus on the difference between SAR image pairs and present a differential attention metric-based network (DAM-Net), to achieve flood detection. By introducing feature interaction during temporal-wise feature representation, we guide the model to focus on changes of interest rather than fully understanding the scene of the image. On the other hand, we devise a class token to capture high-level semantic information about water body changes, increasing the ability to distinguish water body changes and pseudo changes caused by similar signals or speckle noise. To better train and evaluate DAM-Net, we create a large-scale flood detection dataset using Sentinel-1 SAR imagery, namely <em>S1GFloods</em>. This dataset consists of 5,360 image pairs, covering 46 flood events during 2015–2022, and spanning 6 continents of the world. The experimental results on this dataset demonstrate that our method outperforms several advanced change detection methods. DAM-Net achieves 97.8% overall accuracy, 96.5% F1, and 93.2% IoU on the test set. Our dataset and code are available at <span>https://github.com/Tamer-Saleh/S1GFlood-Detection</span><svg><path></path></svg>.</p></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":null,"pages":null},"PeriodicalIF":10.6000,"publicationDate":"2024-05-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271624002168","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
引用次数: 0
Abstract
Flood detection from synthetic aperture radar (SAR) imagery plays an important role in crisis and disaster management. Based on pre- and post-flood SAR images, flooded areas can be extracted by detecting changes of water bodies. Existing state-of-the-art change detection methods primarily target optical image pairs. The nature of SAR images, such as scarce visual information, similar backscatter signals, and ubiquitous speckle noise, pose great challenges to identifying water bodies and mining change features, thus resulting in unsatisfactory performance. Besides, the lack of large-scale annotated datasets hinders the development of accurate flood detection methods. In this paper, we focus on the difference between SAR image pairs and present a differential attention metric-based network (DAM-Net), to achieve flood detection. By introducing feature interaction during temporal-wise feature representation, we guide the model to focus on changes of interest rather than fully understanding the scene of the image. On the other hand, we devise a class token to capture high-level semantic information about water body changes, increasing the ability to distinguish water body changes and pseudo changes caused by similar signals or speckle noise. To better train and evaluate DAM-Net, we create a large-scale flood detection dataset using Sentinel-1 SAR imagery, namely S1GFloods. This dataset consists of 5,360 image pairs, covering 46 flood events during 2015–2022, and spanning 6 continents of the world. The experimental results on this dataset demonstrate that our method outperforms several advanced change detection methods. DAM-Net achieves 97.8% overall accuracy, 96.5% F1, and 93.2% IoU on the test set. Our dataset and code are available at https://github.com/Tamer-Saleh/S1GFlood-Detection.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.