GPR-CUNet: Spatio-Temporal Feature Fusion-Based GPR Forward and Inversion Cycle Network for Root Scene Survey

IF 4.3 2区 综合性期刊 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC IEEE Sensors Journal Pub Date : 2025-01-03 DOI:10.1109/JSEN.2024.3522888
Xiaowei Zhang;Xuan Zhao;Shuang Li;Shenghua Lv;Chen Lin;Jian Wen
{"title":"GPR-CUNet: Spatio-Temporal Feature Fusion-Based GPR Forward and Inversion Cycle Network for Root Scene Survey","authors":"Xiaowei Zhang;Xuan Zhao;Shuang Li;Shenghua Lv;Chen Lin;Jian Wen","doi":"10.1109/JSEN.2024.3522888","DOIUrl":null,"url":null,"abstract":"Ground penetrating radar (GPR) forward and inversion methods are key techniques for studying radar imaging mechanisms and investigating subsurface scenes. Efficiently interpreting radar wave data will facilitate the development of subsurface structure detection applications, especially in the intricate plant root distribution. Existing forward and inversion models are constrained by the highly computational and time-consuming forward process, making it difficult to be applied to complex real-world subsurface scenarios. Inspired by the spatio-temporal properties during radar wave imaging, a spatial and temporal fusion cycle U-shaped model named GPR-CUNet was proposed. The model is more adapted to the transformation between permittivity distribution and GPR B-Scan data in complex environment. First, to extract the spatial and temporal features from the permittivity distribution and radar data, a spatio-temporal feature fusion module (STFM) based on CNN and BiLSTM was designed. Then, for the translation between the permittivity distribution and the radar wave data, two identical U-shaped networks with the STFM were constructed. Finally, guided by predictive consistency and cyclic consistency, a hybrid loss function based on multiscale structural similarity (MS-SSIM) and L1 norm was configured to boost the performance of both the forward and inversion networks. The numerical simulation experiments revealed that the proposed model imparted exceptional performance and efficiency in the prediction of radar wave features and reconstruction of permittivity distribution under complex scenarios. In preburial experiments and field root testing, our inversion model can effectively recover the subsurface root and soil horizons distribution. Accurate permittivity distribution of subsurface scene can provide a theoretical basis for imaging and 3-D reconstruction of the physical media distribution in plant root zones.","PeriodicalId":447,"journal":{"name":"IEEE Sensors Journal","volume":"25 4","pages":"7569-7583"},"PeriodicalIF":4.3000,"publicationDate":"2025-01-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Sensors Journal","FirstCategoryId":"103","ListUrlMain":"https://ieeexplore.ieee.org/document/10824685/","RegionNum":2,"RegionCategory":"综合性期刊","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Ground penetrating radar (GPR) forward and inversion methods are key techniques for studying radar imaging mechanisms and investigating subsurface scenes. Efficiently interpreting radar wave data will facilitate the development of subsurface structure detection applications, especially in the intricate plant root distribution. Existing forward and inversion models are constrained by the highly computational and time-consuming forward process, making it difficult to be applied to complex real-world subsurface scenarios. Inspired by the spatio-temporal properties during radar wave imaging, a spatial and temporal fusion cycle U-shaped model named GPR-CUNet was proposed. The model is more adapted to the transformation between permittivity distribution and GPR B-Scan data in complex environment. First, to extract the spatial and temporal features from the permittivity distribution and radar data, a spatio-temporal feature fusion module (STFM) based on CNN and BiLSTM was designed. Then, for the translation between the permittivity distribution and the radar wave data, two identical U-shaped networks with the STFM were constructed. Finally, guided by predictive consistency and cyclic consistency, a hybrid loss function based on multiscale structural similarity (MS-SSIM) and L1 norm was configured to boost the performance of both the forward and inversion networks. The numerical simulation experiments revealed that the proposed model imparted exceptional performance and efficiency in the prediction of radar wave features and reconstruction of permittivity distribution under complex scenarios. In preburial experiments and field root testing, our inversion model can effectively recover the subsurface root and soil horizons distribution. Accurate permittivity distribution of subsurface scene can provide a theoretical basis for imaging and 3-D reconstruction of the physical media distribution in plant root zones.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
GPR- cunet:基于时空特征融合的GPR根景测量正反演循环网络
探地雷达正反演方法是研究雷达成像机理和调查地下场景的关键技术。有效地解释雷达波数据将促进地下结构探测应用的发展,特别是在复杂的植物根系分布中。现有的正演和反演模型受到计算量大、耗时长的正演过程的限制,难以应用于复杂的真实地下场景。基于雷达波成像的时空特性,提出了一种时空融合周期u型模型GPR-CUNet。该模型更适合于复杂环境下介电常数分布与探地雷达b扫描数据之间的转换。首先,设计了基于CNN和BiLSTM的时空特征融合模块(STFM),从介电常数分布和雷达数据中提取时空特征;然后,为了在介电常数分布和雷达波数据之间进行转换,构造了两个具有STFM的u型网络。最后,以预测一致性和循环一致性为指导,配置基于多尺度结构相似度(MS-SSIM)和L1范数的混合损失函数,提高正反演网络的性能。数值模拟实验表明,该模型在复杂场景下的雷达波特征预测和介电常数分布重建方面具有优异的性能和效率。在埋前试验和田间根系试验中,反演模型能有效地恢复地下根系和土壤层位分布。准确的地下场景介电常数分布可以为植物根区物理介质分布的成像和三维重建提供理论依据。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Sensors Journal
IEEE Sensors Journal 工程技术-工程:电子与电气
CiteScore
7.70
自引率
14.00%
发文量
2058
审稿时长
5.2 months
期刊介绍: The fields of interest of the IEEE Sensors Journal are the theory, design , fabrication, manufacturing and applications of devices for sensing and transducing physical, chemical and biological phenomena, with emphasis on the electronics and physics aspect of sensors and integrated sensors-actuators. IEEE Sensors Journal deals with the following: -Sensor Phenomenology, Modelling, and Evaluation -Sensor Materials, Processing, and Fabrication -Chemical and Gas Sensors -Microfluidics and Biosensors -Optical Sensors -Physical Sensors: Temperature, Mechanical, Magnetic, and others -Acoustic and Ultrasonic Sensors -Sensor Packaging -Sensor Networks -Sensor Applications -Sensor Systems: Signals, Processing, and Interfaces -Actuators and Sensor Power Systems -Sensor Signal Processing for high precision and stability (amplification, filtering, linearization, modulation/demodulation) and under harsh conditions (EMC, radiation, humidity, temperature); energy consumption/harvesting -Sensor Data Processing (soft computing with sensor data, e.g., pattern recognition, machine learning, evolutionary computation; sensor data fusion, processing of wave e.g., electromagnetic and acoustic; and non-wave, e.g., chemical, gravity, particle, thermal, radiative and non-radiative sensor data, detection, estimation and classification based on sensor data) -Sensors in Industrial Practice
期刊最新文献
Dynamic State Estimation of Integrated Energy Systems Based on SCKF Cost-Effective Digilog Chirp Thermal Wave Imaging System for Composite Subsurface Evaluation TS-ResCNN: Efficient Skeleton-Based Action Recognition for Edge IoT Sensor Systems Flexible Piezoelectric Tactile Sensing System for Intelligent Object Recognition A Multi-PPG Patch-Type Flexible Sensor System Recording Blood Pulse Timings on the Forearm as a Potential Indicator of Blood Pressure
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1