Xianping Fu;Wenqiang Qin;Fengqi Li;Fengqiang Xu;Xiaohong Yan
{"title":"CFPNet: Complementary Feature Perception Network for Underwater Image Enhancement","authors":"Xianping Fu;Wenqiang Qin;Fengqi Li;Fengqiang Xu;Xiaohong Yan","doi":"10.1109/JOE.2024.3463838","DOIUrl":null,"url":null,"abstract":"Images shot underwater are usually characterized by global nonuniform information loss due to selective light absorption and scattering, resulting in various degradation problems, such as color distortion and low visibility. Recently, deep learning has drawn much attention in the field of underwater image enhancement (UIE) for its powerful performance. However, most deep learning-based UIE models rely on either pure convolutional neural network (CNN) or pure transformer, which makes it challenging to enhance images while maintaining local representations and global features simultaneously. In this article, we propose a novel complementary feature perception network (CFPNet), which embeds the transformer into the classical CNN-based UNet3+. The core idea is to fuse the advantages of CNN and transformer to obtain satisfactory high-quality underwater images that can naturally perceive local and global features. CFPNet employs a novel dual encoder structure of the CNN and transformer in parallel, while the decoder is composed of one trunk decoder and two auxiliary decoders. First, we propose the regionalized two-stage vision transformer that can progressively eliminate the variable levels of degradation in a coarse-to-fine manner. Second, we design the full-scale feature fusion module to explore sufficient information by merging the multiscale features. In addition, we propose an auxiliary feature guided learning strategy that utilizes reflectance and shading maps to guide the generation of the final results. The advantage of this strategy is to avoid repetitive and ineffective learning of the model, and to accomplish color correction and deblurring tasks more efficiently. Experiments demonstrate that our CFPNet can obtain high-quality underwater images and show superior performance compared to the state-of-the-art UIE methods qualitatively and quantitatively.","PeriodicalId":13191,"journal":{"name":"IEEE Journal of Oceanic Engineering","volume":"50 1","pages":"150-163"},"PeriodicalIF":3.8000,"publicationDate":"2024-11-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Journal of Oceanic Engineering","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/10754901/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, CIVIL","Score":null,"Total":0}
引用次数: 0
Abstract
Images shot underwater are usually characterized by global nonuniform information loss due to selective light absorption and scattering, resulting in various degradation problems, such as color distortion and low visibility. Recently, deep learning has drawn much attention in the field of underwater image enhancement (UIE) for its powerful performance. However, most deep learning-based UIE models rely on either pure convolutional neural network (CNN) or pure transformer, which makes it challenging to enhance images while maintaining local representations and global features simultaneously. In this article, we propose a novel complementary feature perception network (CFPNet), which embeds the transformer into the classical CNN-based UNet3+. The core idea is to fuse the advantages of CNN and transformer to obtain satisfactory high-quality underwater images that can naturally perceive local and global features. CFPNet employs a novel dual encoder structure of the CNN and transformer in parallel, while the decoder is composed of one trunk decoder and two auxiliary decoders. First, we propose the regionalized two-stage vision transformer that can progressively eliminate the variable levels of degradation in a coarse-to-fine manner. Second, we design the full-scale feature fusion module to explore sufficient information by merging the multiscale features. In addition, we propose an auxiliary feature guided learning strategy that utilizes reflectance and shading maps to guide the generation of the final results. The advantage of this strategy is to avoid repetitive and ineffective learning of the model, and to accomplish color correction and deblurring tasks more efficiently. Experiments demonstrate that our CFPNet can obtain high-quality underwater images and show superior performance compared to the state-of-the-art UIE methods qualitatively and quantitatively.
期刊介绍:
The IEEE Journal of Oceanic Engineering (ISSN 0364-9059) is the online-only quarterly publication of the IEEE Oceanic Engineering Society (IEEE OES). The scope of the Journal is the field of interest of the IEEE OES, which encompasses all aspects of science, engineering, and technology that address research, development, and operations pertaining to all bodies of water. This includes the creation of new capabilities and technologies from concept design through prototypes, testing, and operational systems to sense, explore, understand, develop, use, and responsibly manage natural resources.