Semantic segmentation of underwater images based on the improved SegFormer

IF 3 2区 生物学 Q1 MARINE & FRESHWATER BIOLOGY Frontiers in Marine Science Pub Date : 2025-03-11 DOI:10.3389/fmars.2025.1522160
Bowei Chen, Wei Zhao, Qiusheng Zhang, Mingliang Li, Mingyang Qi, You Tang
{"title":"Semantic segmentation of underwater images based on the improved SegFormer","authors":"Bowei Chen, Wei Zhao, Qiusheng Zhang, Mingliang Li, Mingyang Qi, You Tang","doi":"10.3389/fmars.2025.1522160","DOIUrl":null,"url":null,"abstract":"Underwater images segmentation is essential for tasks such as underwater exploration, marine environmental monitoring, and resource development. Nevertheless, given the complexity and variability of the underwater environment, improving model accuracy remains a key challenge in underwater image segmentation tasks. To address these issues, this study presents a high-performance semantic segmentation approach for underwater images based on the standard SegFormer model. First, the Mix Transformer backbone in SegFormer is replaced with a Swin Transformer to enhance feature extraction and facilitate efficient acquisition of global context information. Next, the Efficient Multi-scale Attention (EMA) mechanism is introduced in the backbone’s downsampling stages and the decoder to better capture multi-scale features, further improving segmentation accuracy. Furthermore, a Feature Pyramid Network (FPN) structure is incorporated into the decoder to combine feature maps at multiple resolutions, allowing the model to integrate contextual information effectively, enhancing robustness in complex underwater environments. Testing on the SUIM underwater image dataset shows that the proposed model achieves high performance across multiple metrics: mean Intersection over Union (MIoU) of 77.00%, mean Recall (mRecall) of 85.04%, mean Precision (mPrecision) of 89.03%, and mean F1score (mF1score) of 86.63%. Compared to the standard SegFormer, it demonstrates improvements of 3.73% in MIoU, 1.98% in mRecall, 3.38% in mPrecision, and 2.44% in mF1score, with an increase of 9.89M parameters. The results demonstrate that the proposed method achieves superior segmentation accuracy with minimal additional computation, showcasing high performance in underwater image segmentation.","PeriodicalId":12479,"journal":{"name":"Frontiers in Marine Science","volume":"21 1","pages":""},"PeriodicalIF":3.0000,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in Marine Science","FirstCategoryId":"99","ListUrlMain":"https://doi.org/10.3389/fmars.2025.1522160","RegionNum":2,"RegionCategory":"生物学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MARINE & FRESHWATER BIOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Underwater images segmentation is essential for tasks such as underwater exploration, marine environmental monitoring, and resource development. Nevertheless, given the complexity and variability of the underwater environment, improving model accuracy remains a key challenge in underwater image segmentation tasks. To address these issues, this study presents a high-performance semantic segmentation approach for underwater images based on the standard SegFormer model. First, the Mix Transformer backbone in SegFormer is replaced with a Swin Transformer to enhance feature extraction and facilitate efficient acquisition of global context information. Next, the Efficient Multi-scale Attention (EMA) mechanism is introduced in the backbone’s downsampling stages and the decoder to better capture multi-scale features, further improving segmentation accuracy. Furthermore, a Feature Pyramid Network (FPN) structure is incorporated into the decoder to combine feature maps at multiple resolutions, allowing the model to integrate contextual information effectively, enhancing robustness in complex underwater environments. Testing on the SUIM underwater image dataset shows that the proposed model achieves high performance across multiple metrics: mean Intersection over Union (MIoU) of 77.00%, mean Recall (mRecall) of 85.04%, mean Precision (mPrecision) of 89.03%, and mean F1score (mF1score) of 86.63%. Compared to the standard SegFormer, it demonstrates improvements of 3.73% in MIoU, 1.98% in mRecall, 3.38% in mPrecision, and 2.44% in mF1score, with an increase of 9.89M parameters. The results demonstrate that the proposed method achieves superior segmentation accuracy with minimal additional computation, showcasing high performance in underwater image segmentation.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于改进型 SegFormer 的水下图像语义分割
水下图像分割是水下探测、海洋环境监测和资源开发等任务的重要组成部分。然而,鉴于水下环境的复杂性和可变性,提高模型精度仍然是水下图像分割任务的关键挑战。为了解决这些问题,本研究提出了一种基于标准SegFormer模型的高性能水下图像语义分割方法。首先,将SegFormer中的Mix Transformer主干替换为Swin Transformer,以增强特征提取并促进全局上下文信息的有效获取。其次,在主干的下采样阶段和解码器中引入高效多尺度注意(EMA)机制,以更好地捕获多尺度特征,进一步提高分割精度。此外,在解码器中加入了特征金字塔网络(FPN)结构,以组合多分辨率的特征映射,使模型能够有效地整合上下文信息,增强复杂水下环境中的鲁棒性。在SUIM水下图像数据集上的测试表明,所提出的模型在多个指标上取得了良好的性能:平均交叉点超过联合(Intersection over Union, MIoU)为77.00%,平均召回率(Recall, mRecall)为85.04%,平均精度(Precision, mPrecision)为89.03%,平均F1score (mF1score)为86.63%。与标准的SegFormer相比,MIoU提高了3.73%,mRecall提高了1.98%,mPrecision提高了3.38%,mF1score提高了2.44%,增加了9.89M个参数。结果表明,该方法以最小的额外计算量获得了较高的分割精度,在水下图像分割中具有较高的性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Frontiers in Marine Science
Frontiers in Marine Science Agricultural and Biological Sciences-Aquatic Science
CiteScore
5.10
自引率
16.20%
发文量
2443
审稿时长
14 weeks
期刊介绍: Frontiers in Marine Science publishes rigorously peer-reviewed research that advances our understanding of all aspects of the environment, biology, ecosystem functioning and human interactions with the oceans. Field Chief Editor Carlos M. Duarte at King Abdullah University of Science and Technology Thuwal is supported by an outstanding Editorial Board of international researchers. This multidisciplinary open-access journal is at the forefront of disseminating and communicating scientific knowledge and impactful discoveries to researchers, academics, policy makers and the public worldwide. With the human population predicted to reach 9 billion people by 2050, it is clear that traditional land resources will not suffice to meet the demand for food or energy, required to support high-quality livelihoods. As a result, the oceans are emerging as a source of untapped assets, with new innovative industries, such as aquaculture, marine biotechnology, marine energy and deep-sea mining growing rapidly under a new era characterized by rapid growth of a blue, ocean-based economy. The sustainability of the blue economy is closely dependent on our knowledge about how to mitigate the impacts of the multiple pressures on the ocean ecosystem associated with the increased scale and diversification of industry operations in the ocean and global human pressures on the environment. Therefore, Frontiers in Marine Science particularly welcomes the communication of research outcomes addressing ocean-based solutions for the emerging challenges, including improved forecasting and observational capacities, understanding biodiversity and ecosystem problems, locally and globally, effective management strategies to maintain ocean health, and an improved capacity to sustainably derive resources from the oceans.
期刊最新文献
Multi-marker eDNA metabarcoding reveals significant eukaryotic biodiversity gaps in the Gulf of Gdańsk, Southeastern Baltic Sea Antibacterial activity and mechanism of sanguinarine against Vibrio parahaemolyticus Automated quantification of ecological interactions from video They can’t stand the heat: marine heatwaves and bleaching impair stress responses and reproduction of a photosynthetic symbiont-bearing sea slug Seasonal and interannual variability of the barrier layer in the Bay of Bengal: characteristics and mechanisms
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1