DeepFins: Capturing dynamics in underwater videos for fish detection

IF 7.3 2区 环境科学与生态学 Q1 ECOLOGY Ecological Informatics Pub Date : 2025-01-25 DOI:10.1016/j.ecoinf.2025.103013
Ahsan Jalal , Ahmad Salman , Ajmal Mian , Salman Ghafoor , Faisal Shafait
{"title":"DeepFins: Capturing dynamics in underwater videos for fish detection","authors":"Ahsan Jalal ,&nbsp;Ahmad Salman ,&nbsp;Ajmal Mian ,&nbsp;Salman Ghafoor ,&nbsp;Faisal Shafait","doi":"10.1016/j.ecoinf.2025.103013","DOIUrl":null,"url":null,"abstract":"<div><div>The monitoring of fish in their natural habitat plays a crucial role in anticipating changes within marine ecosystems. Marine scientists have a preference for automated, unrestricted underwater video-based sampling due to its non-invasive nature and its ability to yield desired outcomes more rapidly compared to manual sampling. Generally, research on automated video-based detection using computer vision and machine learning has been confined to controlled environments. Additionally, these solutions encounter difficulties when applied in real-world settings characterized by substantial environmental variability, including issues like poor visibility in unregulated underwater videos, challenges in capturing fish-related visual characteristics, and background interference. In response, we propose a hybrid solution that merges YOLOv11, a popular deep learning based static object detector, with a custom designed lightweight motion-based segmentation model. This approach allows us to simultaneously capture fish dynamics and suppress background interference. The proposed model i.e., DeepFins attains 90.0% F1 Score for fish detection on the OzFish dataset (collected by the Australian Institute of Marine Science). To the best of our knowledge, these results are the most accurate yet, showing about 11% increase over the closest competitor in fish detection tasks on this demanding benchmark OzFish dataset. Moreover, DeepFins achieves an F1 Score of 83.7% on the Fish4Knowledge LifeCLEF 2015 dataset, marking an approximate 4% improvement over the baseline YOLOv11. This positions the proposed model as a highly practical solution for tasks like automated fish sampling and estimating their relative abundance.</div></div>","PeriodicalId":51024,"journal":{"name":"Ecological Informatics","volume":"86 ","pages":"Article 103013"},"PeriodicalIF":7.3000,"publicationDate":"2025-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Ecological Informatics","FirstCategoryId":"93","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1574954125000226","RegionNum":2,"RegionCategory":"环境科学与生态学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ECOLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

The monitoring of fish in their natural habitat plays a crucial role in anticipating changes within marine ecosystems. Marine scientists have a preference for automated, unrestricted underwater video-based sampling due to its non-invasive nature and its ability to yield desired outcomes more rapidly compared to manual sampling. Generally, research on automated video-based detection using computer vision and machine learning has been confined to controlled environments. Additionally, these solutions encounter difficulties when applied in real-world settings characterized by substantial environmental variability, including issues like poor visibility in unregulated underwater videos, challenges in capturing fish-related visual characteristics, and background interference. In response, we propose a hybrid solution that merges YOLOv11, a popular deep learning based static object detector, with a custom designed lightweight motion-based segmentation model. This approach allows us to simultaneously capture fish dynamics and suppress background interference. The proposed model i.e., DeepFins attains 90.0% F1 Score for fish detection on the OzFish dataset (collected by the Australian Institute of Marine Science). To the best of our knowledge, these results are the most accurate yet, showing about 11% increase over the closest competitor in fish detection tasks on this demanding benchmark OzFish dataset. Moreover, DeepFins achieves an F1 Score of 83.7% on the Fish4Knowledge LifeCLEF 2015 dataset, marking an approximate 4% improvement over the baseline YOLOv11. This positions the proposed model as a highly practical solution for tasks like automated fish sampling and estimating their relative abundance.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
DeepFins:捕获动态水下视频的鱼类检测
对鱼类自然栖息地的监测在预测海洋生态系统的变化方面起着至关重要的作用。海洋科学家更喜欢自动的、不受限制的水下视频采样,因为它的非侵入性,而且与人工采样相比,它能够更快地产生预期的结果。通常,使用计算机视觉和机器学习的基于视频的自动检测研究仅限于受控环境。此外,这些解决方案在现实环境中应用时也会遇到困难,包括在不受管制的水下视频中能见度低、捕捉鱼类相关视觉特征的挑战以及背景干扰等问题。为此,我们提出了一种混合解决方案,将流行的基于深度学习的静态目标检测器YOLOv11与定制设计的轻量级基于运动的分割模型相结合。这种方法允许我们同时捕捉鱼类动态和抑制背景干扰。所提出的模型,即DeepFins在OzFish数据集(由澳大利亚海洋科学研究所收集)上的鱼类检测达到90.0%的F1分数。据我们所知,这些结果是迄今为止最准确的,在这个要求苛刻的基准OzFish数据集上,鱼类检测任务比最接近的竞争对手提高了约11%。此外,DeepFins在Fish4Knowledge LifeCLEF 2015数据集上的F1得分为83.7%,比基线YOLOv11提高了约4%。这使得所提出的模型成为自动化鱼类采样和估计其相对丰度等任务的高度实用的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Ecological Informatics
Ecological Informatics 环境科学-生态学
CiteScore
8.30
自引率
11.80%
发文量
346
审稿时长
46 days
期刊介绍: The journal Ecological Informatics is devoted to the publication of high quality, peer-reviewed articles on all aspects of computational ecology, data science and biogeography. The scope of the journal takes into account the data-intensive nature of ecology, the growing capacity of information technology to access, harness and leverage complex data as well as the critical need for informing sustainable management in view of global environmental and climate change. The nature of the journal is interdisciplinary at the crossover between ecology and informatics. It focuses on novel concepts and techniques for image- and genome-based monitoring and interpretation, sensor- and multimedia-based data acquisition, internet-based data archiving and sharing, data assimilation, modelling and prediction of ecological data.
期刊最新文献
Large language vision models for zero-shot handwriting recognition of historical herbarium labels PNW-Cnet: An evolving convolutional neural network to support broad-scale passive acoustic monitoring Copula–information gain-based identification of GPP response thresholds under multiscale agricultural drought Explainable AI-based modeling of chlorophyll-a using water quality and DOM characteristics: Toward interpretable prediction and actionable management strategies Developing a deep learning-based image analysis model for high-throughput micronucleus assays: Genotoxicity as a sediment quality indicator in East Taihu and Yangcheng Lakes, China
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1