SwinVision: Detecting Small Objects in Low-Light Environments

IF 3.6 3区 计算机科学 Q2 COMPUTER SCIENCE, INFORMATION SYSTEMS IEEE Access Pub Date : 2025-03-05 DOI:10.1109/ACCESS.2025.3548151
Tao Dai;Qi Wang;Yuancheng Shen;Shang Gao
{"title":"SwinVision: Detecting Small Objects in Low-Light Environments","authors":"Tao Dai;Qi Wang;Yuancheng Shen;Shang Gao","doi":"10.1109/ACCESS.2025.3548151","DOIUrl":null,"url":null,"abstract":"Neural networks have been widely employed in the field of object detection. Transformers enable effective object detection through global context awareness, modular design, scalability, and adaptability to diverse target scales. However, small object detection requires careful consideration due to its comprehensive computations, data requirements, and real-time performance challenges. To address these issues, we present SwinVision, an innovative framework for small object detection in low-light environments. This research shows a balanced approach between computational efficiency and detection accuracy for advancing object detection in low-light scenarios. Firstly, a Swin Transformer-based computing network is introduced and optimized for object detection in large-scale areas. The framework balances computational power and resource efficiency, surpassing conventional transformers. Secondly, we present the STLE module, which enhances the features of low-light images for beneficial object detection. The last building block is a specialized Swin-based detection block for accurate detection of small, detailed objects in resource-constrained scenarios. Experiments conducted on the VisDrone dataset significantly ameliorated existing methods such as YOLOv8x, with a 6.31% increase in mAP and 12.55% in AP50. SwinVision’s effectiveness in low-light environments, especially with small objects, establishes a foundation for robust detection systems adapting to various environmental challenges.","PeriodicalId":13079,"journal":{"name":"IEEE Access","volume":"13 ","pages":"42797-42812"},"PeriodicalIF":3.6000,"publicationDate":"2025-03-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=10910170","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Access","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10910170/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Neural networks have been widely employed in the field of object detection. Transformers enable effective object detection through global context awareness, modular design, scalability, and adaptability to diverse target scales. However, small object detection requires careful consideration due to its comprehensive computations, data requirements, and real-time performance challenges. To address these issues, we present SwinVision, an innovative framework for small object detection in low-light environments. This research shows a balanced approach between computational efficiency and detection accuracy for advancing object detection in low-light scenarios. Firstly, a Swin Transformer-based computing network is introduced and optimized for object detection in large-scale areas. The framework balances computational power and resource efficiency, surpassing conventional transformers. Secondly, we present the STLE module, which enhances the features of low-light images for beneficial object detection. The last building block is a specialized Swin-based detection block for accurate detection of small, detailed objects in resource-constrained scenarios. Experiments conducted on the VisDrone dataset significantly ameliorated existing methods such as YOLOv8x, with a 6.31% increase in mAP and 12.55% in AP50. SwinVision’s effectiveness in low-light environments, especially with small objects, establishes a foundation for robust detection systems adapting to various environmental challenges.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
SwinVision:在弱光环境中检测小物体
神经网络在目标检测领域得到了广泛的应用。变压器通过全局上下文感知、模块化设计、可扩展性和对不同目标尺度的适应性来实现有效的对象检测。然而,小目标检测由于其计算量大、数据要求高、实时性挑战大,需要慎重考虑。为了解决这些问题,我们提出了SwinVision,这是一种用于低光环境下小物体检测的创新框架。本研究提出了一种平衡计算效率和检测精度的方法来推进低光场景下的目标检测。首先,提出了一种基于Swin变压器的计算网络,并对其进行了优化。该框架平衡了计算能力和资源效率,超越了传统的变压器。其次,我们提出了STLE模块,增强了弱光图像的特征,有利于目标的检测。最后一个构建块是一个专门的基于swan的检测块,用于在资源受限的情况下精确检测小而详细的对象。在VisDrone数据集上进行的实验显著改进了现有的方法,如YOLOv8x, mAP提高了6.31%,AP50提高了12.55%。SwinVision在低光环境中的有效性,特别是小物体,为适应各种环境挑战的强大检测系统奠定了基础。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE Access
IEEE Access COMPUTER SCIENCE, INFORMATION SYSTEMSENGIN-ENGINEERING, ELECTRICAL & ELECTRONIC
CiteScore
9.80
自引率
7.70%
发文量
6673
审稿时长
6 weeks
期刊介绍: IEEE Access® is a multidisciplinary, open access (OA), applications-oriented, all-electronic archival journal that continuously presents the results of original research or development across all of IEEE''s fields of interest. IEEE Access will publish articles that are of high interest to readers, original, technically correct, and clearly presented. Supported by author publication charges (APC), its hallmarks are a rapid peer review and publication process with open access to all readers. Unlike IEEE''s traditional Transactions or Journals, reviews are "binary", in that reviewers will either Accept or Reject an article in the form it is submitted in order to achieve rapid turnaround. Especially encouraged are submissions on: Multidisciplinary topics, or applications-oriented articles and negative results that do not fit within the scope of IEEE''s traditional journals. Practical articles discussing new experiments or measurement techniques, interesting solutions to engineering. Development of new or improved fabrication or manufacturing techniques. Reviews or survey articles of new or evolving fields oriented to assist others in understanding the new area.
期刊最新文献
Low-Cost FPGA-Enhanced CNN Accelerator for Real-Time YOLO Object Detection and Classification A Web-Ready and 5G-Ready Volumetric Video Streaming Platform: A Platform Prototype and Empirical Study Multi-Expert Trajectory Prediction for Highway Weaving Sections Using Conflict Potential Energy and GAN A Hybrid Fractional Chebyshev–Legendre Spectral Collocation Method for Hamilton–Jacobi–Bellman Equations Application-Specific Instruction-Set Processors (ASIPs) for Deep Neural Networks: A Survey
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1