Bounding Box Propagation for Semi-automatic Video Annotation of Nighttime Driving Scenes

Dominik Schörkhuber, Florian Groh, M. Gelautz
{"title":"Bounding Box Propagation for Semi-automatic Video Annotation of Nighttime Driving Scenes","authors":"Dominik Schörkhuber, Florian Groh, M. Gelautz","doi":"10.1109/ISPA52656.2021.9552141","DOIUrl":null,"url":null,"abstract":"Ground-truth annotations are a fundamental requirement for the development of computer vision and deep learning algorithms targeting autonomous driving. Available public datasets have for the most part been recorded in urban settings, while scenes showing countryside roads and nighttime driving conditions are underrepresented in current datasets. In this paper, we present a semi-automated approach for bounding box annotation which was developed in the context of nighttime driving videos. In our three-step approach, we (a) generate trajectory proposals through a tracking-by-detection method, (b) extend and verify object trajectories through single object tracking, and (c) propose a pipeline for efficient semiautomatic annotation of object bounding boxes in videos. We evaluate our approach on the CVL dataset, which focuses on nighttime driving conditions on European countryside roads. We demonstrate the improvements achieved by each processing step, and observe an increase of 23% in recall while precision remains almost constant when compared to the initial tracking-by-detection approach.","PeriodicalId":131088,"journal":{"name":"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 12th International Symposium on Image and Signal Processing and Analysis (ISPA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPA52656.2021.9552141","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Ground-truth annotations are a fundamental requirement for the development of computer vision and deep learning algorithms targeting autonomous driving. Available public datasets have for the most part been recorded in urban settings, while scenes showing countryside roads and nighttime driving conditions are underrepresented in current datasets. In this paper, we present a semi-automated approach for bounding box annotation which was developed in the context of nighttime driving videos. In our three-step approach, we (a) generate trajectory proposals through a tracking-by-detection method, (b) extend and verify object trajectories through single object tracking, and (c) propose a pipeline for efficient semiautomatic annotation of object bounding boxes in videos. We evaluate our approach on the CVL dataset, which focuses on nighttime driving conditions on European countryside roads. We demonstrate the improvements achieved by each processing step, and observe an increase of 23% in recall while precision remains almost constant when compared to the initial tracking-by-detection approach.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于边界盒传播的夜间驾驶场景半自动视频标注
Ground-truth注释是开发针对自动驾驶的计算机视觉和深度学习算法的基本要求。现有的公共数据集大部分是在城市环境中记录的,而显示农村道路和夜间驾驶条件的场景在当前数据集中代表性不足。在本文中,我们提出了一种半自动化的边界框标注方法,该方法是在夜间驾驶视频的背景下开发的。在我们的三步方法中,我们(a)通过检测跟踪方法生成轨迹建议,(b)通过单目标跟踪扩展和验证目标轨迹,以及(c)提出了一个高效半自动标注视频中目标边界框的管道。我们在CVL数据集上评估了我们的方法,该数据集专注于欧洲乡村道路的夜间驾驶条件。我们展示了每个处理步骤所取得的改进,并观察到召回率增加了23%,而精确度与最初的检测跟踪方法相比几乎保持不变。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Bounding Box Propagation for Semi-automatic Video Annotation of Nighttime Driving Scenes Generating Patterns on the Triangular Grid by Cellular Automata including Alternating Use of Two Rules Novel Initial Parameters Computation for EM algorithm-based Univariate Asymmetric Generalized Gaussian Mixture Acoustic Features for Deep Learning-Based Models for Emergency Siren Detection: An Evaluation Study Speech Intelligibility Enhancement using an Optimal Formant Shifting Approach
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1