Driving Behavior Aware Caption Generation for Egocentric Driving Videos Using In-Vehicle Sensors*

Hongkuan Zhang, Koichi Takeda, Ryohei Sasano, Yusuke Adachi, Kento Ohtani
{"title":"Driving Behavior Aware Caption Generation for Egocentric Driving Videos Using In-Vehicle Sensors*","authors":"Hongkuan Zhang, Koichi Takeda, Ryohei Sasano, Yusuke Adachi, Kento Ohtani","doi":"10.1109/ivworkshops54471.2021.9669259","DOIUrl":null,"url":null,"abstract":"Video captioning aims to generate textual descriptions according to the video contents. The risk assessment of autonomous driving vehicles has become essential for an insurance company for providing adequate insurance coverage, in particular, for emerging MaaS business. The insurers need to assess the risk of autonomous driving business plans with a fixed route by analyzing a large number of driving data, including videos recorded by dash cameras and sensor signals. To make the process more efficient, generating captions for driving videos can provide insurers concise information to understand the video contents quickly. A natural problem with driving video captioning is, since the absence of egovehicles in these egocentric videos, descriptions of latent driving behaviors are difficult to be grounded in specific visual cues. To address this issue, we focus on generating driving video captions with accurate behavior descriptions, and propose to incorporate in-vehicle sensors which encapsulate the driving behavior information to assist the caption generation. We evaluate our method on the Japanese driving video captioning dataset called City Traffic, where the results demonstrate the effectiveness of in-vehicle sensors on improving the overall performance of generated captions, especially on generating more accurate descriptions for the driving behaviors.","PeriodicalId":256905,"journal":{"name":"2021 IEEE Intelligent Vehicles Symposium Workshops (IV Workshops)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE Intelligent Vehicles Symposium Workshops (IV Workshops)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ivworkshops54471.2021.9669259","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1

Abstract

Video captioning aims to generate textual descriptions according to the video contents. The risk assessment of autonomous driving vehicles has become essential for an insurance company for providing adequate insurance coverage, in particular, for emerging MaaS business. The insurers need to assess the risk of autonomous driving business plans with a fixed route by analyzing a large number of driving data, including videos recorded by dash cameras and sensor signals. To make the process more efficient, generating captions for driving videos can provide insurers concise information to understand the video contents quickly. A natural problem with driving video captioning is, since the absence of egovehicles in these egocentric videos, descriptions of latent driving behaviors are difficult to be grounded in specific visual cues. To address this issue, we focus on generating driving video captions with accurate behavior descriptions, and propose to incorporate in-vehicle sensors which encapsulate the driving behavior information to assist the caption generation. We evaluate our method on the Japanese driving video captioning dataset called City Traffic, where the results demonstrate the effectiveness of in-vehicle sensors on improving the overall performance of generated captions, especially on generating more accurate descriptions for the driving behaviors.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于车载传感器的自我中心驾驶视频的驾驶行为感知字幕生成*
视频字幕的目的是根据视频内容生成文本描述。自动驾驶汽车的风险评估对于保险公司提供足够的保险范围至关重要,特别是对于新兴的MaaS业务。保险公司需要分析行车记录仪记录的视频和传感器信号等大量驾驶数据,评估固定路线的自动驾驶事业计划的风险。为了提高流程效率,为驾驶视频生成字幕可以为保险公司提供简明的信息,以便快速理解视频内容。驾驶视频字幕的一个自然问题是,由于在这些以自我为中心的视频中没有自我车辆,对潜在驾驶行为的描述很难建立在特定的视觉线索上。为了解决这一问题,我们将重点放在生成具有准确行为描述的驾驶视频字幕上,并提出在车内集成封装驾驶行为信息的传感器来辅助字幕生成。我们在名为City Traffic的日本驾驶视频字幕数据集上评估了我们的方法,结果证明了车载传感器在提高生成字幕的整体性能方面的有效性,特别是在生成更准确的驾驶行为描述方面。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Trajectory Planning with Comfort and Safety in Dynamic Traffic Scenarios for Autonomous Driving Unsupervised Joint Multi-Task Learning of Vision Geometry Tasks An adaptive cooperative adaptive cruise control against varying vehicle loads* Fundamental Design Criteria for Logical Scenarios in Simulation-based Safety Validation of Automated Driving Using Sensor Model Knowledge Parameter-Based Testing and Debugging of Autonomous Driving Systems
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1