Frame of Events: A Low-latency Resource-efficient Approach for Stereo Depth Maps

Shanmuga Venkatachalam, V. Vivekanand, R. Kubendran
{"title":"Frame of Events: A Low-latency Resource-efficient Approach for Stereo Depth Maps","authors":"Shanmuga Venkatachalam, V. Vivekanand, R. Kubendran","doi":"10.1109/ICARA56516.2023.10125817","DOIUrl":null,"url":null,"abstract":"Computer vision traditionally uses cameras that capture visual information as frames at periodic intervals. On the other hand, Dynamic Vision Sensors (DVS) capture temporal contrast (TC) in each pixel asynchronously and stream them serially. This paper proposes a hybrid approach to generate input visual data as ‘frame of events’ for a stereo vision pipeline. We demonstrate that using hybrid vision sensors that produce frames made up of TC events can achieve superior results in terms of low latency, less compute and low memory footprint as compared to the traditional cameras and the event-based DVS. The frame-of-events approach eliminates the latency and memory resources involved in the accumulation of asynchronous events into synchronous frames, while generating acceptable disparity maps for depth estimation. Benchmarking results show that the frame-of-events pipeline outperforms others with the least average latency per frame of 3.8 ms and least average memory usage per frame of 112.4 Kb, which amounts to 7.32% and 9.75% reduction when compared to traditional frame-based pipeline. Hence, the proposed method is suitable for missioncritical robotics applications that involve path planning and localization mapping in a resource-constrained environment, such as drone navigation and autonomous vehicles.","PeriodicalId":443572,"journal":{"name":"2023 9th International Conference on Automation, Robotics and Applications (ICARA)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-02-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 9th International Conference on Automation, Robotics and Applications (ICARA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICARA56516.2023.10125817","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Computer vision traditionally uses cameras that capture visual information as frames at periodic intervals. On the other hand, Dynamic Vision Sensors (DVS) capture temporal contrast (TC) in each pixel asynchronously and stream them serially. This paper proposes a hybrid approach to generate input visual data as ‘frame of events’ for a stereo vision pipeline. We demonstrate that using hybrid vision sensors that produce frames made up of TC events can achieve superior results in terms of low latency, less compute and low memory footprint as compared to the traditional cameras and the event-based DVS. The frame-of-events approach eliminates the latency and memory resources involved in the accumulation of asynchronous events into synchronous frames, while generating acceptable disparity maps for depth estimation. Benchmarking results show that the frame-of-events pipeline outperforms others with the least average latency per frame of 3.8 ms and least average memory usage per frame of 112.4 Kb, which amounts to 7.32% and 9.75% reduction when compared to traditional frame-based pipeline. Hence, the proposed method is suitable for missioncritical robotics applications that involve path planning and localization mapping in a resource-constrained environment, such as drone navigation and autonomous vehicles.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
事件框架:立体深度图的低延迟资源效率方法
计算机视觉传统上使用相机以周期性间隔捕捉视觉信息。另一方面,动态视觉传感器(DVS)异步捕获每个像素的时间对比度(TC)并将其串行传输。本文提出了一种混合方法来生成输入视觉数据作为立体视觉管道的“事件框架”。我们证明,使用混合视觉传感器产生由TC事件组成的帧,与传统相机和基于事件的DVS相比,在低延迟、更少的计算和更低的内存占用方面可以取得更好的结果。事件帧方法消除了将异步事件累积到同步帧中所涉及的延迟和内存资源,同时为深度估计生成可接受的视差图。基准测试结果表明,事件帧管道的性能优于其他管道,每帧的平均延迟最小为3.8 ms,每帧的平均内存使用量最小为112.4 Kb,与传统的基于帧的管道相比,分别减少了7.32%和9.75%。因此,该方法适用于资源受限环境中涉及路径规划和定位映射的任务关键型机器人应用,如无人机导航和自动驾驶汽车。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Fused Swish-ReLU Efficient-Net Model for Deepfakes Detection SensorClouds: A Framework for Real-Time Processing of Multi-modal Sensor Data for Human-Robot-Collaboration Modified Bug Algorithm with Proximity Sensors to Reduce Human-Cobot Collisions Toward Computationally Efficient Path Generation and Push Planning for Robotic Nonprehensile Manipulation Correlation Analysis of Factors Influencing the Motion Planning Accuracy of Articulated Robots
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1