Streaming Quanta Sensors for Online, High-Performance Imaging and Vision

Tianyi Zhang;Matthew Dutson;Vivek Boominathan;Mohit Gupta;Ashok Veeraraghavan
{"title":"Streaming Quanta Sensors for Online, High-Performance Imaging and Vision","authors":"Tianyi Zhang;Matthew Dutson;Vivek Boominathan;Mohit Gupta;Ashok Veeraraghavan","doi":"10.1109/TPAMI.2024.3501154","DOIUrl":null,"url":null,"abstract":"Recently quanta image sensors (QIS) – ultra-fast, zero-read-noise binary image sensors– have demonstrated remarkable imaging capabilities in many challenging scenarios. Despite their potential, the adoption of these sensors is severely hampered by (a) high data rates and (b) the need for new computational pipelines to handle the unconventional raw data. We introduce a simple, low-bandwidth computational pipeline to address these challenges. Our approach is based on a novel streaming representation with a small memory footprint, efficiently capturing intensity information at multiple temporal scales. Updating the representation requires only 24floating-point operations/pixel, which can be efficiently computed online at the native frame rate of the binary frames. We use a neural network operating on this representation to reconstruct videos in real-time (10-30 fps). We illustrate why such representation is well-suited for these emerging sensors, and how it offers low latency and high frame rate while retaining flexibility for downstream computer vision. Our approach results in significant data bandwidth reductions (<inline-formula><tex-math>$\\sim 100\\times$</tex-math></inline-formula>) and real-time image reconstruction and computer vision <inline-formula><tex-math>$-10^{4}\\text{-}10^{5} \\times$</tex-math></inline-formula> reduction in computation than existing state-of-the-art approach (Ma et al. 2020), while maintaining comparable quality. To the best of our knowledge, our approach is the first to achieve online, real-time image reconstruction on QIS.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 3","pages":"1564-1577"},"PeriodicalIF":18.6000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on pattern analysis and machine intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10758928/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Recently quanta image sensors (QIS) – ultra-fast, zero-read-noise binary image sensors– have demonstrated remarkable imaging capabilities in many challenging scenarios. Despite their potential, the adoption of these sensors is severely hampered by (a) high data rates and (b) the need for new computational pipelines to handle the unconventional raw data. We introduce a simple, low-bandwidth computational pipeline to address these challenges. Our approach is based on a novel streaming representation with a small memory footprint, efficiently capturing intensity information at multiple temporal scales. Updating the representation requires only 24floating-point operations/pixel, which can be efficiently computed online at the native frame rate of the binary frames. We use a neural network operating on this representation to reconstruct videos in real-time (10-30 fps). We illustrate why such representation is well-suited for these emerging sensors, and how it offers low latency and high frame rate while retaining flexibility for downstream computer vision. Our approach results in significant data bandwidth reductions ($\sim 100\times$) and real-time image reconstruction and computer vision $-10^{4}\text{-}10^{5} \times$ reduction in computation than existing state-of-the-art approach (Ma et al. 2020), while maintaining comparable quality. To the best of our knowledge, our approach is the first to achieve online, real-time image reconstruction on QIS.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
用于在线高性能成像和视觉的流量子传感器
最近,量子图像传感器(QIS)——一种超快速、零读取噪声的二值图像传感器——在许多具有挑战性的场景中展示了卓越的成像能力。尽管这些传感器具有很大的潜力,但它们的应用受到以下因素的严重阻碍:(a)高数据速率;(b)需要新的计算管道来处理非常规的原始数据。我们引入了一个简单的、低带宽的计算管道来解决这些挑战。我们的方法是基于一种新颖的流表示,具有较小的内存占用,在多个时间尺度上有效地捕获强度信息。更新表示只需要24个浮点运算/像素,这可以在二进制帧的本机帧率下有效地在线计算。我们使用神经网络对该表示进行操作,以实时(10-30 fps)重建视频。我们说明了为什么这种表示非常适合这些新兴的传感器,以及它如何提供低延迟和高帧率,同时保持下游计算机视觉的灵活性。与现有的最先进的方法(Ma et al. 2020)相比,我们的方法显著减少了数据带宽(100倍),实时图像重建和计算机视觉的计算减少了10倍(10倍),同时保持了相当的质量。据我们所知,我们的方法是第一个在QIS上实现在线、实时图像重建的方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
CrossEarth: Geospatial Vision Foundation Model for Domain Generalizable Remote Sensing Semantic Segmentation. Continuous Review and Timely Correction: Enhancing the Resistance to Noisy Labels via Self-Not-True and Class-Wise Distillation. On the Transferability and Discriminability of Representation Learning in Unsupervised Domain Adaptation. Fast Multi-view Discrete Clustering via Spectral Embedding Fusion. GrowSP++: Growing Superpoints and Primitives for Unsupervised 3D Semantic Segmentation.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1