{"title":"Streaming Quanta Sensors for Online, High-Performance Imaging and Vision","authors":"Tianyi Zhang;Matthew Dutson;Vivek Boominathan;Mohit Gupta;Ashok Veeraraghavan","doi":"10.1109/TPAMI.2024.3501154","DOIUrl":null,"url":null,"abstract":"Recently quanta image sensors (QIS) – ultra-fast, zero-read-noise binary image sensors– have demonstrated remarkable imaging capabilities in many challenging scenarios. Despite their potential, the adoption of these sensors is severely hampered by (a) high data rates and (b) the need for new computational pipelines to handle the unconventional raw data. We introduce a simple, low-bandwidth computational pipeline to address these challenges. Our approach is based on a novel streaming representation with a small memory footprint, efficiently capturing intensity information at multiple temporal scales. Updating the representation requires only 24floating-point operations/pixel, which can be efficiently computed online at the native frame rate of the binary frames. We use a neural network operating on this representation to reconstruct videos in real-time (10-30 fps). We illustrate why such representation is well-suited for these emerging sensors, and how it offers low latency and high frame rate while retaining flexibility for downstream computer vision. Our approach results in significant data bandwidth reductions (<inline-formula><tex-math>$\\sim 100\\times$</tex-math></inline-formula>) and real-time image reconstruction and computer vision <inline-formula><tex-math>$-10^{4}\\text{-}10^{5} \\times$</tex-math></inline-formula> reduction in computation than existing state-of-the-art approach (Ma et al. 2020), while maintaining comparable quality. To the best of our knowledge, our approach is the first to achieve online, real-time image reconstruction on QIS.","PeriodicalId":94034,"journal":{"name":"IEEE transactions on pattern analysis and machine intelligence","volume":"47 3","pages":"1564-1577"},"PeriodicalIF":18.6000,"publicationDate":"2024-11-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on pattern analysis and machine intelligence","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10758928/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Recently quanta image sensors (QIS) – ultra-fast, zero-read-noise binary image sensors– have demonstrated remarkable imaging capabilities in many challenging scenarios. Despite their potential, the adoption of these sensors is severely hampered by (a) high data rates and (b) the need for new computational pipelines to handle the unconventional raw data. We introduce a simple, low-bandwidth computational pipeline to address these challenges. Our approach is based on a novel streaming representation with a small memory footprint, efficiently capturing intensity information at multiple temporal scales. Updating the representation requires only 24floating-point operations/pixel, which can be efficiently computed online at the native frame rate of the binary frames. We use a neural network operating on this representation to reconstruct videos in real-time (10-30 fps). We illustrate why such representation is well-suited for these emerging sensors, and how it offers low latency and high frame rate while retaining flexibility for downstream computer vision. Our approach results in significant data bandwidth reductions ($\sim 100\times$) and real-time image reconstruction and computer vision $-10^{4}\text{-}10^{5} \times$ reduction in computation than existing state-of-the-art approach (Ma et al. 2020), while maintaining comparable quality. To the best of our knowledge, our approach is the first to achieve online, real-time image reconstruction on QIS.
最近,量子图像传感器(QIS)——一种超快速、零读取噪声的二值图像传感器——在许多具有挑战性的场景中展示了卓越的成像能力。尽管这些传感器具有很大的潜力,但它们的应用受到以下因素的严重阻碍:(a)高数据速率;(b)需要新的计算管道来处理非常规的原始数据。我们引入了一个简单的、低带宽的计算管道来解决这些挑战。我们的方法是基于一种新颖的流表示,具有较小的内存占用,在多个时间尺度上有效地捕获强度信息。更新表示只需要24个浮点运算/像素,这可以在二进制帧的本机帧率下有效地在线计算。我们使用神经网络对该表示进行操作,以实时(10-30 fps)重建视频。我们说明了为什么这种表示非常适合这些新兴的传感器,以及它如何提供低延迟和高帧率,同时保持下游计算机视觉的灵活性。与现有的最先进的方法(Ma et al. 2020)相比,我们的方法显著减少了数据带宽(100倍),实时图像重建和计算机视觉的计算减少了10倍(10倍),同时保持了相当的质量。据我们所知,我们的方法是第一个在QIS上实现在线、实时图像重建的方法。