首页 > 最新文献

2020 IEEE International Conference on Computational Photography (ICCP)最新文献

英文 中文
Towards Reflectometry from Interreflections 从互反射走向反射测量
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105251
Kfir Shem-Tov, Sai Praveen Bangaru, Anat Levin, Ioannis Gkioulekas
Reflectometry is the task for acquiring the bidirectional reflectance distribution function (BRDFs) of real-world materials. The typical reflectometry pipeline in computer vision, computer graphics, and computational imaging involves capturing images of a convex shape under multiple illumination and imaging conditions; due to the convexity of the shape, which implies that all paths from the light source to the camera perform a single reflection, the intensities in these images can subsequently be analytically mapped to BRDF values. We deviate from this pipeline by investigating the utility of higher-order light transport effects, such as the interreflections arising when illuminating and imaging a concave object, for reflectometry. We show that interreflections provide a rich set of contraints on the unknown BRDF, significantly exceeding those available in equivalent measurements of convex shapes. We develop a differentiable rendering pipeline to solve an inverse rendering problem that uses these constraints to produce high-fidelity BRDF estimates from even a single input image. Finally, we take first steps towards designing new concave shapes that maximize the amount of information about the unknown BRDF available in image measurements. We perform extensive simulations to validate the utility of this reflectometry from interreflections approach.
反射测量是一项获取真实材料的双向反射分布函数(BRDFs)的任务。计算机视觉、计算机图形学和计算成像中的典型反射测量管道涉及在多种照明和成像条件下捕获凸形状的图像;由于形状的凹凸性,这意味着从光源到相机的所有路径都执行一次反射,因此这些图像中的强度随后可以解析映射为BRDF值。我们通过研究高阶光传输效应的效用来偏离这个管道,例如在反射测量中照亮和成像凹物体时产生的相互反射。我们表明,相互反射提供了一组丰富的未知BRDF约束,大大超过了凸形状等效测量中可用的约束。我们开发了一个可微分的渲染管道来解决一个反向渲染问题,该问题使用这些约束从单个输入图像产生高保真的BRDF估计。最后,我们迈出了设计新的凹形状的第一步,这些凹形状可以最大限度地利用图像测量中未知BRDF的信息量。我们进行了大量的模拟来验证这种反射法的实用性。
{"title":"Towards Reflectometry from Interreflections","authors":"Kfir Shem-Tov, Sai Praveen Bangaru, Anat Levin, Ioannis Gkioulekas","doi":"10.1109/ICCP48838.2020.9105251","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105251","url":null,"abstract":"Reflectometry is the task for acquiring the bidirectional reflectance distribution function (BRDFs) of real-world materials. The typical reflectometry pipeline in computer vision, computer graphics, and computational imaging involves capturing images of a convex shape under multiple illumination and imaging conditions; due to the convexity of the shape, which implies that all paths from the light source to the camera perform a single reflection, the intensities in these images can subsequently be analytically mapped to BRDF values. We deviate from this pipeline by investigating the utility of higher-order light transport effects, such as the interreflections arising when illuminating and imaging a concave object, for reflectometry. We show that interreflections provide a rich set of contraints on the unknown BRDF, significantly exceeding those available in equivalent measurements of convex shapes. We develop a differentiable rendering pipeline to solve an inverse rendering problem that uses these constraints to produce high-fidelity BRDF estimates from even a single input image. Finally, we take first steps towards designing new concave shapes that maximize the amount of information about the unknown BRDF available in image measurements. We perform extensive simulations to validate the utility of this reflectometry from interreflections approach.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"96 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123812491","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
The role of Wigner Distribution Function in Non-Line-of-Sight Imaging 维格纳分布函数在非视距成像中的作用
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105266
Xiaochun Liu, A. Velten
Non-Line-of-Sight imaging has been linked to wave diffraction by the recent phasor field method. In wave optics, the Wigner Distribution Function description for an optical imaging system is a powerful analytical tool for modeling the imaging process with geometrical transformations. In this paper, we focus on illustrating the relation between captured signals and hidden objects in the Wigner Distribution domain. The Wigner Distribution Function is usually used together with approximated diffraction propagators, which is fine for most imaging problems. However, these approximated diffraction propagators are not valid for Non-Line-of-Sight imaging scenarios. We show that the exact phasor field propagator (Rayleigh-Sommerfeld Diffraction) does not have a standard geometrical transformation, as compared to approximated diffraction propagators (Fresnel, Fraunhofer diffraction) that can be represented as shearing or rotation in the Wigner Distribution Function domain. Then, we explore differences between the exact and approximated solutions by characterizing errors made in different spatial positions and acquisition methods (confocal, non-confocal scanning). We derive a lateral resolution based on the exact phasor field propagator, which can be used as a reference for theoretical evaluations and comparisons. For targets that lie laterally outside a relay wall, the loss of resolution is geometrically illustrated in the context of the Wigner Distribution Function.
最近的相量场法将非视距成像与波衍射联系起来。在波动光学中,光学成像系统的维格纳分布函数描述是用几何变换对成像过程进行建模的有力分析工具。在本文中,我们着重说明捕获信号和隐藏目标在维格纳分布域中的关系。维格纳分布函数通常与近似衍射传播函数一起使用,这对于大多数成像问题都是很好的。然而,这些近似的衍射传播算子不适用于非视距成像场景。我们表明,与近似的衍射传播子(菲涅耳衍射、弗劳恩霍夫衍射)相比,精确的相量场传播子(瑞利-索默菲尔衍射)没有标准的几何变换,而近似的衍射传播子(菲涅耳衍射)可以在维格纳分布函数域中表示为剪切或旋转。然后,我们通过表征在不同空间位置和采集方法(共聚焦、非共聚焦扫描)下产生的误差来探讨精确解和近似解之间的差异。我们推导了一个基于精确相量场传播子的横向分辨率,可以作为理论评价和比较的参考。对于横向位于中继墙外的目标,分辨率的损失在维格纳分布函数的背景下以几何方式说明。
{"title":"The role of Wigner Distribution Function in Non-Line-of-Sight Imaging","authors":"Xiaochun Liu, A. Velten","doi":"10.1109/ICCP48838.2020.9105266","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105266","url":null,"abstract":"Non-Line-of-Sight imaging has been linked to wave diffraction by the recent phasor field method. In wave optics, the Wigner Distribution Function description for an optical imaging system is a powerful analytical tool for modeling the imaging process with geometrical transformations. In this paper, we focus on illustrating the relation between captured signals and hidden objects in the Wigner Distribution domain. The Wigner Distribution Function is usually used together with approximated diffraction propagators, which is fine for most imaging problems. However, these approximated diffraction propagators are not valid for Non-Line-of-Sight imaging scenarios. We show that the exact phasor field propagator (Rayleigh-Sommerfeld Diffraction) does not have a standard geometrical transformation, as compared to approximated diffraction propagators (Fresnel, Fraunhofer diffraction) that can be represented as shearing or rotation in the Wigner Distribution Function domain. Then, we explore differences between the exact and approximated solutions by characterizing errors made in different spatial positions and acquisition methods (confocal, non-confocal scanning). We derive a lateral resolution based on the exact phasor field propagator, which can be used as a reference for theoretical evaluations and comparisons. For targets that lie laterally outside a relay wall, the loss of resolution is geometrically illustrated in the context of the Wigner Distribution Function.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128059655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
NLDNet++: A Physics Based Single Image Dehazing Network nldnet++:一个基于物理的单图像去雾网络
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105249
Iris Tal, Yael Bekerman, Avi Mor, Lior Knafo, J. Alon, S. Avidan
Deep learning methods for image dehazing achieve impressive results. Yet, the task of collecting ground truth hazy/dehazed image pairs to train the network is cumbersome. We propose to use Non-Local Image Dehazing (NLD), an existing physics based technique, to provide the dehazed image required to training a network. Upon close inspection, we find that NLD suffers from several shortcomings and propose novel extensions to improve it. The new method, termed NLD++, consists of 1) denoising the input image as pre-processing step to avoid noise amplification, 2) introducing a constrained optimization that respects physical constraints. NLD++ produces superior results to NLD at the expense of increased computational cost. To offset that, we propose NLDNet++, a fully convolutional network that is trained on pairs of hazy images and images dehazed by NLD++. This eliminates the need of existing deep learning methods that require hazy/dehazed image pairs that are difficult to obtain. We evaluate the performance of NLDNet++ on standard data sets and find it to compare favorably with existing methods.
图像去雾的深度学习方法取得了令人印象深刻的效果。然而,收集地面真实模糊/去模糊图像对来训练网络的任务很繁琐。我们建议使用非局部图像去雾(NLD),一种现有的基于物理的技术,来提供训练网络所需的去雾图像。经过仔细的检查,我们发现NLD存在一些缺点,并提出了新的扩展来改进它。新方法被称为NLD++,包括:1)对输入图像去噪作为预处理步骤,以避免噪声放大;2)引入尊重物理约束的约束优化。NLD++以增加计算成本为代价,产生优于NLD的结果。为了弥补这一点,我们提出了nldnet++,这是一个完全卷积的网络,它是在模糊图像和nldnet++去雾的图像对上训练的。这消除了现有深度学习方法对难以获得的模糊/去模糊图像对的需求。我们评估了nldnet++在标准数据集上的性能,发现它与现有方法相比具有优势。
{"title":"NLDNet++: A Physics Based Single Image Dehazing Network","authors":"Iris Tal, Yael Bekerman, Avi Mor, Lior Knafo, J. Alon, S. Avidan","doi":"10.1109/ICCP48838.2020.9105249","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105249","url":null,"abstract":"Deep learning methods for image dehazing achieve impressive results. Yet, the task of collecting ground truth hazy/dehazed image pairs to train the network is cumbersome. We propose to use Non-Local Image Dehazing (NLD), an existing physics based technique, to provide the dehazed image required to training a network. Upon close inspection, we find that NLD suffers from several shortcomings and propose novel extensions to improve it. The new method, termed NLD++, consists of 1) denoising the input image as pre-processing step to avoid noise amplification, 2) introducing a constrained optimization that respects physical constraints. NLD++ produces superior results to NLD at the expense of increased computational cost. To offset that, we propose NLDNet++, a fully convolutional network that is trained on pairs of hazy images and images dehazed by NLD++. This eliminates the need of existing deep learning methods that require hazy/dehazed image pairs that are difficult to obtain. We evaluate the performance of NLDNet++ on standard data sets and find it to compare favorably with existing methods.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115531235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Towards Learning-based Inverse Subsurface Scattering 基于学习的逆次表面散射研究
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105209
Chengqian Che, Fujun Luan, Shuang Zhao, K. Bala, Ioannis Gkioulekas
Given images of translucent objects, of unknown shape and lighting, we aim to use learning to infer the optical parameters controlling subsurface scattering of light inside the objects. We introduce a new architecture, the inverse transport network (ITN), that aims to improve generalization of an encoder network to unseen scenes, by connecting it with a physically-accurate, differentiable Monte Carlo renderer capable of estimating image derivatives with respect to scattering material parameters. During training, this combination forces the encoder network to predict parameters that not only match groundtruth values, but also reproduce input images. During testing, the encoder network is used alone, without the renderer, to predict material parameters from a single input image. Drawing insights from the physics of radiative transfer, we additionally use material parameterizations that help reduce estimation errors due to ambiguities in the scattering parameter space. Finally, we augment the training loss with pixelwise weight maps that emphasize the parts of the image most informative about the underlying scattering parameters. We demonstrate that this combination allows neural networks to generalize to scenes with completely unseen geometries and illuminations better than traditional networks, with 38.06% reduced parameter error on average.
给定半透明物体的图像,形状和光照未知,我们的目标是利用学习来推断控制物体内部光的次表面散射的光学参数。我们引入了一种新的架构,即逆传输网络(ITN),旨在通过将编码器网络与能够根据散射材料参数估计图像导数的物理精确、可微的蒙特卡罗渲染器连接起来,提高编码器网络对未见场景的泛化。在训练过程中,这种组合迫使编码器网络预测的参数不仅要匹配真值,而且还要重现输入图像。在测试期间,单独使用编码器网络,而不使用渲染器,从单个输入图像预测材料参数。从辐射传输的物理学中获得见解,我们还使用材料参数化来帮助减少由于散射参数空间中的模糊性而导致的估计误差。最后,我们用像素权重图来增加训练损失,这些权重图强调图像中最能提供潜在散射参数信息的部分。我们证明,这种组合使神经网络能够比传统网络更好地泛化到具有完全看不见的几何形状和光照的场景,平均降低了38.06%的参数误差。
{"title":"Towards Learning-based Inverse Subsurface Scattering","authors":"Chengqian Che, Fujun Luan, Shuang Zhao, K. Bala, Ioannis Gkioulekas","doi":"10.1109/ICCP48838.2020.9105209","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105209","url":null,"abstract":"Given images of translucent objects, of unknown shape and lighting, we aim to use learning to infer the optical parameters controlling subsurface scattering of light inside the objects. We introduce a new architecture, the inverse transport network (ITN), that aims to improve generalization of an encoder network to unseen scenes, by connecting it with a physically-accurate, differentiable Monte Carlo renderer capable of estimating image derivatives with respect to scattering material parameters. During training, this combination forces the encoder network to predict parameters that not only match groundtruth values, but also reproduce input images. During testing, the encoder network is used alone, without the renderer, to predict material parameters from a single input image. Drawing insights from the physics of radiative transfer, we additionally use material parameterizations that help reduce estimation errors due to ambiguities in the scattering parameter space. Finally, we augment the training loss with pixelwise weight maps that emphasize the parts of the image most informative about the underlying scattering parameters. We demonstrate that this combination allows neural networks to generalize to scenes with completely unseen geometries and illuminations better than traditional networks, with 38.06% reduced parameter error on average.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124323912","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Programmable Spectrometry: Per-pixel Material Classification using Learned Spectral Filters 可编程光谱:使用学习光谱滤波器的逐像素材料分类
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105281
Vishwanath Saragadam, Aswin C. Sankaranarayanan
Many materials have distinct spectral profiles, which facilitates estimation of the material composition of a scene by processing its hyperspectral image (HSI). However, this process is inherently wasteful since high-dimensional HSIs are expensive to acquire and only a set of linear projections of the HSI contribute to the classification task. This paper proposes the concept of programmable spectrometry for per-pixel material classification, where instead of sensing the HSI of the scene and then processing it, we optically compute the spectrally-filtered images. This is achieved using a computational camera with a programmable spectral response. Our approach provides gains both in terms of acquisition speed - since only the relevant measurements are acquired - and in signal-to-noise ratio - since we invariably avoid narrowband filters that are light inefficient. Given ample training data, we use learning techniques to identify the bank of spectral profiles that facilitate material classification. We verify the method in simulations, as well as validate our findings using a lab prototype of the camera.
许多材料具有不同的光谱轮廓,这有助于通过处理场景的高光谱图像(HSI)来估计场景的材料组成。然而,这个过程本质上是浪费的,因为高维HSI的获取成本很高,而且只有一组HSI的线性投影有助于分类任务。本文提出了用于逐像素材料分类的可编程光谱的概念,其中不是感知场景的HSI然后对其进行处理,而是光学计算光谱滤波后的图像。这是使用具有可编程光谱响应的计算相机实现的。我们的方法在采集速度(因为只获取相关的测量值)和信噪比(因为我们总是避免光效率低下的窄带滤波器)方面都有好处。给定充足的训练数据,我们使用学习技术来识别光谱轮廓库,从而促进材料分类。我们在模拟中验证了该方法,并使用相机的实验室原型验证了我们的发现。
{"title":"Programmable Spectrometry: Per-pixel Material Classification using Learned Spectral Filters","authors":"Vishwanath Saragadam, Aswin C. Sankaranarayanan","doi":"10.1109/ICCP48838.2020.9105281","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105281","url":null,"abstract":"Many materials have distinct spectral profiles, which facilitates estimation of the material composition of a scene by processing its hyperspectral image (HSI). However, this process is inherently wasteful since high-dimensional HSIs are expensive to acquire and only a set of linear projections of the HSI contribute to the classification task. This paper proposes the concept of programmable spectrometry for per-pixel material classification, where instead of sensing the HSI of the scene and then processing it, we optically compute the spectrally-filtered images. This is achieved using a computational camera with a programmable spectral response. Our approach provides gains both in terms of acquisition speed - since only the relevant measurements are acquired - and in signal-to-noise ratio - since we invariably avoid narrowband filters that are light inefficient. Given ample training data, we use learning techniques to identify the bank of spectral profiles that facilitate material classification. We verify the method in simulations, as well as validate our findings using a lab prototype of the camera.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127314823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Modeling Defocus-Disparity in Dual-Pixel Sensors 双像素传感器离焦视差建模
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105278
Abhijith Punnappurath, Abdullah Abuolaim, M. Afifi, M. S. Brown
Most modern consumer cameras use dual-pixel (DP) sensors that provide two sub-aperture views of the scene in a single photo capture. The DP sensor was designed to assist the camera's autofocus routine, which examines local disparity in the two sub-aperture views to determine which parts of the image are out of focus. Recently, these DP views have been used for tasks beyond autofocus, such as synthetic bokeh, reflection removal, and depth reconstruction. These recent methods treat the two DP views as stereo image pairs and apply stereo matching algorithms to compute local disparity. However, dual-pixel disparity is not caused by view parallax as in stereo, but instead is attributed to defocus blur that occurs in out-of-focus regions in the image. This paper proposes a new parametric point spread function to model the defocus-disparity that occurs on DP sensors. We apply our model to the task of depth estimation from DP data. An important feature of our model is its ability to exploit the symmetry property of the DP blur kernels at each pixel. We leverage this symmetry property to formulate an unsupervised loss function that does not require ground truth depth. We demonstrate our method's effectiveness on both DSLR and smartphone DP data.
大多数现代消费者相机使用双像素(DP)传感器,在一张照片中提供两个子光圈的场景视图。DP传感器的设计是为了协助相机的自动对焦程序,该程序检查两个子光圈视图中的局部差异,以确定图像的哪些部分失焦。最近,这些DP视图已被用于自动对焦以外的任务,如合成散景、反射去除和深度重建。这些方法将两个DP视图视为立体图像对,并应用立体匹配算法计算局部视差。然而,双像素的视差不是由立体视差引起的,而是由于在图像的失焦区域发生的散焦模糊。本文提出了一种新的参数点扩展函数来模拟DP传感器上的离焦视差。我们将该模型应用于DP数据的深度估计任务。我们的模型的一个重要特征是它能够在每个像素上利用DP模糊核的对称性。我们利用这种对称性来形成一个不需要真值深度的无监督损失函数。我们证明了我们的方法在单反和智能手机DP数据上的有效性。
{"title":"Modeling Defocus-Disparity in Dual-Pixel Sensors","authors":"Abhijith Punnappurath, Abdullah Abuolaim, M. Afifi, M. S. Brown","doi":"10.1109/ICCP48838.2020.9105278","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105278","url":null,"abstract":"Most modern consumer cameras use dual-pixel (DP) sensors that provide two sub-aperture views of the scene in a single photo capture. The DP sensor was designed to assist the camera's autofocus routine, which examines local disparity in the two sub-aperture views to determine which parts of the image are out of focus. Recently, these DP views have been used for tasks beyond autofocus, such as synthetic bokeh, reflection removal, and depth reconstruction. These recent methods treat the two DP views as stereo image pairs and apply stereo matching algorithms to compute local disparity. However, dual-pixel disparity is not caused by view parallax as in stereo, but instead is attributed to defocus blur that occurs in out-of-focus regions in the image. This paper proposes a new parametric point spread function to model the defocus-disparity that occurs on DP sensors. We apply our model to the task of depth estimation from DP data. An important feature of our model is its ability to exploit the symmetry property of the DP blur kernels at each pixel. We leverage this symmetry property to formulate an unsupervised loss function that does not require ground truth depth. We demonstrate our method's effectiveness on both DSLR and smartphone DP data.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126178796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Raycast Calibration for Augmented Reality HMDs with Off-Axis Reflective Combiners 带离轴反射组合器的增强现实头戴式显示器的光线投射校准
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105134
Qi Guo, Huixuan Tang, Aaron Schmitz, Wenqi Zhang, Yang Lou, Alexander Fix, S. Lovegrove, H. Strasdat
Augmented reality overlays virtual objects on the real world. To do so, the head mounted display (HMD) needs to be calibrated to establish a mapping between 3D points in the real world with 2D pixels on display panels. This distortion is a high-dimensional function that also depends on pupil position and varifocal settings. We present Raycast calibration, an efficient approach to geometrically calibrate AR displays with off-axis reflective combiners. Our approach requires a small amount of data to estimate a compact, physics-based, and ray-traceable model of the HMD optics. We apply this technique to automatically calibrate an AR prototype with display, SLAM and eye-tracker, without user in the loop.
增强现实技术将虚拟物体叠加在现实世界上。为此,头戴式显示器(HMD)需要进行校准,以便在现实世界中的3D点与显示面板上的2D像素之间建立映射。这种畸变是一种高维函数,也取决于瞳孔位置和变焦设置。我们提出了Raycast校准,这是一种有效的方法,可以用离轴反射组合器对AR显示器进行几何校准。我们的方法需要少量的数据来估计一个紧凑的、基于物理的、光线可追踪的HMD光学模型。我们将该技术应用于具有显示器、SLAM和眼动仪的AR原型的自动校准,而无需用户参与。
{"title":"Raycast Calibration for Augmented Reality HMDs with Off-Axis Reflective Combiners","authors":"Qi Guo, Huixuan Tang, Aaron Schmitz, Wenqi Zhang, Yang Lou, Alexander Fix, S. Lovegrove, H. Strasdat","doi":"10.1109/ICCP48838.2020.9105134","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105134","url":null,"abstract":"Augmented reality overlays virtual objects on the real world. To do so, the head mounted display (HMD) needs to be calibrated to establish a mapping between 3D points in the real world with 2D pixels on display panels. This distortion is a high-dimensional function that also depends on pupil position and varifocal settings. We present Raycast calibration, an efficient approach to geometrically calibrate AR displays with off-axis reflective combiners. Our approach requires a small amount of data to estimate a compact, physics-based, and ray-traceable model of the HMD optics. We apply this technique to automatically calibrate an AR prototype with display, SLAM and eye-tracker, without user in the loop.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130373719","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Awards [3 award winners] 奖项[3名得奖者]
Pub Date : 2020-04-01 DOI: 10.1109/iccp48838.2020.9105216
{"title":"Awards [3 award winners]","authors":"","doi":"10.1109/iccp48838.2020.9105216","DOIUrl":"https://doi.org/10.1109/iccp48838.2020.9105216","url":null,"abstract":"","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"10 6","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114134027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
High Resolution Diffuse Optical Tomography using Short Range Indirect Subsurface Imaging 使用近距离间接地下成像的高分辨率漫射光学层析成像
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105173
Chao Liu, Akash K. Maity, A. Dubrawski, A. Sabharwal, S. Narasimhan
Diffuse optical tomography (DOT) is an approach to recover subsurface structures beneath the skin by measuring light propagation beneath the surface. The method is based on optimizing the difference between the images collected and a forward model that accurately represents diffuse photon propagation within a heterogeneous scattering medium. However, to date, most works have used a few source-detector pairs and recover the medium at only a very low resolution. And increasing the resolution requires prohibitive computations/storage. In this work, we present a fast imaging and algorithm for high resolution diffuse optical tomography with a line imaging and illumination system. Key to our approach is a convolution approximation of the forward heterogeneous scattering model that can be inverted to produce deeper than ever before structured beneath the surface. We show that our proposed method can detect reasonably accurate boundaries and relative depth of heterogeneous structures up to a depth of 8 mm below highly scattering medium such as milk. This work can extend the potential of DOT to recover more intricate structures (vessels, tissue, tumors, etc.) beneath the skin for diagnosing many dermatological and cardio-vascular conditions.
漫射光学层析成像(DOT)是一种通过测量表面下的光传播来恢复皮肤下的地下结构的方法。该方法基于优化所收集图像与准确表示非均匀散射介质中漫射光子传播的正演模型之间的差异。然而,到目前为止,大多数工作都使用了少数源探测器对,并且只能以非常低的分辨率恢复介质。提高分辨率需要令人望而却步的计算/存储。在这项工作中,我们提出了一种具有线成像和照明系统的高分辨率漫射光学层析成像的快速成像和算法。我们方法的关键是前向非均匀散射模型的卷积近似,该模型可以被反转以产生比以往更深的表面下结构。结果表明,该方法可以在高散射介质(如牛奶)下8mm的深度内,较为准确地检测到非均质结构的边界和相对深度。这项工作可以扩展DOT的潜力,以恢复皮肤下更复杂的结构(血管,组织,肿瘤等),用于诊断许多皮肤病和心血管疾病。
{"title":"High Resolution Diffuse Optical Tomography using Short Range Indirect Subsurface Imaging","authors":"Chao Liu, Akash K. Maity, A. Dubrawski, A. Sabharwal, S. Narasimhan","doi":"10.1109/ICCP48838.2020.9105173","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105173","url":null,"abstract":"Diffuse optical tomography (DOT) is an approach to recover subsurface structures beneath the skin by measuring light propagation beneath the surface. The method is based on optimizing the difference between the images collected and a forward model that accurately represents diffuse photon propagation within a heterogeneous scattering medium. However, to date, most works have used a few source-detector pairs and recover the medium at only a very low resolution. And increasing the resolution requires prohibitive computations/storage. In this work, we present a fast imaging and algorithm for high resolution diffuse optical tomography with a line imaging and illumination system. Key to our approach is a convolution approximation of the forward heterogeneous scattering model that can be inverted to produce deeper than ever before structured beneath the surface. We show that our proposed method can detect reasonably accurate boundaries and relative depth of heterogeneous structures up to a depth of 8 mm below highly scattering medium such as milk. This work can extend the potential of DOT to recover more intricate structures (vessels, tissue, tumors, etc.) beneath the skin for diagnosing many dermatological and cardio-vascular conditions.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"78 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126058614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
End-to-End Video Compressive Sensing Using Anderson-Accelerated Unrolled Networks 端到端视频压缩感知使用安德森加速展开网络
Pub Date : 2020-04-01 DOI: 10.1109/ICCP48838.2020.9105237
Yuqi Li, Miao Qi, Rahul Gulve, Mian Wei, R. Genov, Kiriakos N. Kutulakos, W. Heidrich
Compressive imaging systems with spatial-temporal encoding can be used to capture and reconstruct fast-moving objects. The imaging quality highly depends on the choice of encoding masks and reconstruction methods. In this paper, we present a new network architecture to jointly design the encoding masks and the reconstruction method for compressive high-frame-rate imaging. Unlike previous works, the proposed method takes full advantage of denoising prior to provide a promising frame reconstruction. The network is also flexible enough to optimize full-resolution masks and efficient at reconstructing frames. To this end, we develop a new dense network architecture that embeds Anderson acceleration, known from numerical optimization, directly into the neural network architecture. Our experiments show the optimized masks and the dense accelerated network respectively achieve 1.5 dB and 1 dB improvements in PSNR without adding training parameters. The proposed method outperforms other state-of-the-art methods both in simulations and on real hardware. In addition, we set up a coded two-bucket camera for compressive high-frame-rate imaging, which is robust to imaging noise and provides promising results when recovering nearly 1,000 frames per second.
压缩成像系统具有时空编码,可用于捕获和重建快速运动的物体。图像质量在很大程度上取决于编码掩模和重构方法的选择。本文提出了一种新的网络结构,用于压缩高帧率成像的编码掩码和重构方法的联合设计。与以往的工作不同,该方法充分利用了去噪的优势,提供了一个有希望的帧重建。该网络也足够灵活,可以优化全分辨率掩码,并有效地重建帧。为此,我们开发了一种新的密集网络架构,将安德森加速(从数值优化中知道)直接嵌入到神经网络架构中。实验表明,在不增加训练参数的情况下,优化后的掩模和密集加速网络的PSNR分别提高了1.5 dB和1 dB。该方法在仿真和实际硬件上都优于其他先进的方法。此外,我们建立了一个编码的双桶相机用于压缩高帧率成像,该相机对成像噪声具有鲁棒性,并且在恢复近1000帧/秒时提供了令人满意的结果。
{"title":"End-to-End Video Compressive Sensing Using Anderson-Accelerated Unrolled Networks","authors":"Yuqi Li, Miao Qi, Rahul Gulve, Mian Wei, R. Genov, Kiriakos N. Kutulakos, W. Heidrich","doi":"10.1109/ICCP48838.2020.9105237","DOIUrl":"https://doi.org/10.1109/ICCP48838.2020.9105237","url":null,"abstract":"Compressive imaging systems with spatial-temporal encoding can be used to capture and reconstruct fast-moving objects. The imaging quality highly depends on the choice of encoding masks and reconstruction methods. In this paper, we present a new network architecture to jointly design the encoding masks and the reconstruction method for compressive high-frame-rate imaging. Unlike previous works, the proposed method takes full advantage of denoising prior to provide a promising frame reconstruction. The network is also flexible enough to optimize full-resolution masks and efficient at reconstructing frames. To this end, we develop a new dense network architecture that embeds Anderson acceleration, known from numerical optimization, directly into the neural network architecture. Our experiments show the optimized masks and the dense accelerated network respectively achieve 1.5 dB and 1 dB improvements in PSNR without adding training parameters. The proposed method outperforms other state-of-the-art methods both in simulations and on real hardware. In addition, we set up a coded two-bucket camera for compressive high-frame-rate imaging, which is robust to imaging noise and provides promising results when recovering nearly 1,000 frames per second.","PeriodicalId":406823,"journal":{"name":"2020 IEEE International Conference on Computational Photography (ICCP)","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123392188","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
期刊
2020 IEEE International Conference on Computational Photography (ICCP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1