首页 > 最新文献

2011 Conference for Visual Media Production最新文献

英文 中文
Motion Estimation for Regions of Reflections through Layer Separation 基于层分离的反射区域运动估计
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.12
Mohamed A. Elgharib, François Pitié, A. Kokaram
Regions of reflections contain two semi-transparent layers moving over each other. This generates two motion vectors per pel. Current multiple motion estimators either extend the usual brightness consistency assumption to two motions or are based on the Fourier phase shift relationship. Both approaches assume constant motion over at least three frames. As a result they can not handle temporally active motion due to camera shake or acceleration. This paper proposes a new approach for multiple motion estimation by modeling the correct motions as the ones generating the best layer separation of the examined reflection. A Bayesian framework is proposed which then admits a solution using candidate motions generated from KLT trajectories and a layer separation technique. We use novel temporal priors and our results show handling of strong motion inconsistencies and improvements over previous work.
反射区域包含两个相互移动的半透明层。这就产生了两个运动向量。当前的多运动估计要么将通常的亮度一致性假设扩展到两个运动,要么基于傅里叶相移关系。这两种方法都假定在至少三帧内保持恒定运动。因此,它们不能处理由于相机抖动或加速而引起的临时活动运动。本文提出了一种新的多运动估计方法,将正确的运动建模为被检测反射产生最佳层距的运动。提出了一个贝叶斯框架,该框架允许使用由KLT轨迹产生的候选运动和层分离技术进行解决。我们使用新的时间先验,我们的结果显示处理强烈的运动不一致和改进比以前的工作。
{"title":"Motion Estimation for Regions of Reflections through Layer Separation","authors":"Mohamed A. Elgharib, François Pitié, A. Kokaram","doi":"10.1109/CVMP.2011.12","DOIUrl":"https://doi.org/10.1109/CVMP.2011.12","url":null,"abstract":"Regions of reflections contain two semi-transparent layers moving over each other. This generates two motion vectors per pel. Current multiple motion estimators either extend the usual brightness consistency assumption to two motions or are based on the Fourier phase shift relationship. Both approaches assume constant motion over at least three frames. As a result they can not handle temporally active motion due to camera shake or acceleration. This paper proposes a new approach for multiple motion estimation by modeling the correct motions as the ones generating the best layer separation of the examined reflection. A Bayesian framework is proposed which then admits a solution using candidate motions generated from KLT trajectories and a layer separation technique. We use novel temporal priors and our results show handling of strong motion inconsistencies and improvements over previous work.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126788271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Practical Image-Based Relighting and Editing with Spherical-Harmonics and Local Lights 实用的基于图像的重照明和编辑与球谐波和局部光
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.22
Borom Tunwattanapong, A. Ghosh, P. Debevec
We present a practical technique for image-based relighting under environmental illumination which greatly reduces the number of required photographs compared to traditional techniques, while still achieving high quality editable relighting results. The proposed method employs an optimization procedure to combine spherical harmonics, a global lighting basis, with a set of local lights. Our choice of lighting basis captures both low and high frequency components of typical surface reflectance functions while generating close approximations to the ground truth with an order of magnitude less data. This technique benefits the acquisition process by reducing the number of required photographs, while simplifying the modification of reflectance data and enabling artistic lighting edits for post-production effects. Here, we demonstrate two desirable lighting edits, modifying light intensity and angular width, employing the proposed lighting basis.
我们提出了一种实用的环境照明下基于图像的重照明技术,与传统技术相比,该技术大大减少了所需的照片数量,同时仍然可以获得高质量的可编辑重照明结果。该方法采用一种优化程序,将球面谐波(全局照明基础)与一组局部照明相结合。我们选择的照明基础捕获了典型表面反射率函数的低频和高频成分,同时用更少的数量级数据生成接近地面真相的近似。这项技术通过减少所需照片的数量,同时简化反射率数据的修改,并为后期制作效果提供艺术照明编辑,从而使获取过程受益。在这里,我们展示了两种理想的照明编辑,修改光强度和角宽度,采用提出的照明基础。
{"title":"Practical Image-Based Relighting and Editing with Spherical-Harmonics and Local Lights","authors":"Borom Tunwattanapong, A. Ghosh, P. Debevec","doi":"10.1109/CVMP.2011.22","DOIUrl":"https://doi.org/10.1109/CVMP.2011.22","url":null,"abstract":"We present a practical technique for image-based relighting under environmental illumination which greatly reduces the number of required photographs compared to traditional techniques, while still achieving high quality editable relighting results. The proposed method employs an optimization procedure to combine spherical harmonics, a global lighting basis, with a set of local lights. Our choice of lighting basis captures both low and high frequency components of typical surface reflectance functions while generating close approximations to the ground truth with an order of magnitude less data. This technique benefits the acquisition process by reducing the number of required photographs, while simplifying the modification of reflectance data and enabling artistic lighting edits for post-production effects. Here, we demonstrate two desirable lighting edits, modifying light intensity and angular width, employing the proposed lighting basis.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125163756","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Depth Estimation from Three Cameras Using Belief Propagation: 3D Modelling of Sumo Wrestling 基于信念传播的三台摄像机深度估计:相扑三维建模
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.20
Kensuke Ikeya, K. Hisatomi, Miwa Katayama, Y. Iwadate
We propose a method to estimate depth from three wide-baseline camera images using belief propagation. With this method, message propagation is restricted to reduce the effects of boundary overreach, and max and min values and kurtosis of message energy distribution are used to reduce errors caused by large occlusion and texture less areas. In experiments, we focused on scenes of the traditional Japanese sport of sumo and created 3D models from three HD images using our method. We displayed them on a 3D display using the principle of integral photography (IP). We confirmed from the experimental results that our method was effective for estimating depth.
提出了一种基于信念传播的三幅宽基线相机图像深度估计方法。该方法对消息传播进行限制,以减少边界过长带来的影响,并利用消息能量分布的最大值、最小值和峰度来减少大遮挡和少纹理区域带来的误差。在实验中,我们以日本传统相扑运动的场景为研究对象,用我们的方法从三幅高清图像中创建了3D模型。我们利用整体摄影(IP)原理将它们展示在3D显示器上。实验结果表明,该方法对深度估计是有效的。
{"title":"Depth Estimation from Three Cameras Using Belief Propagation: 3D Modelling of Sumo Wrestling","authors":"Kensuke Ikeya, K. Hisatomi, Miwa Katayama, Y. Iwadate","doi":"10.1109/CVMP.2011.20","DOIUrl":"https://doi.org/10.1109/CVMP.2011.20","url":null,"abstract":"We propose a method to estimate depth from three wide-baseline camera images using belief propagation. With this method, message propagation is restricted to reduce the effects of boundary overreach, and max and min values and kurtosis of message energy distribution are used to reduce errors caused by large occlusion and texture less areas. In experiments, we focused on scenes of the traditional Japanese sport of sumo and created 3D models from three HD images using our method. We displayed them on a 3D display using the principle of integral photography (IP). We confirmed from the experimental results that our method was effective for estimating depth.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121359032","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Space-time Editing of 3D Video Sequences 三维视频序列的时空编辑
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.23
M. Tejera, A. Hilton
A shape constrained Laplacian mesh deformation approach is introduced for interactive editing of mesh sequences. This allows low-level constraints, such as foot or hand contact, to be imposed while preserving the natural dynamics of the captured surface. The approach also allows artistic manipulation of motion style to achieve effects such as squash-and-stretch. Interactive editing of key-frames is followed by automatic temporal propagation over a window of frames. User edits are seamlessly integrated into the captured mesh sequence. Three spatio-temporal interpolation methods are evaluated. Results on a variety of real and synthetic sequences demonstrate that the approach enables flexible manipulation of captured 3D video sequences.
提出了一种基于形状约束的拉普拉斯网格变形方法,用于网格序列的交互式编辑。这允许低层次的约束,比如脚或手的接触,在保持捕获表面的自然动态的同时被施加。该方法还允许艺术操作的运动风格,以实现效果,如挤压和拉伸。关键帧的交互式编辑之后是在一个帧窗口上的自动时间传播。用户编辑无缝集成到捕获的网格序列。对三种时空插值方法进行了评价。对各种真实和合成序列的结果表明,该方法能够灵活地操纵捕获的3D视频序列。
{"title":"Space-time Editing of 3D Video Sequences","authors":"M. Tejera, A. Hilton","doi":"10.1109/CVMP.2011.23","DOIUrl":"https://doi.org/10.1109/CVMP.2011.23","url":null,"abstract":"A shape constrained Laplacian mesh deformation approach is introduced for interactive editing of mesh sequences. This allows low-level constraints, such as foot or hand contact, to be imposed while preserving the natural dynamics of the captured surface. The approach also allows artistic manipulation of motion style to achieve effects such as squash-and-stretch. Interactive editing of key-frames is followed by automatic temporal propagation over a window of frames. User edits are seamlessly integrated into the captured mesh sequence. Three spatio-temporal interpolation methods are evaluated. Results on a variety of real and synthetic sequences demonstrate that the approach enables flexible manipulation of captured 3D video sequences.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134194117","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Making of Who Cares? HD Stereoscopic Free Viewpoint Video 制作《谁在乎?》高清立体免费视点视频
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.7
C. Lipski, F. Klose, K. Ruhl, M. Magnor
We present a detailed blueprint of our stereoscopic freeviewpoint video system. Using unsynchronized footage as input, we can render virtual camera paths in the post-production stage. The movement of the virtual camera also extends to the temporal domain, so that slow-motion and freeze-and-rotate shots are possible. As a proof-of-concept, a full length stereoscopic HD music video has been produced using our approach.
我们提出了我们的立体自由视点视频系统的详细蓝图。使用不同步的素材作为输入,我们可以在后期制作阶段渲染虚拟摄像机路径。虚拟摄像机的运动也扩展到时域,因此慢动作和定格旋转的镜头是可能的。作为概念验证,一个完整的立体高清音乐视频已经制作使用我们的方法。
{"title":"Making of Who Cares? HD Stereoscopic Free Viewpoint Video","authors":"C. Lipski, F. Klose, K. Ruhl, M. Magnor","doi":"10.1109/CVMP.2011.7","DOIUrl":"https://doi.org/10.1109/CVMP.2011.7","url":null,"abstract":"We present a detailed blueprint of our stereoscopic freeviewpoint video system. Using unsynchronized footage as input, we can render virtual camera paths in the post-production stage. The movement of the virtual camera also extends to the temporal domain, so that slow-motion and freeze-and-rotate shots are possible. As a proof-of-concept, a full length stereoscopic HD music video has been produced using our approach.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133122514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Efficient Dense Reconstruction from Video 高效的视频密集重建
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.10
Phil Parsonage, A. Hilton, J. Starck
We present a framework for efficient reconstruction of dense scene structure from video. Sequential structure-from-motion recovers camera information from video, providing only sparse 3D points. We build a dense 3D point cloud by performing full-frame tracking and depth estimation across sequences. First, we present a novel algorithm for sequential frame selection to extract a set of key frames with sufficient parallax for accurate depth reconstruction. Second, we introduce a technique for efficient reconstruction using dense tracking with geometrically correct optimisation of depth and orientation. Key frame selection is also performed in optimisation to provide accurate depth reconstruction for different scene elements. We test our work on benchmark footage and scenes containing local non-rigid motion, foreground clutter and occlusions to show comparable performance to state of the art techniques. We also show a substantial increase in speed on real world footage compared to existing methods, when they succeed, and successful reconstructions when they fail.
提出了一种从视频中高效重建密集场景结构的框架。序列运动结构从视频中恢复摄像机信息,只提供稀疏的3D点。我们通过跨序列执行全帧跟踪和深度估计来构建密集的3D点云。首先,我们提出了一种新的连续帧选择算法,以提取一组具有足够视差的关键帧,用于精确的深度重建。其次,我们介绍了一种利用深度和方向几何正确优化的密集跟踪进行高效重建的技术。关键帧选择也进行了优化,以提供准确的深度重建不同的场景元素。我们在包含局部非刚性运动、前景杂波和遮挡的基准镜头和场景上测试我们的工作,以显示与最先进技术相当的性能。我们还显示,与现有方法相比,在真实世界的镜头上,当它们成功时,以及在失败时成功重建时,它们的速度都有了大幅提高。
{"title":"Efficient Dense Reconstruction from Video","authors":"Phil Parsonage, A. Hilton, J. Starck","doi":"10.1109/CVMP.2011.10","DOIUrl":"https://doi.org/10.1109/CVMP.2011.10","url":null,"abstract":"We present a framework for efficient reconstruction of dense scene structure from video. Sequential structure-from-motion recovers camera information from video, providing only sparse 3D points. We build a dense 3D point cloud by performing full-frame tracking and depth estimation across sequences. First, we present a novel algorithm for sequential frame selection to extract a set of key frames with sufficient parallax for accurate depth reconstruction. Second, we introduce a technique for efficient reconstruction using dense tracking with geometrically correct optimisation of depth and orientation. Key frame selection is also performed in optimisation to provide accurate depth reconstruction for different scene elements. We test our work on benchmark footage and scenes containing local non-rigid motion, foreground clutter and occlusions to show comparable performance to state of the art techniques. We also show a substantial increase in speed on real world footage compared to existing methods, when they succeed, and successful reconstructions when they fail.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132032842","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Semantic Kernels Binarized - A Feature Descriptor for Fast and Robust Matching 语义核二值化——一种快速鲁棒匹配的特征描述符
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.11
Frederik Zilly, C. Riechert, P. Eisert, P. Kauff
This paper presents a new approach for feature description used in image processing and robust image recognition algorithms such as 3D camera tracking, view reconstruction or 3D scene analysis. State of the art feature detectors distinguish interest point detection and description. The former is commonly performed in scale space, while the latter is used to describe a normalized support region using histograms of gradients or similar derivatives of the grayscale image patch. This approach has proven to be very successful. However, the descriptors are usually of high dimensionality in order to achieve a high descriptiveness. Against this background, we propose a binarized descriptor which has a low memory usage and good matching performance. The descriptor is composed of binarized responses resulting from a set of folding operations applied to the normalized support region. We demonstrate the real-time capabilities of the feature descriptor in a stereo matching environment.
本文提出了一种新的特征描述方法,用于图像处理和鲁棒图像识别算法,如三维摄像机跟踪,视图重建或三维场景分析。最先进的特征检测器区分兴趣点检测和描述。前者通常在尺度空间中进行,而后者则使用灰度图像patch的梯度直方图或类似导数来描述归一化的支持区域。这种方法已被证明是非常成功的。然而,为了获得高描述性,描述符通常是高维的。在此背景下,我们提出了一种具有低内存占用和良好匹配性能的二值化描述符。描述符由应用于规范化支持区域的一组折叠操作产生的二值化响应组成。我们演示了特征描述符在立体匹配环境中的实时能力。
{"title":"Semantic Kernels Binarized - A Feature Descriptor for Fast and Robust Matching","authors":"Frederik Zilly, C. Riechert, P. Eisert, P. Kauff","doi":"10.1109/CVMP.2011.11","DOIUrl":"https://doi.org/10.1109/CVMP.2011.11","url":null,"abstract":"This paper presents a new approach for feature description used in image processing and robust image recognition algorithms such as 3D camera tracking, view reconstruction or 3D scene analysis. State of the art feature detectors distinguish interest point detection and description. The former is commonly performed in scale space, while the latter is used to describe a normalized support region using histograms of gradients or similar derivatives of the grayscale image patch. This approach has proven to be very successful. However, the descriptors are usually of high dimensionality in order to achieve a high descriptiveness. Against this background, we propose a binarized descriptor which has a low memory usage and good matching performance. The descriptor is composed of binarized responses resulting from a set of folding operations applied to the normalized support region. We demonstrate the real-time capabilities of the feature descriptor in a stereo matching environment.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"673 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116104476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
A Real-time Production Tool for Animated Hand Sketches 动画手绘的实时制作工具
Pub Date : 2011-11-16 DOI: 10.1109/CVMP.2011.17
J. Loviscach
In recent years, the look of hand-drawn sketches has become fashionable in video production. This paper introduces a software tool to produce corresponding videos in real time during lectures, presentations, or in the studio. Currently, two styles are available. In the first style, a hand seems to draw on a whiteboard. In the second style, the presenter seems to stand behind a transparent board, which is simulated with the help of a camera. In both cases, the input stems from a standard graphics tablet. The image of the lecturer's arm is synthesized from photographs and animated through inverse kinematics, the sounds of the pen and the eraser are synthesized from recordings. Auxiliary functions include a ghosted script for the presenter and drag drop of graphical elements prepared in advance.
近年来,手绘草图的外观已经成为视频制作的时尚。本文介绍了一种软件工具,可以在讲座、演示或演播室中实时生成相应的视频。目前有两种风格可供选择。在第一种风格中,一只手似乎在白板上画画。在第二种风格中,演讲者似乎站在一块透明的板后面,这是在相机的帮助下模拟的。在这两种情况下,输入都来自标准的图形平板电脑。讲师手臂的图像是由照片合成的,并通过逆运动学进行动画化,钢笔和橡皮擦的声音是由录音合成的。辅助功能包括用于演示者的虚影脚本和预先准备好的图形元素的拖放。
{"title":"A Real-time Production Tool for Animated Hand Sketches","authors":"J. Loviscach","doi":"10.1109/CVMP.2011.17","DOIUrl":"https://doi.org/10.1109/CVMP.2011.17","url":null,"abstract":"In recent years, the look of hand-drawn sketches has become fashionable in video production. This paper introduces a software tool to produce corresponding videos in real time during lectures, presentations, or in the studio. Currently, two styles are available. In the first style, a hand seems to draw on a whiteboard. In the second style, the presenter seems to stand behind a transparent board, which is simulated with the help of a camera. In both cases, the input stems from a standard graphics tablet. The image of the lecturer's arm is synthesized from photographs and animated through inverse kinematics, the sounds of the pen and the eraser are synthesized from recordings. Auxiliary functions include a ghosted script for the presenter and drag drop of graphical elements prepared in advance.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-11-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128860831","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Head-Mounted Photometric Stereo for Performance Capture 用于性能捕获的头戴式光度立体声
Pub Date : 2010-07-26 DOI: 10.1145/1837026.1837088
Andrew Jones, Graham Fyffe, Xueming Yu, Alex Ma, Jay Busch, M. Bolas, P. Debevec
Head-mounted cameras are an increasingly important tool for capturing facial performances to drive virtual characters. They provide a fixed, unoccluded view of the face, useful for observing motion capture dots or as input to video analysis. However, the 2D imagery captured with these systems is typically affected by ambient light and generally fails to record subtle 3D shape changes as the face performs. We have developed a system that augments a head-mounted camera with LED-based photometric stereo. The system allows observation of the face independent of the ambient light and generates per-pixel surface normals so that the performance is recorded dynamically in 3D. The resulting data can be used for facial relighting or as better input to machine learning algorithms for driving an animated face.
头戴式摄像机是一种越来越重要的工具,用于捕捉面部表演来驱动虚拟角色。它们提供了一个固定的、无遮挡的面部视图,对于观察动作捕捉点或作为视频分析的输入很有用。然而,用这些系统捕获的2D图像通常受到环境光的影响,并且通常无法记录面部执行时微妙的3D形状变化。我们已经开发了一个系统,它可以增强一个头戴式摄像头与基于led的光度立体。该系统允许在不受环境光影响的情况下观察人脸,并生成逐像素的表面法线,以便在3D中动态记录性能。由此产生的数据可以用于面部重新照明,或者作为驱动动画面部的机器学习算法的更好输入。
{"title":"Head-Mounted Photometric Stereo for Performance Capture","authors":"Andrew Jones, Graham Fyffe, Xueming Yu, Alex Ma, Jay Busch, M. Bolas, P. Debevec","doi":"10.1145/1837026.1837088","DOIUrl":"https://doi.org/10.1145/1837026.1837088","url":null,"abstract":"Head-mounted cameras are an increasingly important tool for capturing facial performances to drive virtual characters. They provide a fixed, unoccluded view of the face, useful for observing motion capture dots or as input to video analysis. However, the 2D imagery captured with these systems is typically affected by ambient light and generally fails to record subtle 3D shape changes as the face performs. We have developed a system that augments a head-mounted camera with LED-based photometric stereo. The system allows observation of the face independent of the ambient light and generates per-pixel surface normals so that the performance is recorded dynamically in 3D. The resulting data can be used for facial relighting or as better input to machine learning algorithms for driving an animated face.","PeriodicalId":167135,"journal":{"name":"2011 Conference for Visual Media Production","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-07-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131210935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
期刊
2011 Conference for Visual Media Production
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1