首页 > 最新文献

2020 International Conference on 3D Vision (3DV)最新文献

英文 中文
3DV 2020 Organizing Committee 3DV 2020组委会
Pub Date : 2020-11-01 DOI: 10.1109/3dv50981.2020.00007
{"title":"3DV 2020 Organizing Committee","authors":"","doi":"10.1109/3dv50981.2020.00007","DOIUrl":"https://doi.org/10.1109/3dv50981.2020.00007","url":null,"abstract":"","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130606209","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Cycle-Consistent Generative Rendering for 2D-3D Modality Translation 2D-3D模态翻译的循环一致生成渲染
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00033
Tristan Aumentado-Armstrong, Alex Levinshtein, Stavros Tsogkas, K. Derpanis, A. Jepson
For humans, visual understanding is inherently generative: given a 3D shape, we can postulate how it would look in the world; given a 2D image, we can infer the 3D structure that likely gave rise to it. We can thus translate between the 2D visual and 3D structural modalities of a given object. In the context of computer vision, this corresponds to a learnable module that serves two purposes: (i) generate a realistic rendering of a 3D object (shape-toimage translation) and (ii) infer a realistic 3D shape from an image (image-to-shape translation). In this paper, we learn such a module while being conscious of the difficulties in obtaining large paired 2D-3D datasets. By leveraging generative domain translation methods, we are able to define a learning algorithm that requires only weak supervision, with unpaired data. The resulting model is not only able to perform 3D shape, pose, and texture inference from 2D images, but can also generate novel textured 3D shapes and renders, similar to a graphics pipeline. More specifically, our method (i) infers an explicit 3D mesh representation, (ii) utilizes example shapes to regularize inference, (iii) requires only an image mask (no keypoints or camera extrinsics), and (iv) has generative capabilities. While prior work explores subsets of these properties, their combination is novel. We demonstrate the utility of our learned representation, as well as its performance on image generation and unpaired 3D shape inference tasks.
对于人类来说,视觉理解是天生的:给定一个3D形状,我们可以假设它在世界中的样子;给定一个2D图像,我们可以推断出可能产生它的3D结构。因此,我们可以在给定对象的2D视觉和3D结构模式之间进行转换。在计算机视觉的背景下,这对应于一个可学习的模块,它有两个目的:(i)生成3D对象的逼真渲染(形状到图像的翻译)和(ii)从图像推断出逼真的3D形状(图像到形状的翻译)。在本文中,我们学习了这样一个模块,同时意识到获取大型配对2D-3D数据集的困难。通过利用生成域翻译方法,我们能够定义一种只需要弱监督的学习算法,使用未配对的数据。生成的模型不仅能够从2D图像中执行3D形状、姿态和纹理推断,还可以生成新颖的纹理3D形状和渲染,类似于图形管道。更具体地说,我们的方法(i)推断出一个明确的3D网格表示,(ii)利用示例形状来正则化推理,(iii)只需要一个图像掩模(没有关键点或相机外部),以及(iv)具有生成能力。虽然先前的工作探索了这些属性的子集,但它们的组合是新颖的。我们展示了我们学习到的表示的实用性,以及它在图像生成和未配对的3D形状推理任务上的性能。
{"title":"Cycle-Consistent Generative Rendering for 2D-3D Modality Translation","authors":"Tristan Aumentado-Armstrong, Alex Levinshtein, Stavros Tsogkas, K. Derpanis, A. Jepson","doi":"10.1109/3DV50981.2020.00033","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00033","url":null,"abstract":"For humans, visual understanding is inherently generative: given a 3D shape, we can postulate how it would look in the world; given a 2D image, we can infer the 3D structure that likely gave rise to it. We can thus translate between the 2D visual and 3D structural modalities of a given object. In the context of computer vision, this corresponds to a learnable module that serves two purposes: (i) generate a realistic rendering of a 3D object (shape-toimage translation) and (ii) infer a realistic 3D shape from an image (image-to-shape translation). In this paper, we learn such a module while being conscious of the difficulties in obtaining large paired 2D-3D datasets. By leveraging generative domain translation methods, we are able to define a learning algorithm that requires only weak supervision, with unpaired data. The resulting model is not only able to perform 3D shape, pose, and texture inference from 2D images, but can also generate novel textured 3D shapes and renders, similar to a graphics pipeline. More specifically, our method (i) infers an explicit 3D mesh representation, (ii) utilizes example shapes to regularize inference, (iii) requires only an image mask (no keypoints or camera extrinsics), and (iv) has generative capabilities. While prior work explores subsets of these properties, their combination is novel. We demonstrate the utility of our learned representation, as well as its performance on image generation and unpaired 3D shape inference tasks.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125116934","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Motion Annotation Programs: A Scalable Approach to Annotating Kinematic Articulations in Large 3D Shape Collections 运动注释程序:在大型3D形状集合中注释运动关节的可扩展方法
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00071
Xianghao Xu, David Charatan, Sonia Raychaudhuri, Hanxiao Jiang, Mae Heitmann, Vladimir G. Kim, S. Chaudhuri, M. Savva, Angel X. Chang, Daniel Ritchie
3D models of real-world objects are essential for many applications, including the creation of virtual environments for AI training. To mimic real-world objects in these applications, objects must be annotated with their kinematic mobilities. Annotating kinematic motions is time-consuming, and it is not well-suited to typical crowdsourcing workflows due to the significant domain expertise required. In this paper, we present a system that helps individual expert users rapidly annotate kinematic motions in large 3D shape collections. The organizing concept of our system is motion annotation programs: simple, re-usable procedural rules that generate motion for a given input shape. Our interactive system allows users to author these rules and quickly apply them to collections of functionally-related objects. Using our system, an expert annotated over 1000 joints in under 3 hours. In a user study, participants with no prior experience with our system were able to annotate motions 1.5x faster than with a baseline manual annotation tool.
现实世界物体的3D模型对于许多应用来说是必不可少的,包括为人工智能训练创建虚拟环境。为了在这些应用程序中模拟现实世界中的对象,必须对对象进行运动学移动注释。注释运动学运动是耗时的,并且由于需要大量的领域专业知识,它不适合典型的众包工作流程。在本文中,我们提出了一个系统,可以帮助个人专家用户快速注释大型3D形状集合中的运动学运动。我们系统的组织概念是运动注释程序:简单,可重用的过程规则,为给定的输入形状生成运动。我们的交互式系统允许用户编写这些规则,并快速将它们应用于与功能相关的对象集合。使用我们的系统,专家在3小时内注释了1000多个关节。在一项用户研究中,没有使用我们系统经验的参与者能够比使用基线手动注释工具快1.5倍地注释动作。
{"title":"Motion Annotation Programs: A Scalable Approach to Annotating Kinematic Articulations in Large 3D Shape Collections","authors":"Xianghao Xu, David Charatan, Sonia Raychaudhuri, Hanxiao Jiang, Mae Heitmann, Vladimir G. Kim, S. Chaudhuri, M. Savva, Angel X. Chang, Daniel Ritchie","doi":"10.1109/3DV50981.2020.00071","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00071","url":null,"abstract":"3D models of real-world objects are essential for many applications, including the creation of virtual environments for AI training. To mimic real-world objects in these applications, objects must be annotated with their kinematic mobilities. Annotating kinematic motions is time-consuming, and it is not well-suited to typical crowdsourcing workflows due to the significant domain expertise required. In this paper, we present a system that helps individual expert users rapidly annotate kinematic motions in large 3D shape collections. The organizing concept of our system is motion annotation programs: simple, re-usable procedural rules that generate motion for a given input shape. Our interactive system allows users to author these rules and quickly apply them to collections of functionally-related objects. Using our system, an expert annotated over 1000 joints in under 3 hours. In a user study, participants with no prior experience with our system were able to annotate motions 1.5x faster than with a baseline manual annotation tool.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"240 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114663381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
HyperSLAM: A Generic and Modular Approach to Sensor Fusion and Simultaneous Localization And Mapping in Continuous-Time hyperlam:一种通用和模块化的连续时间传感器融合和同步定位和映射方法
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00108
David Hug, M. Chli
Within recent years, Continuous-Time Simultaneous Localization And Mapping (CTSLAM) formalisms have become subject to increased attention from the scientific community due to their vast potential in facilitating motion corrected feature reprojection and direct unsynchronized multi-rate sensor fusion. They also hold the promise of yielding better estimates in traditional sensor setups (e.g. visual, inertial) when compared to conventional discrete-time approaches. Related works mostly rely on cubic, $C^{2}-$continuous, uniform cumulative B-Splines to exemplify and demonstrate the benefits inherent to continuous-time representations. However, as this type of splines gives rise to continuous trajectories by blending uniformly distributed $mathbb{SE}_{3}$ transformations in time, it is prone to under- or overparametrize underlying motions with varying volatility and prohibits dynamic trajectory refinement or sparsification by design. In light of this, we propose employing a more generalized and efficient non-uniform split interpolation method in $mathbb{R}times mathbb{SU}_{2}times mathbb{R}^{3}$ and commence with development of ‘HyperSLAM’, a generic and modular CTSLAM framework. The efficacy of our approach is exemplified in proof-of-concept simulations based on a visual, monocular setup.
近年来,连续时间同步定位和映射(CTSLAM)形式由于其在促进运动校正特征重投影和直接非同步多速率传感器融合方面的巨大潜力而受到科学界越来越多的关注。与传统的离散时间方法相比,它们还有望在传统传感器设置(例如视觉,惯性)中产生更好的估计。相关工作主要依赖于三次,$C^{2}-$连续,均匀累积b样条来举例说明和演示连续时间表示固有的好处。然而,由于这种类型的样条曲线通过在时间上混合均匀分布的$mathbb{SE}_{3}$变换而产生连续轨迹,它容易使具有不同波动性的底层运动参数化不足或过度,并且禁止通过设计进行动态轨迹细化或稀疏化。鉴于此,我们建议在$mathbb{R}次mathbb{SU}_{2}次mathbb{R}^{3}$中采用一种更通用、更高效的非均匀分割插值方法,并从开发通用、模块化的CTSLAM框架“HyperSLAM”开始。我们的方法的有效性在基于视觉,单目设置的概念验证模拟中得到了例证。
{"title":"HyperSLAM: A Generic and Modular Approach to Sensor Fusion and Simultaneous Localization And Mapping in Continuous-Time","authors":"David Hug, M. Chli","doi":"10.1109/3DV50981.2020.00108","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00108","url":null,"abstract":"Within recent years, Continuous-Time Simultaneous Localization And Mapping (CTSLAM) formalisms have become subject to increased attention from the scientific community due to their vast potential in facilitating motion corrected feature reprojection and direct unsynchronized multi-rate sensor fusion. They also hold the promise of yielding better estimates in traditional sensor setups (e.g. visual, inertial) when compared to conventional discrete-time approaches. Related works mostly rely on cubic, $C^{2}-$continuous, uniform cumulative B-Splines to exemplify and demonstrate the benefits inherent to continuous-time representations. However, as this type of splines gives rise to continuous trajectories by blending uniformly distributed $mathbb{SE}_{3}$ transformations in time, it is prone to under- or overparametrize underlying motions with varying volatility and prohibits dynamic trajectory refinement or sparsification by design. In light of this, we propose employing a more generalized and efficient non-uniform split interpolation method in $mathbb{R}times mathbb{SU}_{2}times mathbb{R}^{3}$ and commence with development of ‘HyperSLAM’, a generic and modular CTSLAM framework. The efficacy of our approach is exemplified in proof-of-concept simulations based on a visual, monocular setup.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"22 14_suppl 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128204516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Precomputed Radiance Transfer for Reflectance and Lighting Estimation 反射和照明估计的预计算辐射转移
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00125
D. Thul, Vagia Tsiminaki, L. Ladicky, M. Pollefeys
Decomposing scenes into reflectance and lighting is an important task for applications such as relighting, image matching or content creation. Advanced light transport effects like occlusion and indirect lighting are often ignored, leading to subpar decompositions in which the albedo needs to compensate for insufficiencies in the estimated shading. We show how to account for these advanced lighting effects by utilizing precomputed radiance transfer to estimate reflectance and lighting. Given the geometry of an object and one or multiple images, our method reconstructs the object’s surface reflectance properties—such as its albedo and glossiness—as well as a colored lighting environment map. Evaluation on synthetic and real data shows that incorporation of indirect light leads to qualitatively and quantitatively improved results.
将场景分解为反射率和光照是重照明、图像匹配或内容创建等应用程序的重要任务。像遮挡和间接照明这样的高级光传输效应经常被忽略,导致反照率需要补偿估计阴影的不足。我们展示了如何利用预先计算的辐射转移来估计反射率和照明来解释这些先进的照明效果。给定物体的几何形状和一张或多张图像,我们的方法重建物体的表面反射率属性——比如反照率和光泽度——以及彩色照明环境图。对合成数据和实际数据的评价表明,间接光的加入导致定性和定量结果的改善。
{"title":"Precomputed Radiance Transfer for Reflectance and Lighting Estimation","authors":"D. Thul, Vagia Tsiminaki, L. Ladicky, M. Pollefeys","doi":"10.1109/3DV50981.2020.00125","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00125","url":null,"abstract":"Decomposing scenes into reflectance and lighting is an important task for applications such as relighting, image matching or content creation. Advanced light transport effects like occlusion and indirect lighting are often ignored, leading to subpar decompositions in which the albedo needs to compensate for insufficiencies in the estimated shading. We show how to account for these advanced lighting effects by utilizing precomputed radiance transfer to estimate reflectance and lighting. Given the geometry of an object and one or multiple images, our method reconstructs the object’s surface reflectance properties—such as its albedo and glossiness—as well as a colored lighting environment map. Evaluation on synthetic and real data shows that incorporation of indirect light leads to qualitatively and quantitatively improved results.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124108510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
High-Dynamic-Range Lighting Estimation From Face Portraits 人脸肖像的高动态范围照明估计
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00045
Alejandro Sztrajman, A. Neophytou, T. Weyrich, Eric Sommerlade
We present a CNN-based method for outdoor highdynamic-range (HDR) environment map prediction from low-dynamic-range (LDR) portrait images. Our method relies on two different CNN architectures, one for light encoding and another for face-to-light prediction. Outdoor lighting is characterised by an extremely high dynamic range, and thus our encoding splits the environment map data between low and high-intensity components, and encodes them using tailored representations. The combination of both network architectures constitutes an end-to-end method for accurate HDR light prediction from faces at real-time rates, inaccessible for previous methods which focused on low dynamic range lighting or relied on non-linear optimisation schemes. We train our networks using both real and synthetic images, we compare our light encoding with other methods for light representation, and we analyse our results for light prediction on real images. We show that our predicted HDR environment maps can be used as accurate illumination sources for scene renderings, with potential applications in 3D object insertion for augmented reality.
本文提出了一种基于cnn的基于低动态范围(LDR)人像图像的户外高动态范围(HDR)环境地图预测方法。我们的方法依赖于两种不同的CNN架构,一种用于光编码,另一种用于对光预测。户外照明的特点是具有极高的动态范围,因此我们的编码将环境地图数据划分为低强度和高强度组件,并使用定制的表示对它们进行编码。这两种网络架构的结合构成了一种端到端的方法,可以实时准确地从面部进行HDR光预测,这是以前专注于低动态范围照明或依赖非线性优化方案的方法无法实现的。我们使用真实图像和合成图像训练我们的网络,将我们的光编码与其他光表示方法进行比较,并分析我们的结果用于真实图像的光预测。我们表明,我们预测的HDR环境地图可以用作场景渲染的精确照明源,在增强现实的3D对象插入中具有潜在的应用。
{"title":"High-Dynamic-Range Lighting Estimation From Face Portraits","authors":"Alejandro Sztrajman, A. Neophytou, T. Weyrich, Eric Sommerlade","doi":"10.1109/3DV50981.2020.00045","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00045","url":null,"abstract":"We present a CNN-based method for outdoor highdynamic-range (HDR) environment map prediction from low-dynamic-range (LDR) portrait images. Our method relies on two different CNN architectures, one for light encoding and another for face-to-light prediction. Outdoor lighting is characterised by an extremely high dynamic range, and thus our encoding splits the environment map data between low and high-intensity components, and encodes them using tailored representations. The combination of both network architectures constitutes an end-to-end method for accurate HDR light prediction from faces at real-time rates, inaccessible for previous methods which focused on low dynamic range lighting or relied on non-linear optimisation schemes. We train our networks using both real and synthetic images, we compare our light encoding with other methods for light representation, and we analyse our results for light prediction on real images. We show that our predicted HDR environment maps can be used as accurate illumination sources for scene renderings, with potential applications in 3D object insertion for augmented reality.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132508346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Semantic Deep Face Models 语义深度人脸模型
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00044
P. Chandran, D. Bradley, M. Gross, T. Beeler
Face models built from 3D face databases are often used in computer vision and graphics tasks such as face reconstruction, replacement, tracking and manipulation. For such tasks, commonly used multi-linear morphable models, which provide semantic control over facial identity and expression, often lack quality and expressivity due to their linear nature. Deep neural networks offer the possibility of non-linear face modeling, where so far most research has focused on generating realistic facial images with less focus on 3D geometry, and methods that do produce geometry have little or no notion of semantic control, thereby limiting their artistic applicability. We present a method for nonlinear 3D face modeling using neural architectures that provides intuitive semantic control over both identity and expression by disentangling these dimensions from each other, essentially combining the benefits of both multi-linear face models and nonlinear deep face networks. The result is a powerful, semantically controllable, nonlinear, parametric face model. We demonstrate the value of our semantic deep face model with applications of 3D face synthesis, facial performance transfer, performance editing, and 2D landmark-based performance retargeting.
从三维人脸数据库建立的人脸模型经常用于计算机视觉和图形任务,如人脸重建、替换、跟踪和操作。对于此类任务,通常使用的多线性变形模型提供了对面部身份和表情的语义控制,但由于其线性特性,往往缺乏质量和表现力。深度神经网络提供了非线性面部建模的可能性,到目前为止,大多数研究都集中在生成逼真的面部图像上,而对3D几何图形的关注较少,而生成几何图形的方法很少或根本没有语义控制的概念,从而限制了它们的艺术适用性。我们提出了一种使用神经结构进行非线性三维人脸建模的方法,该方法通过将这些维度相互分离来提供对身份和表达的直观语义控制,本质上结合了多线性人脸模型和非线性深度人脸网络的优点。结果是一个强大的、语义可控的、非线性的、参数化的人脸模型。我们通过应用3D人脸合成、面部性能转移、性能编辑和基于2D地标的性能重定向来展示我们的语义深度人脸模型的价值。
{"title":"Semantic Deep Face Models","authors":"P. Chandran, D. Bradley, M. Gross, T. Beeler","doi":"10.1109/3DV50981.2020.00044","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00044","url":null,"abstract":"Face models built from 3D face databases are often used in computer vision and graphics tasks such as face reconstruction, replacement, tracking and manipulation. For such tasks, commonly used multi-linear morphable models, which provide semantic control over facial identity and expression, often lack quality and expressivity due to their linear nature. Deep neural networks offer the possibility of non-linear face modeling, where so far most research has focused on generating realistic facial images with less focus on 3D geometry, and methods that do produce geometry have little or no notion of semantic control, thereby limiting their artistic applicability. We present a method for nonlinear 3D face modeling using neural architectures that provides intuitive semantic control over both identity and expression by disentangling these dimensions from each other, essentially combining the benefits of both multi-linear face models and nonlinear deep face networks. The result is a powerful, semantically controllable, nonlinear, parametric face model. We demonstrate the value of our semantic deep face model with applications of 3D face synthesis, facial performance transfer, performance editing, and 2D landmark-based performance retargeting.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130461317","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Learning Distribution Independent Latent Representation for 3D Face Disentanglement 三维人脸解纠缠的学习分布独立潜在表征
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00095
Zihui Zhang, Cuican Yu, Huibin Li, Jian Sun, Feng Liu
Learning disentangled 3D face shape representation is beneficial to face attribute transfer, generation and recognition, etc. In this paper, we propose a novel distribution independence-based method to learn to decompose 3D face shapes. Specifically, we design a variational auto-encoder with Graph Convolutional Network (GCN), namely Mesh-Encoder, to model the distributions of identity and expression representations via variational inference. To disentangle facial expression and identity, we eliminate correlation of the two distributions, and enforce them to be independent by adversarial training. Extensive experiments show that the proposed approach can achieve state-of-the-art results in 3D face shape decomposition and expression transfer. Though focusing on disentanglement, our method also achieves the reconstruction accuracies comparable to the state-of-the-art 3D face reconstruction methods.
学习解纠缠的三维人脸形状表示有利于人脸属性的迁移、生成和识别等。本文提出了一种基于分布独立性的三维人脸形状学习分解方法。具体来说,我们设计了一个基于图卷积网络(GCN)的变分自编码器,即Mesh-Encoder,通过变分推理对恒等式和表达式表示的分布进行建模。为了分离面部表情和身份,我们消除了两种分布的相关性,并通过对抗性训练强制它们独立。大量的实验表明,该方法在三维人脸形状分解和表情传递方面取得了较好的效果。虽然我们的方法侧重于解纠缠,但我们的方法也达到了与最先进的3D人脸重建方法相当的重建精度。
{"title":"Learning Distribution Independent Latent Representation for 3D Face Disentanglement","authors":"Zihui Zhang, Cuican Yu, Huibin Li, Jian Sun, Feng Liu","doi":"10.1109/3DV50981.2020.00095","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00095","url":null,"abstract":"Learning disentangled 3D face shape representation is beneficial to face attribute transfer, generation and recognition, etc. In this paper, we propose a novel distribution independence-based method to learn to decompose 3D face shapes. Specifically, we design a variational auto-encoder with Graph Convolutional Network (GCN), namely Mesh-Encoder, to model the distributions of identity and expression representations via variational inference. To disentangle facial expression and identity, we eliminate correlation of the two distributions, and enforce them to be independent by adversarial training. Extensive experiments show that the proposed approach can achieve state-of-the-art results in 3D face shape decomposition and expression transfer. Though focusing on disentanglement, our method also achieves the reconstruction accuracies comparable to the state-of-the-art 3D face reconstruction methods.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127661047","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
3DV 2020 Program Committee 3DV 2020项目委员会
Pub Date : 2020-11-01 DOI: 10.1109/3dv50981.2020.00008
{"title":"3DV 2020 Program Committee","authors":"","doi":"10.1109/3dv50981.2020.00008","DOIUrl":"https://doi.org/10.1109/3dv50981.2020.00008","url":null,"abstract":"","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131377788","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast Discontinuity-Aware Subpixel Correspondence in Structured Light 结构光中快速间断感知亚像素对应
Pub Date : 2020-11-01 DOI: 10.1109/3DV50981.2020.00121
Nicolas Hurtubise, S. Roy
Structured light-based 3D scanning presents various challenges. While robustness to indirect illumination has been the subject of recent research, little has been said about discontinuities. This paper proposes a new discontinuity-aware algorithm for estimating structured light correspondences with subpixel accuracy. The algorithm is not only robust to common structured light problems, such as indirect lighting effects, but also identifies discontinuities explicitly. This results in a significant reduction of reconstruction artifacts at objects borders, an omnipresent problem of structured light methods, especially those relying on direct decoding. Our method is faster than previously proposed robust subpixel methods, has been tested on synthetic as well as real data and shows a significant improvement on measurement at discontinuities when compared with other state-of-the-art methods.
基于结构光的三维扫描提出了各种挑战。虽然对间接照明的稳健性一直是最近研究的主题,但关于不连续的研究很少。本文提出了一种新的亚像素精度结构光对应估计的间断感知算法。该算法不仅对常见的结构光问题(如间接照明效应)具有鲁棒性,而且可以明确地识别不连续性。这大大减少了物体边界的重建伪影,这是结构光方法普遍存在的问题,特别是那些依赖于直接解码的方法。我们的方法比以前提出的鲁棒亚像素方法更快,已经在合成和真实数据上进行了测试,与其他最先进的方法相比,在不连续处的测量显示出显着的改进。
{"title":"Fast Discontinuity-Aware Subpixel Correspondence in Structured Light","authors":"Nicolas Hurtubise, S. Roy","doi":"10.1109/3DV50981.2020.00121","DOIUrl":"https://doi.org/10.1109/3DV50981.2020.00121","url":null,"abstract":"Structured light-based 3D scanning presents various challenges. While robustness to indirect illumination has been the subject of recent research, little has been said about discontinuities. This paper proposes a new discontinuity-aware algorithm for estimating structured light correspondences with subpixel accuracy. The algorithm is not only robust to common structured light problems, such as indirect lighting effects, but also identifies discontinuities explicitly. This results in a significant reduction of reconstruction artifacts at objects borders, an omnipresent problem of structured light methods, especially those relying on direct decoding. Our method is faster than previously proposed robust subpixel methods, has been tested on synthetic as well as real data and shows a significant improvement on measurement at discontinuities when compared with other state-of-the-art methods.","PeriodicalId":293399,"journal":{"name":"2020 International Conference on 3D Vision (3DV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2020-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129176845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2020 International Conference on 3D Vision (3DV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1