首页 > 最新文献

2019 International Conference on 3D Immersion (IC3D)最新文献

英文 中文
Local-Convexity Reinforcement for Scene Reconstruction from Sparse Point Clouds 稀疏点云场景重建的局部凸性增强
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975900
M. Lhuillier
Several methods reconstruct surfaces from sparse point clouds that are estimated from images. Most of them build 3D Delaunay triangulation of the points and compute occupancy labeling of the tetrahedra thanks to visibility information and surface constraints. However their most notable errors are falsely-labeled freespace tetrahedra. We present labeling corrections of these errors based on a new shape constraint: local-convexity. In the simplest case, this means that a freespace tetrahedron of the Delaunay is relabeled matter if its size is small enough and all its vertices are in matter tetrahedra. The allowed corrections are more important in the vertical direction than in the horizontal ones to take into account the anisotropy of usual scenes. In the experiments, our corrections improve the results of previous surface reconstruction methods applied to videos taken by a consumer 360 camera.
有几种方法是从图像中估计的稀疏点云重建表面。它们大多建立点的三维Delaunay三角剖分,并利用可见性信息和表面约束计算四面体的占用标记。然而,他们最显著的错误是错误地标记了自由空间四面体。我们提出了一种新的形状约束:局部凸性对这些误差进行标记校正。在最简单的情况下,这意味着如果一个Delaunay的自由空间四面体的尺寸足够小,并且它的所有顶点都在物质四面体中,那么它就被重新标记为物质。考虑到通常场景的各向异性,在垂直方向上允许的校正比在水平方向上更重要。在实验中,我们的修正改进了先前用于消费者360相机拍摄的视频的表面重建方法的结果。
{"title":"Local-Convexity Reinforcement for Scene Reconstruction from Sparse Point Clouds","authors":"M. Lhuillier","doi":"10.1109/IC3D48390.2019.8975900","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975900","url":null,"abstract":"Several methods reconstruct surfaces from sparse point clouds that are estimated from images. Most of them build 3D Delaunay triangulation of the points and compute occupancy labeling of the tetrahedra thanks to visibility information and surface constraints. However their most notable errors are falsely-labeled freespace tetrahedra. We present labeling corrections of these errors based on a new shape constraint: local-convexity. In the simplest case, this means that a freespace tetrahedron of the Delaunay is relabeled matter if its size is small enough and all its vertices are in matter tetrahedra. The allowed corrections are more important in the vertical direction than in the horizontal ones to take into account the anisotropy of usual scenes. In the experiments, our corrections improve the results of previous surface reconstruction methods applied to videos taken by a consumer 360 camera.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124349341","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A Rich Stereoscopic 3D High Dynamic Range Image & Video Database of Natural Scenes 一个丰富的立体3D高动态范围自然场景图像和视频数据库
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975903
Aditya Wadaskar, Mansi Sharma, Rohan Lal
The consumer market of High Dynamic Range (HDR) displays and cameras is blooming rapidly with the advent of 3D video and display technologies. Specialised agencies like Moving Picture Experts Group and International Telecommunication Union are demanding the standardization of latest display advancements. Lack of sufficient experimental data is a major bottleneck for the development of preliminary research efforts in 3D HDR video technology. We propose to make publicly available to the research community, a diversified database of Stereoscopic 3D HDR images and videos, captured within the beautiful campus of Indian Institute of Technology, Madras, which is blessed with rich flora and fauna, and is home to several rare wildlife species. Further, we have described the procedure of capturing, aligning, calibrating and post-processing of 3D images and videos. We have discussed research opportunities and challenges, and the potential use cases of HDR stereo 3D applications and depth-from-HDR aspects.
随着3D视频和显示技术的出现,高动态范围(HDR)显示器和相机的消费市场正在迅速发展。像移动图像专家组和国际电信联盟这样的专门机构要求对最新的显示技术进行标准化。缺乏足够的实验数据是制约3D HDR视频技术前期研究工作开展的主要瓶颈。我们建议向研究界公开提供一个立体3D HDR图像和视频的多样化数据库,这些图像和视频是在印度理工学院美丽的马德拉斯校园内拍摄的,那里有丰富的动植物,是几种稀有野生动物的家园。此外,我们还描述了三维图像和视频的捕获、对准、校准和后处理过程。我们讨论了研究机遇和挑战,以及HDR立体3D应用和深度的潜在用例。
{"title":"A Rich Stereoscopic 3D High Dynamic Range Image & Video Database of Natural Scenes","authors":"Aditya Wadaskar, Mansi Sharma, Rohan Lal","doi":"10.1109/IC3D48390.2019.8975903","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975903","url":null,"abstract":"The consumer market of High Dynamic Range (HDR) displays and cameras is blooming rapidly with the advent of 3D video and display technologies. Specialised agencies like Moving Picture Experts Group and International Telecommunication Union are demanding the standardization of latest display advancements. Lack of sufficient experimental data is a major bottleneck for the development of preliminary research efforts in 3D HDR video technology. We propose to make publicly available to the research community, a diversified database of Stereoscopic 3D HDR images and videos, captured within the beautiful campus of Indian Institute of Technology, Madras, which is blessed with rich flora and fauna, and is home to several rare wildlife species. Further, we have described the procedure of capturing, aligning, calibrating and post-processing of 3D images and videos. We have discussed research opportunities and challenges, and the potential use cases of HDR stereo 3D applications and depth-from-HDR aspects.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132459683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Process for the Semi-Automated Generation of Life-Sized, Interactive 3D Character Models for Holographic Projection 用于全息投影的半自动化生成真人大小的交互式3D角色模型的过程
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975993
Xinyu Huang, J. Twycross, Fridolin Wild
By mixing digital data into the real world, Augmented Reality (AR) can deliver potent immersive and interactive experience to its users. In many application contexts, this requires the capability to deploy animated, high fidelity 3D character models. In this paper, we propose a novel approach to efficiently transform – using 3D scanning – an actor to a photorealistic, animated character. This generated 3D assistant must be able to move to perform recorded motion capture data, and it must be able to generate dialogue with lip sync to naturally interact with the users. The approach we propose for creating these virtual AR assistants utilizes photogrammetric scanning, motion capture, and free viewpoint video for their integration in Unity. We deploy the Occipital Structure sensor to acquire static high-resolution textured surfaces, and a Vicon motion capture system to track series of movements. The proposed capturing process consists of the steps scanning, reconstruction with Wrap 3 and Maya, editing texture maps to reduce artefacts with Photoshop, and rigging with Maya and Motion Builder to render the models fit for animation and lip-sync using LipSyncPro. We test the approach in Unity by scanning two human models with 23 captured animations each. Our findings indicate that the major factors affecting the result quality are environment setup, lighting, and processing constraints.
通过将数字数据混合到现实世界中,增强现实(AR)可以为用户提供强大的沉浸式互动体验。在许多应用环境中,这需要部署动画、高保真3D角色模型的能力。在本文中,我们提出了一种新的方法来有效地转换-使用3D扫描-一个演员到一个逼真的动画人物。这个生成的3D助手必须能够移动来执行记录的动作捕捉数据,并且它必须能够生成与口型同步的对话,以便与用户自然互动。我们提出的创建这些虚拟AR助手的方法利用摄影测量扫描,动作捕捉和免费视点视频将其集成在Unity中。我们使用枕结构传感器来获取静态高分辨率纹理表面,并使用Vicon运动捕捉系统来跟踪一系列运动。提出的捕获过程包括步骤扫描,重建与Wrap 3和玛雅,编辑纹理映射,以减少与Photoshop的人工制品,并与Maya和Motion Builder索具,以渲染模型适合动画和口型同步使用LipSyncPro。我们在Unity中通过扫描两个具有23个捕获动画的人体模型来测试该方法。我们的研究结果表明,影响结果质量的主要因素是环境设置,照明和处理约束。
{"title":"A Process for the Semi-Automated Generation of Life-Sized, Interactive 3D Character Models for Holographic Projection","authors":"Xinyu Huang, J. Twycross, Fridolin Wild","doi":"10.1109/IC3D48390.2019.8975993","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975993","url":null,"abstract":"By mixing digital data into the real world, Augmented Reality (AR) can deliver potent immersive and interactive experience to its users. In many application contexts, this requires the capability to deploy animated, high fidelity 3D character models. In this paper, we propose a novel approach to efficiently transform – using 3D scanning – an actor to a photorealistic, animated character. This generated 3D assistant must be able to move to perform recorded motion capture data, and it must be able to generate dialogue with lip sync to naturally interact with the users. The approach we propose for creating these virtual AR assistants utilizes photogrammetric scanning, motion capture, and free viewpoint video for their integration in Unity. We deploy the Occipital Structure sensor to acquire static high-resolution textured surfaces, and a Vicon motion capture system to track series of movements. The proposed capturing process consists of the steps scanning, reconstruction with Wrap 3 and Maya, editing texture maps to reduce artefacts with Photoshop, and rigging with Maya and Motion Builder to render the models fit for animation and lip-sync using LipSyncPro. We test the approach in Unity by scanning two human models with 23 captured animations each. Our findings indicate that the major factors affecting the result quality are environment setup, lighting, and processing constraints.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"271 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133274399","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
IC3D 2019 Technical Program Committee IC3D 2019技术计划委员会
Pub Date : 2019-12-01 DOI: 10.1109/ic3d48390.2019.8976002
{"title":"IC3D 2019 Technical Program Committee","authors":"","doi":"10.1109/ic3d48390.2019.8976002","DOIUrl":"https://doi.org/10.1109/ic3d48390.2019.8976002","url":null,"abstract":"","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122530808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting Forward & Backward Facial Depth Maps From a Single RGB Image For Mobile 3d AR Application 从移动3d AR应用程序的单个RGB图像预测向前和向后面部深度图
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975899
P. Avinash, Mansi Sharma
Cheap and fast 3D asset creation to enable AR/VR applications is a fast growing domain. This paper addresses a significant problem of reconstructing complete 3D information of a face in near real-time speed on a mobile phone. We propose a novel deep learning based solution to predict robust depth maps of a face, one forward facing and the other backward facing, from a single image from the wild. A critical contribution is that the proposed network is capable of learning the depths of the occluded part of the face too. This is achieved by training a fully convolutional neural network to learn the dual (forward and backward) depth maps, with a common encoder and two separate decoders. The 300W-LP, a cloud point dataset, is used to compute the required dual depth maps from the training data. The code and results will be made available at project page.
廉价和快速的3D资产创建使AR/VR应用程序是一个快速增长的领域。本文解决了在手机上以接近实时的速度重建人脸完整三维信息的重要问题。我们提出了一种新的基于深度学习的解决方案来预测人脸的鲁棒深度图,一个面向前,另一个面向后,来自野外的单个图像。一个关键的贡献是,所提出的网络也能够学习人脸遮挡部分的深度。这是通过训练一个全卷积神经网络来学习双(前向和后向)深度图,使用一个通用编码器和两个独立的解码器来实现的。使用云点数据集300W-LP从训练数据中计算所需的双深度图。代码和结果将在项目页面上提供。
{"title":"Predicting Forward & Backward Facial Depth Maps From a Single RGB Image For Mobile 3d AR Application","authors":"P. Avinash, Mansi Sharma","doi":"10.1109/IC3D48390.2019.8975899","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975899","url":null,"abstract":"Cheap and fast 3D asset creation to enable AR/VR applications is a fast growing domain. This paper addresses a significant problem of reconstructing complete 3D information of a face in near real-time speed on a mobile phone. We propose a novel deep learning based solution to predict robust depth maps of a face, one forward facing and the other backward facing, from a single image from the wild. A critical contribution is that the proposed network is capable of learning the depths of the occluded part of the face too. This is achieved by training a fully convolutional neural network to learn the dual (forward and backward) depth maps, with a common encoder and two separate decoders. The 300W-LP, a cloud point dataset, is used to compute the required dual depth maps from the training data. The code and results will be made available at project page.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123818483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2019 International Conference on 3D Immersion (IC3D)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1