首页 > 最新文献

2019 International Conference on 3D Immersion (IC3D)最新文献

英文 中文
A Rich Stereoscopic 3D High Dynamic Range Image & Video Database of Natural Scenes 一个丰富的立体3D高动态范围自然场景图像和视频数据库
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975903
Aditya Wadaskar, Mansi Sharma, Rohan Lal
The consumer market of High Dynamic Range (HDR) displays and cameras is blooming rapidly with the advent of 3D video and display technologies. Specialised agencies like Moving Picture Experts Group and International Telecommunication Union are demanding the standardization of latest display advancements. Lack of sufficient experimental data is a major bottleneck for the development of preliminary research efforts in 3D HDR video technology. We propose to make publicly available to the research community, a diversified database of Stereoscopic 3D HDR images and videos, captured within the beautiful campus of Indian Institute of Technology, Madras, which is blessed with rich flora and fauna, and is home to several rare wildlife species. Further, we have described the procedure of capturing, aligning, calibrating and post-processing of 3D images and videos. We have discussed research opportunities and challenges, and the potential use cases of HDR stereo 3D applications and depth-from-HDR aspects.
随着3D视频和显示技术的出现,高动态范围(HDR)显示器和相机的消费市场正在迅速发展。像移动图像专家组和国际电信联盟这样的专门机构要求对最新的显示技术进行标准化。缺乏足够的实验数据是制约3D HDR视频技术前期研究工作开展的主要瓶颈。我们建议向研究界公开提供一个立体3D HDR图像和视频的多样化数据库,这些图像和视频是在印度理工学院美丽的马德拉斯校园内拍摄的,那里有丰富的动植物,是几种稀有野生动物的家园。此外,我们还描述了三维图像和视频的捕获、对准、校准和后处理过程。我们讨论了研究机遇和挑战,以及HDR立体3D应用和深度的潜在用例。
{"title":"A Rich Stereoscopic 3D High Dynamic Range Image & Video Database of Natural Scenes","authors":"Aditya Wadaskar, Mansi Sharma, Rohan Lal","doi":"10.1109/IC3D48390.2019.8975903","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975903","url":null,"abstract":"The consumer market of High Dynamic Range (HDR) displays and cameras is blooming rapidly with the advent of 3D video and display technologies. Specialised agencies like Moving Picture Experts Group and International Telecommunication Union are demanding the standardization of latest display advancements. Lack of sufficient experimental data is a major bottleneck for the development of preliminary research efforts in 3D HDR video technology. We propose to make publicly available to the research community, a diversified database of Stereoscopic 3D HDR images and videos, captured within the beautiful campus of Indian Institute of Technology, Madras, which is blessed with rich flora and fauna, and is home to several rare wildlife species. Further, we have described the procedure of capturing, aligning, calibrating and post-processing of 3D images and videos. We have discussed research opportunities and challenges, and the potential use cases of HDR stereo 3D applications and depth-from-HDR aspects.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132459683","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Process for the Semi-Automated Generation of Life-Sized, Interactive 3D Character Models for Holographic Projection 用于全息投影的半自动化生成真人大小的交互式3D角色模型的过程
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975993
Xinyu Huang, J. Twycross, Fridolin Wild
By mixing digital data into the real world, Augmented Reality (AR) can deliver potent immersive and interactive experience to its users. In many application contexts, this requires the capability to deploy animated, high fidelity 3D character models. In this paper, we propose a novel approach to efficiently transform – using 3D scanning – an actor to a photorealistic, animated character. This generated 3D assistant must be able to move to perform recorded motion capture data, and it must be able to generate dialogue with lip sync to naturally interact with the users. The approach we propose for creating these virtual AR assistants utilizes photogrammetric scanning, motion capture, and free viewpoint video for their integration in Unity. We deploy the Occipital Structure sensor to acquire static high-resolution textured surfaces, and a Vicon motion capture system to track series of movements. The proposed capturing process consists of the steps scanning, reconstruction with Wrap 3 and Maya, editing texture maps to reduce artefacts with Photoshop, and rigging with Maya and Motion Builder to render the models fit for animation and lip-sync using LipSyncPro. We test the approach in Unity by scanning two human models with 23 captured animations each. Our findings indicate that the major factors affecting the result quality are environment setup, lighting, and processing constraints.
通过将数字数据混合到现实世界中,增强现实(AR)可以为用户提供强大的沉浸式互动体验。在许多应用环境中,这需要部署动画、高保真3D角色模型的能力。在本文中,我们提出了一种新的方法来有效地转换-使用3D扫描-一个演员到一个逼真的动画人物。这个生成的3D助手必须能够移动来执行记录的动作捕捉数据,并且它必须能够生成与口型同步的对话,以便与用户自然互动。我们提出的创建这些虚拟AR助手的方法利用摄影测量扫描,动作捕捉和免费视点视频将其集成在Unity中。我们使用枕结构传感器来获取静态高分辨率纹理表面,并使用Vicon运动捕捉系统来跟踪一系列运动。提出的捕获过程包括步骤扫描,重建与Wrap 3和玛雅,编辑纹理映射,以减少与Photoshop的人工制品,并与Maya和Motion Builder索具,以渲染模型适合动画和口型同步使用LipSyncPro。我们在Unity中通过扫描两个具有23个捕获动画的人体模型来测试该方法。我们的研究结果表明,影响结果质量的主要因素是环境设置,照明和处理约束。
{"title":"A Process for the Semi-Automated Generation of Life-Sized, Interactive 3D Character Models for Holographic Projection","authors":"Xinyu Huang, J. Twycross, Fridolin Wild","doi":"10.1109/IC3D48390.2019.8975993","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975993","url":null,"abstract":"By mixing digital data into the real world, Augmented Reality (AR) can deliver potent immersive and interactive experience to its users. In many application contexts, this requires the capability to deploy animated, high fidelity 3D character models. In this paper, we propose a novel approach to efficiently transform – using 3D scanning – an actor to a photorealistic, animated character. This generated 3D assistant must be able to move to perform recorded motion capture data, and it must be able to generate dialogue with lip sync to naturally interact with the users. The approach we propose for creating these virtual AR assistants utilizes photogrammetric scanning, motion capture, and free viewpoint video for their integration in Unity. We deploy the Occipital Structure sensor to acquire static high-resolution textured surfaces, and a Vicon motion capture system to track series of movements. The proposed capturing process consists of the steps scanning, reconstruction with Wrap 3 and Maya, editing texture maps to reduce artefacts with Photoshop, and rigging with Maya and Motion Builder to render the models fit for animation and lip-sync using LipSyncPro. We test the approach in Unity by scanning two human models with 23 captured animations each. Our findings indicate that the major factors affecting the result quality are environment setup, lighting, and processing constraints.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"271 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133274399","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Hybrid Approach to Wide Baseline View Synthesis with Convolutional Neural Networks 基于卷积神经网络的宽基线视图混合合成方法
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8976000
Nour Hobloss, Andrei I. Purica, A. Fiandrotti, Marco Cagnazzo, R. Cozot, W. Hamidouche
Convolutional Neural Networks (CNN) have been recently employed for implementing complete end-to-end view synthesis architectures, from reference view warping to target view blending while dealing with occlusions as well. However, the convolutional sizes filters must increase with the distance between reference views, making all-convolutional approaches prohibitively complex for wide baseline setups. In this work we propose a hybrid approach to view synthesis where we first warp the reference views resolving the occlusions, and then we train a simpler convolutional architecture for blending the preprocessed views. By warping the reference views, we reduce the equivalent distance between reference views, allowing the use of smaller convolutional filters and thus lower network complexity. We experimentally show that our method performs favorably against both traditional and convolutional synthesis methods while retaining lower complexity with respect to the latter.
卷积神经网络(CNN)最近被用于实现完整的端到端视图合成架构,从参考视图扭曲到目标视图混合,同时处理遮挡。然而,卷积大小过滤器必须随着参考视图之间的距离而增加,这使得全卷积方法对于宽基线设置来说过于复杂。在这项工作中,我们提出了一种混合的视图合成方法,我们首先扭曲参考视图来解决遮挡,然后我们训练一个更简单的卷积架构来混合预处理视图。通过扭曲参考视图,我们减少了参考视图之间的等效距离,允许使用更小的卷积过滤器,从而降低了网络的复杂性。我们的实验表明,我们的方法优于传统和卷积合成方法,同时保持较低的复杂性相对于后者。
{"title":"A Hybrid Approach to Wide Baseline View Synthesis with Convolutional Neural Networks","authors":"Nour Hobloss, Andrei I. Purica, A. Fiandrotti, Marco Cagnazzo, R. Cozot, W. Hamidouche","doi":"10.1109/IC3D48390.2019.8976000","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8976000","url":null,"abstract":"Convolutional Neural Networks (CNN) have been recently employed for implementing complete end-to-end view synthesis architectures, from reference view warping to target view blending while dealing with occlusions as well. However, the convolutional sizes filters must increase with the distance between reference views, making all-convolutional approaches prohibitively complex for wide baseline setups. In this work we propose a hybrid approach to view synthesis where we first warp the reference views resolving the occlusions, and then we train a simpler convolutional architecture for blending the preprocessed views. By warping the reference views, we reduce the equivalent distance between reference views, allowing the use of smaller convolutional filters and thus lower network complexity. We experimentally show that our method performs favorably against both traditional and convolutional synthesis methods while retaining lower complexity with respect to the latter.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130114438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
IC3D 2019 Technical Program Committee IC3D 2019技术计划委员会
Pub Date : 2019-12-01 DOI: 10.1109/ic3d48390.2019.8976002
{"title":"IC3D 2019 Technical Program Committee","authors":"","doi":"10.1109/ic3d48390.2019.8976002","DOIUrl":"https://doi.org/10.1109/ic3d48390.2019.8976002","url":null,"abstract":"","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122530808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Predicting Forward & Backward Facial Depth Maps From a Single RGB Image For Mobile 3d AR Application 从移动3d AR应用程序的单个RGB图像预测向前和向后面部深度图
Pub Date : 2019-12-01 DOI: 10.1109/IC3D48390.2019.8975899
P. Avinash, Mansi Sharma
Cheap and fast 3D asset creation to enable AR/VR applications is a fast growing domain. This paper addresses a significant problem of reconstructing complete 3D information of a face in near real-time speed on a mobile phone. We propose a novel deep learning based solution to predict robust depth maps of a face, one forward facing and the other backward facing, from a single image from the wild. A critical contribution is that the proposed network is capable of learning the depths of the occluded part of the face too. This is achieved by training a fully convolutional neural network to learn the dual (forward and backward) depth maps, with a common encoder and two separate decoders. The 300W-LP, a cloud point dataset, is used to compute the required dual depth maps from the training data. The code and results will be made available at project page.
廉价和快速的3D资产创建使AR/VR应用程序是一个快速增长的领域。本文解决了在手机上以接近实时的速度重建人脸完整三维信息的重要问题。我们提出了一种新的基于深度学习的解决方案来预测人脸的鲁棒深度图,一个面向前,另一个面向后,来自野外的单个图像。一个关键的贡献是,所提出的网络也能够学习人脸遮挡部分的深度。这是通过训练一个全卷积神经网络来学习双(前向和后向)深度图,使用一个通用编码器和两个独立的解码器来实现的。使用云点数据集300W-LP从训练数据中计算所需的双深度图。代码和结果将在项目页面上提供。
{"title":"Predicting Forward & Backward Facial Depth Maps From a Single RGB Image For Mobile 3d AR Application","authors":"P. Avinash, Mansi Sharma","doi":"10.1109/IC3D48390.2019.8975899","DOIUrl":"https://doi.org/10.1109/IC3D48390.2019.8975899","url":null,"abstract":"Cheap and fast 3D asset creation to enable AR/VR applications is a fast growing domain. This paper addresses a significant problem of reconstructing complete 3D information of a face in near real-time speed on a mobile phone. We propose a novel deep learning based solution to predict robust depth maps of a face, one forward facing and the other backward facing, from a single image from the wild. A critical contribution is that the proposed network is capable of learning the depths of the occluded part of the face too. This is achieved by training a fully convolutional neural network to learn the dual (forward and backward) depth maps, with a common encoder and two separate decoders. The 300W-LP, a cloud point dataset, is used to compute the required dual depth maps from the training data. The code and results will be made available at project page.","PeriodicalId":344706,"journal":{"name":"2019 International Conference on 3D Immersion (IC3D)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123818483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2019 International Conference on 3D Immersion (IC3D)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1