Wanzhang Li;Fukun Yin;Wen Liu;Yiying Yang;Xin Chen;Biao Jiang;Gang Yu;Jiayuan Fan
{"title":"无界-GS:利用混合表示法扩展三维高斯拼接,实现无界大规模场景重建","authors":"Wanzhang Li;Fukun Yin;Wen Liu;Yiying Yang;Xin Chen;Biao Jiang;Gang Yu;Jiayuan Fan","doi":"10.1109/LRA.2024.3494652","DOIUrl":null,"url":null,"abstract":"Modeling large-scale scenes from multi-view images is challenging due to the trade-off dilemma between visual quality and computational cost. Existing NeRF-based methods have made advancements in neural implicit representation through volumetric ray-marching, but still struggle to deal with cubically growing sampling space in large-scale scenes. Fortunately, the rendering approach based on 3D Gaussian splatting (3DGS) has shown promising results, inspiring further exploration in the splatting setting. However, 3DGS has the limitation of inadequate Gaussian points for modeling distant backgrounds, leading to “splotchy” artifacts. To address this problem, we introduce a novel hybrid neural representation called Unbounded 3D Gaussian. For foreground area, we employs an explicit 3D Gaussian representation to efficiently model the geometry and appearance through splatting weighted Gaussians. For far-away background, we additionally introduce an implicit module comprising Multi-layer Perceptions (MLPs) to directly predict far-away background colors from positional encodings of view positions and ray directions. Furthermore, we design a seamless blending mechanism between the color predictions of the explicit splatting and implicit branches to reconstruct holistic scenes. Extensive experiments demonstrate that our proposed Unbounded-GS inherits the advantages of both faster convergence and high-fidelity rendering quality.","PeriodicalId":13241,"journal":{"name":"IEEE Robotics and Automation Letters","volume":"9 12","pages":"11529-11536"},"PeriodicalIF":4.6000,"publicationDate":"2024-11-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Unbounded-GS: Extending 3D Gaussian Splatting With Hybrid Representation for Unbounded Large-Scale Scene Reconstruction\",\"authors\":\"Wanzhang Li;Fukun Yin;Wen Liu;Yiying Yang;Xin Chen;Biao Jiang;Gang Yu;Jiayuan Fan\",\"doi\":\"10.1109/LRA.2024.3494652\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Modeling large-scale scenes from multi-view images is challenging due to the trade-off dilemma between visual quality and computational cost. Existing NeRF-based methods have made advancements in neural implicit representation through volumetric ray-marching, but still struggle to deal with cubically growing sampling space in large-scale scenes. Fortunately, the rendering approach based on 3D Gaussian splatting (3DGS) has shown promising results, inspiring further exploration in the splatting setting. However, 3DGS has the limitation of inadequate Gaussian points for modeling distant backgrounds, leading to “splotchy” artifacts. To address this problem, we introduce a novel hybrid neural representation called Unbounded 3D Gaussian. For foreground area, we employs an explicit 3D Gaussian representation to efficiently model the geometry and appearance through splatting weighted Gaussians. For far-away background, we additionally introduce an implicit module comprising Multi-layer Perceptions (MLPs) to directly predict far-away background colors from positional encodings of view positions and ray directions. Furthermore, we design a seamless blending mechanism between the color predictions of the explicit splatting and implicit branches to reconstruct holistic scenes. Extensive experiments demonstrate that our proposed Unbounded-GS inherits the advantages of both faster convergence and high-fidelity rendering quality.\",\"PeriodicalId\":13241,\"journal\":{\"name\":\"IEEE Robotics and Automation Letters\",\"volume\":\"9 12\",\"pages\":\"11529-11536\"},\"PeriodicalIF\":4.6000,\"publicationDate\":\"2024-11-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Robotics and Automation Letters\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10747249/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ROBOTICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Robotics and Automation Letters","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10747249/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ROBOTICS","Score":null,"Total":0}
Unbounded-GS: Extending 3D Gaussian Splatting With Hybrid Representation for Unbounded Large-Scale Scene Reconstruction
Modeling large-scale scenes from multi-view images is challenging due to the trade-off dilemma between visual quality and computational cost. Existing NeRF-based methods have made advancements in neural implicit representation through volumetric ray-marching, but still struggle to deal with cubically growing sampling space in large-scale scenes. Fortunately, the rendering approach based on 3D Gaussian splatting (3DGS) has shown promising results, inspiring further exploration in the splatting setting. However, 3DGS has the limitation of inadequate Gaussian points for modeling distant backgrounds, leading to “splotchy” artifacts. To address this problem, we introduce a novel hybrid neural representation called Unbounded 3D Gaussian. For foreground area, we employs an explicit 3D Gaussian representation to efficiently model the geometry and appearance through splatting weighted Gaussians. For far-away background, we additionally introduce an implicit module comprising Multi-layer Perceptions (MLPs) to directly predict far-away background colors from positional encodings of view positions and ray directions. Furthermore, we design a seamless blending mechanism between the color predictions of the explicit splatting and implicit branches to reconstruct holistic scenes. Extensive experiments demonstrate that our proposed Unbounded-GS inherits the advantages of both faster convergence and high-fidelity rendering quality.
期刊介绍:
The scope of this journal is to publish peer-reviewed articles that provide a timely and concise account of innovative research ideas and application results, reporting significant theoretical findings and application case studies in areas of robotics and automation.