{"title":"视图合成采样:从局部光场融合到神经辐射场及其他","authors":"Ravi Ramamoorthi","doi":"arxiv-2408.04586","DOIUrl":null,"url":null,"abstract":"Capturing and rendering novel views of complex real-world scenes is a\nlong-standing problem in computer graphics and vision, with applications in\naugmented and virtual reality, immersive experiences and 3D photography. The\nadvent of deep learning has enabled revolutionary advances in this area,\nclassically known as image-based rendering. However, previous approaches\nrequire intractably dense view sampling or provide little or no guidance for\nhow users should sample views of a scene to reliably render high-quality novel\nviews. Local light field fusion proposes an algorithm for practical view\nsynthesis from an irregular grid of sampled views that first expands each\nsampled view into a local light field via a multiplane image scene\nrepresentation, then renders novel views by blending adjacent local light\nfields. Crucially, we extend traditional plenoptic sampling theory to derive a\nbound that specifies precisely how densely users should sample views of a given\nscene when using our algorithm. We achieve the perceptual quality of Nyquist\nrate view sampling while using up to 4000x fewer views. Subsequent developments\nhave led to new scene representations for deep learning with view synthesis,\nnotably neural radiance fields, but the problem of sparse view synthesis from a\nsmall number of images has only grown in importance. We reprise some of the\nrecent results on sparse and even single image view synthesis, while posing the\nquestion of whether prescriptive sampling guidelines are feasible for the new\ngeneration of image-based rendering algorithms.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":"8 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Sampling for View Synthesis: From Local Light Field Fusion to Neural Radiance Fields and Beyond\",\"authors\":\"Ravi Ramamoorthi\",\"doi\":\"arxiv-2408.04586\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Capturing and rendering novel views of complex real-world scenes is a\\nlong-standing problem in computer graphics and vision, with applications in\\naugmented and virtual reality, immersive experiences and 3D photography. The\\nadvent of deep learning has enabled revolutionary advances in this area,\\nclassically known as image-based rendering. However, previous approaches\\nrequire intractably dense view sampling or provide little or no guidance for\\nhow users should sample views of a scene to reliably render high-quality novel\\nviews. Local light field fusion proposes an algorithm for practical view\\nsynthesis from an irregular grid of sampled views that first expands each\\nsampled view into a local light field via a multiplane image scene\\nrepresentation, then renders novel views by blending adjacent local light\\nfields. Crucially, we extend traditional plenoptic sampling theory to derive a\\nbound that specifies precisely how densely users should sample views of a given\\nscene when using our algorithm. We achieve the perceptual quality of Nyquist\\nrate view sampling while using up to 4000x fewer views. Subsequent developments\\nhave led to new scene representations for deep learning with view synthesis,\\nnotably neural radiance fields, but the problem of sparse view synthesis from a\\nsmall number of images has only grown in importance. We reprise some of the\\nrecent results on sparse and even single image view synthesis, while posing the\\nquestion of whether prescriptive sampling guidelines are feasible for the new\\ngeneration of image-based rendering algorithms.\",\"PeriodicalId\":501174,\"journal\":{\"name\":\"arXiv - CS - Graphics\",\"volume\":\"8 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-08-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"arXiv - CS - Graphics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/arxiv-2408.04586\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Graphics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2408.04586","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Sampling for View Synthesis: From Local Light Field Fusion to Neural Radiance Fields and Beyond
Capturing and rendering novel views of complex real-world scenes is a
long-standing problem in computer graphics and vision, with applications in
augmented and virtual reality, immersive experiences and 3D photography. The
advent of deep learning has enabled revolutionary advances in this area,
classically known as image-based rendering. However, previous approaches
require intractably dense view sampling or provide little or no guidance for
how users should sample views of a scene to reliably render high-quality novel
views. Local light field fusion proposes an algorithm for practical view
synthesis from an irregular grid of sampled views that first expands each
sampled view into a local light field via a multiplane image scene
representation, then renders novel views by blending adjacent local light
fields. Crucially, we extend traditional plenoptic sampling theory to derive a
bound that specifies precisely how densely users should sample views of a given
scene when using our algorithm. We achieve the perceptual quality of Nyquist
rate view sampling while using up to 4000x fewer views. Subsequent developments
have led to new scene representations for deep learning with view synthesis,
notably neural radiance fields, but the problem of sparse view synthesis from a
small number of images has only grown in importance. We reprise some of the
recent results on sparse and even single image view synthesis, while posing the
question of whether prescriptive sampling guidelines are feasible for the new
generation of image-based rendering algorithms.