UPST-NeRF: Universal Photorealistic Style Transfer of Neural Radiance Fields for 3D Scene

Yaosen Chen;Qi Yuan;Zhiqiang Li;Yuegen Liu;Wei Wang;Chaoping Xie;Xuming Wen;Qien Yu
{"title":"UPST-NeRF: Universal Photorealistic Style Transfer of Neural Radiance Fields for 3D Scene","authors":"Yaosen Chen;Qi Yuan;Zhiqiang Li;Yuegen Liu;Wei Wang;Chaoping Xie;Xuming Wen;Qien Yu","doi":"10.1109/TVCG.2024.3378692","DOIUrl":null,"url":null,"abstract":"Photorealistic stylization of 3D scenes aims to generate photorealistic images from arbitrary novel views according to a given style image, while ensuring consistency when rendering video from different viewpoints. Some existing stylization methods using neural radiance fields can effectively predict stylized scenes by combining the features of the style image with multi-view images to train 3D scenes. However, these methods generate novel view images that contain undesirable artifacts. In addition, they cannot achieve universal photorealistic stylization for a 3D scene. Therefore, a stylization image needs to retrain a 3D scene representation network based on a neural radiation field. We propose a novel photorealistic 3D scene stylization transfer framework to address these issues. It can realize photorealistic 3D scene style transfer with a 2D style image for novel view video rendering. We first pre-trained a 2D photorealistic style transfer network, which can satisfy the photorealistic style transfer between any content image and style image. Then, we use voxel features to optimize a 3D scene and obtain the geometric representation of the scene. Finally, we jointly optimize a hypernetwork to realize the photorealistic style transfer of arbitrary style images. In the transfer stage, we use a pre-trained 2D photorealistic network to constrain the photorealistic style of different views and different style images in the 3D scene. The experimental results show that our method not only realizes the 3D photorealistic style transfer of arbitrary style images, but also outperforms the existing methods in terms of visual quality and consistency.","PeriodicalId":94035,"journal":{"name":"IEEE transactions on visualization and computer graphics","volume":"31 4","pages":"2045-2057"},"PeriodicalIF":6.5000,"publicationDate":"2024-03-19","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on visualization and computer graphics","FirstCategoryId":"1085","ListUrlMain":"https://ieeexplore.ieee.org/document/10475593/","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Photorealistic stylization of 3D scenes aims to generate photorealistic images from arbitrary novel views according to a given style image, while ensuring consistency when rendering video from different viewpoints. Some existing stylization methods using neural radiance fields can effectively predict stylized scenes by combining the features of the style image with multi-view images to train 3D scenes. However, these methods generate novel view images that contain undesirable artifacts. In addition, they cannot achieve universal photorealistic stylization for a 3D scene. Therefore, a stylization image needs to retrain a 3D scene representation network based on a neural radiation field. We propose a novel photorealistic 3D scene stylization transfer framework to address these issues. It can realize photorealistic 3D scene style transfer with a 2D style image for novel view video rendering. We first pre-trained a 2D photorealistic style transfer network, which can satisfy the photorealistic style transfer between any content image and style image. Then, we use voxel features to optimize a 3D scene and obtain the geometric representation of the scene. Finally, we jointly optimize a hypernetwork to realize the photorealistic style transfer of arbitrary style images. In the transfer stage, we use a pre-trained 2D photorealistic network to constrain the photorealistic style of different views and different style images in the 3D scene. The experimental results show that our method not only realizes the 3D photorealistic style transfer of arbitrary style images, but also outperforms the existing methods in terms of visual quality and consistency.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
UPST-NeRF: 三维场景神经辐射场的通用逼真风格转移
逼真的3D场景风格化旨在根据给定的风格图像从任意新颖的视角生成逼真的图像,同时保证从不同视点渲染视频时的一致性。现有的一些基于神经辐射场的风格化方法,通过将风格化图像的特征与多视图图像相结合来训练3D场景,可以有效地预测风格化场景。然而,这些方法生成的新视图图像包含不需要的工件。此外,它们无法实现3D场景的通用逼真风格。因此,风格化图像需要重新训练基于神经辐射场的三维场景表示网络。为了解决这些问题,我们提出了一种新的逼真的3D场景风格化传输框架。它可以实现逼真的3D场景风格与2D风格图像的转换,实现新颖的视角视频渲染。我们首先预训练了一个2D逼真风格转移网络,该网络可以满足任意内容图像和风格图像之间的逼真风格转移。然后,利用体素特征对三维场景进行优化,得到场景的几何表示。最后,我们共同优化了一个超网络,实现了任意风格图像的逼真风格转换。在迁移阶段,我们使用预训练好的2D真实感网络来约束3D场景中不同视图和不同风格图像的真实感风格。实验结果表明,该方法不仅实现了任意风格图像的三维真实感风格转换,而且在视觉质量和一致性方面优于现有方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
HYVE: Hybrid Vertex Encoder for Neural Distance Fields. Errata to "DiffCap: Diffusion-Based Real-Time Human Motion Capture Using Sparse IMUs and a Monocular Camera". Crossing the Chasm: Bridging Visual Augmentations and Designer Intent. IMAGGarment: Fine-Grained Garment Generation for Controllable Fashion Design. Seeing Avatars during Social Interaction in VR May Enhance Inter-Brain Synchronization.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1