FreqSpace-NeRF: A fourier-enhanced Neural Radiance Fields method via dual-domain contrastive learning for novel view synthesis

IF 2.8 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Computers & Graphics-Uk Pub Date : 2025-04-01 Epub Date: 2025-01-20 DOI:10.1016/j.cag.2025.104171
Xiaosheng Yu , Xiaolei Tian , Jubo Chen , Ying Wang
{"title":"FreqSpace-NeRF: A fourier-enhanced Neural Radiance Fields method via dual-domain contrastive learning for novel view synthesis","authors":"Xiaosheng Yu ,&nbsp;Xiaolei Tian ,&nbsp;Jubo Chen ,&nbsp;Ying Wang","doi":"10.1016/j.cag.2025.104171","DOIUrl":null,"url":null,"abstract":"<div><div>Inspired by Neural Radiance Field’s (NeRF) groundbreaking success in novel view synthesis, current methods mostly employ variants of various deep neural network architectures, and use the combination of multi-scale feature maps with the target viewpoint to synthesize novel views. However, these methods only consider spatial domain features, inevitably leading to the loss of some details and edge information. To address these issues, this paper innovatively proposes the FreqSpace-NeRF (FS-NeRF), aiming to significantly enhance the rendering fidelity and generalization performance of NeRF in complex scenes by integrating the unique advantages of spectral domain and spatial domain deep neural networks, and combining contrastive learning driven data augmentation techniques. Specifically, the core contribution of this method lies in designing a dual-stream network architecture: on one hand, capturing global frequency features through Fourier transformation; on the other hand, finely refining local details using well-established spatial domain convolutional neural networks. Moreover, to ensure the model can more acutely distinguish subtle differences between different views, we propose two loss functions: Frequency-Space Contrastive Entropy Loss (FSCE Loss) and Adaptive Spectral Contrastive Loss (ASC Loss). This innovation aims to more effectively guide the data flow and focuses on minimizing the frequency discrepancies between different views. By comprehensively utilizing the fusion of spectral and spatial domain features along with contrastive learning, FS-NeRF achieves significant performance improvements in scene reconstruction tasks. Extensive qualitative and quantitative evaluations confirm that our method surpasses current state-of-the-art (SOTA) models in this field.</div></div>","PeriodicalId":50628,"journal":{"name":"Computers & Graphics-Uk","volume":"127 ","pages":"Article 104171"},"PeriodicalIF":2.8000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Computers & Graphics-Uk","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S009784932500010X","RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/20 0:00:00","PubModel":"Epub","JCR":"Q2","JCRName":"COMPUTER SCIENCE, SOFTWARE ENGINEERING","Score":null,"Total":0}
引用次数: 0

Abstract

Inspired by Neural Radiance Field’s (NeRF) groundbreaking success in novel view synthesis, current methods mostly employ variants of various deep neural network architectures, and use the combination of multi-scale feature maps with the target viewpoint to synthesize novel views. However, these methods only consider spatial domain features, inevitably leading to the loss of some details and edge information. To address these issues, this paper innovatively proposes the FreqSpace-NeRF (FS-NeRF), aiming to significantly enhance the rendering fidelity and generalization performance of NeRF in complex scenes by integrating the unique advantages of spectral domain and spatial domain deep neural networks, and combining contrastive learning driven data augmentation techniques. Specifically, the core contribution of this method lies in designing a dual-stream network architecture: on one hand, capturing global frequency features through Fourier transformation; on the other hand, finely refining local details using well-established spatial domain convolutional neural networks. Moreover, to ensure the model can more acutely distinguish subtle differences between different views, we propose two loss functions: Frequency-Space Contrastive Entropy Loss (FSCE Loss) and Adaptive Spectral Contrastive Loss (ASC Loss). This innovation aims to more effectively guide the data flow and focuses on minimizing the frequency discrepancies between different views. By comprehensively utilizing the fusion of spectral and spatial domain features along with contrastive learning, FS-NeRF achieves significant performance improvements in scene reconstruction tasks. Extensive qualitative and quantitative evaluations confirm that our method surpasses current state-of-the-art (SOTA) models in this field.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
FreqSpace-NeRF:一种基于双域对比学习的傅里叶增强神经辐射场方法
受神经辐射场(Neural Radiance Field, NeRF)在新颖视图合成方面突破性成功的启发,目前的方法大多采用各种深度神经网络架构的变体,并将多尺度特征图与目标视点相结合来合成新颖视图。然而,这些方法只考虑空间域特征,不可避免地导致一些细节和边缘信息的丢失。针对这些问题,本文创新性地提出了FreqSpace-NeRF (FS-NeRF),旨在通过整合谱域和空间域深度神经网络的独特优势,结合对比学习驱动的数据增强技术,显著提高NeRF在复杂场景下的渲染保真度和泛化性能。具体来说,该方法的核心贡献在于设计了一种双流网络架构:一方面,通过傅里叶变换捕获全局频率特征;另一方面,使用完善的空间域卷积神经网络精细地细化局部细节。此外,为了确保模型能够更敏锐地区分不同视图之间的细微差异,我们提出了两个损失函数:频率-空间对比熵损失(FSCE loss)和自适应频谱对比损失(ASC loss)。这种创新旨在更有效地指导数据流,并专注于最小化不同视图之间的频率差异。FS-NeRF通过综合利用光谱和空间域特征的融合以及对比学习,在场景重建任务中实现了显著的性能提升。广泛的定性和定量评估证实,我们的方法在该领域超越了当前最先进的(SOTA)模型。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Computers & Graphics-Uk
Computers & Graphics-Uk 工程技术-计算机:软件工程
CiteScore
5.30
自引率
12.00%
发文量
173
审稿时长
38 days
期刊介绍: Computers & Graphics is dedicated to disseminate information on research and applications of computer graphics (CG) techniques. The journal encourages articles on: 1. Research and applications of interactive computer graphics. We are particularly interested in novel interaction techniques and applications of CG to problem domains. 2. State-of-the-art papers on late-breaking, cutting-edge research on CG. 3. Information on innovative uses of graphics principles and technologies. 4. Tutorial papers on both teaching CG principles and innovative uses of CG in education.
期刊最新文献
QuantizationGS: Differentiable quantization model for 3D Gaussian Splatting compression PS-GS: Gaussian splatting for multi-view photometric stereo LSIST-Net: A lightweight stereoscopic image style transfer network for view consistency A systematic review on throwing in virtual environments FC-Font: Few-shot diffusion-based font generation via frequency-domain style compensation
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1