首页 > 最新文献

Computer Graphics Forum最新文献

英文 中文
MF-SDF: Neural Implicit Surface Reconstruction using Mixed Incident Illumination and Fourier Feature Optimization MF-SDF:基于混合入射照明和傅里叶特征优化的神经隐式曲面重建
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70244
Xueyang Zhou, Xukun Shen, Yong Hu

The utilization of neural implicit surface as a geometry representation has proven to be an effective multi-view surface reconstruction method. Despite the promising results achieved, reconstructing geometry from objects in real-world scenes remains challenging due to the interaction between surface materials and complex ambient light, as well as shadow effects caused by self-occlusion, making it a highly ill-posed problem. To address this challenge, we propose MF-SDF, a method that use a hybrid neural network and spherical gaussian representation to model environmental lighting, so that the model can express the situation of multiple light sources including directional light (such as outdoor sunlight) in real-world scenarios. Benefit from this, our method effectively reconstructs coherent surfaces and accurately locates the shadow location on the surface. Furthermore, we adopt a shadow aware multi-view photometric consistency loss, which mitigates the erroneous reconstruction results of previous methods on surfaces containing shadows, thereby improve the overall smoothness of the surface. Additionally, unlike previous approaches that directly optimize spatial features, we propose a Fourier feature optimization method that directly optimizes the tensorial feature in the frequency domain. By optimizing the high-frequency components, this approach further enhances the details of surface reconstruction. Finally, through experiments, we demonstrate that our method outperforms existing methods in terms of reconstruction accuracy on real captured data.

利用神经隐式曲面作为几何表示已被证明是一种有效的多视图曲面重建方法。尽管取得了令人鼓舞的结果,但由于表面材料和复杂环境光之间的相互作用,以及由自遮挡引起的阴影效应,在现实世界场景中重建物体的几何形状仍然具有挑战性,使其成为一个高度病态的问题。为了解决这一挑战,我们提出了MF-SDF,这是一种使用混合神经网络和球面高斯表示来模拟环境照明的方法,因此该模型可以表达现实场景中包括定向光(如室外阳光)在内的多个光源的情况。利用这一点,我们的方法有效地重建了相干表面,并准确地定位了表面上的阴影位置。此外,我们采用了阴影感知的多视点光度一致性损失,减轻了以往方法在含有阴影的表面上的错误重建结果,从而提高了表面的整体光滑度。此外,与以往直接优化空间特征的方法不同,我们提出了一种直接优化频域张量特征的傅里叶特征优化方法。该方法通过对高频分量的优化,进一步增强了表面重构的细节。最后,通过实验,我们证明了我们的方法在真实捕获数据的重建精度方面优于现有方法。
{"title":"MF-SDF: Neural Implicit Surface Reconstruction using Mixed Incident Illumination and Fourier Feature Optimization","authors":"Xueyang Zhou,&nbsp;Xukun Shen,&nbsp;Yong Hu","doi":"10.1111/cgf.70244","DOIUrl":"https://doi.org/10.1111/cgf.70244","url":null,"abstract":"<p>The utilization of neural implicit surface as a geometry representation has proven to be an effective multi-view surface reconstruction method. Despite the promising results achieved, reconstructing geometry from objects in real-world scenes remains challenging due to the interaction between surface materials and complex ambient light, as well as shadow effects caused by self-occlusion, making it a highly ill-posed problem. To address this challenge, we propose MF-SDF, a method that use a hybrid neural network and spherical gaussian representation to model environmental lighting, so that the model can express the situation of multiple light sources including directional light (such as outdoor sunlight) in real-world scenarios. Benefit from this, our method effectively reconstructs coherent surfaces and accurately locates the shadow location on the surface. Furthermore, we adopt a shadow aware multi-view photometric consistency loss, which mitigates the erroneous reconstruction results of previous methods on surfaces containing shadows, thereby improve the overall smoothness of the surface. Additionally, unlike previous approaches that directly optimize spatial features, we propose a Fourier feature optimization method that directly optimizes the tensorial feature in the frequency domain. By optimizing the high-frequency components, this approach further enhances the details of surface reconstruction. Finally, through experiments, we demonstrate that our method outperforms existing methods in terms of reconstruction accuracy on real captured data.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297345","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Swept Volume Computation with Enhanced Geometric Detail Preservation 增强几何细节保存的扫描体积计算
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70238
Pengfei Wang, Yuexin Yang, Shuangmin Chen, Shiqing Xin, Changhe Tu, Wenping Wang

Swept volume computation—the determination of regions occupied by moving objects—is essential in graphics, robotics, and manufacturing. Existing approaches either explicitly track surfaces, suffering from robustness issues under complex interactions, or employ implicit representations that trade off geometric fidelity and face optimization difficulties. We propose a novel inversion of motion perspective: rather than tracking object motion, we fix the object and trace spatial points backward in time, reducing complex trajectories to efficiently linearizable point motions. Based on this, we introduce a multi-field tetrahedral framework that maintains multiple distance fileds per element, preserving fine geometric details at trajectory intersections where single-field methods fail. Our method robustly computes swept volumes for diverse motions, including translations and screw motions, and enables practical applications in path planning and collision detection.

扫描体积计算——确定移动物体所占据的区域——在图形学、机器人和制造业中是必不可少的。现有的方法要么显式地跟踪表面,在复杂的相互作用下存在鲁棒性问题,要么采用隐式表示来权衡几何保真度并面临优化困难。我们提出了一种新的运动视角反转:我们不跟踪物体运动,而是固定物体并在时间上向后跟踪空间点,将复杂的轨迹简化为有效的线性点运动。在此基础上,我们引入了一种多场四面体框架,该框架可以维持每个元素的多个距离场,并在单场方法无法实现的轨迹交叉处保留精细的几何细节。我们的方法可以鲁棒地计算各种运动的扫描体积,包括平移和螺旋运动,并在路径规划和碰撞检测中实现实际应用。
{"title":"Swept Volume Computation with Enhanced Geometric Detail Preservation","authors":"Pengfei Wang,&nbsp;Yuexin Yang,&nbsp;Shuangmin Chen,&nbsp;Shiqing Xin,&nbsp;Changhe Tu,&nbsp;Wenping Wang","doi":"10.1111/cgf.70238","DOIUrl":"https://doi.org/10.1111/cgf.70238","url":null,"abstract":"<p>Swept volume computation—the determination of regions occupied by moving objects—is essential in graphics, robotics, and manufacturing. Existing approaches either explicitly track surfaces, suffering from robustness issues under complex interactions, or employ implicit representations that trade off geometric fidelity and face optimization difficulties. We propose a novel inversion of motion perspective: rather than tracking object motion, we fix the object and trace spatial points backward in time, reducing complex trajectories to efficiently linearizable point motions. Based on this, we introduce a multi-field tetrahedral framework that maintains multiple distance fileds per element, preserving fine geometric details at trajectory intersections where single-field methods fail. Our method robustly computes swept volumes for diverse motions, including translations and screw motions, and enables practical applications in path planning and collision detection.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
SPG: Style-Prompting Guidance for Style-Specific Content Creation SPG:特定样式内容创建的样式提示指南
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70251
Qian Liang, Zichong Chen, Yang Zhou, Hui Huang

Although recent text-to-image (T2I) diffusion models excel at aligning generated images with textual prompts, controlling the visual style of the output remains a challenging task. In this work, we propose Style-Prompting Guidance (SPG), a novel sampling strategy for style-specific image generation. SPG constructs a style noise vector and leverages its directional deviation from unconditional noise to guide the diffusion process toward the target style distribution. By integrating SPG with Classifier-Free Guidance (CFG), our method achieves both semantic fidelity and style consistency. SPG is simple, robust, and compatible with controllable frameworks like ControlNet and IPAdapter, making it practical and widely applicable. Extensive experiments demonstrate the effectiveness and generality of our approach compared to state-of-the-art methods. Code is available at https://github.com/Rumbling281441/SPG.

尽管最近的文本到图像(tt2i)扩散模型擅长将生成的图像与文本提示对齐,但控制输出的视觉风格仍然是一项具有挑战性的任务。在这项工作中,我们提出了风格提示指导(SPG),这是一种用于特定风格图像生成的新颖采样策略。SPG构造一个风格噪声向量,并利用其与无条件噪声的方向性偏差来引导扩散过程向目标风格分布方向发展。通过将SPG与无分类器制导(CFG)相结合,我们的方法实现了语义保真度和风格一致性。SPG简单、健壮,并与ControlNet和IPAdapter等可控框架兼容,具有实用性和广泛的适用性。与最先进的方法相比,大量的实验证明了我们的方法的有效性和普遍性。代码可从https://github.com/Rumbling281441/SPG获得。
{"title":"SPG: Style-Prompting Guidance for Style-Specific Content Creation","authors":"Qian Liang,&nbsp;Zichong Chen,&nbsp;Yang Zhou,&nbsp;Hui Huang","doi":"10.1111/cgf.70251","DOIUrl":"https://doi.org/10.1111/cgf.70251","url":null,"abstract":"<p>Although recent text-to-image (T2I) diffusion models excel at aligning generated images with textual prompts, controlling the visual style of the output remains a challenging task. In this work, we propose Style-Prompting Guidance (SPG), a novel sampling strategy for style-specific image generation. SPG constructs a style noise vector and leverages its directional deviation from unconditional noise to guide the diffusion process toward the target style distribution. By integrating SPG with Classifier-Free Guidance (CFG), our method achieves both semantic fidelity and style consistency. SPG is simple, robust, and compatible with controllable frameworks like ControlNet and IPAdapter, making it practical and widely applicable. Extensive experiments demonstrate the effectiveness and generality of our approach compared to state-of-the-art methods. Code is available at https://github.com/Rumbling281441/SPG.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gaussians on their Way: Wasserstein-Constrained 4D Gaussian Splatting with State-Space Modeling 高斯在他们的道路上:wasserstein约束的四维高斯溅射与状态空间建模
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70271
J. Deng, P. Shi, Y. Luo, Q. Li

Dynamic scene rendering has taken a leap forward with the rise of 4D Gaussian Splatting, but there is still one elusive challenge: how to make 3D Gaussians move through time as naturally as they would in the real world, all while keeping the motion smooth and consistent. In this paper, we present an approach that blends state-space modeling with Wasserstein geometry, enabling a more fluid and coherent representation of dynamic scenes. We introduce a State Consistency Filter that merges prior predictions with the current observations, enabling Gaussians to maintain coherent trajectories over time. We also employ Wasserstein Consistency Constraint to ensure smooth, consistent updates of Gaussian parameters, reducing motion artifacts. Lastly, we leverage Wasserstein geometry to capture both translational motion and shape deformations, creating a more geometrically consistent model for dynamic scenes. Our approach models the evolution of Gaussians along geodesics on the manifold of Gaussian distributions, achieving smoother, more realistic motion and stronger temporal coherence. Experimental results show consistent improvements in rendering quality and efficiency.

(see https://www.acm.org/publications/class-2012)

动态场景渲染已经采取了一个飞跃与四维高斯飞溅的兴起,但仍然有一个难以捉摸的挑战:如何使3D高斯移动通过时间自然,因为他们会在现实世界中,同时保持运动的平滑和一致。在本文中,我们提出了一种混合状态空间建模和沃瑟斯坦几何的方法,使动态场景的表示更加流畅和连贯。我们引入了一个状态一致性过滤器,将先前的预测与当前的观测合并,使高斯函数能够随着时间的推移保持一致的轨迹。我们还使用Wasserstein一致性约束来确保高斯参数的平滑一致更新,减少运动伪像。最后,我们利用Wasserstein几何来捕获平移运动和形状变形,为动态场景创建更几何一致的模型。我们的方法在高斯分布的流形上模拟高斯分布沿着测地线的演化,实现更平滑、更真实的运动和更强的时间相干性。实验结果表明,改进后的渲染质量和效率得到了一致的提高。(见https://www.acm.org/publications/class - 2012)
{"title":"Gaussians on their Way: Wasserstein-Constrained 4D Gaussian Splatting with State-Space Modeling","authors":"J. Deng,&nbsp;P. Shi,&nbsp;Y. Luo,&nbsp;Q. Li","doi":"10.1111/cgf.70271","DOIUrl":"https://doi.org/10.1111/cgf.70271","url":null,"abstract":"<p>Dynamic scene rendering has taken a leap forward with the rise of 4D Gaussian Splatting, but there is still one elusive challenge: how to make 3D Gaussians move through time as naturally as they would in the real world, all while keeping the motion smooth and consistent. In this paper, we present an approach that blends state-space modeling with Wasserstein geometry, enabling a more fluid and coherent representation of dynamic scenes. We introduce a State Consistency Filter that merges prior predictions with the current observations, enabling Gaussians to maintain coherent trajectories over time. We also employ Wasserstein Consistency Constraint to ensure smooth, consistent updates of Gaussian parameters, reducing motion artifacts. Lastly, we leverage Wasserstein geometry to capture both translational motion and shape deformations, creating a more geometrically consistent model for dynamic scenes. Our approach models the evolution of Gaussians along geodesics on the manifold of Gaussian distributions, achieving smoother, more realistic motion and stronger temporal coherence. Experimental results show consistent improvements in rendering quality and efficiency.</p><p>(see https://www.acm.org/publications/class-2012)</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297026","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
LucidFusion: Reconstructing 3D Gaussians with Arbitrary Unposed Images LucidFusion:重建三维高斯任意未置图像
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70227
Hao He, Yixun Liang, Luozhou Wang, Yuanhao Cai, Xinli Xu, Haoxiang Guo, Xiang Wen, Yingcong Chen

Recent large reconstruction models have made notable progress in generating high-quality 3D objects from single images. However, current reconstruction methods often rely on explicit camera pose estimation or fixed viewpoints, restricting their flexibility and practical applicability. We reformulate 3D reconstruction as image-to-image translation and introduce the Relative Coordinate Map (RCM), which aligns multiple unposed images to a “main” view without pose estimation. While RCM simplifies the process, its lack of global 3D supervision can yield noisy outputs. To address this, we propose Relative Coordinate Gaussians (RCG) as an extension to RCM, which treats each pixel's coordinates as a Gaussian center and employs differentiable rasterization for consistent geometry and pose recovery. Our LucidFusion framework handles an arbitrary number of unposed inputs, producing robust 3D reconstructions within seconds and paving the way for more flexible, pose-free 3D pipelines.

最近的大型重建模型在从单个图像生成高质量3D物体方面取得了显著进展。然而,目前的重建方法往往依赖于明确的相机姿态估计或固定视点,限制了其灵活性和实用性。我们将3D重建重新表述为图像到图像的转换,并引入了相对坐标图(RCM),它将多个未放置的图像对齐到“主”视图,而不需要姿态估计。虽然RCM简化了过程,但它缺乏全局3D监督,可能会产生噪声输出。为了解决这个问题,我们提出了相对坐标高斯(RCG)作为RCM的扩展,它将每个像素的坐标作为高斯中心,并采用可微光栅化来实现一致的几何形状和姿态恢复。我们的LucidFusion框架处理任意数量的未放置的输入,在几秒钟内产生强大的3D重建,并为更灵活,无姿势的3D管道铺平道路。
{"title":"LucidFusion: Reconstructing 3D Gaussians with Arbitrary Unposed Images","authors":"Hao He,&nbsp;Yixun Liang,&nbsp;Luozhou Wang,&nbsp;Yuanhao Cai,&nbsp;Xinli Xu,&nbsp;Haoxiang Guo,&nbsp;Xiang Wen,&nbsp;Yingcong Chen","doi":"10.1111/cgf.70227","DOIUrl":"https://doi.org/10.1111/cgf.70227","url":null,"abstract":"<p>Recent large reconstruction models have made notable progress in generating high-quality 3D objects from single images. However, current reconstruction methods often rely on explicit camera pose estimation or fixed viewpoints, restricting their flexibility and practical applicability. We reformulate 3D reconstruction as image-to-image translation and introduce the Relative Coordinate Map (RCM), which aligns multiple unposed images to a “main” view without pose estimation. While RCM simplifies the process, its lack of global 3D supervision can yield noisy outputs. To address this, we propose Relative Coordinate Gaussians (RCG) as an extension to RCM, which treats each pixel's coordinates as a Gaussian center and employs differentiable rasterization for consistent geometry and pose recovery. Our LucidFusion framework handles an arbitrary number of unposed inputs, producing robust 3D reconstructions within seconds and paving the way for more flexible, pose-free 3D pipelines.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FAHNet: Accurate and Robust Normal Estimation for Point Clouds via Frequency-Aware Hierarchical Geometry 基于频率感知层次几何的点云准确鲁棒正态估计
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70264
Chengwei Wang, Wenming Wu, Yue Fei, Gaofeng Zhang, Liping Zheng

Point cloud normal estimation underpins many 3D vision and graphics applications. Precise normal estimation in regions of sharp curvature and high-frequency variation remains a major bottleneck; existing learning-based methods still struggle to isolate fine geometry details under noise and uneven sampling. We present FAHNet, a novel frequency-aware hierarchical network that precisely tackles those challenges. Our Frequency-Aware Hierarchical Geometry (FAHG) feature extraction module selectively amplifies and merges cross-scale cues, ensuring that both fine-grained local features and sharp structures are faithfully represented. Crucially, a dedicated Frequency-Aware geometry enhancement (FA) branch intensifies sensitivity to abrupt normal transitions and sharp features, preventing the common over-smoothing limitation. Extensive experiments on synthetic benchmarks (PCPNet, FamousShape) and real-world scans (SceneNN) demonstrate that FAHNet outperforms state-of-the-art approaches in normal estimation accuracy. Ablation studies further quantify the contribution of each component, and downstream surface reconstruction results validate the practical impact of our design.

点云法向估计是许多3D视觉和图形应用的基础。在急剧曲率和高频变化区域精确的正态估计仍然是一个主要的瓶颈;现有的基于学习的方法仍然难以在噪声和不均匀采样的情况下分离出精细的几何细节。我们提出了FAHNet,一种新颖的频率感知分层网络,可以精确地解决这些挑战。我们的频率感知分层几何(FAHG)特征提取模块选择性地放大和合并跨尺度线索,确保细粒度的局部特征和尖锐结构都得到忠实的表示。至关重要的是,一个专用的频率感知几何增强(FA)分支增强了对突然正常过渡和尖锐特征的敏感性,防止了常见的过度平滑限制。在合成基准测试(PCPNet、FamousShape)和真实世界扫描(SceneNN)上进行的大量实验表明,FAHNet在正常估计精度方面优于最先进的方法。消融研究进一步量化了每个组件的贡献,下游表面重建结果验证了我们设计的实际影响。
{"title":"FAHNet: Accurate and Robust Normal Estimation for Point Clouds via Frequency-Aware Hierarchical Geometry","authors":"Chengwei Wang,&nbsp;Wenming Wu,&nbsp;Yue Fei,&nbsp;Gaofeng Zhang,&nbsp;Liping Zheng","doi":"10.1111/cgf.70264","DOIUrl":"https://doi.org/10.1111/cgf.70264","url":null,"abstract":"<p>Point cloud normal estimation underpins many 3D vision and graphics applications. Precise normal estimation in regions of sharp curvature and high-frequency variation remains a major bottleneck; existing learning-based methods still struggle to isolate fine geometry details under noise and uneven sampling. We present FAHNet, a novel frequency-aware hierarchical network that precisely tackles those challenges. Our Frequency-Aware Hierarchical Geometry (FAHG) feature extraction module selectively amplifies and merges cross-scale cues, ensuring that both fine-grained local features and sharp structures are faithfully represented. Crucially, a dedicated Frequency-Aware geometry enhancement (FA) branch intensifies sensitivity to abrupt normal transitions and sharp features, preventing the common over-smoothing limitation. Extensive experiments on synthetic benchmarks (PCPNet, FamousShape) and real-world scans (SceneNN) demonstrate that FAHNet outperforms state-of-the-art approaches in normal estimation accuracy. Ablation studies further quantify the contribution of each component, and downstream surface reconstruction results validate the practical impact of our design.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297126","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Uncertainty-Aware Adjustment via Learnable Coefficients for Detailed 3D Reconstruction of Clothed Humans from Single Images 基于可学习系数的不确定性感知调整在单幅图像中对穿衣服的人进行详细的三维重建
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70239
Yadan Yang, Yunze Li, Fangli Ying, Aniwat Phaphuangwittayakul, Riyad Dhuny

Although single-image 3D human reconstruction has made significant progress in recent years, few of the current state-of-the-art methods can accurately restore the appearance and geometric details of loose clothing. To achieve high-quality reconstruction of a human body wearing loose clothing, we propose a learnable dynamic adjustment framework that integrates side-view features and the uncertainty of the parametric human body model to adaptively regulate its reliability based on the clothing type. Specifically, we first adopt the Vision Transformer model as an encoder to capture the image features of the input image, and then employ SMPL-X to decouple the side-view body features. Secondly, to reduce the limitations imposed by the regularization of the parametric model, particularly for loose garments, we introduce a learnable coefficient to reduce the reliance on SMPL-X. This strategy effectively accommodates the large deformations caused by loose clothing, thereby accurately expressing the posture and clothing in the image. To evaluate the effectiveness, we validate our method on the public CLOTH4D and Cape datasets, and the experimental results demonstrate better performance compared to existing approaches. The code is available at https://github.com/yyd0613/CoRe-Human.

尽管近年来单图像三维人体重建取得了重大进展,但目前最先进的方法很少能准确地恢复宽松服装的外观和几何细节。为了实现宽松服装人体的高质量重建,我们提出了一种可学习的动态调整框架,该框架将侧视图特征和参数化人体模型的不确定性相结合,根据服装类型自适应调节其可靠性。具体而言,我们首先采用Vision Transformer模型作为编码器捕获输入图像的图像特征,然后使用SMPL-X对侧视体特征进行解耦。其次,为了减少参数化模型的正则化所带来的限制,特别是对于宽松的服装,我们引入了一个可学习系数来减少对SMPL-X的依赖。这一策略有效地容纳了由于服装宽松造成的较大变形,从而准确地表达了图像中的姿势和服装。为了评估该方法的有效性,我们在CLOTH4D和Cape公共数据集上进行了验证,实验结果表明,与现有方法相比,该方法具有更好的性能。代码可在https://github.com/yyd0613/CoRe-Human上获得。
{"title":"Uncertainty-Aware Adjustment via Learnable Coefficients for Detailed 3D Reconstruction of Clothed Humans from Single Images","authors":"Yadan Yang,&nbsp;Yunze Li,&nbsp;Fangli Ying,&nbsp;Aniwat Phaphuangwittayakul,&nbsp;Riyad Dhuny","doi":"10.1111/cgf.70239","DOIUrl":"https://doi.org/10.1111/cgf.70239","url":null,"abstract":"<p>Although single-image 3D human reconstruction has made significant progress in recent years, few of the current state-of-the-art methods can accurately restore the appearance and geometric details of loose clothing. To achieve high-quality reconstruction of a human body wearing loose clothing, we propose a learnable dynamic adjustment framework that integrates side-view features and the uncertainty of the parametric human body model to adaptively regulate its reliability based on the clothing type. Specifically, we first adopt the Vision Transformer model as an encoder to capture the image features of the input image, and then employ SMPL-X to decouple the side-view body features. Secondly, to reduce the limitations imposed by the regularization of the parametric model, particularly for loose garments, we introduce a learnable coefficient to reduce the reliance on SMPL-X. This strategy effectively accommodates the large deformations caused by loose clothing, thereby accurately expressing the posture and clothing in the image. To evaluate the effectiveness, we validate our method on the public CLOTH4D and Cape datasets, and the experimental results demonstrate better performance compared to existing approaches. The code is available at https://github.com/yyd0613/CoRe-Human.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297322","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Gaussian Splatting for Large-Scale Aerial Scene Reconstruction From Ultra-High-Resolution Images 高斯溅射用于超高分辨率图像的大规模航拍场景重建
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70265
Qiulin Sun, Wei Lai, Yixian Li, Yanci Zhang

Using 3D Gaussian splatting to reconstruct large-scale aerial scenes from ultra-high-resolution images is still a challenge problem because of two memory bottlenecks - excessive Gaussian primitives and the tensor sizes for ultra-high-resolution images. In this paper, we propose a task partitioning algorithm that operates in both object and image space to generate a set of small-scale subtasks. Each subtask's memory footprints is strictly limited, enabling training on a single high-end consumer-grade GPU. More specifically, Gaussian primitives are clustered into blocks in object space, and the input images are partitioned into sub-images according to the projected footprints of these blocks. This dual-space partitioning significantly reduces training memory requirements. During subtask training, we propose a depth comparison method to generate a mask map for each sub-image. This mask map isolates pixels primarily contributed by the Gaussian primitives of the current subtask, excluding all other pixels from training. Experimental results demonstrate that our method successfully achieves large-scale aerial scene reconstruction using 9K resolution images on a single RTX 4090 GPU. The novel views synthesized by our method retain significantly more details than those from current state-of-the-art methods.

利用三维高斯溅射从超高分辨率图像中重建大规模航拍场景仍然是一个具有挑战性的问题,因为存在两个内存瓶颈-高斯基元过多和超高分辨率图像的张量大小。在本文中,我们提出了一种任务划分算法,该算法同时在物体和图像空间中运行,以生成一组小规模的子任务。每个子任务的内存占用都受到严格限制,因此可以在单个高端消费级GPU上进行训练。更具体地说,高斯原语在目标空间中聚类成块,并根据这些块的投影足迹将输入图像划分为子图像。这种双空间分区显著降低了训练内存需求。在子任务训练中,我们提出了一种深度比较的方法来生成每个子图像的掩码映射。这个掩码映射隔离了主要由当前子任务的高斯原语贡献的像素,排除了所有其他像素的训练。实验结果表明,该方法在单个RTX 4090 GPU上成功实现了使用9K分辨率图像的大规模航拍场景重建。用我们的方法合成的新视图比目前最先进的方法保留了更多的细节。
{"title":"Gaussian Splatting for Large-Scale Aerial Scene Reconstruction From Ultra-High-Resolution Images","authors":"Qiulin Sun,&nbsp;Wei Lai,&nbsp;Yixian Li,&nbsp;Yanci Zhang","doi":"10.1111/cgf.70265","DOIUrl":"https://doi.org/10.1111/cgf.70265","url":null,"abstract":"<p>Using 3D Gaussian splatting to reconstruct large-scale aerial scenes from ultra-high-resolution images is still a challenge problem because of two memory bottlenecks - excessive Gaussian primitives and the tensor sizes for ultra-high-resolution images. In this paper, we propose a task partitioning algorithm that operates in both object and image space to generate a set of small-scale subtasks. Each subtask's memory footprints is strictly limited, enabling training on a single high-end consumer-grade GPU. More specifically, Gaussian primitives are clustered into blocks in object space, and the input images are partitioned into sub-images according to the projected footprints of these blocks. This dual-space partitioning significantly reduces training memory requirements. During subtask training, we propose a depth comparison method to generate a mask map for each sub-image. This mask map isolates pixels primarily contributed by the Gaussian primitives of the current subtask, excluding all other pixels from training. Experimental results demonstrate that our method successfully achieves large-scale aerial scene reconstruction using 9K resolution images on a single RTX 4090 GPU. The novel views synthesized by our method retain significantly more details than those from current state-of-the-art methods.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Self-Supervised Humidity-Controllable Garment Simulation via Capillary Bridge Modeling 基于毛细管桥模型的自监督湿度控制服装仿真
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70236
M. Shi, X. Wang, J. Zhang, L. Gao, D. Zhu, H. Zhang

Simulating wet clothing remains a significant challenge due to the complex physical interactions between moist fabric and the human body, compounded by the lack of dedicated datasets for training data-driven models. Existing self-supervised approaches struggle to capture moisture-induced dynamics such as skin adhesion, anisotropic surface resistance, and non-linear wrinkling, leading to limited accuracy and efficiency. To address this, we present SHGS, a novel self-supervised framework for humidity-controllable clothing simulation grounded in the physical modeling of capillary bridges that form between fabric and skin. We abstract the forces induced by wetness into two physically motivated components: a normal adhesive force derived from Laplace pressure and a tangential shear-resistance force that opposes relative motion along the fabric surface. By formulating these forces as potential energy for conservative effects and as mechanical work for non-conservative effects, we construct a physics-consistent wetness loss. This enables self-supervised training without requiring labeled data of wet clothing. Our humidity-sensitive dynamics are driven by a multi-layer graph neural network, which facilitates a smooth and physically realistic transition between different moisture levels. This architecture decouples the garment's dynamics in wet and dry states through a local weight interpolation mechanism, adjusting the fabric's behavior in response to varying humidity conditions. Experiments demonstrate that SHGS outperforms existing methods in both visual fidelity and computational efficiency, marking a significant advancement in realistic wet-cloth simulation.

由于潮湿织物与人体之间复杂的物理相互作用,再加上缺乏训练数据驱动模型的专用数据集,模拟湿衣服仍然是一个重大挑战。现有的自监督方法难以捕捉水分引起的动态,如皮肤粘附、各向异性表面阻力和非线性起皱,导致精度和效率有限。为了解决这个问题,我们提出了SHGS,这是一种基于织物和皮肤之间形成的毛细血管桥的物理建模的湿度可控服装模拟的新型自监督框架。我们将由湿度引起的力抽象为两个物理驱动的分量:由拉普拉斯压力导出的法向附着力和反对沿织物表面相对运动的切向剪切阻力。通过将这些力表述为保守效应的势能和非保守效应的机械功,我们构建了一个物理一致的湿损失。这使得自我监督训练不需要湿衣服的标签数据。我们的湿度敏感动态是由多层图形神经网络驱动的,它促进了不同湿度水平之间的平滑和物理上真实的过渡。这种结构通过局部重量插值机制来解耦服装在干湿状态下的动态,根据不同的湿度条件调整织物的行为。实验表明,SHGS在视觉保真度和计算效率方面都优于现有方法,标志着现实湿布模拟的重大进步。
{"title":"Self-Supervised Humidity-Controllable Garment Simulation via Capillary Bridge Modeling","authors":"M. Shi,&nbsp;X. Wang,&nbsp;J. Zhang,&nbsp;L. Gao,&nbsp;D. Zhu,&nbsp;H. Zhang","doi":"10.1111/cgf.70236","DOIUrl":"https://doi.org/10.1111/cgf.70236","url":null,"abstract":"<p>Simulating wet clothing remains a significant challenge due to the complex physical interactions between moist fabric and the human body, compounded by the lack of dedicated datasets for training data-driven models. Existing self-supervised approaches struggle to capture moisture-induced dynamics such as skin adhesion, anisotropic surface resistance, and non-linear wrinkling, leading to limited accuracy and efficiency. To address this, we present SHGS, a novel self-supervised framework for humidity-controllable clothing simulation grounded in the physical modeling of capillary bridges that form between fabric and skin. We abstract the forces induced by wetness into two physically motivated components: a normal adhesive force derived from Laplace pressure and a tangential shear-resistance force that opposes relative motion along the fabric surface. By formulating these forces as potential energy for conservative effects and as mechanical work for non-conservative effects, we construct a physics-consistent wetness loss. This enables self-supervised training without requiring labeled data of wet clothing. Our humidity-sensitive dynamics are driven by a multi-layer graph neural network, which facilitates a smooth and physically realistic transition between different moisture levels. This architecture decouples the garment's dynamics in wet and dry states through a local weight interpolation mechanism, adjusting the fabric's behavior in response to varying humidity conditions. Experiments demonstrate that SHGS outperforms existing methods in both visual fidelity and computational efficiency, marking a significant advancement in realistic wet-cloth simulation.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297028","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FlatCAD: Fast Curvature Regularization of Neural SDFs for CAD Models 平面CAD: CAD模型神经sdf的快速曲率正则化
IF 2.9 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-10-11 DOI: 10.1111/cgf.70249
Haotian Yin, Aleksander Plocharski, Michal Jan Wlodarczyk, Mikolaj Kida, Przemyslaw Musialski

Neural signed-distance fields (SDFs) are a versatile backbone for neural geometry representation, but enforcing CAD-style developability usually requires Gaussian-curvature penalties with full Hessian evaluation and second-order differentiation, which are costly in memory and time. We introduce an off-diagonal Weingarten loss that regularizes only the mixed shape operator term that represents the gap between principal curvatures and flattens the surface. We present two variants: a finite-difference version using six SDF evaluations plus one gradient, and an auto-diff version using a single Hessian-vector product. Both converge to the exact mixed term and preserve the intended geometric properties without assembling the full Hessian. On the ABC benchmarks the losses match or exceed Hessian-based baselines while cutting GPU memory and training time by roughly a factor of two. The method is drop-in and framework-agnostic, enabling scalable curvature-aware SDF learning for engineering-grade shape reconstruction. Our code is available at https://flatcad.github.io/.

神经符号距离场(sdf)是神经几何表示的通用主干,但是实现cad风格的可开发性通常需要高斯曲率惩罚和完全的Hessian求值和二阶微分,这在内存和时间上都很昂贵。我们引入了一个非对角线Weingarten损失,它只正则化了表示主曲率之间的间隙和表面平坦的混合形状算子项。我们提出了两个变体:使用六个SDF评估加一个梯度的有限差分版本,以及使用单个Hessian-vector积的自动差分版本。两者都收敛到精确的混合项,并保留了预期的几何性质,而无需组装完整的黑森函数。在ABC基准测试中,损失达到或超过了基于hessian的基准,同时将GPU内存和训练时间减少了大约两倍。该方法是插入式的,与框架无关,可以为工程级形状重建提供可扩展的曲率感知SDF学习。我们的代码可在https://flatcad.github.io/上获得。
{"title":"FlatCAD: Fast Curvature Regularization of Neural SDFs for CAD Models","authors":"Haotian Yin,&nbsp;Aleksander Plocharski,&nbsp;Michal Jan Wlodarczyk,&nbsp;Mikolaj Kida,&nbsp;Przemyslaw Musialski","doi":"10.1111/cgf.70249","DOIUrl":"https://doi.org/10.1111/cgf.70249","url":null,"abstract":"<p>Neural signed-distance fields (SDFs) are a versatile backbone for neural geometry representation, but enforcing CAD-style developability usually requires Gaussian-curvature penalties with full Hessian evaluation and second-order differentiation, which are costly in memory and time. We introduce an off-diagonal Weingarten loss that regularizes only the mixed shape operator term that represents the gap between principal curvatures and flattens the surface. We present two variants: a finite-difference version using six SDF evaluations plus one gradient, and an auto-diff version using a single Hessian-vector product. Both converge to the exact mixed term and preserve the intended geometric properties without assembling the full Hessian. On the ABC benchmarks the losses match or exceed Hessian-based baselines while cutting GPU memory and training time by roughly a factor of two. The method is drop-in and framework-agnostic, enabling scalable curvature-aware SDF learning for engineering-grade shape reconstruction. Our code is available at https://flatcad.github.io/.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Graphics Forum
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1