Jiangbei Hu, Ben Fei, Baixin Xu, Fei Hou, Shengfa Wang, Na Lei, Weidong Yang, Chen Qian, Ying He
Topological properties play a crucial role in the analysis, reconstruction, and generation of 3D shapes. Yet, most existing research focuses primarily on geometric features, due to the lack of effective representations for topology. In this paper, we introduce TopoGen, a method that extracts both discrete and continuous topological descriptors–Betti numbers and persistence points–using persistent homology. These features provide robust characterizations of 3D shapes in terms of their topology. We incorporate them as conditional guidance in generative models for 3D shape synthesis, enabling topology-aware generation from diverse inputs such as sparse and partial point clouds, as well as sketches. Furthermore, by modifying persistence points, we can explicitly control and alter the topology of generated shapes. Experimental results demonstrate that TopoGen enhances both diversity and controllability in 3D generation by embedding global topological structure into the synthesis process.
{"title":"TopoGen: Topology-Aware 3D Generation with Persistence Points","authors":"Jiangbei Hu, Ben Fei, Baixin Xu, Fei Hou, Shengfa Wang, Na Lei, Weidong Yang, Chen Qian, Ying He","doi":"10.1111/cgf.70257","DOIUrl":"https://doi.org/10.1111/cgf.70257","url":null,"abstract":"<p>Topological properties play a crucial role in the analysis, reconstruction, and generation of 3D shapes. Yet, most existing research focuses primarily on geometric features, due to the lack of effective representations for topology. In this paper, we introduce <i>TopoGen</i>, a method that extracts both discrete and continuous topological descriptors–Betti numbers and persistence points–using persistent homology. These features provide robust characterizations of 3D shapes in terms of their topology. We incorporate them as conditional guidance in generative models for 3D shape synthesis, enabling topology-aware generation from diverse inputs such as sparse and partial point clouds, as well as sketches. Furthermore, by modifying persistence points, we can explicitly control and alter the topology of generated shapes. Experimental results demonstrate that TopoGen enhances both diversity and controllability in 3D generation by embedding global topological structure into the synthesis process.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297024","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Recent deep learning methods can generate diverse graphic design layouts efficiently. However, these methods often create layouts with flaws, such as misalignment, unwanted overlaps, and unsatisfied containment. To tackle this issue, we propose an optimization-based method called LayoutRectifier, which gracefully rectifies auto-generated graphic design layouts to reduce these flaws while minimizing deviation from the generated layout. The core of our method is a two-stage optimization. First, we utilize grid systems, which professional designers commonly use to organize elements, to mitigate misalignments through discrete search. Second, we introduce a novel box containment function designed to adjust the positions and sizes of the layout elements, preventing unwanted overlapping and promoting desired containment. We evaluate our method on content-agnostic and content-aware layout generation tasks and achieve better-quality layouts that are more suitable for downstream graphic design tasks. Our method complements learning-based layout generation methods and does not require additional training.
{"title":"LayoutRectifier: An Optimization-based Post-processing for Graphic Design Layout Generation","authors":"I-Chao Shen, Ariel Shamir, Takeo Igarashi","doi":"10.1111/cgf.70273","DOIUrl":"https://doi.org/10.1111/cgf.70273","url":null,"abstract":"<p>Recent deep learning methods can generate diverse graphic design layouts efficiently. However, these methods often create layouts with flaws, such as misalignment, unwanted overlaps, and unsatisfied containment. To tackle this issue, we propose an optimization-based method called LayoutRectifier, which gracefully rectifies auto-generated graphic design layouts to reduce these flaws while minimizing deviation from the generated layout. The core of our method is a two-stage optimization. First, we utilize grid systems, which professional designers commonly use to organize elements, to mitigate misalignments through discrete search. Second, we introduce a novel box containment function designed to adjust the positions and sizes of the layout elements, preventing unwanted overlapping and promoting desired containment. We evaluate our method on content-agnostic and content-aware layout generation tasks and achieve better-quality layouts that are more suitable for downstream graphic design tasks. Our method complements learning-based layout generation methods and does not require additional training.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.70273","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297130","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Point cloud completion is vital for accurate 3D reconstruction, yet real world scans frequently exhibit large structural gaps that compromise recovery. Meanwhile, in 2D vision, VAR (Visual Auto-Regression) has demonstrated that a coarse-to-fine “next-scale prediction” can significantly improve generation quality, inference speed, and generalization. Because this coarse-to-fine approach closely aligns with the progressive nature of filling missing geometry in point clouds, we were inspired to develop PARC (Patch-Aware Coarse-to-Fine Refinement Completion), a two-stage multimodal framework specifically designed for handling missing structures. In the pretraining stage, PARC leverages complete point clouds alongside a Patch-Aware Coarse-to-Fine Refinement (PAR) strategy and a Mixture-of-Experts (MoE) architecture to generate high-quality local fragments, thereby improving geometric structure understanding and feature representation quality. During finetuning, the model is adapted to partial scans, further enhancing its resilience to incomplete inputs. To address remaining uncertainties in areas with missing structure, we introduce a dual-branch architecture that incorporates image cues: point cloud and image features are extracted independently and then fused via the MoE with an alignment loss, allowing complementary modalities to guide reconstruction in occluded or missing regions. Experiments conducted on the ShapeNet-ViPC dataset show that PARC has achieved highly competitive performance. Code is available at https://github.com/caiyujiaocyj/PARC.
{"title":"PARC: A Two-Stage Multi-Modal Framework for Point Cloud Completion","authors":"Yujiao Cai, Yuhao Su","doi":"10.1111/cgf.70266","DOIUrl":"https://doi.org/10.1111/cgf.70266","url":null,"abstract":"<p>Point cloud completion is vital for accurate 3D reconstruction, yet real world scans frequently exhibit large structural gaps that compromise recovery. Meanwhile, in 2D vision, VAR (Visual Auto-Regression) has demonstrated that a coarse-to-fine “next-scale prediction” can significantly improve generation quality, inference speed, and generalization. Because this coarse-to-fine approach closely aligns with the progressive nature of filling missing geometry in point clouds, we were inspired to develop PARC (Patch-Aware Coarse-to-Fine Refinement Completion), a two-stage multimodal framework specifically designed for handling missing structures. In the pretraining stage, PARC leverages complete point clouds alongside a Patch-Aware Coarse-to-Fine Refinement (PAR) strategy and a Mixture-of-Experts (MoE) architecture to generate high-quality local fragments, thereby improving geometric structure understanding and feature representation quality. During finetuning, the model is adapted to partial scans, further enhancing its resilience to incomplete inputs. To address remaining uncertainties in areas with missing structure, we introduce a dual-branch architecture that incorporates image cues: point cloud and image features are extracted independently and then fused via the MoE with an alignment loss, allowing complementary modalities to guide reconstruction in occluded or missing regions. Experiments conducted on the ShapeNet-ViPC dataset show that PARC has achieved highly competitive performance. Code is available at https://github.com/caiyujiaocyj/PARC.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297134","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Reducing the triangle count in complex 3D models is a basic geometry preprocessing step in graphics pipelines such as efficient rendering and interactive editing. However, most existing mesh simplification methods exhibit a few issues. Firstly, they often lead to self-intersections during decimation, a major issue for applications such as 3D printing and soft-body simulation. Second, to perform simplification on a mesh in the wild, one would first need to perform re-meshing, which often suffers from surface shifts and losses of sharp features. Finally, existing re-meshing and simplification methods can take minutes when processing large-scale meshes, limiting their applications in practice. To address the challenges, we introduce a novel GPU-based mesh optimization approach containing three key components: (1) a parallel re-meshing algorithm to turn meshes in the wild into watertight, manifold, and intersection-free ones, and reduce the prevalence of poorly shaped triangles; (2) a robust parallel simplification algorithm with intersection-free guarantees; (3) an optimization-based safe projection algorithm to realign the simplified mesh with the input, eliminating the surface shift introduced by re-meshing and recovering the original sharp features. The algorithm demonstrates remarkable efficiency, simplifying a 2-million-face mesh to 20k triangles in 3 seconds on RTX4090. We evaluated the approach on the Thingi10K dataset and showcased its exceptional performance in geometry preservation and speed. https://seonghunn.github.io/pamo/
{"title":"PaMO: Parallel Mesh Optimization for Intersection-Free Low-Poly Modeling on the GPU","authors":"Seonghun Oh, Xiaodi Yuan, Xinyue Wei, Ruoxi Shi, Fanbo Xiang, Minghua Liu, Hao Su","doi":"10.1111/cgf.70267","DOIUrl":"https://doi.org/10.1111/cgf.70267","url":null,"abstract":"<p>Reducing the triangle count in complex 3D models is a basic geometry preprocessing step in graphics pipelines such as efficient rendering and interactive editing. However, most existing mesh simplification methods exhibit a few issues. Firstly, they often lead to self-intersections during decimation, a major issue for applications such as 3D printing and soft-body simulation. Second, to perform simplification on a mesh in the wild, one would first need to perform re-meshing, which often suffers from surface shifts and losses of sharp features. Finally, existing re-meshing and simplification methods can take minutes when processing large-scale meshes, limiting their applications in practice. To address the challenges, we introduce a novel GPU-based mesh optimization approach containing three key components: (1) a parallel re-meshing algorithm to turn meshes in the wild into watertight, manifold, and intersection-free ones, and reduce the prevalence of poorly shaped triangles; (2) a robust parallel simplification algorithm with intersection-free guarantees; (3) an optimization-based safe projection algorithm to realign the simplified mesh with the input, eliminating the surface shift introduced by re-meshing and recovering the original sharp features. The algorithm demonstrates remarkable efficiency, simplifying a 2-million-face mesh to 20k triangles in 3 seconds on RTX4090. We evaluated the approach on the Thingi10K dataset and showcased its exceptional performance in geometry preservation and speed. https://seonghunn.github.io/pamo/</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Materials exhibit geometric structures across mesoscopic to microscopic scales, influencing macroscale properties such as appearance, mechanical strength, and thermal behavior. Capturing and modeling these multiscale structures is challenging but essential for computer graphics, engineering, and materials science. We present a framework inspired by hypertexture methods, using implicit surfaces and sphere tracing to synthesize multiscale structures on the fly without precomputation. This framework models volumetric materials with particulate, fibrous, porous, and laminar structures, allowing control over size, shape, density, distribution, and orientation. We enhance structural diversity by superimposing implicit periodic functions while improving computational efficiency. The framework also supports spatially varying particulate media, particle agglomeration, and piling on convex and concave structures, such as rock formations (mesoscale), without explicit simulation. We demonstrate its potential in the appearance modeling of volumetric materials and investigate how spatially varying properties affect the perceived macroscale appearance. As a proof of concept, we show that microstructures created by our framework can be reconstructed from image and distance values defined by implicit surfaces, using both first-order and gradient-free optimization methods.
{"title":"Procedural Multiscale Geometry Modeling using Implicit Surfaces","authors":"Bojja Venu, Adam Bosak, Juan Raúl Padrón-Griffe","doi":"10.1111/cgf.70233","DOIUrl":"https://doi.org/10.1111/cgf.70233","url":null,"abstract":"<p>Materials exhibit geometric structures across mesoscopic to microscopic scales, influencing macroscale properties such as appearance, mechanical strength, and thermal behavior. Capturing and modeling these multiscale structures is challenging but essential for computer graphics, engineering, and materials science. We present a framework inspired by hypertexture methods, using implicit surfaces and sphere tracing to synthesize multiscale structures on the fly without precomputation. This framework models volumetric materials with particulate, fibrous, porous, and laminar structures, allowing control over size, shape, density, distribution, and orientation. We enhance structural diversity by superimposing implicit periodic functions while improving computational efficiency. The framework also supports spatially varying particulate media, particle agglomeration, and piling on convex and concave structures, such as rock formations (mesoscale), without explicit simulation. We demonstrate its potential in the appearance modeling of volumetric materials and investigate how spatially varying properties affect the perceived macroscale appearance. As a proof of concept, we show that microstructures created by our framework can be reconstructed from image and distance values defined by implicit surfaces, using both first-order and gradient-free optimization methods.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://onlinelibrary.wiley.com/doi/epdf/10.1111/cgf.70233","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Control variates are a variance-reduction technique for Monte Carlo integration. The principle involves approximating the integrand by a function that can be analytically integrated, and integrating using the Monte Carlo method only the residual difference between the integrand and the approximation, to obtain an unbiased estimate. Neural networks are universal approx-imators that could potentially be used as a control variate. However, the challenge lies in the analytic integration, which is not possible in general. In this manuscript, we study one of the simplest neural network models, the multilayered perceptron (MLP) with continuous piecewise linear activation functions, and its possible analytic integration. We propose an integration method based on integration domain subdivision, employing techniques from computational geometry to solve this problem in 2D. We demonstrate that an MLP can be used as a control variate in combination with our integration method, showing applications in the light transport simulation.
{"title":"Geometric Integration for Neural Control Variates","authors":"D. Meister, T. Harada","doi":"10.1111/cgf.70275","DOIUrl":"https://doi.org/10.1111/cgf.70275","url":null,"abstract":"<p>Control variates are a variance-reduction technique for Monte Carlo integration. The principle involves approximating the integrand by a function that can be analytically integrated, and integrating using the Monte Carlo method only the residual difference between the integrand and the approximation, to obtain an unbiased estimate. Neural networks are universal approx-imators that could potentially be used as a control variate. However, the challenge lies in the analytic integration, which is not possible in general. In this manuscript, we study one of the simplest neural network models, the multilayered perceptron (MLP) with continuous piecewise linear activation functions, and its possible analytic integration. We propose an integration method based on integration domain subdivision, employing techniques from computational geometry to solve this problem in 2D. We demonstrate that an MLP can be used as a control variate in combination with our integration method, showing applications in the light transport simulation.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297321","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
S. Q. Wang, W. B. Wu, M. Shi, Z. X. Li, Q. Wang, D. M. Zhu
Reconstructing underwater object geometry from multi-view images is a long-standing challenge in computer graphics, primarily due to image degradation caused by underwater scattering, blur, and color shift. These degradations severely impair feature extraction and multi-view consistency. Existing methods typically rely on pre-trained image enhancement models as a preprocessing step, but often struggle with robustness under varying water conditions. To overcome these limitations, we propose WaterGS, a novel framework for underwater surface reconstruction that jointly recovers accurate 3D geometry and restores true object colors. The core of our approach lies in introducing a Physically-Based imaging model into the rendering process of 2D Gaussian Splatting. This enables accurate separation of true object colors from water-induced distortions, thereby facilitating more robust photometric alignment and denser geometric reconstruction across views. Building upon this improved photometric consistency, we further introduce a Gaussian bundle adjustment scheme guided by our physical model to jointly optimize camera poses and geometry, enhancing reconstruction accuracy. Extensive experiments on synthetic and real-world datasets show that WaterGS achieves robust, high-fidelity reconstruction directly from raw underwater images, outperforming prior approaches in both geometric accuracy and visual consistency.
{"title":"WaterGS: Physically-Based Imaging in Gaussian Splatting for Underwater Scene Reconstruction","authors":"S. Q. Wang, W. B. Wu, M. Shi, Z. X. Li, Q. Wang, D. M. Zhu","doi":"10.1111/cgf.70270","DOIUrl":"https://doi.org/10.1111/cgf.70270","url":null,"abstract":"<p>Reconstructing underwater object geometry from multi-view images is a long-standing challenge in computer graphics, primarily due to image degradation caused by underwater scattering, blur, and color shift. These degradations severely impair feature extraction and multi-view consistency. Existing methods typically rely on pre-trained image enhancement models as a preprocessing step, but often struggle with robustness under varying water conditions. To overcome these limitations, we propose WaterGS, a novel framework for underwater surface reconstruction that jointly recovers accurate 3D geometry and restores true object colors. The core of our approach lies in introducing a Physically-Based imaging model into the rendering process of 2D Gaussian Splatting. This enables accurate separation of true object colors from water-induced distortions, thereby facilitating more robust photometric alignment and denser geometric reconstruction across views. Building upon this improved photometric consistency, we further introduce a Gaussian bundle adjustment scheme guided by our physical model to jointly optimize camera poses and geometry, enhancing reconstruction accuracy. Extensive experiments on synthetic and real-world datasets show that WaterGS achieves robust, high-fidelity reconstruction directly from raw underwater images, outperforming prior approaches in both geometric accuracy and visual consistency.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Dabeen Park, Junsuh Park, Jooeun Son, Seungyoung Lee, Joo-Ho Lee
Decomposing environmental lighting and materials is challenging as they are tightly intertwined and integrated over the hemisphere. In order to precisely decouple them, the lighting representation must represent general image features such as object boundaries or texture contrast, called light edges, which are often neglected in the existing inverse rendering methods. In this paper, we propose an inverse rendering method that efficiently captures light edges. We introduce a triangle mesh-based light representation that can express light edges by aligning triangle edges with light edges. We exploit the linearly transformed cosines as BRDF approximations to efficiently compute environmental lighting with our light representation. Our edge-aware inverse rendering precisely decouples distributions of reflectance and lighting through differentiable rendering by jointly reconstructing light edges and estimating the BRDF parameters. Our experiments, including various material/scene settings and ablation studies, demonstrate the reconstruction performance and computational efficiency of our method.
{"title":"LTC-IR: Multiview Edge-Aware Inverse Rendering with Linearly Transformed Cosines","authors":"Dabeen Park, Junsuh Park, Jooeun Son, Seungyoung Lee, Joo-Ho Lee","doi":"10.1111/cgf.70262","DOIUrl":"https://doi.org/10.1111/cgf.70262","url":null,"abstract":"<p>Decomposing environmental lighting and materials is challenging as they are tightly intertwined and integrated over the hemisphere. In order to precisely decouple them, the lighting representation must represent general image features such as object boundaries or texture contrast, called light edges, which are often neglected in the existing inverse rendering methods. In this paper, we propose an inverse rendering method that efficiently captures light edges. We introduce a triangle mesh-based light representation that can express light edges by aligning triangle edges with light edges. We exploit the linearly transformed cosines as BRDF approximations to efficiently compute environmental lighting with our light representation. Our edge-aware inverse rendering precisely decouples distributions of reflectance and lighting through differentiable rendering by jointly reconstructing light edges and estimating the BRDF parameters. Our experiments, including various material/scene settings and ablation studies, demonstrate the reconstruction performance and computational efficiency of our method.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pierre Hubert-Brierre, Eric Guérin, Adrien Peytavie, Eric Galin
Processing and particularly visualizing implicit surfaces remains computationally intensive when dealing with complex objects built from construction trees. We introduce optimization nodes to reduce the computational cost of the field function evaluation for hierarchical construction trees, while preserving the Lipschitz or conservative properties of the function. Our goal is to propose acceleration nodes directly embedded in the construction tree, and avoid external, accompanying data-structures such as octrees. We present proxy and continuous level of detail nodes to reduce the overall evaluation cost, along with a normal warping technique that enhances surface details with negligible computational overhead. Our approach is compatible with existing algorithms that aim at reducing the number of function calls. We validate our methods by computing timings as well as the average cost for traversing the tree and evaluating the signed distance field at a given point in space. Our method speeds-up signed distance field evaluation by up to three orders or magnitude, and applies both to ray-surface intersection computation in Sphere Tracing applications, and to polygonization algorithms.
{"title":"Accelerating Signed Distance Functions","authors":"Pierre Hubert-Brierre, Eric Guérin, Adrien Peytavie, Eric Galin","doi":"10.1111/cgf.70258","DOIUrl":"https://doi.org/10.1111/cgf.70258","url":null,"abstract":"<p>Processing and particularly visualizing implicit surfaces remains computationally intensive when dealing with complex objects built from construction trees. We introduce optimization nodes to reduce the computational cost of the field function evaluation for hierarchical construction trees, while preserving the Lipschitz or conservative properties of the function. Our goal is to propose acceleration nodes directly embedded in the construction tree, and avoid external, accompanying data-structures such as octrees. We present proxy and continuous level of detail nodes to reduce the overall evaluation cost, along with a normal warping technique that enhances surface details with negligible computational overhead. Our approach is compatible with existing algorithms that aim at reducing the number of function calls. We validate our methods by computing timings as well as the average cost for traversing the tree and evaluating the signed distance field at a given point in space. Our method speeds-up signed distance field evaluation by up to three orders or magnitude, and applies both to ray-surface intersection computation in Sphere Tracing applications, and to polygonization algorithms.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Although significant progress has been made in monocular video-based 3D human pose estimation, existing methods lack guidance from fine-grained high-level prior knowledge such as action semantics and camera viewpoints, leading to significant challenges for pose reconstruction accuracy under scenarios with severely missing visual features, i.e., complex occlusion situations. We identify that the 3D human pose estimation task fundamentally constitutes a canonical inverse problem, and propose a motion-semantics-based diffusion(MS-Diff) framework to address this issue by incorporating high-level motion semantics with spectral feature regularization to eliminate interference noise in complex scenes and improve estimation accuracy. Specifically, we design a Multimodal Diffusion Interaction (MDI) module that incorporates motion semantics including action categories and camera viewpoints into the diffusion process, establishing semantic-visual feature alignment through a cross-modal mechanism to resolve pose ambiguities and effectively handle occlusions. Additionally, we leverage a Spectral Convolutional Regularization (SCR) module that implements adaptive filtering in the frequency domain to selectively suppress noise components. Extensive experiments on large-scale public datasets Human3.6M and MPI-INF-3DHP demonstrate that our method achieves state-of-the-art performance.
{"title":"Text-Guided Diffusion with Spectral Convolution for 3D Human Pose Estimation","authors":"Liyuan Shi, Suping Wu, Sheng Yang, Weibin Qiu, Dong Qiang, Jiarui Zhao","doi":"10.1111/cgf.70263","DOIUrl":"https://doi.org/10.1111/cgf.70263","url":null,"abstract":"<p>Although significant progress has been made in monocular video-based 3D human pose estimation, existing methods lack guidance from fine-grained high-level prior knowledge such as action semantics and camera viewpoints, leading to significant challenges for pose reconstruction accuracy under scenarios with severely missing visual features, i.e., complex occlusion situations. We identify that the 3D human pose estimation task fundamentally constitutes a canonical inverse problem, and propose a motion-semantics-based diffusion(MS-Diff) framework to address this issue by incorporating high-level motion semantics with spectral feature regularization to eliminate interference noise in complex scenes and improve estimation accuracy. Specifically, we design a Multimodal Diffusion Interaction (MDI) module that incorporates motion semantics including action categories and camera viewpoints into the diffusion process, establishing semantic-visual feature alignment through a cross-modal mechanism to resolve pose ambiguities and effectively handle occlusions. Additionally, we leverage a Spectral Convolutional Regularization (SCR) module that implements adaptive filtering in the frequency domain to selectively suppress noise components. Extensive experiments on large-scale public datasets Human3.6M and MPI-INF-3DHP demonstrate that our method achieves state-of-the-art performance.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"44 7","pages":""},"PeriodicalIF":2.9,"publicationDate":"2025-10-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145297324","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}