首页 > 最新文献

ACM Transactions on Graphics最新文献

英文 中文
Lifted Surfacing of Generalized Sweep Volumes 广义扫描体的抬升表面
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763360
Yiwen Ju, Qingnan Zhou, Xingyi Du, Nathan Carr, Tao Ju
Computing the boundary surface of the 3D volume swept by a rigid or deforming solid remains a challenging problem in geometric modeling. Existing approaches are often limited to sweeping rigid shapes, cannot guarantee a watertight surface, or struggle with modeling the intricate geometric features (e.g., sharp creases and narrow gaps) and topological features (e.g., interior voids). We make the observation that the sweep boundary is a subset of the projection of the intersection of two implicit surfaces in a higher dimension, and we derive a characterization of the subset using winding numbers. These insights lead to a general algorithm for any sweep represented as a smooth time-varying implicit function satisfying a genericity assumption, and it produces a watertight and intersection-free surface that better approximates the geometric and topological features than existing methods.
在几何建模中,计算被刚性或变形实体扫过的三维体的边界面一直是一个具有挑战性的问题。现有的方法通常仅限于扫描刚性形状,不能保证水密表面,或者难以模拟复杂的几何特征(例如,尖锐的折痕和狭窄的间隙)和拓扑特征(例如,内部空隙)。我们观察到扫描边界是两个隐式曲面交点在高维上的投影的子集,并且我们用圈数推导了子集的表征。这些见解导致了一种用于任何扫描的通用算法,该算法表示为满足一般假设的光滑时变隐式函数,并且它产生一个水密和无相交的表面,比现有方法更好地接近几何和拓扑特征。
{"title":"Lifted Surfacing of Generalized Sweep Volumes","authors":"Yiwen Ju, Qingnan Zhou, Xingyi Du, Nathan Carr, Tao Ju","doi":"10.1145/3763360","DOIUrl":"https://doi.org/10.1145/3763360","url":null,"abstract":"Computing the boundary surface of the 3D volume swept by a rigid or deforming solid remains a challenging problem in geometric modeling. Existing approaches are often limited to sweeping rigid shapes, cannot guarantee a watertight surface, or struggle with modeling the intricate geometric features (e.g., sharp creases and narrow gaps) and topological features (e.g., interior voids). We make the observation that the sweep boundary is a subset of the projection of the intersection of two implicit surfaces in a higher dimension, and we derive a characterization of the subset using winding numbers. These insights lead to a general algorithm for any sweep represented as a smooth time-varying implicit function satisfying a genericity assumption, and it produces a watertight and intersection-free surface that better approximates the geometric and topological features than existing methods.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"33 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Aerial Path Planning for Urban Geometry and Texture Co-Capture 城市几何与纹理协同捕获的空中路径规划
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763292
Weidan Xiong, Bochuan Zeng, Ziyu Hu, Jianwei Guo, Ke Xie, Hui Huang
Recent advances in image acquisition and scene reconstruction have enabled the generation of high-quality structural urban scene geometry, given sufficient site information. However, current capture techniques often overlook the crucial importance of texture quality, resulting in noticeable visual artifacts in the textured models. In this work, we introduce the urban geometry and texture co-capture problem under limited prior knowledge before a site visit. The only inputs are a 2D building contour map of the target area and a safe flying altitude above the buildings. We propose an innovative aerial path planning framework designed to co-capture images for reconstructing both structured geometry and high-fidelity textures. To evaluate and guide view planning, we introduce a comprehensive texture quality assessment system, including two novel metrics tailored for building facades. Firstly, our method generates high-quality vertical dipping views and horizontal planar views to effectively capture both geometric and textural details. A multi-objective optimization strategy is then proposed to jointly maximize texture fidelity, improve geometric accuracy, and minimize the cost associated with aerial views. Furthermore, we present a sequential path planning algorithm that accounts for texture consistency during image capture. Extensive experiments on large-scale synthetic and real-world urban datasets demonstrate that our approach effectively produces image sets suitable for concurrent geometric and texture reconstruction, enabling the creation of realistic, textured scene proxies at low operational cost.
在图像采集和场景重建方面的最新进展,使得在给定足够的场地信息的情况下,能够生成高质量的城市场景几何结构。然而,目前的捕获技术往往忽视了纹理质量的重要性,导致纹理模型中明显的视觉伪影。在这项工作中,我们介绍了在实地考察前有限先验知识下的城市几何和纹理共同捕获问题。唯一的输入是目标区域的二维建筑等高线地图和建筑物上方的安全飞行高度。我们提出了一种创新的空中路径规划框架,旨在共同捕获图像以重建结构几何和高保真纹理。为了评估和指导景观规划,我们引入了一个全面的纹理质量评估系统,包括为建筑立面量身定制的两个新指标。首先,我们的方法生成高质量的垂直倾斜视图和水平平面视图,以有效捕获几何和纹理细节。然后提出了一种多目标优化策略,以最大限度地提高纹理保真度,提高几何精度,并最小化与鸟瞰图相关的成本。此外,我们提出了一种顺序路径规划算法,该算法考虑了图像捕获过程中的纹理一致性。在大规模合成和现实世界城市数据集上进行的大量实验表明,我们的方法有效地生成了适合并行几何和纹理重建的图像集,从而能够以低运营成本创建逼真的纹理场景代理。
{"title":"Aerial Path Planning for Urban Geometry and Texture Co-Capture","authors":"Weidan Xiong, Bochuan Zeng, Ziyu Hu, Jianwei Guo, Ke Xie, Hui Huang","doi":"10.1145/3763292","DOIUrl":"https://doi.org/10.1145/3763292","url":null,"abstract":"Recent advances in image acquisition and scene reconstruction have enabled the generation of high-quality structural urban scene geometry, given sufficient site information. However, current capture techniques often overlook the crucial importance of texture quality, resulting in noticeable visual artifacts in the textured models. In this work, we introduce the urban <jats:italic toggle=\"yes\">geometry and texture co-capture</jats:italic> problem under limited prior knowledge before a site visit. The only inputs are a 2D building contour map of the target area and a safe flying altitude above the buildings. We propose an innovative aerial path planning framework designed to co-capture images for reconstructing both structured geometry and high-fidelity textures. To evaluate and guide view planning, we introduce a comprehensive texture quality assessment system, including two novel metrics tailored for building facades. Firstly, our method generates high-quality vertical dipping views and horizontal planar views to effectively capture both geometric and textural details. A multi-objective optimization strategy is then proposed to jointly maximize texture fidelity, improve geometric accuracy, and minimize the cost associated with aerial views. Furthermore, we present a sequential path planning algorithm that accounts for texture consistency during image capture. Extensive experiments on large-scale synthetic and real-world urban datasets demonstrate that our approach effectively produces image sets suitable for concurrent geometric and texture reconstruction, enabling the creation of realistic, textured scene proxies at low operational cost.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"1 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673765","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Detail Enhanced Gaussian Splatting for Large-Scale Volumetric Capture 细节增强高斯飞溅大规模体积捕获
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763336
Julien Philip, Li Ma, Pascal Clausen, Wenqi Xian, Ahmet Levent Taşel, Mingming He, Xueming Yu, David M. George, Ning Yu, Oliver Pilarski, Paul Debevec
We present a unique system for large-scale, multi-performer, high resolution 4D volumetric capture providing realistic free-viewpoint video up to and including 4K resolution facial closeups. To achieve this, we employ a novel volumetric capture, reconstruction and rendering pipeline based on Dynamic Gaussian Splatting and Diffusion-based Detail Enhancement. We design our pipeline specifically to meet the demands of high-end media production. We employ two capture rigs: the Scene Rig , which captures multi-actor performances at a resolution which falls short of 4K production quality, and the Face Rig , which records high-fidelity single-actor facial detail to serve as a reference for detail enhancement. We first reconstruct dynamic performances from the Scene Rig using 4D Gaussian Splatting, incorporating new model designs and training strategies to improve reconstruction, dynamic range, and rendering quality. Then to render high-quality images for facial closeups, we introduce a diffusion-based detail enhancement model. This model is fine-tuned with high-fidelity data from the same actors recorded in the Face Rig. We train on paired data generated from low- and high-quality Gaussian Splatting (GS) models, using the low-quality input to match the quality of the Scene Rig , with the high-quality GS as ground truth. Our results demonstrate the effectiveness of this pipeline in bridging the gap between the scalable performance capture of a large-scale rig and the high-resolution standards required for film and media production.
我们提出了一个独特的系统,用于大规模,多表演者,高分辨率4D体捕捉,提供逼真的自由视点视频,包括4K分辨率面部特写。为了实现这一点,我们采用了一种新的基于动态高斯飞溅和基于扩散的细节增强的体积捕获,重建和渲染管道。我们专门设计我们的管道来满足高端媒体制作的需求。我们采用了两个捕捉钻机:场景钻机,它捕捉多演员表演的分辨率低于4K生产质量,和脸钻机,它记录高保真的单演员面部细节,作为细节增强的参考。我们首先使用4D高斯飞溅从场景钻机重建动态性能,结合新的模型设计和训练策略,以提高重建,动态范围和渲染质量。然后,我们引入了一种基于扩散的细节增强模型来渲染高质量的面部特写图像。该模型使用来自面部钻机中记录的相同演员的高保真数据进行微调。我们对从低质量和高质量高斯飞溅(GS)模型生成的配对数据进行训练,使用低质量输入来匹配场景钻机的质量,以高质量的GS作为地面真实值。我们的研究结果表明,该管道在弥合大型钻井平台的可扩展性能捕获与电影和媒体制作所需的高分辨率标准之间的差距方面是有效的。
{"title":"Detail Enhanced Gaussian Splatting for Large-Scale Volumetric Capture","authors":"Julien Philip, Li Ma, Pascal Clausen, Wenqi Xian, Ahmet Levent Taşel, Mingming He, Xueming Yu, David M. George, Ning Yu, Oliver Pilarski, Paul Debevec","doi":"10.1145/3763336","DOIUrl":"https://doi.org/10.1145/3763336","url":null,"abstract":"We present a unique system for large-scale, multi-performer, high resolution 4D volumetric capture providing realistic free-viewpoint video up to and including 4K resolution facial closeups. To achieve this, we employ a novel volumetric capture, reconstruction and rendering pipeline based on Dynamic Gaussian Splatting and Diffusion-based Detail Enhancement. We design our pipeline specifically to meet the demands of high-end media production. We employ two capture rigs: the <jats:italic toggle=\"yes\">Scene Rig</jats:italic> , which captures multi-actor performances at a resolution which falls short of 4K production quality, and the <jats:italic toggle=\"yes\">Face Rig</jats:italic> , which records high-fidelity single-actor facial detail to serve as a reference for detail enhancement. We first reconstruct dynamic performances from the <jats:italic toggle=\"yes\">Scene Rig</jats:italic> using 4D Gaussian Splatting, incorporating new model designs and training strategies to improve reconstruction, dynamic range, and rendering quality. Then to render high-quality images for facial closeups, we introduce a diffusion-based detail enhancement model. This model is fine-tuned with high-fidelity data from the same actors recorded in the <jats:italic toggle=\"yes\">Face Rig.</jats:italic> We train on paired data generated from low- and high-quality Gaussian Splatting (GS) models, using the low-quality input to match the quality of the <jats:italic toggle=\"yes\">Scene Rig</jats:italic> , with the high-quality GS as ground truth. Our results demonstrate the effectiveness of this pipeline in bridging the gap between the scalable performance capture of a large-scale rig and the high-resolution standards required for film and media production.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"168 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673767","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BSP-OT: Sparse transport plans between discrete measures in loglinear time BSP-OT:在线性时间内离散度量之间的稀疏传输计划
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763281
Baptiste Genest, Nicolas Bonneel, Vincent Nivoliers, David Coeurjolly
To solve the optimal transport problem between two uniform discrete measures of the same size, one seeks a bijective assignment that minimizes some matching cost. For this task, exact algorithms are intractable for large problems, while approximate ones may lose the bijectivity of the assignment. We address this issue and the more general cases of non-uniform discrete measures with different total masses, where partial transport may be desirable. The core of our algorithm is a variant of the Quicksort algorithm that provides an efficient strategy to randomly explore many relevant and easy-to-compute couplings, by matching BSP trees in loglinear time. The couplings we obtain are as sparse as possible, in the sense that they provide bijections, injective partial matchings or sparse couplings depending on the nature of the matched measures. To improve the transport cost, we propose efficient strategies to merge k sparse couplings into a higher quality one. For k = 64, we obtain transport plans with typically less than 1% of relative error in a matter of seconds between hundreds of thousands of points in 3D on the CPU. We demonstrate how these high-quality approximations can drastically speed-up usual pipelines involving optimal transport, such as shape interpolation, intrinsic manifold sampling, color transfer, topological data analysis, rigid partial registration of point clouds and image stippling.
为了解决两个相同尺寸的统一离散度量之间的最优运输问题,人们寻求一个使某些匹配成本最小化的双目标分配。对于这个任务,精确算法对于大问题是难以处理的,而近似算法可能会失去分配的客观性。我们解决了这个问题,以及具有不同总质量的非均匀离散测度的更一般的情况,其中部分输运可能是可取的。我们算法的核心是快速排序算法的一种变体,它提供了一种有效的策略,通过在对数时间内匹配BSP树来随机探索许多相关且易于计算的耦合。我们得到的耦合是尽可能稀疏的,在某种意义上,它们提供双射、内射部分匹配或稀疏耦合,这取决于匹配度量的性质。为了提高传输成本,我们提出了将k个稀疏耦合合并为一个更高质量的稀疏耦合的有效策略。对于k = 64,我们在CPU上的数十万个3D点之间获得传输计划,通常在几秒钟内相对误差小于1%。我们展示了这些高质量的近似如何极大地加快涉及最佳传输的常规管道,如形状插值、内在流形采样、颜色转移、拓扑数据分析、点云的刚性部分配准和图像点点。
{"title":"BSP-OT: Sparse transport plans between discrete measures in loglinear time","authors":"Baptiste Genest, Nicolas Bonneel, Vincent Nivoliers, David Coeurjolly","doi":"10.1145/3763281","DOIUrl":"https://doi.org/10.1145/3763281","url":null,"abstract":"To solve the optimal transport problem between two uniform discrete measures of the same size, one seeks a bijective assignment that minimizes some matching cost. For this task, exact algorithms are intractable for large problems, while approximate ones may lose the bijectivity of the assignment. We address this issue and the more general cases of non-uniform discrete measures with different total masses, where partial transport may be desirable. The core of our algorithm is a variant of the Quicksort algorithm that provides an efficient strategy to randomly explore many relevant and easy-to-compute couplings, by matching BSP trees in loglinear time. The couplings we obtain are as sparse as possible, in the sense that they provide bijections, injective partial matchings or sparse couplings depending on the nature of the matched measures. To improve the transport cost, we propose efficient strategies to merge <jats:italic toggle=\"yes\">k</jats:italic> sparse couplings into a higher quality one. For <jats:italic toggle=\"yes\">k =</jats:italic> 64, we obtain transport plans with typically less than 1% of relative error in a matter of seconds between hundreds of thousands of points in 3D on the CPU. We demonstrate how these high-quality approximations can drastically speed-up usual pipelines involving optimal transport, such as shape interpolation, intrinsic manifold sampling, color transfer, topological data analysis, rigid partial registration of point clouds and image stippling.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"30 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Jump Restore Light Transport 跳跃恢复轻型运输
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763286
Sascha Holl, Gurprit Singh, Hans-Peter Seidel
Markov chain Monte Carlo (MCMC) algorithms are indispensable when sampling from a complex, high-dimensional distribution by a conventional method is intractable. Even though MCMC is a powerful tool, it is also hard to control and tune in practice. Simultaneously achieving both rapid local exploration of the state space and efficient global discovery of the target distribution is a challenging task. In this work, we introduce a novel continuous-time MCMC formulation to the computer science community. Generalizing existing work from the statistics community, we propose a novel framework for adjusting an arbitrary family of Markov processes - used for local exploration of the state space only - to an overall process which is invariant with respect to a target distribution. To demonstrate the potential of our framework, we focus on a simple, but yet insightful, application in light transport simulation. As a by-product, we introduce continuous-time MCMC sampling to the computer graphics community. We show how any existing MCMC-based light transport algorithm can be seamlessly integrated into our framework. We prove empirically and theoretically that the integrated version is superior to the ordinary algorithm. In fact, our approach will convert any existing algorithm into a highly parallelizable variant with shorter running time, smaller error and less variance.
当用传统方法难以从复杂的高维分布中采样时,马尔可夫链蒙特卡罗(MCMC)算法必不可少。尽管MCMC是一个强大的工具,但在实践中也很难控制和调整。同时实现状态空间的快速局部探测和目标分布的高效全局发现是一项具有挑战性的任务。在这项工作中,我们向计算机科学界介绍了一种新的连续时间MCMC公式。在推广统计界现有工作的基础上,我们提出了一个新的框架,用于将任意一组马尔可夫过程(仅用于状态空间的局部探索)调整为一个相对于目标分布不变的整体过程。为了展示我们的框架的潜力,我们将重点放在一个简单但富有洞察力的轻传输模拟应用上。作为副产品,我们向计算机图形界介绍了连续时间MCMC采样。我们展示了如何将任何现有的基于mcmc的光传输算法无缝集成到我们的框架中。从理论上和经验上证明了综合算法优于普通算法。实际上,我们的方法可以将任何现有的算法转换为具有更短运行时间、更小误差和更少方差的高度并行化的变体。
{"title":"Jump Restore Light Transport","authors":"Sascha Holl, Gurprit Singh, Hans-Peter Seidel","doi":"10.1145/3763286","DOIUrl":"https://doi.org/10.1145/3763286","url":null,"abstract":"Markov chain Monte Carlo (MCMC) algorithms are indispensable when sampling from a complex, high-dimensional distribution by a conventional method is intractable. Even though MCMC is a powerful tool, it is also hard to control and tune in practice. Simultaneously achieving both rapid <jats:italic toggle=\"yes\">local exploration</jats:italic> of the state space and efficient <jats:italic toggle=\"yes\">global discovery</jats:italic> of the target distribution is a challenging task. In this work, we introduce a novel continuous-time MCMC formulation to the computer science community. Generalizing existing work from the statistics community, we propose a novel framework for <jats:italic toggle=\"yes\">adjusting</jats:italic> an arbitrary family of Markov processes - used for local exploration of the state space only - to an overall process which is invariant with respect to a target distribution. To demonstrate the potential of our framework, we focus on a simple, but yet insightful, application in light transport simulation. As a by-product, we introduce continuous-time MCMC sampling to the computer graphics community. We show how any existing MCMC-based light transport algorithm can be seamlessly integrated into our framework. We prove empirically and theoretically that the integrated version is superior to the ordinary algorithm. In fact, our approach will convert any existing algorithm into a highly <jats:italic toggle=\"yes\">parallelizable</jats:italic> variant with shorter running time, smaller error and less variance.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"12 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Consecutive Frame Extrapolation with Predictive Sparse Shading 预测稀疏阴影的连续帧外推
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763363
Zhizhen Wu, Zhe Cao, Yazhen Yuan, Zhilong Yuan, Rui Wang, Yuchi Huo
The demand for high-frame-rate rendering keeps increasing in modern displays. Existing frame generation and super-resolution techniques accelerate rendering by reducing rendering samples across space or time. However, they rely on a uniform sampling reduction strategy, which undersamples areas with complex details or dynamic shading. To address this, we propose to sparsely shade critical areas while reusing generated pixels in low-variation areas for neural extrapolation. Specifically, we introduce the Predictive Error-Flow-eXtrapolation Network (EFXNet)-an architecture that predicts extrapolation errors, estimates flows, and extrapolates frames at once. Firstly, EFXNet leverages temporal coherence to predict extrapolation error and guide the sparse shading of dynamic areas. In addition, EFXNet employs a target-grid correlation module to estimate robust optical flows from pixel correlations rather than pixel values. Finally, EFXNet uses dedicated motion representations for the historical geometric and lighting components, respectively, to extrapolate temporally stable frames. Extensive experimental results show that, compared with state-of-the-art methods, our frame extrapolation method exhibits superior visual quality and temporal stability under a low rendering budget.
现代显示器对高帧率渲染的需求不断增加。现有的帧生成和超分辨率技术通过减少跨空间或时间的渲染样本来加速渲染。然而,它们依赖于统一的采样减少策略,该策略对具有复杂细节或动态阴影的区域进行了欠采样。为了解决这个问题,我们建议对关键区域进行稀疏阴影处理,同时在低变化区域重复使用生成的像素进行神经外推。具体来说,我们介绍了预测错误流外推网络(EFXNet)——一种预测外推错误、估计流和一次外推帧的架构。首先,EFXNet利用时间相干性预测外推误差并指导动态区域的稀疏阴影。此外,EFXNet采用目标网格相关模块,从像素相关性而不是像素值来估计鲁棒光流。最后,EFXNet分别为历史几何和照明组件使用专用的运动表示来推断暂时稳定的帧。大量的实验结果表明,与目前最先进的方法相比,我们的帧外推方法在较低的渲染预算下具有更好的视觉质量和时间稳定性。
{"title":"Consecutive Frame Extrapolation with Predictive Sparse Shading","authors":"Zhizhen Wu, Zhe Cao, Yazhen Yuan, Zhilong Yuan, Rui Wang, Yuchi Huo","doi":"10.1145/3763363","DOIUrl":"https://doi.org/10.1145/3763363","url":null,"abstract":"The demand for high-frame-rate rendering keeps increasing in modern displays. Existing frame generation and super-resolution techniques accelerate rendering by reducing rendering samples across space or time. However, they rely on a uniform sampling reduction strategy, which undersamples areas with complex details or dynamic shading. To address this, we propose to sparsely shade critical areas while reusing generated pixels in low-variation areas for neural extrapolation. Specifically, we introduce the Predictive Error-Flow-eXtrapolation Network (EFXNet)-an architecture that predicts extrapolation errors, estimates flows, and extrapolates frames at once. Firstly, EFXNet leverages temporal coherence to predict extrapolation error and guide the sparse shading of dynamic areas. In addition, EFXNet employs a target-grid correlation module to estimate robust optical flows from pixel correlations rather than pixel values. Finally, EFXNet uses dedicated motion representations for the historical geometric and lighting components, respectively, to extrapolate temporally stable frames. Extensive experimental results show that, compared with state-of-the-art methods, our frame extrapolation method exhibits superior visual quality and temporal stability under a low rendering budget.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"21 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673714","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Sampling for Discontinuities in Differentiable Shaders 可微分着色器中不连续点的自动采样
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763291
Yash Belhe, Ishit Mehta, Wesley Chang, Iliyan Georgiev, Michael Gharbi, Ravi Ramamoorthi, Tzu-Mao Li
We present a novel method to differentiate integrals of discontinuous functions, which are common in inverse graphics, computer vision, and machine learning applications. Previous methods either require specialized routines to sample the discontinuous boundaries of predetermined primitives, or use reparameterization techniques that suffer from high variance. In contrast, our method handles general discontinuous functions, expressed as shader programs, without requiring manually specified boundary sampling routines. We achieve this through a program transformation that converts discontinuous functions into piecewise constant ones, enabling efficient boundary sampling through a novel segment snapping technique, and accurate derivatives at the boundary by simply comparing values on both sides of the discontinuity. Our method handles both explicit boundaries (polygons, ellipses, Bézier curves) and implicit ones (neural networks, noise-based functions, swept surfaces). We demonstrate that our system supports a wide range of applications, including painterly rendering, raster image fitting, constructive solid geometry, swept surfaces, mosaicing, and ray marching.
我们提出了一种新的方法来微分不连续函数的积分,这在逆图,计算机视觉和机器学习应用中很常见。以前的方法要么需要专门的例程来采样预定原语的不连续边界,要么使用高方差的重新参数化技术。相比之下,我们的方法处理一般的不连续函数,表示为着色器程序,不需要手动指定边界采样例程。我们通过将不连续函数转换为分段常数函数的程序转换来实现这一点,通过一种新的片段捕捉技术实现有效的边界采样,并通过简单地比较不连续两侧的值来实现边界上的精确导数。我们的方法可以处理显式边界(多边形、椭圆、bsamzier曲线)和隐式边界(神经网络、基于噪声的函数、扫描表面)。我们证明了我们的系统支持广泛的应用,包括绘画渲染,光栅图像拟合,构造实体几何,扫描表面,马赛克和射线推进。
{"title":"Automatic Sampling for Discontinuities in Differentiable Shaders","authors":"Yash Belhe, Ishit Mehta, Wesley Chang, Iliyan Georgiev, Michael Gharbi, Ravi Ramamoorthi, Tzu-Mao Li","doi":"10.1145/3763291","DOIUrl":"https://doi.org/10.1145/3763291","url":null,"abstract":"We present a novel method to differentiate integrals of discontinuous functions, which are common in inverse graphics, computer vision, and machine learning applications. Previous methods either require specialized routines to sample the discontinuous boundaries of predetermined primitives, or use reparameterization techniques that suffer from high variance. In contrast, our method handles general discontinuous functions, expressed as shader programs, without requiring manually specified boundary sampling routines. We achieve this through a program transformation that converts discontinuous functions into piecewise constant ones, enabling efficient boundary sampling through a novel segment snapping technique, and accurate derivatives at the boundary by simply comparing values on both sides of the discontinuity. Our method handles both explicit boundaries (polygons, ellipses, Bézier curves) and implicit ones (neural networks, noise-based functions, swept surfaces). We demonstrate that our system supports a wide range of applications, including painterly rendering, raster image fitting, constructive solid geometry, swept surfaces, mosaicing, and ray marching.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"125 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust Derivative Estimation with Walk on Stars 星上行走的鲁棒导数估计
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763333
Zihan Yu, Rohan Sawhney, Bailey Miller, Lifan Wu, Shuang Zhao
Monte Carlo methods based on the walk on spheres (WoS) algorithm offer a parallel, progressive, and output-sensitive approach for solving partial differential equations (PDEs) in complex geometric domains. Building on this foundation, the walk on stars (WoSt) method generalizes WoS to support mixed Dirichlet, Neumann, and Robin boundary conditions. However, accurately computing spatial derivatives of PDE solutions remains a major challenge: existing methods exhibit high variance and bias near the domain boundary, especially in Neumann-dominated problems. We address this limitation with a new extension of WoSt specifically designed for derivative estimation. Our method reformulates the boundary integral equation (BIE) for Poisson PDEs by directly leveraging the harmonicity of spatial derivatives. Combined with a tailored random-walk sampling scheme and an unbiased early termination strategy, we achieve significantly improved accuracy in derivative estimates near the Neumann boundary. We further demonstrate the effectiveness of our approach across various tasks, including recovering the non-unique solution to a pure Neumann problem with reduced bias and variance, constructing divergence-free vector fields, and optimizing parametrically defined boundaries under PDE constraints.
基于球上行走(WoS)算法的蒙特卡罗方法为求解复杂几何域的偏微分方程(PDEs)提供了一种并行、渐进和输出敏感的方法。在此基础上,星光漫步(WoSt)方法将WoS推广到支持Dirichlet、Neumann和Robin混合边界条件。然而,精确计算PDE解的空间导数仍然是一个主要挑战:现有方法在域边界附近表现出高方差和偏差,特别是在neumann主导问题中。我们通过专门为导数估计设计的WoSt的新扩展来解决这一限制。我们的方法通过直接利用空间导数的调和性,重新表述了泊松偏微分方程的边界积分方程(BIE)。结合量身定制的随机漫步抽样方案和无偏早期终止策略,我们在Neumann边界附近的导数估计中获得了显着提高的准确性。我们进一步证明了我们的方法在各种任务中的有效性,包括恢复具有减少偏差和方差的纯诺伊曼问题的非唯一解,构造无散度向量场,以及在PDE约束下优化参数定义边界。
{"title":"Robust Derivative Estimation with Walk on Stars","authors":"Zihan Yu, Rohan Sawhney, Bailey Miller, Lifan Wu, Shuang Zhao","doi":"10.1145/3763333","DOIUrl":"https://doi.org/10.1145/3763333","url":null,"abstract":"Monte Carlo methods based on the walk on spheres (WoS) algorithm offer a parallel, progressive, and output-sensitive approach for solving partial differential equations (PDEs) in complex geometric domains. Building on this foundation, the walk on stars (WoSt) method generalizes WoS to support mixed Dirichlet, Neumann, and Robin boundary conditions. However, accurately computing spatial derivatives of PDE solutions remains a major challenge: existing methods exhibit high variance and bias near the domain boundary, especially in Neumann-dominated problems. We address this limitation with a new extension of WoSt specifically designed for derivative estimation. Our method reformulates the boundary integral equation (BIE) for Poisson PDEs by directly leveraging the harmonicity of spatial derivatives. Combined with a tailored random-walk sampling scheme and an unbiased early termination strategy, we achieve significantly improved accuracy in derivative estimates near the Neumann boundary. We further demonstrate the effectiveness of our approach across various tasks, including recovering the non-unique solution to a pure Neumann problem with reduced bias and variance, constructing divergence-free vector fields, and optimizing parametrically defined boundaries under PDE constraints.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"33 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673854","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Evaluating and Sampling Glinty NDFs in Constant Time 常数时间内粘性ndf的评估与采样
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763282
Pauli Kemppinen, Loïs Paulin, Théo Thonat, Jean-Marc Thiery, Jaakko Lehtinen, Tamy Boubekeur
Geometric features between the micro and macro scales produce an expressive family of visual effects grouped under the term "glints". Efficiently rendering these effects amounts to finding the highlights caused by the geometry under each pixel. To allow for fast rendering, we represent our faceted geometry as a 4D point process on an implicit multiscale grid, designed to efficiently find the facets most likely to cause a highlight. The facets' normals are generated to match a given micro-facet normal distribution such as Trowbridge-Reitz (GGX) or Beckmann, to which our model converges under increasing surface area. Our method is simple to implement, memory-and-precomputation-free, allows for importance sampling and covers a wide range of different appearances such as anisotropic as well as individually colored particles. We provide a base implementation as a standalone fragment shader.
微观和宏观尺度之间的几何特征产生了一系列富有表现力的视觉效果,统称为“闪烁”。有效地渲染这些效果相当于在每个像素下找到由几何形状引起的高光。为了允许快速渲染,我们在隐式多尺度网格上将我们的面几何表示为4D点过程,旨在有效地找到最有可能引起高亮的面。facet的法线是为了匹配给定的微facet正态分布(如Trowbridge-Reitz (GGX)或Beckmann)而生成的,我们的模型在增加表面积的情况下收敛于此。我们的方法易于实现,不需要内存和预计算,允许重要采样,并涵盖了各种不同的外观,如各向异性和单独着色的颗粒。我们提供了一个基本实现作为一个独立的片段着色器。
{"title":"Evaluating and Sampling Glinty NDFs in Constant Time","authors":"Pauli Kemppinen, Loïs Paulin, Théo Thonat, Jean-Marc Thiery, Jaakko Lehtinen, Tamy Boubekeur","doi":"10.1145/3763282","DOIUrl":"https://doi.org/10.1145/3763282","url":null,"abstract":"Geometric features between the micro and macro scales produce an expressive family of visual effects grouped under the term \"glints\". Efficiently rendering these effects amounts to finding the highlights caused by the geometry under each pixel. To allow for fast rendering, we represent our faceted geometry as a 4D point process on an implicit multiscale grid, designed to efficiently find the facets most likely to cause a highlight. The facets' normals are generated to match a given micro-facet normal distribution such as Trowbridge-Reitz (GGX) or Beckmann, to which our model converges under increasing surface area. Our method is simple to implement, memory-and-precomputation-free, allows for importance sampling and covers a wide range of different appearances such as anisotropic as well as individually colored particles. We provide a base implementation as a standalone fragment shader.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"1 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673860","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ConsiStyle: Style Diversity in Training-Free Consistent T2I Generation 一致性:不需要训练的一致性T2I世代的风格多样性
IF 6.2 1区 计算机科学 Q1 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2025-12-04 DOI: 10.1145/3763303
Yohai Mazuz, Janna Bruner, Lior Wolf
In text-to-image models, consistent character generation is the task of achieving text alignment while maintaining the subject's appearance across different prompts. However, since style and appearance are often entangled, the existing methods struggle to preserve consistent subject characteristics while adhering to varying style prompts. Current approaches for consistent text-to-image generation typically rely on large-scale fine-tuning on curated image sets or per-subject optimization, which either fail to generalize across prompts or do not align well with textual descriptions. Meanwhile, training-free methods often fail to maintain subject consistency across different styles. In this work, we introduce a training-free method that, for the first time, jointly achieves style preservation and subject consistency across varied styles. The attention matrices are manipulated such that Queries and Keys are obtained from the anchor image(s) that are used to define the subject, while the Values are imported from a parallel copy that is not subject-anchored. Additionally, cross-image components are added to the self-attention mechanism by expanding the Key and Value matrices. To do without shifting from the target style, we align the statistics of the Value matrices. As is demonstrated in a comprehensive battery of qualitative and quantitative experiments, our method effectively decouples style from subject appearance and enables faithful generation of text-aligned images with consistent characters across diverse styles. Code will be available at our project page: jbruner23.github.io/consistyle.
在文本到图像模型中,一致的字符生成任务是实现文本对齐,同时在不同的提示中保持主题的外观。然而,由于风格和外观经常纠缠在一起,现有的方法很难保持一致的主题特征,同时坚持不同的风格提示。当前用于一致的文本到图像生成的方法通常依赖于对策划的图像集或每个主题的优化进行大规模微调,这些方法要么无法在提示间进行泛化,要么不能很好地与文本描述对齐。同时,无训练的方法往往不能保持不同风格的主题一致性。在这项工作中,我们引入了一种无需训练的方法,首次在不同风格之间共同实现了风格保留和主题一致性。对注意力矩阵进行操作,以便从用于定义主题的锚定图像中获得查询和键,而从非主题锚定的并行副本中导入值。此外,通过扩展键和值矩阵,将交叉图像组件添加到自关注机制中。为了不改变目标样式,我们对齐了Value矩阵的统计数据。正如定性和定量实验所证明的那样,我们的方法有效地将风格与主题外观分离开来,并能够忠实地生成具有不同风格的一致字符的文本对齐图像。代码将在我们的项目页面上提供:jbruner23.github.io/consistyle。
{"title":"ConsiStyle: Style Diversity in Training-Free Consistent T2I Generation","authors":"Yohai Mazuz, Janna Bruner, Lior Wolf","doi":"10.1145/3763303","DOIUrl":"https://doi.org/10.1145/3763303","url":null,"abstract":"In text-to-image models, consistent character generation is the task of achieving text alignment while maintaining the subject's appearance across different prompts. However, since style and appearance are often entangled, the existing methods struggle to preserve consistent subject characteristics while adhering to varying style prompts. Current approaches for consistent text-to-image generation typically rely on large-scale fine-tuning on curated image sets or per-subject optimization, which either fail to generalize across prompts or do not align well with textual descriptions. Meanwhile, training-free methods often fail to maintain subject consistency across different styles. In this work, we introduce a training-free method that, for the first time, jointly achieves style preservation and subject consistency across varied styles. The attention matrices are manipulated such that Queries and Keys are obtained from the anchor image(s) that are used to define the subject, while the Values are imported from a parallel copy that is not subject-anchored. Additionally, cross-image components are added to the self-attention mechanism by expanding the Key and Value matrices. To do without shifting from the target style, we align the statistics of the Value matrices. As is demonstrated in a comprehensive battery of qualitative and quantitative experiments, our method effectively decouples style from subject appearance and enables faithful generation of text-aligned images with consistent characters across diverse styles. Code will be available at our project page: jbruner23.github.io/consistyle.","PeriodicalId":50913,"journal":{"name":"ACM Transactions on Graphics","volume":"4 1","pages":""},"PeriodicalIF":6.2,"publicationDate":"2025-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145673863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
ACM Transactions on Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1