首页 > 最新文献

Computer Graphics Forum最新文献

英文 中文
GSEditPro: 3D Gaussian Splatting Editing with Attention-based Progressive Localization GSEditPro:利用基于注意力的渐进定位进行三维高斯拼接编辑
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-11-04 DOI: 10.1111/cgf.15215
Y. Sun, R. Tian, X. Han, X. Liu, Y. Zhang, K. Xu

With the emergence of large-scale Text-to-Image(T2I) models and implicit 3D representations like Neural Radiance Fields (NeRF), many text-driven generative editing methods based on NeRF have appeared. However, the implicit encoding of geometric and textural information poses challenges in accurately locating and controlling objects during editing. Recently, significant advancements have been made in the editing methods of 3D Gaussian Splatting, a real-time rendering technology that relies on explicit representation. However, these methods still suffer from issues including inaccurate localization and limited manipulation over editing. To tackle these challenges, we propose GSEditPro, a novel 3D scene editing framework which allows users to perform various creative and precise editing using text prompts only. Leveraging the explicit nature of the 3D Gaussian distribution, we introduce an attention-based progressive localization module to add semantic labels to each Gaussian during rendering. This enables precise localization on editing areas by classifying Gaussians based on their relevance to the editing prompts derived from cross-attention layers of the T2I model. Furthermore, we present an innovative editing optimization method based on 3D Gaussian Splatting, obtaining stable and refined editing results through the guidance of Score Distillation Sampling and pseudo ground truth. We prove the efficacy of our method through extensive experiments.

随着大规模文本到图像(T2I)模型和神经辐射场(NeRF)等隐式三维表示法的出现,出现了许多基于 NeRF 的文本驱动生成编辑方法。然而,几何和纹理信息的隐式编码给编辑过程中准确定位和控制对象带来了挑战。最近,三维高斯拼接的编辑方法取得了重大进展,这是一种依赖于显式表示的实时渲染技术。然而,这些方法仍然存在定位不准确、编辑操作受限等问题。为了应对这些挑战,我们提出了 GSEditPro,这是一种新颖的三维场景编辑框架,用户只需使用文本提示即可进行各种创造性的精确编辑。利用三维高斯分布的显式特性,我们引入了基于注意力的渐进式定位模块,在渲染过程中为每个高斯添加语义标签。这样就能根据高斯与 T2I 模型交叉注意力层中的编辑提示的相关性对高斯进行分类,从而实现编辑区域的精确定位。此外,我们还提出了一种基于三维高斯拼接的创新编辑优化方法,通过分数蒸馏采样和伪地面实况的指导,获得稳定而精细的编辑结果。我们通过大量实验证明了我们方法的有效性。
{"title":"GSEditPro: 3D Gaussian Splatting Editing with Attention-based Progressive Localization","authors":"Y. Sun,&nbsp;R. Tian,&nbsp;X. Han,&nbsp;X. Liu,&nbsp;Y. Zhang,&nbsp;K. Xu","doi":"10.1111/cgf.15215","DOIUrl":"https://doi.org/10.1111/cgf.15215","url":null,"abstract":"<p>With the emergence of large-scale Text-to-Image(T2I) models and implicit 3D representations like Neural Radiance Fields (NeRF), many text-driven generative editing methods based on NeRF have appeared. However, the implicit encoding of geometric and textural information poses challenges in accurately locating and controlling objects during editing. Recently, significant advancements have been made in the editing methods of 3D Gaussian Splatting, a real-time rendering technology that relies on explicit representation. However, these methods still suffer from issues including inaccurate localization and limited manipulation over editing. To tackle these challenges, we propose GSEditPro, a novel 3D scene editing framework which allows users to perform various creative and precise editing using text prompts only. Leveraging the explicit nature of the 3D Gaussian distribution, we introduce an attention-based progressive localization module to add semantic labels to each Gaussian during rendering. This enables precise localization on editing areas by classifying Gaussians based on their relevance to the editing prompts derived from cross-attention layers of the T2I model. Furthermore, we present an innovative editing optimization method based on 3D Gaussian Splatting, obtaining stable and refined editing results through the guidance of Score Distillation Sampling and pseudo ground truth. We prove the efficacy of our method through extensive experiments.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiscale Spectral Manifold Wavelet Regularizer for Unsupervised Deep Functional Maps 用于无监督深度函数图谱的多尺度光谱频谱小波规整器
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-11-04 DOI: 10.1111/cgf.15230
Shengjun Liu, Jing Meng, Ling Hu, Yueyu Guo, Xinru Liu, Xiaoxia Yang, Haibo Wang, Qinsong Li

In deep functional maps, the regularizer computing the functional map is especially crucial for ensuring the global consistency of the computed pointwise map. As the regularizers integrated into deep learning should be differentiable, it is not trivial to incorporate informative axiomatic structural constraints into the deep functional map, such as the orientation-preserving term. Although commonly used regularizers include the Laplacian-commutativity term and the resolvent Laplacian commutativity term, these are limited to single-scale analysis for capturing geometric information. To this end, we propose a novel and theoretically well-justified regularizer commuting the functional map with the multiscale spectral manifold wavelet operator. This regularizer enhances the isometric constraints of the functional map and is conducive to providing it with better structural properties with multiscale analysis. Furthermore, we design an unsupervised deep functional map with the regularizer in a fully differentiable way. The quantitative and qualitative comparisons with several existing techniques on the (near-)isometric and non-isometric datasets show our method's superior accuracy and generalization capabilities. Additionally, we illustrate that our regularizer can be easily inserted into other functional map methods and improve their accuracy.

在深度函数图谱中,计算函数图谱的正则化器对于确保所计算的点阵图谱的全局一致性尤为重要。由于集成到深度学习中的正则应该是可微分的,因此将信息公理结构约束(如方向保持项)纳入深度函数图并非易事。虽然常用的正则包括拉普拉斯换向项和解析拉普拉斯换向项,但这些正则仅限于捕捉几何信息的单尺度分析。为此,我们提出了一种新颖的、理论上合理的正则表达式,将函数图与多尺度谱流形小波算子相换算。该正则化器增强了函数图的等距约束,有利于通过多尺度分析为函数图提供更好的结构特性。此外,我们还以完全可微分的方式设计了带有正则化器的无监督深度函数图。在(近)等距和非等距数据集上与几种现有技术进行的定量和定性比较表明,我们的方法具有卓越的准确性和泛化能力。此外,我们还说明,我们的正则化器可以很容易地插入到其他函数图方法中,并提高它们的准确性。
{"title":"Multiscale Spectral Manifold Wavelet Regularizer for Unsupervised Deep Functional Maps","authors":"Shengjun Liu,&nbsp;Jing Meng,&nbsp;Ling Hu,&nbsp;Yueyu Guo,&nbsp;Xinru Liu,&nbsp;Xiaoxia Yang,&nbsp;Haibo Wang,&nbsp;Qinsong Li","doi":"10.1111/cgf.15230","DOIUrl":"https://doi.org/10.1111/cgf.15230","url":null,"abstract":"<p>In deep functional maps, the regularizer computing the functional map is especially crucial for ensuring the global consistency of the computed pointwise map. As the regularizers integrated into deep learning should be differentiable, it is not trivial to incorporate informative axiomatic structural constraints into the deep functional map, such as the orientation-preserving term. Although commonly used regularizers include the Laplacian-commutativity term and the resolvent Laplacian commutativity term, these are limited to single-scale analysis for capturing geometric information. To this end, we propose a novel and theoretically well-justified regularizer commuting the functional map with the multiscale spectral manifold wavelet operator. This regularizer enhances the isometric constraints of the functional map and is conducive to providing it with better structural properties with multiscale analysis. Furthermore, we design an unsupervised deep functional map with the regularizer in a fully differentiable way. The quantitative and qualitative comparisons with several existing techniques on the (near-)isometric and non-isometric datasets show our method's superior accuracy and generalization capabilities. Additionally, we illustrate that our regularizer can be easily inserted into other functional map methods and improve their accuracy.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664656","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distinguishing Structures from Textures by Patch-based Contrasts around Pixels for High-quality and Efficient Texture filtering 通过像素周围基于斑块的对比度从纹理中区分结构,实现高质量、高效率的纹理过滤
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-11-04 DOI: 10.1111/cgf.15212
Shengchun Wang, Panpan Xu, Fei Hou, Wencheng Wang, Chong Zhao

It is still challenging with existing methods to distinguish structures from texture details, and so preventing texture filtering. Considering that the textures on both sides of a structural edge always differ much from each other in appearances, we determine whether a pixel is on a structure edge by exploiting the appearance contrast between patches around the pixel, and further propose an efficient implementation method. We demonstrate that our proposed method is more effective than existing methods to distinguish structures from texture details, and our required patches for texture measurement can be smaller than the used patches in existing methods by at least half. Thus, we can improve texture filtering on both quality and efficiency, as shown by the experimental results, e.g., we can handle the textured images with a resolution of 800 × 600 pixels in real-time. (The code is available at https://github.com/hefengxiyulu/MLPC)

现有方法仍难以区分结构和纹理细节,因此无法进行纹理过滤。考虑到结构边缘两侧的纹理在外观上总是相差很大,我们通过利用像素周围斑块的外观对比来判断像素是否位于结构边缘,并进一步提出了一种高效的实现方法。结果表明,与现有方法相比,我们提出的方法能更有效地从纹理细节中区分出结构,而且我们测量纹理所需的斑块比现有方法所用的斑块至少小一半。因此,正如实验结果所示,我们可以在质量和效率两方面提高纹理过滤效果,例如,我们可以实时处理分辨率为 800 × 600 像素的纹理图像。(代码见 https://github.com/hefengxiyulu/MLPC)
{"title":"Distinguishing Structures from Textures by Patch-based Contrasts around Pixels for High-quality and Efficient Texture filtering","authors":"Shengchun Wang,&nbsp;Panpan Xu,&nbsp;Fei Hou,&nbsp;Wencheng Wang,&nbsp;Chong Zhao","doi":"10.1111/cgf.15212","DOIUrl":"https://doi.org/10.1111/cgf.15212","url":null,"abstract":"<p>It is still challenging with existing methods to distinguish structures from texture details, and so preventing texture filtering. Considering that the textures on both sides of a structural edge always differ much from each other in appearances, we determine whether a pixel is on a structure edge by exploiting the appearance contrast between patches around the pixel, and further propose an efficient implementation method. We demonstrate that our proposed method is more effective than existing methods to distinguish structures from texture details, and our required patches for texture measurement can be smaller than the used patches in existing methods by at least half. Thus, we can improve texture filtering on both quality and efficiency, as shown by the experimental results, e.g., we can handle the textured images with a resolution of 800 × 600 pixels in real-time. (The code is available at https://github.com/hefengxiyulu/MLPC)</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142664657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Ray Tracing Animated Displaced Micro-Meshes 光线追踪动画位移微切口
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-30 DOI: 10.1111/cgf.15225
Holger Gruen, Carsten Benthin, Andrew Kensler, Joshua Barczak, David McAllister

We present a new method that allows efficient ray tracing of virtually artefact-free animated displaced micro-meshes (DMMs) [MMT23] and preserves their low memory footprint and low BVH build and update cost. DMMs allow for compact representation of micro-triangle geometry through hierarchical encoding of displacements. Displacements are computed with respect to a coarse base mesh and are used to displace new vertices introduced during 1 : 4 subdivision of the base mesh. Applying non-rigid transformation to the base mesh can result in silhouette and normal artefacts (see Figure 1) during animation. We propose an approach which prevents these artefacts by interpolating transformation matrices before applying them to the DMM representation. Our interpolation-based algorithm does not change DMM data structures and it allows for efficient bounding of animated micro-triangle geometry which is essential for fast tessellation-free ray tracing of animated DMMs.

我们提出了一种新方法,可对几乎无伪影的动画位移微模型(DMMs)[MMT23]进行高效光线追踪,并保留其低内存占用率和低 BVH 构建与更新成本。DMM 通过对位移进行分层编码,可以紧凑地表示微三角形几何图形。位移是相对于粗略的基础网格计算的,用于对基础网格进行 1 : 4 细分时引入的新顶点进行位移。在动画制作过程中,对基本网格进行非刚性变换可能会产生轮廓和法线伪影(见图 1)。我们提出了一种方法,在将变换矩阵应用到 DMM 表示之前对其进行插值,从而防止出现这些假象。我们基于插值的算法不会改变 DMM 的数据结构,而且可以高效地限定动画微三角形几何图形,这对于快速无细分光线追踪动画 DMM 至关重要。
{"title":"Ray Tracing Animated Displaced Micro-Meshes","authors":"Holger Gruen,&nbsp;Carsten Benthin,&nbsp;Andrew Kensler,&nbsp;Joshua Barczak,&nbsp;David McAllister","doi":"10.1111/cgf.15225","DOIUrl":"https://doi.org/10.1111/cgf.15225","url":null,"abstract":"<p>We present a new method that allows efficient ray tracing of virtually artefact-free animated displaced micro-meshes (DMMs) [MMT23] and preserves their low memory footprint and low BVH build and update cost. DMMs allow for compact representation of micro-triangle geometry through hierarchical encoding of displacements. Displacements are computed with respect to a coarse base mesh and are used to displace new vertices introduced during <i>1 : 4</i> subdivision of the base mesh. Applying non-rigid transformation to the base mesh can result in silhouette and normal artefacts (see Figure 1) during animation. We propose an approach which prevents these artefacts by interpolating transformation matrices before applying them to the DMM representation. Our interpolation-based algorithm does not change DMM data structures and it allows for efficient bounding of animated micro-triangle geometry which is essential for fast tessellation-free ray tracing of animated DMMs.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Anisotropic Specular Image-Based Lighting Based on BRDF Major Axis Sampling 基于 BRDF 主轴采样的各向异性镜面图像照明
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-30 DOI: 10.1111/cgf.15233
Giovanni Cocco, Cédric Zanni, Xavier Chermain

Anisotropic specular appearances are ubiquitous in the environment: brushed stainless steel pans, kettles, elevator walls, fur, or scratched plastics. Real-time rendering of these materials with image-based lighting is challenging due to the complex shape of the bidirectional reflectance distribution function (BRDF). We propose an anisotropic specular image-based lighting method that can serve as a drop-in replacement for the standard bent normal technique [Rev11]. Our method yields more realistic results with a 50% increase in computation time of the previous technique, using the same high dynamic range (HDR) preintegrated environment image. We use several environment samples positioned along the major axis of the specular microfacet BRDF. We derive an analytic formula to determine the two closest and two farthest points from the reflected direction on an approximation of the BRDF confidence region boundary. The two farthest points define the BRDF major axis, while the two closest points are used to approximate the BRDF width. The environment level of detail is derived from the BRDF width and the distance between the samples. We extensively compare our method with the bent normal technique and the ground truth using the GGX specular BRDF.

各向异性镜面外观在环境中无处不在:拉丝不锈钢锅、水壶、电梯墙壁、毛皮或刮伤的塑料。由于双向反射分布函数(BRDF)形状复杂,使用基于图像的照明技术实时渲染这些材料具有挑战性。我们提出了一种基于各向异性镜面反射图像的照明方法,可以直接替代标准的弯曲法线技术 [Rev11]。使用相同的高动态范围(HDR)预集成环境图像,我们的方法能产生更逼真的结果,而计算时间只比前一种技术增加 50%。我们使用沿镜面微面 BRDF 主轴定位的多个环境样本。我们推导出一个解析公式,用于确定 BRDF 置信区域边界近似值上距离反射方向最近和最远的两个点。最远的两个点定义 BRDF 主轴,而最近的两个点则用于近似 BRDF 宽度。环境详细程度由 BRDF 宽度和样本之间的距离得出。我们使用 GGX 镜面 BRDF 将我们的方法与弯曲法线技术和地面实况进行了广泛比较。
{"title":"Anisotropic Specular Image-Based Lighting Based on BRDF Major Axis Sampling","authors":"Giovanni Cocco,&nbsp;Cédric Zanni,&nbsp;Xavier Chermain","doi":"10.1111/cgf.15233","DOIUrl":"https://doi.org/10.1111/cgf.15233","url":null,"abstract":"<p>Anisotropic specular appearances are ubiquitous in the environment: brushed stainless steel pans, kettles, elevator walls, fur, or scratched plastics. Real-time rendering of these materials with image-based lighting is challenging due to the complex shape of the bidirectional reflectance distribution function (BRDF). We propose an anisotropic specular image-based lighting method that can serve as a drop-in replacement for the standard bent normal technique [Rev11]. Our method yields more realistic results with a 50% increase in computation time of the previous technique, using the same high dynamic range (HDR) preintegrated environment image. We use several environment samples positioned along the major axis of the specular microfacet BRDF. We derive an analytic formula to determine the two closest and two farthest points from the reflected direction on an approximation of the BRDF confidence region boundary. The two farthest points define the BRDF major axis, while the two closest points are used to approximate the BRDF width. The environment level of detail is derived from the BRDF width and the distance between the samples. We extensively compare our method with the bent normal technique and the ground truth using the GGX specular BRDF.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665194","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Variable offsets and processing of implicit forms toward the adaptive synthesis and analysis of heterogeneous conforming microstructure 可变偏移和隐式处理,实现异质保形微结构的自适应合成和分析
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-30 DOI: 10.1111/cgf.15224
Q. Y. Hong, P. Antolin, G. Elber, M.-S. Kim

The synthesis of porous, lattice, or microstructure geometries has captured the attention of many researchers in recent years. Implicit forms, such as triply periodic minimal surfaces (TPMS) has captured a significant attention, recently, as tiles in lattices, partially because implicit forms have the potential for synthesizing with ease more complex topologies of tiles, compared to parametric forms. In this work, we show how variable offsets of implicit forms could be used in lattice design as well as lattice analysis, while graded wall and edge thicknesses could be fully controlled in the lattice and even vary within a single tile. As a result, (geometrically) heterogeneous lattices could be created and adapted to follow analysis results while maintaining continuity between adjacent tiles. We demonstrate this ability on several 3D models, including TPMS.

近年来,多孔、晶格或微结构几何形状的合成吸引了许多研究人员的目光。隐含形式,如三重周期性极小表面(TPMS),作为晶格中的瓦片,最近吸引了大量关注,部分原因是与参数形式相比,隐含形式有可能更容易合成更复杂的瓦片拓扑结构。在这项研究中,我们展示了如何在晶格设计和晶格分析中使用隐含形式的可变偏移,同时在晶格中完全控制分级的壁厚和边厚,甚至在单个瓦片中也可以变化。因此,可以创建(几何)异质晶格,并根据分析结果进行调整,同时保持相邻晶格之间的连续性。我们在多个三维模型(包括 TPMS)上演示了这种能力。
{"title":"Variable offsets and processing of implicit forms toward the adaptive synthesis and analysis of heterogeneous conforming microstructure","authors":"Q. Y. Hong,&nbsp;P. Antolin,&nbsp;G. Elber,&nbsp;M.-S. Kim","doi":"10.1111/cgf.15224","DOIUrl":"https://doi.org/10.1111/cgf.15224","url":null,"abstract":"<p>The synthesis of porous, lattice, or microstructure geometries has captured the attention of many researchers in recent years. Implicit forms, such as triply periodic minimal surfaces (TPMS) has captured a significant attention, recently, as tiles in lattices, partially because implicit forms have the potential for synthesizing with ease more complex topologies of tiles, compared to parametric forms. In this work, we show how variable offsets of implicit forms could be used in lattice design as well as lattice analysis, while graded wall and edge thicknesses could be fully controlled in the lattice and even vary within a single tile. As a result, (geometrically) heterogeneous lattices could be created and adapted to follow analysis results while maintaining continuity between adjacent tiles. We demonstrate this ability on several 3D models, including TPMS.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MISNeR: Medical Implicit Shape Neural Representation for Image Volume Visualisation MISNeR:用于图像体积可视化的医学隐含形状神经表示法
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-30 DOI: 10.1111/cgf.15222
G. Jin, Y. Jung, L. Bi, J. Kim

Three-dimensional visualisation of mesh reconstruction of the medical images is commonly used for various clinical applications including pre / post-surgical planning. Such meshes are conventionally generated by extracting the surface from volumetric segmentation masks. Therefore, they have inherent limitations of staircase artefacts due to their anisotropic voxel dimensions. The time-consuming process for manual refinement to remove artefacts and/or the isolated regions further adds to these limitations. Methods for directly generating meshes from volumetric data by template deformation are often limited to simple topological structures, and methods that use implicit functions for continuous surfaces, do not achieve the level of mesh reconstruction accuracy when compared to segmentation-based methods. In this study, we address these limitations by combining the implicit function representation with a multi-level deep learning architecture. We introduce a novel multi-level local feature sampling component which leverages the spatial features for the implicit function regression to enhance the segmentation result. We further introduce a shape boundary estimator that accelerates the explicit mesh reconstruction by minimising the number of the signed distance queries during model inference. The result is a multi-level deep learning network that directly regresses the implicit function from medical image volumes to a continuous surface model, which can be used for mesh reconstruction from arbitrary high volume resolution to minimise staircase artefacts. We evaluated our method using pelvic computed tomography (CT) dataset from two public sources with varying z-axis resolutions. We show that our method minimised the staircase artefacts while achieving comparable results in surface accuracy when compared to the state-of-the-art segmentation algorithms. Furthermore, our method was 9 times faster in volume reconstruction than comparable implicit shape representation networks.

医学图像的三维可视化网格重建通常用于各种临床应用,包括手术前/后规划。这些网格通常是通过从体积分割掩膜中提取表面而生成的。因此,由于其各向异性的体素尺寸,它们具有阶梯假象的固有局限性。为去除伪影和/或孤立区域而进行的耗时的手动细化过程进一步增加了这些局限性。通过模板变形从体积数据中直接生成网格的方法通常局限于简单的拓扑结构,而使用隐式函数生成连续曲面的方法与基于分割的方法相比,无法达到网格重建的精度水平。在本研究中,我们通过将隐函数表示法与多层次深度学习架构相结合来解决这些局限性。我们引入了一个新颖的多层次局部特征采样组件,利用隐函数回归的空间特征来增强分割结果。我们还进一步引入了形状边界估计器,通过在模型推理过程中尽量减少带符号距离查询的次数来加速显式网格重建。由此产生的多层次深度学习网络可直接将医学影像体积中的隐函数回归到连续曲面模型,该模型可用于任意高体积分辨率的网格重建,从而最大限度地减少阶梯伪影。我们使用两个公开来源的骨盆计算机断层扫描(CT)数据集对我们的方法进行了评估,这些数据集的 Z 轴分辨率各不相同。结果表明,与最先进的分割算法相比,我们的方法最大限度地减少了阶梯伪影,同时在表面精度方面取得了相当的结果。此外,我们的方法在体积重建方面比同类隐式形状表示网络快 9 倍。
{"title":"MISNeR: Medical Implicit Shape Neural Representation for Image Volume Visualisation","authors":"G. Jin,&nbsp;Y. Jung,&nbsp;L. Bi,&nbsp;J. Kim","doi":"10.1111/cgf.15222","DOIUrl":"https://doi.org/10.1111/cgf.15222","url":null,"abstract":"<p>Three-dimensional visualisation of mesh reconstruction of the medical images is commonly used for various clinical applications including pre / post-surgical planning. Such meshes are conventionally generated by extracting the surface from volumetric segmentation masks. Therefore, they have inherent limitations of staircase artefacts due to their anisotropic voxel dimensions. The time-consuming process for manual refinement to remove artefacts and/or the isolated regions further adds to these limitations. Methods for directly generating meshes from volumetric data by template deformation are often limited to simple topological structures, and methods that use implicit functions for continuous surfaces, do not achieve the level of mesh reconstruction accuracy when compared to segmentation-based methods. In this study, we address these limitations by combining the implicit function representation with a multi-level deep learning architecture. We introduce a novel multi-level local feature sampling component which leverages the spatial features for the implicit function regression to enhance the segmentation result. We further introduce a shape boundary estimator that accelerates the explicit mesh reconstruction by minimising the number of the signed distance queries during model inference. The result is a multi-level deep learning network that directly regresses the implicit function from medical image volumes to a continuous surface model, which can be used for mesh reconstruction from arbitrary high volume resolution to minimise staircase artefacts. We evaluated our method using pelvic computed tomography (CT) dataset from two public sources with varying z-axis resolutions. We show that our method minimised the staircase artefacts while achieving comparable results in surface accuracy when compared to the state-of-the-art segmentation algorithms. Furthermore, our method was 9 times faster in volume reconstruction than comparable implicit shape representation networks.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FSH3D: 3D Representation via Fibonacci Spherical Harmonics FSH3D:通过斐波那契球面谐波进行 3D 表示
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-24 DOI: 10.1111/cgf.15231
Zikuan Li, Anyi Huang, Wenru Jia, Qiaoyun Wu, Mingqiang Wei, Jun Wang

Spherical harmonics are a favorable technique for 3D representation, employing a frequency-based approach through the spherical harmonic transform (SHT). Typically, SHT is performed using equiangular sampling grids. However, these grids are non-uniform on spherical surfaces and exhibit local anisotropy, a common limitation in existing spherical harmonic decomposition methods. This paper proposes a 3D representation method using Fibonacci Spherical Harmonics (FSH3D). We introduce a spherical Fibonacci grid (SFG), which is more uniform than equiangular grids for SHT in the frequency domain. Our method employs analytical weights for SHT on SFG, effectively assigning sampling errors to spherical harmonic degrees higher than the recovered band-limited function. This provides a novel solution for spherical harmonic transformation on non-equiangular grids. The key advantages of our FSH3D method include: 1) With the same number of sampling points, SFG captures more features without bias compared to equiangular grids; 2) The root mean square error of 32-degree spherical harmonic coefficients is reduced by approximately 34.6% for SFG compared to equiangular grids; and 3) FSH3D offers more stable frequency domain representations, especially for rotating functions. FSH3D enhances the stability of frequency domain representations under rotational transformations. Its application in 3D shape reconstruction and 3D shape classification results in more accurate and robust representations. Our code is publicly available at https://github.com/Miraclelzk/Fibonacci-Spherical-Harmonics.

球面谐波是三维表示的一种有利技术,它通过球面谐波变换(SHT)采用基于频率的方法。通常,SHT 采用等角采样网格。然而,这些网格在球面上是不均匀的,并表现出局部各向异性,这是现有球谐波分解方法的一个共同局限。本文提出了一种使用斐波那契球面谐波(FSH3D)的三维表示方法。我们引入了一种球形斐波那契网格(SFG),它比等边网格更均匀,可用于频域中的 SHT。我们的方法采用了 SFG 上 SHT 的分析权重,有效地将采样误差分配给高于恢复带限函数的球谐波度。这为非等边网格上的球谐波变换提供了一种新的解决方案。我们的 FSH3D 方法的主要优势包括1) 与等边网格相比,在相同的采样点数下,SFG 能捕捉到更多的特征而不会产生偏差;2) 与等边网格相比,SFG 的 32 度球谐波系数的均方根误差降低了约 34.6%;3) FSH3D 提供了更稳定的频域表示,尤其是对于旋转函数。FSH3D 增强了频域表示在旋转变换下的稳定性。它在三维形状重建和三维形状分类中的应用能带来更准确、更稳健的表示。我们的代码可在 https://github.com/Miraclelzk/Fibonacci-Spherical-Harmonics 公开获取。
{"title":"FSH3D: 3D Representation via Fibonacci Spherical Harmonics","authors":"Zikuan Li,&nbsp;Anyi Huang,&nbsp;Wenru Jia,&nbsp;Qiaoyun Wu,&nbsp;Mingqiang Wei,&nbsp;Jun Wang","doi":"10.1111/cgf.15231","DOIUrl":"https://doi.org/10.1111/cgf.15231","url":null,"abstract":"<p>Spherical harmonics are a favorable technique for 3D representation, employing a frequency-based approach through the spherical harmonic transform (SHT). Typically, SHT is performed using equiangular sampling grids. However, these grids are non-uniform on spherical surfaces and exhibit local anisotropy, a common limitation in existing spherical harmonic decomposition methods. This paper proposes a 3D representation method using Fibonacci Spherical Harmonics (FSH3D). We introduce a spherical Fibonacci grid (SFG), which is more uniform than equiangular grids for SHT in the frequency domain. Our method employs analytical weights for SHT on SFG, effectively assigning sampling errors to spherical harmonic degrees higher than the recovered band-limited function. This provides a novel solution for spherical harmonic transformation on non-equiangular grids. The key advantages of our FSH3D method include: 1) With the same number of sampling points, SFG captures more features without bias compared to equiangular grids; 2) The root mean square error of 32-degree spherical harmonic coefficients is reduced by approximately 34.6% for SFG compared to equiangular grids; and 3) FSH3D offers more stable frequency domain representations, especially for rotating functions. FSH3D enhances the stability of frequency domain representations under rotational transformations. Its application in 3D shape reconstruction and 3D shape classification results in more accurate and robust representations. Our code is publicly available at https://github.com/Miraclelzk/Fibonacci-Spherical-Harmonics.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665131","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Disk B-spline on 𝕊2: A Skeleton-based Region Representation on 2-Sphere ᵔ2上的圆盘B样条曲线:基于骨架的2球面区域表示法
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-24 DOI: 10.1111/cgf.15239
Chunhao Zheng, Yuming Zhao, Zhongke Wu, Xingce Wang

Due to the widespread applications of 2-dimensional spherical designs, there has been an increasing requirement of modeling on the 𝕊2 manifold in recent years. Due to the non-Euclidean nature of the sphere, it has some challenges to find a method to represent 2D regions on 𝕊2 manifold. In this paper, a skeleton-based representation method of regions on 𝕊2, disk B-spline(DBSC) on 𝕊2 is proposed. Firstly, we give the definition and basic algorithms of DBSC on 𝕊2. Then we provide the calculation method of DBSC on 𝕊2, which includes calculating the boundary points, internal points and their corresponding derivatives. Based on that, we give some modeling methods of DBSC on 𝕊2, including approximation, deformation. In the end, some stunning application examples of DBSC on 𝕊2 are shown. This work lays a theoretical foundation for further applications of DBSC on 𝕊2.

由于二维球面设计的广泛应用,近年来对ᵔ2 流形建模的要求越来越高。由于球体的非欧几里得性质,要找到一种在ᵔ2 流形上表示二维区域的方法有一定的难度。本文提出了一种基于骨架的ᵔ2 上区域表示方法--ᵔ2 上的圆盘 B 样条(DBSC)。首先,我们给出了ᵔ2 上 DBSC 的定义和基本算法。然后,我们给出了ᵔ2 上 DBSC 的计算方法,包括边界点、内部点及其相应导数的计算。在此基础上,我们给出了ᵔ2 上 DBSC 的一些建模方法,包括近似、变形等。最后,展示了一些令人惊叹的ᵔ2 上 DBSC 的应用实例。这项工作为 DBSC 在ᵔ2 上的进一步应用奠定了理论基础。
{"title":"Disk B-spline on 𝕊2: A Skeleton-based Region Representation on 2-Sphere","authors":"Chunhao Zheng,&nbsp;Yuming Zhao,&nbsp;Zhongke Wu,&nbsp;Xingce Wang","doi":"10.1111/cgf.15239","DOIUrl":"https://doi.org/10.1111/cgf.15239","url":null,"abstract":"<p>Due to the widespread applications of 2-dimensional spherical designs, there has been an increasing requirement of modeling on the 𝕊<sup>2</sup> manifold in recent years. Due to the non-Euclidean nature of the sphere, it has some challenges to find a method to represent 2D regions on 𝕊<sup>2</sup> manifold. In this paper, a skeleton-based representation method of regions on 𝕊<sup>2</sup>, disk B-spline(DBSC) on 𝕊<sup>2</sup> is proposed. Firstly, we give the definition and basic algorithms of DBSC on 𝕊<sup>2</sup>. Then we provide the calculation method of DBSC on 𝕊<sup>2</sup>, which includes calculating the boundary points, internal points and their corresponding derivatives. Based on that, we give some modeling methods of DBSC on 𝕊<sup>2</sup>, including approximation, deformation. In the end, some stunning application examples of DBSC on 𝕊<sup>2</sup> are shown. This work lays a theoretical foundation for further applications of DBSC on 𝕊<sup>2</sup>.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665136","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Exploring Fast and Flexible Zero-Shot Low-Light Image/Video Enhancement 探索快速灵活的零镜头低照度图像/视频增强技术
IF 2.7 4区 计算机科学 Q2 COMPUTER SCIENCE, SOFTWARE ENGINEERING Pub Date : 2024-10-24 DOI: 10.1111/cgf.15210
Xianjun Han, Taoli Bao, Hongyu Yang

Low-light image/video enhancement is a challenging task when images or video are captured under harsh lighting conditions. Existing methods mostly formulate this task as an image-to-image conversion task via supervised or unsupervised learning. However, such conversion methods require an extremely large amount of data for training, whether paired or unpaired. In addition, these methods are restricted to specific training data, making it difficult for the trained model to enhance other types of images or video. In this paper, we explore a novel, fast and flexible, zero-shot, low-light image or video enhancement framework. Without relying on prior training or relationships among neighboring frames, we are committed to estimating the illumination of the input image/frame by a well-designed network. The proposed zero-shot, low-light image/video enhancement architecture includes illumination estimation and residual correction modules. The network architecture is very concise and does not require any paired or unpaired data during training, which allows low-light enhancement to be performed with several simple iterations. Despite its simplicity, we show that the method is fast and generalizes well to diverse lighting conditions. Many experiments on various images and videos qualitatively and quantitatively demonstrate the advantages of our method over state-of-the-art methods.

低照度图像/视频增强是一项具有挑战性的任务,因为图像或视频是在恶劣的照明条件下拍摄的。现有的方法大多通过有监督或无监督学习将这项任务表述为图像到图像的转换任务。然而,这类转换方法需要大量数据进行训练,无论是配对数据还是非配对数据。此外,这些方法仅限于特定的训练数据,使得训练好的模型难以增强其他类型的图像或视频。在本文中,我们探索了一种新颖、快速、灵活、零镜头、低照度图像或视频增强框架。在不依赖事先训练或相邻帧之间关系的情况下,我们致力于通过精心设计的网络来估计输入图像/帧的光照度。所提出的零镜头、低照度图像/视频增强架构包括照度估计和残差校正模块。该网络架构非常简洁,在训练过程中不需要任何配对或非配对数据,因此只需进行几次简单的迭代即可实现弱光增强。尽管方法简单,但我们发现该方法不仅速度快,而且能很好地适应各种照明条件。在各种图像和视频上进行的大量实验从定性和定量两方面证明了我们的方法比最先进的方法更具优势。
{"title":"Exploring Fast and Flexible Zero-Shot Low-Light Image/Video Enhancement","authors":"Xianjun Han,&nbsp;Taoli Bao,&nbsp;Hongyu Yang","doi":"10.1111/cgf.15210","DOIUrl":"https://doi.org/10.1111/cgf.15210","url":null,"abstract":"<p>Low-light image/video enhancement is a challenging task when images or video are captured under harsh lighting conditions. Existing methods mostly formulate this task as an image-to-image conversion task via supervised or unsupervised learning. However, such conversion methods require an extremely large amount of data for training, whether paired or unpaired. In addition, these methods are restricted to specific training data, making it difficult for the trained model to enhance other types of images or video. In this paper, we explore a novel, fast and flexible, zero-shot, low-light image or video enhancement framework. Without relying on prior training or relationships among neighboring frames, we are committed to estimating the illumination of the input image/frame by a well-designed network. The proposed zero-shot, low-light image/video enhancement architecture includes illumination estimation and residual correction modules. The network architecture is very concise and does not require any paired or unpaired data during training, which allows low-light enhancement to be performed with several simple iterations. Despite its simplicity, we show that the method is fast and generalizes well to diverse lighting conditions. Many experiments on various images and videos qualitatively and quantitatively demonstrate the advantages of our method over state-of-the-art methods.</p>","PeriodicalId":10687,"journal":{"name":"Computer Graphics Forum","volume":"43 7","pages":""},"PeriodicalIF":2.7,"publicationDate":"2024-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142665175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
Computer Graphics Forum
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1