Pub Date : 2024-05-03DOI: 10.1016/j.cagd.2024.102290
Juan Gerardo Alcázar , Miroslav Lávička , Jan Vršek
In this paper, we address the computation of the symmetries of polynomial (and thus also rational) planar vector fields using elements from Computer Algebra. We show that they can be recovered from the symmetries of the roots of an associated univariate complex polynomial which is constructed as a generator of a certain elimination ideal. Computing symmetries of the roots of the auxiliary polynomial is a task considerably simpler than the original problem, which can be done efficiently working with classical Computer Algebra tools. Special cases, in which the group of symmetries of the polynomial roots is infinite, are separately considered and investigated. The presented theory is complemented by illustrative examples. The main steps of the procedure for investigating the symmetries of a given polynomial vector field are summarized in a flow chart for clarity.
{"title":"Symmetries of planar algebraic vector fields","authors":"Juan Gerardo Alcázar , Miroslav Lávička , Jan Vršek","doi":"10.1016/j.cagd.2024.102290","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102290","url":null,"abstract":"<div><p>In this paper, we address the computation of the symmetries of polynomial (and thus also rational) planar vector fields using elements from Computer Algebra. We show that they can be recovered from the symmetries of the roots of an associated univariate complex polynomial which is constructed as a generator of a certain elimination ideal. Computing symmetries of the roots of the auxiliary polynomial is a task considerably simpler than the original problem, which can be done efficiently working with classical Computer Algebra tools. Special cases, in which the group of symmetries of the polynomial roots is infinite, are separately considered and investigated. The presented theory is complemented by illustrative examples. The main steps of the procedure for investigating the symmetries of a given polynomial vector field are summarized in a flow chart for clarity.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102290"},"PeriodicalIF":1.5,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140824644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-03DOI: 10.1016/j.cagd.2024.102338
Kaikai Qin, Yajuan Li, Chongyang Deng
In recent years, there has been growing interest in the representation of volumes within the field of geometric modeling (GM). While polygonal patches for surface modeling have been extensively studied, there has been little focus on the representation of polyhedral volumes. Inspired by the polygonal representation of the Generalized Bézier (GB) patch proposed by Várady et al. (2016), this paper introduces a novel method for polyhedral volumetric modeling called the Generalized Bézier (GB) volume.
GB volumes are defined over simple convex polyhedra using generalized barycentric coordinates (GBCs), with the control nets which are a direct generalization of those of tensor-product Bézier volumes. GB volumes can be smoothly connected to adjacent tensor-product Bézier or GB volumes with or continuity. Besides, when the parametric polyhedron becomes a prism, the GB volume also degenerates into a tensor-product form. We provide some practical examples to demonstrate the advantages of GB volumes. Suggestions for future work are also discussed.
{"title":"Generalized Bézier volumes over simple convex polyhedra","authors":"Kaikai Qin, Yajuan Li, Chongyang Deng","doi":"10.1016/j.cagd.2024.102338","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102338","url":null,"abstract":"<div><p>In recent years, there has been growing interest in the representation of volumes within the field of geometric modeling (GM). While <em>polygonal</em> patches for surface modeling have been extensively studied, there has been little focus on the representation of <em>polyhedral</em> volumes. Inspired by the polygonal representation of the <em>Generalized Bézier (GB)</em> patch proposed by <span>Várady et al. (2016)</span>, this paper introduces a novel method for polyhedral volumetric modeling called the Generalized Bézier (GB) volume.</p><p>GB volumes are defined over <em>simple convex polyhedra</em> using <em>generalized barycentric coordinates</em> (GBCs), with the control nets which are a direct generalization of those of tensor-product Bézier volumes. GB volumes can be smoothly connected to adjacent tensor-product Bézier or GB volumes with <span><math><msup><mrow><mi>G</mi></mrow><mrow><mn>1</mn></mrow></msup></math></span> or <span><math><msup><mrow><mi>G</mi></mrow><mrow><mn>2</mn></mrow></msup></math></span> continuity. Besides, when the parametric polyhedron becomes a prism, the GB volume also degenerates into a tensor-product form. We provide some practical examples to demonstrate the advantages of GB volumes. Suggestions for future work are also discussed.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102338"},"PeriodicalIF":1.5,"publicationDate":"2024-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140878476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-05-01DOI: 10.1016/j.cagd.2024.102311
Lipeng Gu , Xuefeng Yan , Liangliang Nan , Dingkun Zhu , Honghua Chen , Weiming Wang , Mingqiang Wei
The conventional wisdom in point cloud analysis predominantly explores 3D geometries. It is often achieved through the introduction of intricate learnable geometric extractors in the encoder or by deepening networks with repeated blocks. However, these methods contain a significant number of learnable parameters, resulting in substantial computational costs and imposing memory burdens on CPU/GPU. Moreover, they are primarily tailored for object-level point cloud classification and segmentation tasks, with limited extensions to crucial scene-level applications, such as autonomous driving. To this end, we introduce PointeNet, an efficient network designed specifically for point cloud analysis. PointeNet distinguishes itself with its lightweight architecture, low training cost, and plug-and-play capability, while also effectively capturing representative features. The network consists of a Multivariate Geometric Encoding (MGE) module and an optional Distance-aware Semantic Enhancement (DSE) module. MGE employs operations of sampling, grouping, pooling, and multivariate geometric aggregation to lightweightly capture and adaptively aggregate multivariate geometric features, providing a comprehensive depiction of 3D geometries. DSE, designed for real-world autonomous driving scenarios, enhances the semantic perception of point clouds, particularly for distant points. Our method demonstrates flexibility by seamlessly integrating with a classification/segmentation head or embedding into off-the-shelf 3D object detection networks, achieving notable performance improvements at a minimal cost. Extensive experiments on object-level datasets, including ModelNet40, ScanObjectNN, ShapeNetPart, and the scene-level dataset KITTI, demonstrate the superior performance of PointeNet over state-of-the-art methods in point cloud analysis. Notably, PointeNet outperforms PointMLP with significantly fewer parameters on ModelNet40, ScanObjectNN, and ShapeNetPart, and achieves a substantial improvement of over 2% in for PointRCNN on KITTI with a minimal parameter cost of 1.4 million. Code is publicly available at https://github.com/lipeng-gu/PointeNet.
{"title":"PointeNet: A lightweight framework for effective and efficient point cloud analysis","authors":"Lipeng Gu , Xuefeng Yan , Liangliang Nan , Dingkun Zhu , Honghua Chen , Weiming Wang , Mingqiang Wei","doi":"10.1016/j.cagd.2024.102311","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102311","url":null,"abstract":"<div><p>The conventional wisdom in point cloud analysis predominantly explores 3D geometries. It is often achieved through the introduction of intricate learnable geometric extractors in the encoder or by deepening networks with repeated blocks. However, these methods contain a significant number of learnable parameters, resulting in substantial computational costs and imposing memory burdens on CPU/GPU. Moreover, they are primarily tailored for object-level point cloud classification and segmentation tasks, with limited extensions to crucial scene-level applications, such as autonomous driving. To this end, we introduce <strong>PointeNet</strong>, an <strong>e</strong>fficient <strong>net</strong>work designed specifically for <strong>point</strong> cloud analysis. PointeNet distinguishes itself with its lightweight architecture, low training cost, and plug-and-play capability, while also effectively capturing representative features. The network consists of a Multivariate Geometric Encoding (MGE) module and an <em>optional</em> Distance-aware Semantic Enhancement (DSE) module. MGE employs operations of sampling, grouping, pooling, and multivariate geometric aggregation to lightweightly capture and adaptively aggregate multivariate geometric features, providing a comprehensive depiction of 3D geometries. DSE, designed for real-world autonomous driving scenarios, enhances the semantic perception of point clouds, particularly for distant points. Our method demonstrates flexibility by seamlessly integrating with a classification/segmentation head or embedding into off-the-shelf 3D object detection networks, achieving notable performance improvements at a minimal cost. Extensive experiments on object-level datasets, including ModelNet40, ScanObjectNN, ShapeNetPart, and the scene-level dataset KITTI, demonstrate the superior performance of PointeNet over state-of-the-art methods in point cloud analysis. Notably, PointeNet outperforms PointMLP with significantly fewer parameters on ModelNet40, ScanObjectNN, and ShapeNetPart, and achieves a substantial improvement of over 2% in <span><math><mn>3</mn><mi>D</mi><mspace></mspace><mi>A</mi><msub><mrow><mi>P</mi></mrow><mrow><mi>R</mi><mn>40</mn></mrow></msub></math></span> for PointRCNN on KITTI with a minimal parameter cost of 1.4 million. Code is publicly available at <span>https://github.com/lipeng-gu/PointeNet</span><svg><path></path></svg>.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"110 ","pages":"Article 102311"},"PeriodicalIF":1.5,"publicationDate":"2024-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140822385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-30DOI: 10.1016/j.cagd.2024.102327
Hanxiao Wang , Mingyang Zhao , Yiqun Wang , Weize Quan , Dong-Ming Yan
Computer-Aided Design (CAD) software remains a pivotal tool in modern engineering and manufacturing, driving the design of a diverse range of products. In this work, we introduce VQ-CAD, the first CAD generation model based on Denoising Diffusion Probabilistic Models. This model utilizes a vector quantized diffusion model, employing multiple hierarchical codebooks generated through VQ-VAE. This integration not only offers a novel perspective on CAD model generation but also achieves state-of-the-art performance in 3D CAD model creation in a fully automatic fashion. Our model is able to recognize and incorporate implicit design constraints by simply forgoing traditional data augmentation. Furthermore, by melding our approach with CLIP, we significantly simplify the existing design process, directly generate CAD command sequences from initial design concepts represented by text or sketches, capture design intentions, and ensure designs adhere to implicit constraints.
{"title":"VQ-CAD: Computer-Aided Design model generation with vector quantized diffusion","authors":"Hanxiao Wang , Mingyang Zhao , Yiqun Wang , Weize Quan , Dong-Ming Yan","doi":"10.1016/j.cagd.2024.102327","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102327","url":null,"abstract":"<div><p>Computer-Aided Design (CAD) software remains a pivotal tool in modern engineering and manufacturing, driving the design of a diverse range of products. In this work, we introduce VQ-CAD, the first CAD generation model based on <em>Denoising Diffusion Probabilistic Models</em>. This model utilizes a vector quantized diffusion model, employing multiple hierarchical codebooks generated through VQ-VAE. This integration not only offers a novel perspective on CAD model generation but also achieves state-of-the-art performance in 3D CAD model creation in a fully automatic fashion. Our model is able to recognize and incorporate implicit design constraints by simply forgoing traditional data augmentation. Furthermore, by melding our approach with CLIP, we significantly simplify the existing design process, directly generate CAD command sequences from initial design concepts represented by <em>text</em> or <em>sketches</em>, capture design intentions, and ensure designs adhere to implicit constraints.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102327"},"PeriodicalIF":1.5,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140894801","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-30DOI: 10.1016/j.cagd.2024.102328
Shu Tang , Jiabin He , Shuli Yang , Xu Gong , Hongxing Qin
Although significant advances have been made in the field of multi-view 3D reconstruction using implicit neural field-based methods, existing reconstruction methods overlook the estimation of the material information (e.g. the base color, albedo, roughness, and metallic) during the learning process. In this paper, we propose a novel differentiable rendering framework, named as material NueS (M-NeuS), for simultaneously achieving precise surface reconstruction and competitive material estimation. For surface reconstruction, we perform multi-view geometry optimization by proposing an enhanced-low-to-high frequency encoding registration strategy (EFERS) and a second-order interpolated signed distance function (SI-SDF) for precise details and outline reconstruction. For material estimation, inspired by the NeuS, we first propose a volume-rendering-based material estimation strategy (VMES) to estimate the base color, albedo, roughness, and metallic accurately. And then, different from most material estimation methods that need ground-truth geometric priors, we use the geometry information reconstructed in the surface reconstruction stage and the directions of incidence from different viewpoints to model a neural light field, which can extract the lighting information from image observations. Next, the extracted lighting and the estimated base color, albedo, roughness, and metallic are optimized by the physics-based rendering equation. Extensive experiments demonstrate that our M-NeuS can not only reconstruct more precise geometry surface than existing state-of-the-art (SOTA) reconstruction methods but also can estimate competitive material information: the base color, albedo, roughness, and metallic.
{"title":"M-NeuS: Volume rendering based surface reconstruction and material estimation","authors":"Shu Tang , Jiabin He , Shuli Yang , Xu Gong , Hongxing Qin","doi":"10.1016/j.cagd.2024.102328","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102328","url":null,"abstract":"<div><p>Although significant advances have been made in the field of multi-view 3D reconstruction using implicit neural field-based methods, existing reconstruction methods overlook the estimation of the material information (e.g. the base color, albedo, roughness, and metallic) during the learning process. In this paper, we propose a novel differentiable rendering framework, named as material NueS (M-NeuS), for simultaneously achieving precise surface reconstruction and competitive material estimation. For surface reconstruction, we perform multi-view geometry optimization by proposing an enhanced-low-to-high frequency encoding registration strategy (EFERS) and a second-order interpolated signed distance function (SI-SDF) for precise details and outline reconstruction. For material estimation, inspired by the NeuS, we first propose a volume-rendering-based material estimation strategy (VMES) to estimate the base color, albedo, roughness, and metallic accurately. And then, different from most material estimation methods that need ground-truth geometric priors, we use the geometry information reconstructed in the surface reconstruction stage and the directions of incidence from different viewpoints to model a neural light field, which can extract the lighting information from image observations. Next, the extracted lighting and the estimated base color, albedo, roughness, and metallic are optimized by the physics-based rendering equation. Extensive experiments demonstrate that our M-NeuS can not only reconstruct more precise geometry surface than existing state-of-the-art (SOTA) reconstruction methods but also can estimate competitive material information: the base color, albedo, roughness, and metallic.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102328"},"PeriodicalIF":1.5,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140842884","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
One of the main challenges of indoor scene synthesis is preserving the functionality of synthesized scenes to create practical and usable indoor environments. Function groups exhibit the capability of balancing the global structure and local scenes of an indoor space. In this paper, we propose a function-centric indoor scene synthesis framework, named FuncScene. Our key idea is to use function groups as an intermedium to connect the local scenes and the global structure, thus achieving a coarse-to-fine indoor scene synthesis while maintaining the functionality and practicality of synthesized scenes. Indoor scenes are synthesized by first generating function groups using generative models and then instantiating by searching and matching the specific function groups from a dataset. The proposed framework also makes it easier to achieve multi-level generation control of scene synthesis, which was challenging for previous works. Extensive experiments on various indoor scene synthesis tasks demonstrate the validity of our method. Qualitative and quantitative evaluations show the proposed framework outperforms the existing state-of-the-art.
{"title":"FuncScene: Function-centric indoor scene synthesis via a variational autoencoder framework","authors":"Wenjie Min, Wenming Wu, Gaofeng Zhang, Liping Zheng","doi":"10.1016/j.cagd.2024.102319","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102319","url":null,"abstract":"<div><p>One of the main challenges of indoor scene synthesis is preserving the functionality of synthesized scenes to create practical and usable indoor environments. Function groups exhibit the capability of balancing the global structure and local scenes of an indoor space. In this paper, we propose a function-centric indoor scene synthesis framework, named FuncScene. Our key idea is to use function groups as an intermedium to connect the local scenes and the global structure, thus achieving a coarse-to-fine indoor scene synthesis while maintaining the functionality and practicality of synthesized scenes. Indoor scenes are synthesized by first generating function groups using generative models and then instantiating by searching and matching the specific function groups from a dataset. The proposed framework also makes it easier to achieve multi-level generation control of scene synthesis, which was challenging for previous works. Extensive experiments on various indoor scene synthesis tasks demonstrate the validity of our method. Qualitative and quantitative evaluations show the proposed framework outperforms the existing state-of-the-art.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102319"},"PeriodicalIF":1.5,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140824645","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-29DOI: 10.1016/j.cagd.2024.102310
Chiara Fuda, Kai Hormann
The generalization of barycentric coordinates to arbitrary simple polygons with more than three vertices has been a subject of study for a long time. Among the different constructions proposed, mean value coordinates have emerged as a popular choice, particularly due to their suitability for the non-convex setting. Since their introduction, they have found applications in numerous fields, and several equivalent formulas for their evaluation have been presented in the literature. However, so far, there has been no study regarding their numerical stability. In this paper, we aim to investigate the numerical stability of the algorithms that compute mean value coordinates. We show that all the known methods exhibit instability in some regions of the domain. To address this problem, we introduce a new formula for computing mean value coordinates, explain how to implement it, and formally prove that our new algorithm provides a stable evaluation of mean value coordinates. We validate our results through numerical experiments.
{"title":"A new stable method to compute mean value coordinates","authors":"Chiara Fuda, Kai Hormann","doi":"10.1016/j.cagd.2024.102310","DOIUrl":"https://doi.org/10.1016/j.cagd.2024.102310","url":null,"abstract":"<div><p>The generalization of barycentric coordinates to arbitrary simple polygons with more than three vertices has been a subject of study for a long time. Among the different constructions proposed, mean value coordinates have emerged as a popular choice, particularly due to their suitability for the non-convex setting. Since their introduction, they have found applications in numerous fields, and several equivalent formulas for their evaluation have been presented in the literature. However, so far, there has been no study regarding their numerical stability. In this paper, we aim to investigate the numerical stability of the algorithms that compute mean value coordinates. We show that all the known methods exhibit instability in some regions of the domain. To address this problem, we introduce a new formula for computing mean value coordinates, explain how to implement it, and formally prove that our new algorithm provides a stable evaluation of mean value coordinates. We validate our results through numerical experiments.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102310"},"PeriodicalIF":1.5,"publicationDate":"2024-04-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S016783962400044X/pdfft?md5=153410286c615fea62e2d69c27e8fedf&pid=1-s2.0-S016783962400044X-main.pdf","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140824642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"OA","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-25DOI: 10.1016/j.cagd.2024.102316
Xiyu Bao , Meng Qi , Chenglei Yang , Wei Gai
In motion planning, two-dimensional (2D) splinegons are typically used to represent the contours of 2D objects. In general, a 2D splinegon must be pre-decomposed to support rapid queries of the shortest paths or visibility. Herein, we propose a new region decomposition strategy, known as the Voronoi-based decomposition (VBD), based on the Voronoi diagram of curved boundary-segment generators (either convex or concave). The number of regions in the VBD is O(n+). Compared with the well-established horizontal visibility decomposition (HVD), whose complexity is O(n+), the VBD decomposition generally contains less regions because ≤, where n is the number of the vertices of the input splinegon, and and are the number of inserted vertices at the boundary. We systematically discuss the usage of VBD. Based on the VBD, the shortest path tree (SPT) can be computed in linear time. Statistics show that the VBD performs faster than HVD in SPT computations. Additionally, based on the SPT, we design algorithms that can rapidly compute the visibility between two points, the visible area of a point/line-segment, and the shortest path between two points.
{"title":"Voronoi-based splinegon decomposition and shortest-path tree computation","authors":"Xiyu Bao , Meng Qi , Chenglei Yang , Wei Gai","doi":"10.1016/j.cagd.2024.102316","DOIUrl":"10.1016/j.cagd.2024.102316","url":null,"abstract":"<div><p>In motion planning, two-dimensional (2D) splinegons are typically used to represent the contours of 2D objects. In general, a 2D splinegon must be pre-decomposed to support rapid queries of the shortest paths or visibility. Herein, we propose a new region decomposition strategy, known as the Voronoi-based decomposition (VBD), based on the Voronoi diagram of curved boundary-segment generators (either convex or concave). The number of regions in the VBD is O(<em>n</em>+<span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span>). Compared with the well-established horizontal visibility decomposition (HVD), whose complexity is O(<em>n</em>+<span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span>), the VBD decomposition generally contains less regions because <span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span>≤<span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span>, where <em>n</em> is the number of the vertices of the input splinegon, and <span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>1</mn></mrow></msub></math></span> and <span><math><msub><mrow><mi>c</mi></mrow><mrow><mn>2</mn></mrow></msub></math></span> are the number of inserted vertices at the boundary. We systematically discuss the usage of VBD. Based on the VBD, the shortest path tree (SPT) can be computed in linear time. Statistics show that the VBD performs faster than HVD in SPT computations. Additionally, based on the SPT, we design algorithms that can rapidly compute the visibility between two points, the visible area of a point/line-segment, and the shortest path between two points.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102316"},"PeriodicalIF":1.5,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140758403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-25DOI: 10.1016/j.cagd.2024.102326
Zitong He , Peisheng Zhuo , Hongwei Lin , Junfei Dai
In recent years, with the rapid development of the computer aided design and computer graphics, a large number of 3D models have emerged, making it a challenge to quickly find models of interest. As a concise and informative representation of 3D models, shape descriptors are a key factor in achieving effective retrieval. In this paper, we propose a novel global descriptor for 3D models that incorporates both geometric and topological information. We refer to this descriptor as the persistent heat kernel signature descriptor (PHKS). Constructed by concatenating our isometry-invariant geometric descriptor with topological descriptor, PHKS possesses high recognition ability, while remaining insensitive to noise and can be efficiently calculated. Retrieval experiments of 3D models on the benchmark datasets show considerable performance gains of the proposed method compared to other descriptors based on HKS and advanced topological descriptors.
{"title":"3D shape descriptor design based on HKS and persistent homology with stability analysis","authors":"Zitong He , Peisheng Zhuo , Hongwei Lin , Junfei Dai","doi":"10.1016/j.cagd.2024.102326","DOIUrl":"10.1016/j.cagd.2024.102326","url":null,"abstract":"<div><p>In recent years, with the rapid development of the computer aided design and computer graphics, a large number of 3D models have emerged, making it a challenge to quickly find models of interest. As a concise and informative representation of 3D models, shape descriptors are a key factor in achieving effective retrieval. In this paper, we propose a novel global descriptor for 3D models that incorporates both geometric and topological information. We refer to this descriptor as the <em>persistent heat kernel signature descriptor</em> (PHKS). Constructed by concatenating our isometry-invariant geometric descriptor with topological descriptor, PHKS possesses high recognition ability, while remaining insensitive to noise and can be efficiently calculated. Retrieval experiments of 3D models on the benchmark datasets show considerable performance gains of the proposed method compared to other descriptors based on HKS and advanced topological descriptors.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102326"},"PeriodicalIF":1.5,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140760451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Pub Date : 2024-04-25DOI: 10.1016/j.cagd.2024.102315
Yueji Ma , Yanzun Meng , Dong Xiao , Zuoqiang Shi , Bin Wang
In this paper, we propose a novel surface reconstruction method for unoriented points by establishing and solving a nonlinear equation system. By treating normals as unknown parameters and imposing the conditions that the implicit field is constant and its gradients parallel to the normals on the input point cloud, we establish a nonlinear equation system involving the oriented normals. To simplify the system, we transform it into a 0-1 integer programming problem solely focusing on orientation by incorporating inconsistent oriented normal information through PCA. We solve the simplified problem using flipping-based iterative algorithms and propose two novel criteria for flipping based on theoretical analysis.
Extensive experiments on renowned datasets demonstrate that our flipping-based method with wavelet surface reconstruction achieves state-of-the-art results in orientation and reconstruction. Furthermore, it exhibits linear computational and storage complexity by leveraging the orthogonality and compact support properties of wavelet bases. The source code is available at https://github.com/mayueji/FISR_code.
{"title":"Flipping-based iterative surface reconstruction for unoriented points","authors":"Yueji Ma , Yanzun Meng , Dong Xiao , Zuoqiang Shi , Bin Wang","doi":"10.1016/j.cagd.2024.102315","DOIUrl":"10.1016/j.cagd.2024.102315","url":null,"abstract":"<div><p>In this paper, we propose a novel surface reconstruction method for unoriented points by establishing and solving a nonlinear equation system. By treating normals as unknown parameters and imposing the conditions that the implicit field is constant and its gradients parallel to the normals on the input point cloud, we establish a nonlinear equation system involving the oriented normals. To simplify the system, we transform it into a 0-1 integer programming problem solely focusing on orientation by incorporating inconsistent oriented normal information through PCA. We solve the simplified problem using flipping-based iterative algorithms and propose two novel criteria for flipping based on theoretical analysis.</p><p>Extensive experiments on renowned datasets demonstrate that our flipping-based method with wavelet surface reconstruction achieves state-of-the-art results in orientation and reconstruction. Furthermore, it exhibits linear computational and storage complexity by leveraging the orthogonality and compact support properties of wavelet bases. The source code is available at <span>https://github.com/mayueji/FISR_code</span><svg><path></path></svg>.</p></div>","PeriodicalId":55226,"journal":{"name":"Computer Aided Geometric Design","volume":"111 ","pages":"Article 102315"},"PeriodicalIF":1.5,"publicationDate":"2024-04-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140775858","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":4,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}