首页 > 最新文献

arXiv - CS - Graphics最新文献

英文 中文
Fine-grained Metrics for Point Cloud Semantic Segmentation 点云语义分割的细粒度指标
Pub Date : 2024-07-31 DOI: arxiv-2407.21289
Zhuheng Lu, Ting Wu, Yuewei Dai, Weiqing Li, Zhiyong Su
Two forms of imbalances are commonly observed in point cloud semanticsegmentation datasets: (1) category imbalances, where certain objects are moreprevalent than others; and (2) size imbalances, where certain objects occupymore points than others. Because of this, the majority of categories and largeobjects are favored in the existing evaluation metrics. This paper suggestsfine-grained mIoU and mAcc for a more thorough assessment of point cloudsegmentation algorithms in order to address these issues. Richer statisticalinformation is provided for models and datasets by these fine-grained metrics,which also lessen the bias of current semantic segmentation metrics towardslarge objects. The proposed metrics are used to train and assess varioussemantic segmentation algorithms on three distinct indoor and outdoor semanticsegmentation datasets.
在点云语义分割数据集中通常会观察到两种形式的不平衡:(1) 类别不平衡,即某些物体比其他物体更普遍;(2) 大小不平衡,即某些物体比其他物体占据更多的点。正因为如此,在现有的评估指标中,大多数类别和大型对象都受到了青睐。本文建议采用精细度的 mIoU 和 mAcc 对点云分割算法进行更全面的评估,以解决这些问题。这些细粒度指标为模型和数据集提供了更丰富的统计信息,同时也减少了当前语义分割指标对大型物体的偏见。所提出的指标被用于在三个不同的室内和室外语义分割数据集上训练和评估各种语义分割算法。
{"title":"Fine-grained Metrics for Point Cloud Semantic Segmentation","authors":"Zhuheng Lu, Ting Wu, Yuewei Dai, Weiqing Li, Zhiyong Su","doi":"arxiv-2407.21289","DOIUrl":"https://doi.org/arxiv-2407.21289","url":null,"abstract":"Two forms of imbalances are commonly observed in point cloud semantic\u0000segmentation datasets: (1) category imbalances, where certain objects are more\u0000prevalent than others; and (2) size imbalances, where certain objects occupy\u0000more points than others. Because of this, the majority of categories and large\u0000objects are favored in the existing evaluation metrics. This paper suggests\u0000fine-grained mIoU and mAcc for a more thorough assessment of point cloud\u0000segmentation algorithms in order to address these issues. Richer statistical\u0000information is provided for models and datasets by these fine-grained metrics,\u0000which also lessen the bias of current semantic segmentation metrics towards\u0000large objects. The proposed metrics are used to train and assess various\u0000semantic segmentation algorithms on three distinct indoor and outdoor semantic\u0000segmentation datasets.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863840","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Comparative Study of Neural Surface Reconstruction for Scientific Visualization 用于科学可视化的神经表面重构比较研究
Pub Date : 2024-07-30 DOI: arxiv-2407.20868
Siyuan Yao, Weixi Song, Chaoli Wang
This comparative study evaluates various neural surface reconstructionmethods, particularly focusing on their implications for scientificvisualization through reconstructing 3D surfaces via multi-view renderingimages. We categorize ten methods into neural radiance fields and neuralimplicit surfaces, uncovering the benefits of leveraging distance functions(i.e., SDFs and UDFs) to enhance the accuracy and smoothness of thereconstructed surfaces. Our findings highlight the efficiency and quality ofNeuS2 for reconstructing closed surfaces and identify NeUDF as a promisingcandidate for reconstructing open surfaces despite some limitations. By sharingour benchmark dataset, we invite researchers to test the performance of theirmethods, contributing to the advancement of surface reconstruction solutionsfor scientific visualization.
这项比较研究评估了各种神经曲面重建方法,尤其关注它们通过多视角渲染图像重建三维曲面对科学可视化的影响。我们将十种方法分为神经辐射场和神经隐式曲面,揭示了利用距离函数(即 SDF 和 UDF)来提高所构建曲面的准确性和平滑度的好处。我们的研究结果凸显了 NeuS2 重构封闭曲面的效率和质量,并确定 NeUDF 是重构开放曲面的理想候选方案,尽管存在一些局限性。通过分享我们的基准数据集,我们邀请研究人员测试他们方法的性能,为科学可视化的曲面重建解决方案的进步做出贡献。
{"title":"A Comparative Study of Neural Surface Reconstruction for Scientific Visualization","authors":"Siyuan Yao, Weixi Song, Chaoli Wang","doi":"arxiv-2407.20868","DOIUrl":"https://doi.org/arxiv-2407.20868","url":null,"abstract":"This comparative study evaluates various neural surface reconstruction\u0000methods, particularly focusing on their implications for scientific\u0000visualization through reconstructing 3D surfaces via multi-view rendering\u0000images. We categorize ten methods into neural radiance fields and neural\u0000implicit surfaces, uncovering the benefits of leveraging distance functions\u0000(i.e., SDFs and UDFs) to enhance the accuracy and smoothness of the\u0000reconstructed surfaces. Our findings highlight the efficiency and quality of\u0000NeuS2 for reconstructing closed surfaces and identify NeUDF as a promising\u0000candidate for reconstructing open surfaces despite some limitations. By sharing\u0000our benchmark dataset, we invite researchers to test the performance of their\u0000methods, contributing to the advancement of surface reconstruction solutions\u0000for scientific visualization.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863844","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Monocular Human-Object Reconstruction in the Wild 野外单目人-物重构
Pub Date : 2024-07-30 DOI: arxiv-2407.20566
Chaofan Huo, Ye Shi, Jingya Wang
Learning the prior knowledge of the 3D human-object spatial relation iscrucial for reconstructing human-object interaction from images andunderstanding how humans interact with objects in 3D space. Previous workslearn this prior from datasets collected in controlled environments, but due tothe diversity of domains, they struggle to generalize to real-world scenarios.To overcome this limitation, we present a 2D-supervised method that learns the3D human-object spatial relation prior purely from 2D images in the wild. Ourmethod utilizes a flow-based neural network to learn the prior distribution ofthe 2D human-object keypoint layout and viewports for each image in thedataset. The effectiveness of the prior learned from 2D images is demonstratedon the human-object reconstruction task by applying the prior to tune therelative pose between the human and the object during the post-optimizationstage. To validate and benchmark our method on in-the-wild images, we collectthe WildHOI dataset from the YouTube website, which consists of variousinteractions with 8 objects in real-world scenarios. We conduct the experimentson the indoor BEHAVE dataset and the outdoor WildHOI dataset. The results showthat our method achieves almost comparable performance with fully 3D supervisedmethods on the BEHAVE dataset, even if we have only utilized the 2D layoutinformation, and outperforms previous methods in terms of generality andinteraction diversity on in-the-wild images.
学习三维人-物空间关系的先验知识对于从图像中重建人-物互动以及理解人类如何在三维空间中与物体互动至关重要。为了克服这一局限性,我们提出了一种二维监督方法,该方法纯粹从野外的二维图像中学习三维人-物空间关系先验知识。我们的方法利用基于流的神经网络来学习数据集中每幅图像的二维人-物关键点布局和视口的先验分布。在后优化阶段,通过应用先验来调整人与物体之间的相对姿态,证明了从二维图像中学到的先验在人-物重建任务中的有效性。为了在野外图像上对我们的方法进行验证和基准测试,我们从 YouTube 网站上收集了 WildHOI 数据集,其中包括在真实世界场景中与 8 个物体的各种交互。我们在室内 BEHAVE 数据集和室外 WildHOI 数据集上进行了实验。结果表明,在 BEHAVE 数据集上,即使我们只利用了二维布局信息,我们的方法也取得了与全三维监督方法几乎相当的性能;在野外图像上,我们的方法在通用性和交互多样性方面优于之前的方法。
{"title":"Monocular Human-Object Reconstruction in the Wild","authors":"Chaofan Huo, Ye Shi, Jingya Wang","doi":"arxiv-2407.20566","DOIUrl":"https://doi.org/arxiv-2407.20566","url":null,"abstract":"Learning the prior knowledge of the 3D human-object spatial relation is\u0000crucial for reconstructing human-object interaction from images and\u0000understanding how humans interact with objects in 3D space. Previous works\u0000learn this prior from datasets collected in controlled environments, but due to\u0000the diversity of domains, they struggle to generalize to real-world scenarios.\u0000To overcome this limitation, we present a 2D-supervised method that learns the\u00003D human-object spatial relation prior purely from 2D images in the wild. Our\u0000method utilizes a flow-based neural network to learn the prior distribution of\u0000the 2D human-object keypoint layout and viewports for each image in the\u0000dataset. The effectiveness of the prior learned from 2D images is demonstrated\u0000on the human-object reconstruction task by applying the prior to tune the\u0000relative pose between the human and the object during the post-optimization\u0000stage. To validate and benchmark our method on in-the-wild images, we collect\u0000the WildHOI dataset from the YouTube website, which consists of various\u0000interactions with 8 objects in real-world scenarios. We conduct the experiments\u0000on the indoor BEHAVE dataset and the outdoor WildHOI dataset. The results show\u0000that our method achieves almost comparable performance with fully 3D supervised\u0000methods on the BEHAVE dataset, even if we have only utilized the 2D layout\u0000information, and outperforms previous methods in terms of generality and\u0000interaction diversity on in-the-wild images.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
StackFLOW: Monocular Human-Object Reconstruction by Stacked Normalizing Flow with Offset StackFLOW:通过带偏移的叠加归一化流进行单目人-物重构
Pub Date : 2024-07-30 DOI: arxiv-2407.20545
Chaofan Huo, Ye Shi, Yuexin Ma, Lan Xu, Jingyi Yu, Jingya Wang
Modeling and capturing the 3D spatial arrangement of the human and the objectis the key to perceiving 3D human-object interaction from monocular images. Inthis work, we propose to use the Human-Object Offset between anchors which aredensely sampled from the surface of human mesh and object mesh to representhuman-object spatial relation. Compared with previous works which use contactmap or implicit distance filed to encode 3D human-object spatial relations, ourmethod is a simple and efficient way to encode the highly detailed spatialcorrelation between the human and object. Based on this representation, wepropose Stacked Normalizing Flow (StackFLOW) to infer the posteriordistribution of human-object spatial relations from the image. During theoptimization stage, we finetune the human body pose and object 6D pose bymaximizing the likelihood of samples based on this posterior distribution andminimizing the 2D-3D corresponding reprojection loss. Extensive experimentalresults show that our method achieves impressive results on two challengingbenchmarks, BEHAVE and InterCap datasets.
建模和捕捉人与物体的三维空间排列是通过单目图像感知三维人与物体交互的关键。在这项工作中,我们建议使用从人类网格和物体网格表面密集采样的锚点之间的人-物偏移来表示人-物空间关系。与之前使用接触图或隐式距离锉来编码三维人-物空间关系的工作相比,我们的方法是一种简单而有效的方法来编码人与物体之间高度精细的空间关系。在此基础上,我们提出了堆栈归一化流(Stacked Normalizing Flow,StackFLOW)来推断图像中人-物空间关系的后向分布。在优化阶段,我们根据后验分布最大化样本的可能性,并最小化 2D-3D 相应的重投影损失,从而对人体姿态和物体的 6D 姿态进行微调。广泛的实验结果表明,我们的方法在 BEHAVE 和 InterCap 数据集这两个具有挑战性的基准测试中取得了令人印象深刻的结果。
{"title":"StackFLOW: Monocular Human-Object Reconstruction by Stacked Normalizing Flow with Offset","authors":"Chaofan Huo, Ye Shi, Yuexin Ma, Lan Xu, Jingyi Yu, Jingya Wang","doi":"arxiv-2407.20545","DOIUrl":"https://doi.org/arxiv-2407.20545","url":null,"abstract":"Modeling and capturing the 3D spatial arrangement of the human and the object\u0000is the key to perceiving 3D human-object interaction from monocular images. In\u0000this work, we propose to use the Human-Object Offset between anchors which are\u0000densely sampled from the surface of human mesh and object mesh to represent\u0000human-object spatial relation. Compared with previous works which use contact\u0000map or implicit distance filed to encode 3D human-object spatial relations, our\u0000method is a simple and efficient way to encode the highly detailed spatial\u0000correlation between the human and object. Based on this representation, we\u0000propose Stacked Normalizing Flow (StackFLOW) to infer the posterior\u0000distribution of human-object spatial relations from the image. During the\u0000optimization stage, we finetune the human body pose and object 6D pose by\u0000maximizing the likelihood of samples based on this posterior distribution and\u0000minimizing the 2D-3D corresponding reprojection loss. Extensive experimental\u0000results show that our method achieves impressive results on two challenging\u0000benchmarks, BEHAVE and InterCap datasets.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
colorspace: A Python Toolbox for Manipulating and Assessing Colors and Palettes 颜色空间用于操作和评估颜色与调色板的 Python 工具箱
Pub Date : 2024-07-29 DOI: arxiv-2407.19921
Reto Stauffer, Achim Zeileis
The Python colorspace package provides a toolbox for mapping betweendifferent color spaces which can then be used to generate a wide range ofperceptually-based color palettes for qualitative or quantitative (sequentialor diverging) information. These palettes (as well as any other sets of colors)can be visualized, assessed, and manipulated in various ways, e.g., by colorswatches, emulating the effects of color vision deficiencies, or depicting theperceptual properties. Finally, the color palettes generated by the package canbe easily integrated into standard visualization workflows in Python, e.g.,using matplotlib, seaborn, or plotly.
Python 色彩空间软件包提供了一个在不同色彩空间之间进行映射的工具箱,可用于生成各种基于感知的调色板,以获取定性或定量(连续或发散)信息。这些调色板(以及任何其他颜色集)可以通过各种方式进行可视化、评估和操作,例如,通过配色、模拟色觉缺陷的影响或描述感知特性。最后,软件包生成的调色板可以轻松集成到 Python 的标准可视化工作流中,例如使用 matplotlib、seaborn 或 plotly。
{"title":"colorspace: A Python Toolbox for Manipulating and Assessing Colors and Palettes","authors":"Reto Stauffer, Achim Zeileis","doi":"arxiv-2407.19921","DOIUrl":"https://doi.org/arxiv-2407.19921","url":null,"abstract":"The Python colorspace package provides a toolbox for mapping between\u0000different color spaces which can then be used to generate a wide range of\u0000perceptually-based color palettes for qualitative or quantitative (sequential\u0000or diverging) information. These palettes (as well as any other sets of colors)\u0000can be visualized, assessed, and manipulated in various ways, e.g., by color\u0000swatches, emulating the effects of color vision deficiencies, or depicting the\u0000perceptual properties. Finally, the color palettes generated by the package can\u0000be easily integrated into standard visualization workflows in Python, e.g.,\u0000using matplotlib, seaborn, or plotly.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Structure-Aware Simplification for Hypergraph Visualization 超图可视化的结构感知简化
Pub Date : 2024-07-29 DOI: arxiv-2407.19621
Peter Oliver, Eugene Zhang, Yue Zhang
Hypergraphs provide a natural way to represent polyadic relationships innetwork data. For large hypergraphs, it is often difficult to visually detectstructures within the data. Recently, a scalable polygon-based visualizationapproach was developed allowing hypergraphs with thousands of hyperedges to besimplified and examined at different levels of detail. However, this approachis not guaranteed to eliminate all of the visual clutter caused by unavoidableoverlaps. Furthermore, meaningful structures can be lost at simplified scales,making their interpretation unreliable. In this paper, we define hypergraphstructures using the bipartite graph representation, allowing us to decomposethe hypergraph into a union of structures including topological blocks,bridges, and branches, and to identify exactly where unavoidable overlaps mustoccur. We also introduce a set of topology preserving and topology alteringatomic operations, enabling the preservation of important structures whilereducing unavoidable overlaps to improve visual clarity and interpretability insimplified scales. We demonstrate our approach in several real-worldapplications.
超图为表示网络数据中的多向关系提供了一种自然的方法。对于大型超图,通常很难直观地发现数据中的结构。最近,开发出了一种基于多边形的可扩展可视化方法,可以简化具有数千个超节点的超图,并以不同的详细程度对其进行检查。然而,这种方法并不能保证消除不可避免的重叠造成的所有视觉混乱。此外,有意义的结构可能会在简化的尺度上丢失,从而使其解释变得不可靠。在本文中,我们使用二方图表示法定义了超图结构,从而可以将超图分解为包括拓扑块、桥和分支在内的结构联盟,并准确识别不可避免的重叠必须出现在哪里。我们还引入了一组拓扑保留和拓扑改变原子操作,从而在保留重要结构的同时减少不可避免的重叠,提高视觉清晰度和简化尺度下的可解释性。我们在几个实际应用中演示了我们的方法。
{"title":"Structure-Aware Simplification for Hypergraph Visualization","authors":"Peter Oliver, Eugene Zhang, Yue Zhang","doi":"arxiv-2407.19621","DOIUrl":"https://doi.org/arxiv-2407.19621","url":null,"abstract":"Hypergraphs provide a natural way to represent polyadic relationships in\u0000network data. For large hypergraphs, it is often difficult to visually detect\u0000structures within the data. Recently, a scalable polygon-based visualization\u0000approach was developed allowing hypergraphs with thousands of hyperedges to be\u0000simplified and examined at different levels of detail. However, this approach\u0000is not guaranteed to eliminate all of the visual clutter caused by unavoidable\u0000overlaps. Furthermore, meaningful structures can be lost at simplified scales,\u0000making their interpretation unreliable. In this paper, we define hypergraph\u0000structures using the bipartite graph representation, allowing us to decompose\u0000the hypergraph into a union of structures including topological blocks,\u0000bridges, and branches, and to identify exactly where unavoidable overlaps must\u0000occur. We also introduce a set of topology preserving and topology altering\u0000atomic operations, enabling the preservation of important structures while\u0000reducing unavoidable overlaps to improve visual clarity and interpretability in\u0000simplified scales. We demonstrate our approach in several real-world\u0000applications.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863851","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
From Flat to Spatial: Comparison of 4 methods constructing 3D, 2 and 1/2D Models from 2D Plans with neural networks 从平面到空间:利用神经网络从二维平面图构建三维、二维和 1/2D 模型的四种方法比较
Pub Date : 2024-07-29 DOI: arxiv-2407.19970
Jacob Sam, Karan Patel, Mike Saad
In the field of architecture, the conversion of single images into 2 and 1/2Dand 3D meshes is a promising technology that enhances design visualization andefficiency. This paper evaluates four innovative methods: "One-2-3-45," "CRM:Single Image to 3D Textured Mesh with Convolutional Reconstruction Model,""Instant Mesh," and "Image-to-Mesh." These methods are at the forefront of thistechnology, focusing on their applicability in architectural design andvisualization. They streamline the creation of 3D architectural models,enabling rapid prototyping and detailed visualization from minimal initialinputs, such as photographs or simple sketches.One-2-3-45 leverages adiffusion-based approach to generate multi-view reconstructions, ensuring highgeometric fidelity and texture quality. CRM utilizes a convolutional network tointegrate geometric priors into its architecture, producing detailed andtextured meshes quickly and efficiently. Instant Mesh combines the strengths ofmulti-view diffusion and sparse-view models to offer speed and scalability,suitable for diverse architectural projects. Image-to-Mesh leverages agenerative adversarial network (GAN) to produce 3D meshes from single images,focusing on maintaining high texture fidelity and geometric accuracy byincorporating image and depth map data into its training process. It uses ahybrid approach that combines voxel-based representations with surfacereconstruction techniques to ensure detailed and realistic 3D models.Thiscomparative study highlights each method's contribution to reducing designcycle times, improving accuracy, and enabling flexible adaptations to variousarchitectural styles and requirements. By providing architects with powerfultools for rapid visualization and iteration, these advancements in 3D meshgeneration are set to revolutionize architectural practices.
在建筑领域,将单幅图像转换为二维、1/2 维和三维网格是一项很有前途的技术,可提高设计的可视化和效率。本文评估了四种创新方法:"One-2-3-45"、"CRM:利用卷积重建模型将单张图像转换为三维纹理网格"、"即时网格 "和 "图像到网格"。这些方法都处于该技术的前沿,重点关注其在建筑设计和可视化方面的适用性。One-2-3-45 利用基于扩散的方法生成多视角重建,确保了高几何保真度和纹理质量。CRM 利用卷积网络将几何先验整合到其架构中,快速高效地生成细节丰富、纹理清晰的网格。Instant Mesh 结合了多视图扩散和稀疏视图模型的优势,速度快,可扩展性强,适用于各种建筑项目。Image-to-Mesh 利用生成对抗网络 (GAN) 从单张图像生成三维网格,通过将图像和深度图数据纳入训练过程,重点保持高纹理保真度和几何精度。该比较研究强调了每种方法在缩短设计周期时间、提高精度以及灵活适应各种建筑风格和要求方面的贡献。通过为建筑师提供快速可视化和迭代的强大工具,这些三维网格生成技术的进步必将彻底改变建筑实践。
{"title":"From Flat to Spatial: Comparison of 4 methods constructing 3D, 2 and 1/2D Models from 2D Plans with neural networks","authors":"Jacob Sam, Karan Patel, Mike Saad","doi":"arxiv-2407.19970","DOIUrl":"https://doi.org/arxiv-2407.19970","url":null,"abstract":"In the field of architecture, the conversion of single images into 2 and 1/2D\u0000and 3D meshes is a promising technology that enhances design visualization and\u0000efficiency. This paper evaluates four innovative methods: \"One-2-3-45,\" \"CRM:\u0000Single Image to 3D Textured Mesh with Convolutional Reconstruction Model,\"\u0000\"Instant Mesh,\" and \"Image-to-Mesh.\" These methods are at the forefront of this\u0000technology, focusing on their applicability in architectural design and\u0000visualization. They streamline the creation of 3D architectural models,\u0000enabling rapid prototyping and detailed visualization from minimal initial\u0000inputs, such as photographs or simple sketches.One-2-3-45 leverages a\u0000diffusion-based approach to generate multi-view reconstructions, ensuring high\u0000geometric fidelity and texture quality. CRM utilizes a convolutional network to\u0000integrate geometric priors into its architecture, producing detailed and\u0000textured meshes quickly and efficiently. Instant Mesh combines the strengths of\u0000multi-view diffusion and sparse-view models to offer speed and scalability,\u0000suitable for diverse architectural projects. Image-to-Mesh leverages a\u0000generative adversarial network (GAN) to produce 3D meshes from single images,\u0000focusing on maintaining high texture fidelity and geometric accuracy by\u0000incorporating image and depth map data into its training process. It uses a\u0000hybrid approach that combines voxel-based representations with surface\u0000reconstruction techniques to ensure detailed and realistic 3D models.This\u0000comparative study highlights each method's contribution to reducing design\u0000cycle times, improving accuracy, and enabling flexible adaptations to various\u0000architectural styles and requirements. By providing architects with powerful\u0000tools for rapid visualization and iteration, these advancements in 3D mesh\u0000generation are set to revolutionize architectural practices.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863849","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Physically-based Path Tracer using WebGPU and OpenPBR 使用 WebGPU 和 OpenPBR 的物理路径追踪器
Pub Date : 2024-07-29 DOI: arxiv-2407.19977
Simon Stucki, Philipp Ackermann
This work presents a web-based, open-source path tracer for renderingphysically-based 3D scenes using WebGPU and the OpenPBR surface shading model.While rasterization has been the dominant real-time rendering technique on theweb since WebGL's introduction in 2011, it struggles with global illumination.This necessitates more complex techniques, often relying on pregeneratedartifacts to attain the desired level of visual fidelity. Path tracinginherently addresses these limitations but at the cost of increased renderingtime. Our work focuses on industrial applications where highly customizableproducts are common and real-time performance is not critical. We leverageWebGPU to implement path tracing on the web, integrating the OpenPBR standardfor physically-based material representation. The result is a near real-timepath tracer capable of rendering high-fidelity 3D scenes directly in webbrowsers, eliminating the need for pregenerated assets. Our implementationdemonstrates the potential of WebGPU for advanced rendering techniques andopens new possibilities for web-based 3D visualization in industrialapplications.
自 2011 年 WebGL 推出以来,光栅化一直是网络上最主要的实时渲染技术,但它在全局光照问题上却举步维艰,这就需要采用更复杂的技术,通常需要依赖预生成的人工图像才能达到理想的视觉保真度。路径跟踪从本质上解决了这些限制,但代价是增加了渲染时间。我们的工作侧重于工业应用,在这些应用中,高度可定制的产品很常见,实时性并不重要。我们利用 WebGPU 在网络上实现路径跟踪,并集成了基于物理材料表示的 OpenPBR 标准。其结果是一个接近实时的路径跟踪器,能够直接在网页浏览器中渲染高保真三维场景,无需预生成资产。我们的实现展示了 WebGPU 在高级渲染技术方面的潜力,并为工业应用中基于网络的三维可视化开辟了新的可能性。
{"title":"Physically-based Path Tracer using WebGPU and OpenPBR","authors":"Simon Stucki, Philipp Ackermann","doi":"arxiv-2407.19977","DOIUrl":"https://doi.org/arxiv-2407.19977","url":null,"abstract":"This work presents a web-based, open-source path tracer for rendering\u0000physically-based 3D scenes using WebGPU and the OpenPBR surface shading model.\u0000While rasterization has been the dominant real-time rendering technique on the\u0000web since WebGL's introduction in 2011, it struggles with global illumination.\u0000This necessitates more complex techniques, often relying on pregenerated\u0000artifacts to attain the desired level of visual fidelity. Path tracing\u0000inherently addresses these limitations but at the cost of increased rendering\u0000time. Our work focuses on industrial applications where highly customizable\u0000products are common and real-time performance is not critical. We leverage\u0000WebGPU to implement path tracing on the web, integrating the OpenPBR standard\u0000for physically-based material representation. The result is a near real-time\u0000path tracer capable of rendering high-fidelity 3D scenes directly in web\u0000browsers, eliminating the need for pregenerated assets. Our implementation\u0000demonstrates the potential of WebGPU for advanced rendering techniques and\u0000opens new possibilities for web-based 3D visualization in industrial\u0000applications.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
textsc{Perm}: A Parametric Representation for Multi-Style 3D Hair Modeling textsc{Perm}:多风格三维发型建模的参数表示法
Pub Date : 2024-07-28 DOI: arxiv-2407.19451
Chengan He, Xin Sun, Zhixin Shu, Fujun Luan, Sören Pirk, Jorge Alejandro Amador Herrera, Dominik L. Michels, Tuanfeng Y. Wang, Meng Zhang, Holly Rushmeier, Yi Zhou
We present textsc{Perm}, a learned parametric model of human 3D hairdesigned to facilitate various hair-related applications. Unlike previous workthat jointly models the global hair shape and local strand details, we proposeto disentangle them using a PCA-based strand representation in the frequencydomain, thereby allowing more precise editing and output control. Specifically,we leverage our strand representation to fit and decompose hair geometrytextures into low- to high-frequency hair structures. These decomposed texturesare later parameterized with different generative models, emulating commonstages in the hair modeling process. We conduct extensive experiments tovalidate the architecture design of textsc{Perm}, and finally deploy thetrained model as a generic prior to solve task-agnostic problems, furthershowcasing its flexibility and superiority in tasks such as 3D hairparameterization, hairstyle interpolation, single-view hair reconstruction, andhair-conditioned image generation. Our code and data will be available at:url{https://github.com/c-he/perm}.
我们展示了人类三维头发的学习参数模型 textsc{Perm},该模型旨在促进各种与头发相关的应用。与以往联合建模全局发丝形状和局部发丝细节的工作不同,我们建议使用基于 PCA 的频域发丝表示法将它们分开,从而实现更精确的编辑和输出控制。具体来说,我们利用头发丝表示法将头发几何纹理拟合并分解为低频到高频的头发结构。这些分解后的纹理随后用不同的生成模型进行参数化,模拟头发建模过程中的常见阶段。我们进行了大量实验来验证 textsc{Perm} 的架构设计,最后将训练好的模型作为通用先验模型来解决与任务无关的问题,进一步展示了它在三维毛发参数化、发型插值、单视角毛发重建和毛发条件图像生成等任务中的灵活性和优越性。我们的代码和数据将发布在:url{https://github.com/c-he/perm}。
{"title":"textsc{Perm}: A Parametric Representation for Multi-Style 3D Hair Modeling","authors":"Chengan He, Xin Sun, Zhixin Shu, Fujun Luan, Sören Pirk, Jorge Alejandro Amador Herrera, Dominik L. Michels, Tuanfeng Y. Wang, Meng Zhang, Holly Rushmeier, Yi Zhou","doi":"arxiv-2407.19451","DOIUrl":"https://doi.org/arxiv-2407.19451","url":null,"abstract":"We present textsc{Perm}, a learned parametric model of human 3D hair\u0000designed to facilitate various hair-related applications. Unlike previous work\u0000that jointly models the global hair shape and local strand details, we propose\u0000to disentangle them using a PCA-based strand representation in the frequency\u0000domain, thereby allowing more precise editing and output control. Specifically,\u0000we leverage our strand representation to fit and decompose hair geometry\u0000textures into low- to high-frequency hair structures. These decomposed textures\u0000are later parameterized with different generative models, emulating common\u0000stages in the hair modeling process. We conduct extensive experiments to\u0000validate the architecture design of textsc{Perm}, and finally deploy the\u0000trained model as a generic prior to solve task-agnostic problems, further\u0000showcasing its flexibility and superiority in tasks such as 3D hair\u0000parameterization, hairstyle interpolation, single-view hair reconstruction, and\u0000hair-conditioned image generation. Our code and data will be available at:\u0000url{https://github.com/c-he/perm}.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
FreeShell: A Context-Free 4D Printing Technique for Fabricating Complex 3D Triangle Mesh Shells FreeShell:用于制造复杂三维三角网格壳体的无上下文 4D 打印技术
Pub Date : 2024-07-28 DOI: arxiv-2407.19533
Chao Yuan, Nan Cao, Xuejiao Ma, Shengqi Dang
Freeform thin-shell surfaces are critical in various fields, but theirfabrication is complex and costly. Traditional methods are wasteful and requirecustom molds, while 3D printing needs extensive support structures andpost-processing. Thermoshrinkage actuated 4D printing is an effective methodthrough flat structures fabricating 3D shell. However, existing research facesissues related to precise deformation and limited robustness. Addressing theseissues is challenging due to three key factors: (1) Difficulty in finding auniversal method to control deformation across different materials; (2)Variability in deformation influenced by factors such as printing speed, layerthickness, and heating temperature; (3) Environmental factors affecting thedeformation process. To overcome these challenges, we introduce FreeShell, arobust 4D printing technique that uses thermoshrinkage to create precise 3Dshells. This method prints triangular tiles connected by shrinkable connectorsfrom a single material. Upon heating, the connectors shrink, moving the tilesto form the desired 3D shape, simplifying fabrication and reducing material andenvironment dependency. An optimized algorithm for flattening 3D meshes ensuresprecision in printing. FreeShell demonstrates its effectiveness through variousexamples and experiments, showcasing accuracy, robustness, and strength,representing advancement in fabricating complex freeform surfaces.
自由形态薄壳表面在各个领域都至关重要,但其制造工艺复杂且成本高昂。传统方法既浪费又需要定制模具,而三维打印则需要大量的支撑结构和后处理。热收缩驱动 4D 打印是一种通过平面结构制造 3D 外壳的有效方法。然而,现有研究面临着与精确变形和有限鲁棒性相关的问题。由于以下三个关键因素,解决这些问题具有挑战性:(1) 难以找到一种通用的方法来控制不同材料的变形;(2) 受打印速度、层厚度和加热温度等因素的影响,变形存在差异;(3) 环境因素影响变形过程。为了克服这些挑战,我们引入了 FreeShell,一种利用热收缩技术创建精确三维外壳的稳健 4D 打印技术。这种方法用单一材料打印出由可收缩连接器连接的三角形瓦片。加热时,连接器收缩,移动瓷砖以形成所需的三维形状,从而简化了制造过程,降低了对材料和环境的依赖性。扁平化三维网格的优化算法可确保打印精度。FreeShell 通过各种示例和实验证明了其有效性,展示了精确性、稳健性和强度,代表了制造复杂自由曲面的进步。
{"title":"FreeShell: A Context-Free 4D Printing Technique for Fabricating Complex 3D Triangle Mesh Shells","authors":"Chao Yuan, Nan Cao, Xuejiao Ma, Shengqi Dang","doi":"arxiv-2407.19533","DOIUrl":"https://doi.org/arxiv-2407.19533","url":null,"abstract":"Freeform thin-shell surfaces are critical in various fields, but their\u0000fabrication is complex and costly. Traditional methods are wasteful and require\u0000custom molds, while 3D printing needs extensive support structures and\u0000post-processing. Thermoshrinkage actuated 4D printing is an effective method\u0000through flat structures fabricating 3D shell. However, existing research faces\u0000issues related to precise deformation and limited robustness. Addressing these\u0000issues is challenging due to three key factors: (1) Difficulty in finding a\u0000universal method to control deformation across different materials; (2)\u0000Variability in deformation influenced by factors such as printing speed, layer\u0000thickness, and heating temperature; (3) Environmental factors affecting the\u0000deformation process. To overcome these challenges, we introduce FreeShell, a\u0000robust 4D printing technique that uses thermoshrinkage to create precise 3D\u0000shells. This method prints triangular tiles connected by shrinkable connectors\u0000from a single material. Upon heating, the connectors shrink, moving the tiles\u0000to form the desired 3D shape, simplifying fabrication and reducing material and\u0000environment dependency. An optimized algorithm for flattening 3D meshes ensures\u0000precision in printing. FreeShell demonstrates its effectiveness through various\u0000examples and experiments, showcasing accuracy, robustness, and strength,\u0000representing advancement in fabricating complex freeform surfaces.","PeriodicalId":501174,"journal":{"name":"arXiv - CS - Graphics","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-07-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141863853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
arXiv - CS - Graphics
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1