Qendrim Schreiber , Nicola Wolpert , Elmar Schömer
{"title":"METNet:分割三维纹理城市场景的网格探索方法","authors":"Qendrim Schreiber , Nicola Wolpert , Elmar Schömer","doi":"10.1016/j.isprsjprs.2024.10.020","DOIUrl":null,"url":null,"abstract":"<div><div>In this work, we present the neural network Mesh Exploring Tensor Net (METNet) for the segmentation of 3D urban scenes, that operates directly on textured meshes. Since triangular meshes have a very irregular structure, many existing approaches change the input by sampling evenly distributed point clouds on the meshes. The resulting simplified representation of the urban scenes has the advantage that state-of-the-art neural network architectures for point clouds can be used for the segmentation task. The disadvantages are that crucial geodesic information is lost and for a sufficiently good approximation of the scene many points are often necessary. Since memory on the GPU is limited, the consequence is that only small areas can be examined locally.</div><div>To overcome these limitations, METNet generates its input directly from the textured triangular mesh. It applies a new star-shaped exploration strategy, starting from a triangle on the mesh and expanding in various directions. This way, a vertex based regular local tensor from the unstructured mesh is generated, which we call a Mesh Exploring Tensor (MET). By expanding on the mesh, a MET maintains the information about the connectivity and distance of vertices along the surface of the mesh. It also effectively captures the characteristics of large regions. Our new architecture, METNet is optimized for processing METs as input. The regular structure of the input allows METNet the use of established convolutional neural networks.</div><div>Experimental results, conducted on two urban textured mesh benchmarks, demonstrate that METNet surpasses the performance of previous state-of-the-art techniques. METNet improves the previous state-of-the-art results by 8.6% in terms of mean IoU (intersection over union) on the SUM dataset compared to the second-best method PSSNet, and by 3.2% in terms of mean F1 score on the H3D dataset compared to the second-best method ifp-SCN. Our source code is available at <span><span>https://github.com/QenSchr/METNet</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50269,"journal":{"name":"ISPRS Journal of Photogrammetry and Remote Sensing","volume":"218 ","pages":"Pages 498-509"},"PeriodicalIF":10.6000,"publicationDate":"2024-11-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"METNet: A mesh exploring approach for segmenting 3D textured urban scenes\",\"authors\":\"Qendrim Schreiber , Nicola Wolpert , Elmar Schömer\",\"doi\":\"10.1016/j.isprsjprs.2024.10.020\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>In this work, we present the neural network Mesh Exploring Tensor Net (METNet) for the segmentation of 3D urban scenes, that operates directly on textured meshes. Since triangular meshes have a very irregular structure, many existing approaches change the input by sampling evenly distributed point clouds on the meshes. The resulting simplified representation of the urban scenes has the advantage that state-of-the-art neural network architectures for point clouds can be used for the segmentation task. The disadvantages are that crucial geodesic information is lost and for a sufficiently good approximation of the scene many points are often necessary. Since memory on the GPU is limited, the consequence is that only small areas can be examined locally.</div><div>To overcome these limitations, METNet generates its input directly from the textured triangular mesh. It applies a new star-shaped exploration strategy, starting from a triangle on the mesh and expanding in various directions. This way, a vertex based regular local tensor from the unstructured mesh is generated, which we call a Mesh Exploring Tensor (MET). By expanding on the mesh, a MET maintains the information about the connectivity and distance of vertices along the surface of the mesh. It also effectively captures the characteristics of large regions. Our new architecture, METNet is optimized for processing METs as input. The regular structure of the input allows METNet the use of established convolutional neural networks.</div><div>Experimental results, conducted on two urban textured mesh benchmarks, demonstrate that METNet surpasses the performance of previous state-of-the-art techniques. METNet improves the previous state-of-the-art results by 8.6% in terms of mean IoU (intersection over union) on the SUM dataset compared to the second-best method PSSNet, and by 3.2% in terms of mean F1 score on the H3D dataset compared to the second-best method ifp-SCN. Our source code is available at <span><span>https://github.com/QenSchr/METNet</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50269,\"journal\":{\"name\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"volume\":\"218 \",\"pages\":\"Pages 498-509\"},\"PeriodicalIF\":10.6000,\"publicationDate\":\"2024-11-21\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"ISPRS Journal of Photogrammetry and Remote Sensing\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0924271624003976\",\"RegionNum\":1,\"RegionCategory\":\"地球科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"GEOGRAPHY, PHYSICAL\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"ISPRS Journal of Photogrammetry and Remote Sensing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0924271624003976","RegionNum":1,"RegionCategory":"地球科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"GEOGRAPHY, PHYSICAL","Score":null,"Total":0}
METNet: A mesh exploring approach for segmenting 3D textured urban scenes
In this work, we present the neural network Mesh Exploring Tensor Net (METNet) for the segmentation of 3D urban scenes, that operates directly on textured meshes. Since triangular meshes have a very irregular structure, many existing approaches change the input by sampling evenly distributed point clouds on the meshes. The resulting simplified representation of the urban scenes has the advantage that state-of-the-art neural network architectures for point clouds can be used for the segmentation task. The disadvantages are that crucial geodesic information is lost and for a sufficiently good approximation of the scene many points are often necessary. Since memory on the GPU is limited, the consequence is that only small areas can be examined locally.
To overcome these limitations, METNet generates its input directly from the textured triangular mesh. It applies a new star-shaped exploration strategy, starting from a triangle on the mesh and expanding in various directions. This way, a vertex based regular local tensor from the unstructured mesh is generated, which we call a Mesh Exploring Tensor (MET). By expanding on the mesh, a MET maintains the information about the connectivity and distance of vertices along the surface of the mesh. It also effectively captures the characteristics of large regions. Our new architecture, METNet is optimized for processing METs as input. The regular structure of the input allows METNet the use of established convolutional neural networks.
Experimental results, conducted on two urban textured mesh benchmarks, demonstrate that METNet surpasses the performance of previous state-of-the-art techniques. METNet improves the previous state-of-the-art results by 8.6% in terms of mean IoU (intersection over union) on the SUM dataset compared to the second-best method PSSNet, and by 3.2% in terms of mean F1 score on the H3D dataset compared to the second-best method ifp-SCN. Our source code is available at https://github.com/QenSchr/METNet.
期刊介绍:
The ISPRS Journal of Photogrammetry and Remote Sensing (P&RS) serves as the official journal of the International Society for Photogrammetry and Remote Sensing (ISPRS). It acts as a platform for scientists and professionals worldwide who are involved in various disciplines that utilize photogrammetry, remote sensing, spatial information systems, computer vision, and related fields. The journal aims to facilitate communication and dissemination of advancements in these disciplines, while also acting as a comprehensive source of reference and archive.
P&RS endeavors to publish high-quality, peer-reviewed research papers that are preferably original and have not been published before. These papers can cover scientific/research, technological development, or application/practical aspects. Additionally, the journal welcomes papers that are based on presentations from ISPRS meetings, as long as they are considered significant contributions to the aforementioned fields.
In particular, P&RS encourages the submission of papers that are of broad scientific interest, showcase innovative applications (especially in emerging fields), have an interdisciplinary focus, discuss topics that have received limited attention in P&RS or related journals, or explore new directions in scientific or professional realms. It is preferred that theoretical papers include practical applications, while papers focusing on systems and applications should include a theoretical background.