Jiongchao Jin, Huanqiang Xu, Pengliang Ji, Zehao Tang, Z. Xiong
{"title":"基于零件的三维形状检索循环多视图聚合网络","authors":"Jiongchao Jin, Huanqiang Xu, Pengliang Ji, Zehao Tang, Z. Xiong","doi":"10.1109/ICCSMT54525.2021.00065","DOIUrl":null,"url":null,"abstract":"We propose the Part-based Recurrent Multi-view Aggregation network(PREMA) to eliminate the detrimental effects of the practical view defects, such as insufficient view numbers, occlusions or background clutters, and also enhance the discriminative ability of shape representations. Inspired by the fact that human recognize an object mainly by its discriminant parts, we define the multi-view coherent part(MCP), a discriminant part reoccurring in different views. Our PREMA can reliably locate and effectively utilize MCPs to build robust shape representations. Comprehensively, we design a novel Regional Attention Unit(RAU) in PREMA to compute the confidence map for each view, and extract MCPs by applying those maps to view features. PREMA accentuates MCPs via correlating features of different views, and aggregates the part-aware features for shape representation. Finally, we show extensive evaluations to demonstrate that our method achieves the state-of-the-art accuracy for 3D shape retrieval on ModelNet-40 and ShapeNetCore-55 datasets.","PeriodicalId":304337,"journal":{"name":"2021 2nd International Conference on Computer Science and Management Technology (ICCSMT)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"PREMA: Part-based REcurrent Multi-view Aggregation Network for 3D Shape Retrieval\",\"authors\":\"Jiongchao Jin, Huanqiang Xu, Pengliang Ji, Zehao Tang, Z. Xiong\",\"doi\":\"10.1109/ICCSMT54525.2021.00065\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We propose the Part-based Recurrent Multi-view Aggregation network(PREMA) to eliminate the detrimental effects of the practical view defects, such as insufficient view numbers, occlusions or background clutters, and also enhance the discriminative ability of shape representations. Inspired by the fact that human recognize an object mainly by its discriminant parts, we define the multi-view coherent part(MCP), a discriminant part reoccurring in different views. Our PREMA can reliably locate and effectively utilize MCPs to build robust shape representations. Comprehensively, we design a novel Regional Attention Unit(RAU) in PREMA to compute the confidence map for each view, and extract MCPs by applying those maps to view features. PREMA accentuates MCPs via correlating features of different views, and aggregates the part-aware features for shape representation. Finally, we show extensive evaluations to demonstrate that our method achieves the state-of-the-art accuracy for 3D shape retrieval on ModelNet-40 and ShapeNetCore-55 datasets.\",\"PeriodicalId\":304337,\"journal\":{\"name\":\"2021 2nd International Conference on Computer Science and Management Technology (ICCSMT)\",\"volume\":\"1 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-11-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 2nd International Conference on Computer Science and Management Technology (ICCSMT)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICCSMT54525.2021.00065\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 2nd International Conference on Computer Science and Management Technology (ICCSMT)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCSMT54525.2021.00065","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
提出了基于部分的循环多视图聚合网络(PREMA),以消除实际视图缺陷(如视图数不足、遮挡或背景杂波)的不利影响,并增强形状表示的判别能力。受人类主要通过物体的判别部分来识别物体的启发,我们定义了多视图连贯部分(multi-view coherent part, MCP),即在不同视图中重复出现的判别部分。我们的PREMA可以可靠地定位和有效地利用mcp来构建鲁棒的形状表示。综合而言,我们在PREMA中设计了一种新的区域注意单元(Regional Attention Unit, RAU)来计算每个视图的置信度图,并通过将这些图应用于视图特征来提取mcp。PREMA通过关联不同视图的特征来强调mcp,并将部件感知特征聚合在一起进行形状表示。最后,我们进行了广泛的评估,以证明我们的方法在ModelNet-40和ShapeNetCore-55数据集上实现了最先进的3D形状检索精度。
PREMA: Part-based REcurrent Multi-view Aggregation Network for 3D Shape Retrieval
We propose the Part-based Recurrent Multi-view Aggregation network(PREMA) to eliminate the detrimental effects of the practical view defects, such as insufficient view numbers, occlusions or background clutters, and also enhance the discriminative ability of shape representations. Inspired by the fact that human recognize an object mainly by its discriminant parts, we define the multi-view coherent part(MCP), a discriminant part reoccurring in different views. Our PREMA can reliably locate and effectively utilize MCPs to build robust shape representations. Comprehensively, we design a novel Regional Attention Unit(RAU) in PREMA to compute the confidence map for each view, and extract MCPs by applying those maps to view features. PREMA accentuates MCPs via correlating features of different views, and aggregates the part-aware features for shape representation. Finally, we show extensive evaluations to demonstrate that our method achieves the state-of-the-art accuracy for 3D shape retrieval on ModelNet-40 and ShapeNetCore-55 datasets.