AAGNet:基于属性感知图的网络,用于模糊行人再识别

IF 4.3 2区 计算机科学 Q1 ENGINEERING, ELECTRICAL & ELECTRONIC IEEE Transactions on Consumer Electronics Pub Date : 2024-09-03 DOI:10.1109/TCE.2024.3453890
Shihong Yao;Keyu Pan;Tao Wang;Zhigao Zheng;Jing Jin;Chuli Hu
{"title":"AAGNet:基于属性感知图的网络,用于模糊行人再识别","authors":"Shihong Yao;Keyu Pan;Tao Wang;Zhigao Zheng;Jing Jin;Chuli Hu","doi":"10.1109/TCE.2024.3453890","DOIUrl":null,"url":null,"abstract":"In large consumer sites, pedestrian re-identification (Re-ID) has the potential to enhance identify loyal consumers and create a more enjoyable shopping experience. Current Re-ID models always rely on some certain pedestrian feature descriptors, including body parts matching and pose key points, to extract part-level features. However, occlusion always causes a tremendous amount of noise and affects the feature representation, thereby significantly degrading the performance of those models. To address this problem, we propose an attribute-aware graph-based network (AAGNet) for Occluded Re-ID. Specifically, we develop a part-attribute feature extractor that maps the manually labeled pedestrian features into word vectors, and combines them with specific body part to obtain both attribute features and part features. The weight information of body parts and attributes are learned through graph convolution networks. Moreover, we introduce an occluded Re-ID dataset called Occluded-Market that can support the subsequent studies of occluded Re-ID. Comparative experimental results evidently demonstrate that the AAGNet shows superior performance in terms of accuracy, efficiency, and robustness on two open-source data sets. Our study can provide data and methodological support for further research on the occluded Re-ID and technological baseline for Re-ID-based commercial analytic applications in large consumer sites. The dataset is available at: github.com/Occluded_Market.","PeriodicalId":13208,"journal":{"name":"IEEE Transactions on Consumer Electronics","volume":"70 4","pages":"6580-6588"},"PeriodicalIF":4.3000,"publicationDate":"2024-09-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AAGNet: Attribute-Aware Graph-Based Network for Occluded Pedestrian Re-Identification\",\"authors\":\"Shihong Yao;Keyu Pan;Tao Wang;Zhigao Zheng;Jing Jin;Chuli Hu\",\"doi\":\"10.1109/TCE.2024.3453890\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In large consumer sites, pedestrian re-identification (Re-ID) has the potential to enhance identify loyal consumers and create a more enjoyable shopping experience. Current Re-ID models always rely on some certain pedestrian feature descriptors, including body parts matching and pose key points, to extract part-level features. However, occlusion always causes a tremendous amount of noise and affects the feature representation, thereby significantly degrading the performance of those models. To address this problem, we propose an attribute-aware graph-based network (AAGNet) for Occluded Re-ID. Specifically, we develop a part-attribute feature extractor that maps the manually labeled pedestrian features into word vectors, and combines them with specific body part to obtain both attribute features and part features. The weight information of body parts and attributes are learned through graph convolution networks. Moreover, we introduce an occluded Re-ID dataset called Occluded-Market that can support the subsequent studies of occluded Re-ID. Comparative experimental results evidently demonstrate that the AAGNet shows superior performance in terms of accuracy, efficiency, and robustness on two open-source data sets. Our study can provide data and methodological support for further research on the occluded Re-ID and technological baseline for Re-ID-based commercial analytic applications in large consumer sites. The dataset is available at: github.com/Occluded_Market.\",\"PeriodicalId\":13208,\"journal\":{\"name\":\"IEEE Transactions on Consumer Electronics\",\"volume\":\"70 4\",\"pages\":\"6580-6588\"},\"PeriodicalIF\":4.3000,\"publicationDate\":\"2024-09-03\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Consumer Electronics\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10663678/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"ENGINEERING, ELECTRICAL & ELECTRONIC\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Consumer Electronics","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10663678/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

摘要

在大型消费场所,行人再识别(Re-ID)有可能提高识别忠诚的消费者,创造更愉快的购物体验。目前的Re-ID模型总是依赖于一些特定的行人特征描述符,包括身体部位匹配和姿势关键点,来提取部分级特征。然而,遮挡总是会产生大量的噪声,影响特征的表示,从而大大降低了这些模型的性能。为了解决这个问题,我们提出了一个属性感知的基于图的网络(AAGNet)来处理闭塞的Re-ID。具体来说,我们开发了一个部分属性特征提取器,将手动标记的行人特征映射到词向量中,并将它们与特定的身体部位结合起来,获得属性特征和部分特征。通过图卷积网络学习身体部位和属性的权重信息。此外,我们还引入了一个名为occlded - market的闭塞Re-ID数据集,可以支持闭塞Re-ID的后续研究。对比实验结果表明,在两个开源数据集上,AAGNet在准确率、效率和鲁棒性方面表现出优异的性能。我们的研究可以为进一步研究被遮挡的Re-ID和基于Re-ID的大型消费者站点商业分析应用的技术基线提供数据和方法支持。该数据集可从github.com/Occluded_Market获取。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
AAGNet: Attribute-Aware Graph-Based Network for Occluded Pedestrian Re-Identification
In large consumer sites, pedestrian re-identification (Re-ID) has the potential to enhance identify loyal consumers and create a more enjoyable shopping experience. Current Re-ID models always rely on some certain pedestrian feature descriptors, including body parts matching and pose key points, to extract part-level features. However, occlusion always causes a tremendous amount of noise and affects the feature representation, thereby significantly degrading the performance of those models. To address this problem, we propose an attribute-aware graph-based network (AAGNet) for Occluded Re-ID. Specifically, we develop a part-attribute feature extractor that maps the manually labeled pedestrian features into word vectors, and combines them with specific body part to obtain both attribute features and part features. The weight information of body parts and attributes are learned through graph convolution networks. Moreover, we introduce an occluded Re-ID dataset called Occluded-Market that can support the subsequent studies of occluded Re-ID. Comparative experimental results evidently demonstrate that the AAGNet shows superior performance in terms of accuracy, efficiency, and robustness on two open-source data sets. Our study can provide data and methodological support for further research on the occluded Re-ID and technological baseline for Re-ID-based commercial analytic applications in large consumer sites. The dataset is available at: github.com/Occluded_Market.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
7.70
自引率
9.30%
发文量
59
审稿时长
3.3 months
期刊介绍: The main focus for the IEEE Transactions on Consumer Electronics is the engineering and research aspects of the theory, design, construction, manufacture or end use of mass market electronics, systems, software and services for consumers.
期刊最新文献
2024 Index IEEE Transactions on Consumer Electronics Vol. 70 Table of Contents Guest Editorial Consumer-Driven Energy-Efficient WSNs Architecture for Personalization and Contextualization in E-Commerce Systems IEEE Consumer Technology Society Officers and Committee Chairs Energy-Efficient Secure Architecture For Personalization E-Commerce WSN
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1