{"title":"通过信息解缠实现可见光-红外线人员再识别的模态偏差校准网络","authors":"Haojie Liu;Hao Luo;Xiantao Peng;Wei Jiang","doi":"10.1109/TCSS.2024.3398696","DOIUrl":null,"url":null,"abstract":"Visible–infrared person reidentification (VI-ReID) in social surveillance systems involves analyzing social behavior using nonoverlapping cross-modality camera sets. It often has poor retrieval performance under modality gap. One way to alleviate such the modality discrepancy is to learn shared person features that are generalizable across different modalities. However, because of significant differences in color between the visible and infrared images, the learned share features are always inclined to specific information of corresponding modality. To this end, we propose a modality bias calibration network (MBCNet) that filters out identity-irrelevant interference and recalibrates the learned modality-shared features. Specifically, to emphasize the modality-shared cues, we employ a feature decomposition module in the feature-level to filter out style variations and extract identity-relevant discriminative cues from the residual feature. In order to achieve a better disentanglement, a dual ranking entropy constraint is further proposed to ensure that the learned features contain only identity-relevant information and discard style-relevant information. Simultaneously, we design a decorrelated orthogonality Loss to ensure the disentangled features are not correlated with each other. Through comprehensive experiments, we demonstrate that MBCNet significantly improves the cross-modality retrieval performance in social surveillance systems and effectively addresses the modality bias training issue.","PeriodicalId":13044,"journal":{"name":"IEEE Transactions on Computational Social Systems","volume":"11 5","pages":"6925-6938"},"PeriodicalIF":4.5000,"publicationDate":"2024-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Modality Bias Calibration Network via Information Disentanglement for Visible–Infrared Person Reidentification\",\"authors\":\"Haojie Liu;Hao Luo;Xiantao Peng;Wei Jiang\",\"doi\":\"10.1109/TCSS.2024.3398696\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Visible–infrared person reidentification (VI-ReID) in social surveillance systems involves analyzing social behavior using nonoverlapping cross-modality camera sets. It often has poor retrieval performance under modality gap. One way to alleviate such the modality discrepancy is to learn shared person features that are generalizable across different modalities. However, because of significant differences in color between the visible and infrared images, the learned share features are always inclined to specific information of corresponding modality. To this end, we propose a modality bias calibration network (MBCNet) that filters out identity-irrelevant interference and recalibrates the learned modality-shared features. Specifically, to emphasize the modality-shared cues, we employ a feature decomposition module in the feature-level to filter out style variations and extract identity-relevant discriminative cues from the residual feature. In order to achieve a better disentanglement, a dual ranking entropy constraint is further proposed to ensure that the learned features contain only identity-relevant information and discard style-relevant information. Simultaneously, we design a decorrelated orthogonality Loss to ensure the disentangled features are not correlated with each other. Through comprehensive experiments, we demonstrate that MBCNet significantly improves the cross-modality retrieval performance in social surveillance systems and effectively addresses the modality bias training issue.\",\"PeriodicalId\":13044,\"journal\":{\"name\":\"IEEE Transactions on Computational Social Systems\",\"volume\":\"11 5\",\"pages\":\"6925-6938\"},\"PeriodicalIF\":4.5000,\"publicationDate\":\"2024-06-20\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"IEEE Transactions on Computational Social Systems\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://ieeexplore.ieee.org/document/10566608/\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, CYBERNETICS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Computational Social Systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10566608/","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
Modality Bias Calibration Network via Information Disentanglement for Visible–Infrared Person Reidentification
Visible–infrared person reidentification (VI-ReID) in social surveillance systems involves analyzing social behavior using nonoverlapping cross-modality camera sets. It often has poor retrieval performance under modality gap. One way to alleviate such the modality discrepancy is to learn shared person features that are generalizable across different modalities. However, because of significant differences in color between the visible and infrared images, the learned share features are always inclined to specific information of corresponding modality. To this end, we propose a modality bias calibration network (MBCNet) that filters out identity-irrelevant interference and recalibrates the learned modality-shared features. Specifically, to emphasize the modality-shared cues, we employ a feature decomposition module in the feature-level to filter out style variations and extract identity-relevant discriminative cues from the residual feature. In order to achieve a better disentanglement, a dual ranking entropy constraint is further proposed to ensure that the learned features contain only identity-relevant information and discard style-relevant information. Simultaneously, we design a decorrelated orthogonality Loss to ensure the disentangled features are not correlated with each other. Through comprehensive experiments, we demonstrate that MBCNet significantly improves the cross-modality retrieval performance in social surveillance systems and effectively addresses the modality bias training issue.
期刊介绍:
IEEE Transactions on Computational Social Systems focuses on such topics as modeling, simulation, analysis and understanding of social systems from the quantitative and/or computational perspective. "Systems" include man-man, man-machine and machine-machine organizations and adversarial situations as well as social media structures and their dynamics. More specifically, the proposed transactions publishes articles on modeling the dynamics of social systems, methodologies for incorporating and representing socio-cultural and behavioral aspects in computational modeling, analysis of social system behavior and structure, and paradigms for social systems modeling and simulation. The journal also features articles on social network dynamics, social intelligence and cognition, social systems design and architectures, socio-cultural modeling and representation, and computational behavior modeling, and their applications.