FMCNet+: Feature-Level Modality Compensation for Visible-Infrared Person Re-Identification

IF 8.9 1区 计算机科学 Q1 COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE IEEE transactions on neural networks and learning systems Pub Date : 2024-10-15 DOI:10.1109/TNNLS.2024.3453292
Ruida Xi;Nianchang Huang;Changzhou Lai;Qiang Zhang;Jungong Han
{"title":"FMCNet+: Feature-Level Modality Compensation for Visible-Infrared Person Re-Identification","authors":"Ruida Xi;Nianchang Huang;Changzhou Lai;Qiang Zhang;Jungong Han","doi":"10.1109/TNNLS.2024.3453292","DOIUrl":null,"url":null,"abstract":"For visible-infrared person re-identification (VI-ReID), current models that compensate modality-specific information strive to generate missing modality images from existing ones to bridge the cross-modality discrepancies. Despite that, those generated images often suffer from low qualities due to the significant modality gap and include interfering information, e.g., inconsistent colors, thus severely degrading the subsequent VI-ReID performance. Alternatively, we propose a feature-level modality compensation network, i.e., FMCNet+, for VI-ReID in this article as an improved version of our previous work (FMCNet). The core of FMCNet+ is to compensate for the missing modality-specific information at the feature level, rather than at the image level, enabling our model to generate more person-related and discriminative modality-specific features for VI-ReID. Concretely, FMCNet+ aims to progressively generate missing modality-specific features by fully exploring the relationships among single-modality features, modality-shared features, and modality-specific features, instead of directly generating them through a generative adversarial way as in the previous FMCNet. To this end, three modules, i.e., single-modality feature decomposition (SFD), modality characteristic dictionary learning (MCDL), and missing modality-specific feature compensation (MMFC), are incorporated in FMCNet+. Experimental results demonstrate the superiority of our proposed FMCNet+ over existing ones, especially for those that compensate for modality-specific information at the image level. Our intriguing findings highlight the necessity of feature-level modality compensation in VI-ReID. Our code and pre-trained models will be released on <uri>https://github.com/jssyzsfzy/FMCNet_series</uri>.","PeriodicalId":13303,"journal":{"name":"IEEE transactions on neural networks and learning systems","volume":"36 7","pages":"13247-13261"},"PeriodicalIF":8.9000,"publicationDate":"2024-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE transactions on neural networks and learning systems","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10718348/","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0

Abstract

For visible-infrared person re-identification (VI-ReID), current models that compensate modality-specific information strive to generate missing modality images from existing ones to bridge the cross-modality discrepancies. Despite that, those generated images often suffer from low qualities due to the significant modality gap and include interfering information, e.g., inconsistent colors, thus severely degrading the subsequent VI-ReID performance. Alternatively, we propose a feature-level modality compensation network, i.e., FMCNet+, for VI-ReID in this article as an improved version of our previous work (FMCNet). The core of FMCNet+ is to compensate for the missing modality-specific information at the feature level, rather than at the image level, enabling our model to generate more person-related and discriminative modality-specific features for VI-ReID. Concretely, FMCNet+ aims to progressively generate missing modality-specific features by fully exploring the relationships among single-modality features, modality-shared features, and modality-specific features, instead of directly generating them through a generative adversarial way as in the previous FMCNet. To this end, three modules, i.e., single-modality feature decomposition (SFD), modality characteristic dictionary learning (MCDL), and missing modality-specific feature compensation (MMFC), are incorporated in FMCNet+. Experimental results demonstrate the superiority of our proposed FMCNet+ over existing ones, especially for those that compensate for modality-specific information at the image level. Our intriguing findings highlight the necessity of feature-level modality compensation in VI-ReID. Our code and pre-trained models will be released on https://github.com/jssyzsfzy/FMCNet_series.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
FMCNet$+$:用于可见光-红外线人员再识别的特征级模态补偿
对于可见-红外人体再识别(VI-ReID),当前补偿模态特定信息的模型努力从现有模态图像中生成缺失的模态图像,以弥合跨模态差异。尽管如此,这些生成的图像往往由于模态差距较大而质量较低,并且包含干扰信息,例如颜色不一致,从而严重降低了随后的VI-ReID性能。另外,我们在本文中为VI-ReID提出了一个特征级模态补偿网络,即FMCNet+,作为我们之前工作(FMCNet)的改进版本。FMCNet+的核心是在特征级而不是在图像级补偿缺失的特定于模态的信息,使我们的模型能够为VI-ReID生成更多与人相关且具有判别性的特定于模态的特征。具体而言,FMCNet+旨在通过充分探索单模态特征、模态共享特征和模态特定特征之间的关系,逐步生成缺失的模态特定特征,而不是像以前的FMCNet那样,通过生成对抗的方式直接生成。为此,FMCNet+中加入了三个模块,即单模态特征分解(SFD)、模态特征字典学习(MCDL)和缺失模态特定特征补偿(MMFC)。实验结果表明,我们提出的FMCNet+优于现有的FMCNet+,特别是在图像级补偿模态特定信息的FMCNet+。我们有趣的发现强调了VI-ReID中特征级模态补偿的必要性。我们的代码和预训练模型将在https://github.com/jssyzsfzy/FMCNet_series上发布。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
IEEE transactions on neural networks and learning systems
IEEE transactions on neural networks and learning systems COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE-COMPUTER SCIENCE, HARDWARE & ARCHITECTURE
CiteScore
23.80
自引率
9.60%
发文量
2102
审稿时长
3-8 weeks
期刊介绍: The focus of IEEE Transactions on Neural Networks and Learning Systems is to present scholarly articles discussing the theory, design, and applications of neural networks as well as other learning systems. The journal primarily highlights technical and scientific research in this domain.
期刊最新文献
A Deep Neural Network Optimization Framework Based on Optimal Transport Bridge Feature Selection and Sparse Representation. A Dual-Network Framework With Adversarial GMM Augmentation and Frequency-Mamba Fusion for Hyperspectral Target Detection. Disentangled Generative Graph Representation Learning Adaptive Prototype-Guided Personalized Propagation for Heterophilic Graphs With Missing Data. Causal Counterfactual Inference Network for Video Object State Changes in Open-World Scenarios.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1