Semantic-Consistent Deep Quantization for Cross-modal Retrieval

Liya Ma, N. Zhang, Kuang-I Shu, Xitao Zou
{"title":"Semantic-Consistent Deep Quantization for Cross-modal Retrieval","authors":"Liya Ma, N. Zhang, Kuang-I Shu, Xitao Zou","doi":"10.1109/ICICIP53388.2021.9642180","DOIUrl":null,"url":null,"abstract":"With making up for the deficiency of the constraint representation capability of hashing codes for high-dimensional data, the quantization method has been found to generally perform better in cross-modal similarity retrieval research. However, in current quantization approaches, the codebook, as the most critical basis for quantization, is still in a passive status and detached from the learning framework. To improve the initiative of codebook, we propose a semantic-consistent deep quantization (SCDQ), which is the first scheme to integrate quantization into deep network learning in an end-to-end fashion. Specifically, two classifiers following the deep representation learning networks are formulated to produce the class-wise abstract patterns with the help of label alignment. Meanwhile, our approach learns a collaborative codebook for both modalities, which embeds bimodality semantic consistent information in codewords and bridges the relationship between the patterns in classifiers and codewords in codebook. By designing a novel algorithm architecture and codebook update strategy, SCDQ enables effective and efficient cross-modal retrieval in an asymmetric way. Extensive experiments on two benchmark datasets demonstrate that SCDQ yields optimal cross-modal retrieval performance and outperforms several state of-the-art cross-modal retrieval methods.","PeriodicalId":435799,"journal":{"name":"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)","volume":"154 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 11th International Conference on Intelligent Control and Information Processing (ICICIP)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICICIP53388.2021.9642180","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

With making up for the deficiency of the constraint representation capability of hashing codes for high-dimensional data, the quantization method has been found to generally perform better in cross-modal similarity retrieval research. However, in current quantization approaches, the codebook, as the most critical basis for quantization, is still in a passive status and detached from the learning framework. To improve the initiative of codebook, we propose a semantic-consistent deep quantization (SCDQ), which is the first scheme to integrate quantization into deep network learning in an end-to-end fashion. Specifically, two classifiers following the deep representation learning networks are formulated to produce the class-wise abstract patterns with the help of label alignment. Meanwhile, our approach learns a collaborative codebook for both modalities, which embeds bimodality semantic consistent information in codewords and bridges the relationship between the patterns in classifiers and codewords in codebook. By designing a novel algorithm architecture and codebook update strategy, SCDQ enables effective and efficient cross-modal retrieval in an asymmetric way. Extensive experiments on two benchmark datasets demonstrate that SCDQ yields optimal cross-modal retrieval performance and outperforms several state of-the-art cross-modal retrieval methods.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
跨模态检索的语义一致深度量化
量化方法弥补了哈希码对高维数据约束表示能力的不足,在跨模态相似性检索研究中普遍表现较好。然而,在目前的量化方法中,码本作为量化最关键的基础,仍然处于被动状态,脱离了学习框架。为了提高码本的主动性,我们提出了语义一致深度量化(SCDQ),这是第一个将量化以端到端方式集成到深度网络学习中的方案。具体来说,两个分类器遵循深度表示学习网络,在标签对齐的帮助下产生类智能抽象模式。同时,我们的方法学习了一个两模态的协作码本,它在码字中嵌入了双模态语义一致信息,并在分类器模式和码字之间架起了桥梁。通过设计一种新颖的算法架构和码本更新策略,SCDQ能够以不对称的方式实现高效的跨模态检索。在两个基准数据集上的大量实验表明,SCDQ产生了最佳的跨模态检索性能,并且优于几种最先进的跨模态检索方法。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
A novel RBF neural network based recognition of human upper limb active motion intention Time-Varying Polar Decomposition by Continuous-Time Model and Discrete-Time Algorithm of Zeroing Neural Network Using Zhang Time Discretization (ZTD) Integrated Res2Net combined with Seesaw loss for Long-Tailed PCG signal classification On Pinning Synchronization of An Array of Linearly Coupled Dynamical Network Design and Implementation of Braking Control for Hybrid Electric Vehicles
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1