Yingfan Liu , Xiaotian Qiao , Zhaoqing Liu , Xiaofang Xia , Yinlong Zhang , Jiangtao Cui
{"title":"Deep multi-negative supervised hashing for large-scale image retrieval","authors":"Yingfan Liu , Xiaotian Qiao , Zhaoqing Liu , Xiaofang Xia , Yinlong Zhang , Jiangtao Cui","doi":"10.1016/j.eswa.2024.125795","DOIUrl":null,"url":null,"abstract":"<div><div>Deep hashing has been widely applied in large-scale image retrieval due to its robust retrieval performance and high efficiency in computation/storage. Most deep supervised hashing methods exploit pairwise sampling or triplet sampling for optimization. However, these methods can only leverage the intrinsic structure information of a small subset of images, resulting in a suboptimal retrieval space for deep hashing methods. For addressing this limitation, we propose a novel deep multi-negative supervised hashing (DMNSH) method whose basic idea is to sample a positive and multiple negatives for an anchor, in order to leverage more structure and supervise information during the training process. For improving the training efficiency of the convolutional neural networks (CNN) for large-scale image retrieval, the DMNSH method adopts a mini-batch optimization strategy. A sample reusing strategy is proposed to construct multi-negative tuples efficiently with limited training images in the mini-batch during each round of optimization. To perform multi-negative learning, we further design a multi-negative loss function in which hashing codes are relaxed as CNN output features. By minimizing this multi-negative loss function, the similarity semantics of images are preserved, and the quantization errors of relaxing are minimized. For making the optimization process more stable, an adaptive margin is further incorporated into the loss function to improve the retrieval performance. The stochastic gradient descent and backpropagation strategies are employed to optimize the CNN parameters. Experimental results on three popular deep hashing datasets demonstrate that DMNSH significantly outperforms existing state-of-the-art hashing methods in terms of both precision and efficiency.</div></div>","PeriodicalId":50461,"journal":{"name":"Expert Systems with Applications","volume":"264 ","pages":"Article 125795"},"PeriodicalIF":7.5000,"publicationDate":"2024-11-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Expert Systems with Applications","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0957417424026629","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Deep hashing has been widely applied in large-scale image retrieval due to its robust retrieval performance and high efficiency in computation/storage. Most deep supervised hashing methods exploit pairwise sampling or triplet sampling for optimization. However, these methods can only leverage the intrinsic structure information of a small subset of images, resulting in a suboptimal retrieval space for deep hashing methods. For addressing this limitation, we propose a novel deep multi-negative supervised hashing (DMNSH) method whose basic idea is to sample a positive and multiple negatives for an anchor, in order to leverage more structure and supervise information during the training process. For improving the training efficiency of the convolutional neural networks (CNN) for large-scale image retrieval, the DMNSH method adopts a mini-batch optimization strategy. A sample reusing strategy is proposed to construct multi-negative tuples efficiently with limited training images in the mini-batch during each round of optimization. To perform multi-negative learning, we further design a multi-negative loss function in which hashing codes are relaxed as CNN output features. By minimizing this multi-negative loss function, the similarity semantics of images are preserved, and the quantization errors of relaxing are minimized. For making the optimization process more stable, an adaptive margin is further incorporated into the loss function to improve the retrieval performance. The stochastic gradient descent and backpropagation strategies are employed to optimize the CNN parameters. Experimental results on three popular deep hashing datasets demonstrate that DMNSH significantly outperforms existing state-of-the-art hashing methods in terms of both precision and efficiency.
期刊介绍:
Expert Systems With Applications is an international journal dedicated to the exchange of information on expert and intelligent systems used globally in industry, government, and universities. The journal emphasizes original papers covering the design, development, testing, implementation, and management of these systems, offering practical guidelines. It spans various sectors such as finance, engineering, marketing, law, project management, information management, medicine, and more. The journal also welcomes papers on multi-agent systems, knowledge management, neural networks, knowledge discovery, data mining, and other related areas, excluding applications to military/defense systems.