{"title":"基于深度无监督关键帧哈希的近重复视频检索","authors":"Wenhao Zhao, Shijiao Yang, Mengqun Jin","doi":"10.1109/CSE53436.2021.00021","DOIUrl":null,"url":null,"abstract":"In recent years, original videos are often re-edited, modified and redistributed, which not only cause copyright problems, but also deteriorate users’ experience. Near-duplicate video retrieval based on learning to hash has been widely concerned by people. However, there are still two major defects with existing methods. Firstly, the information capacity of hash code needs to be maximized. Secondly, the retrieval efficiency of partially repeated video is insufficient. In this paper, we propose a near-duplicate video retrieval method based on key frame hashing to improve retrieval performance. We design the semi-distributed hash layer to force the distribution of the continuous key frame hash code to approach the optimal distribution, i.e., the half-half distribution. By minimizing the semantic loss, quantization loss, and bit uncorrelated loss, we train our model to generate compact binary hash codes. To retrieve partially repeated videos, the proposed video subsequence matching method can accurately locate the near-duplicate fragments between the queried video and the target video. Experiments on two public datasets present that the mean average precision (MAP) of our hashing method is 0.63, which effectively improves the accuracy of video retrieval.","PeriodicalId":6838,"journal":{"name":"2021 IEEE 24th International Conference on Computational Science and Engineering (CSE)","volume":"117 1","pages":"80-86"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Near-duplicate Video Retrieval Based on Deep Unsupervised Key Frame Hashing\",\"authors\":\"Wenhao Zhao, Shijiao Yang, Mengqun Jin\",\"doi\":\"10.1109/CSE53436.2021.00021\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In recent years, original videos are often re-edited, modified and redistributed, which not only cause copyright problems, but also deteriorate users’ experience. Near-duplicate video retrieval based on learning to hash has been widely concerned by people. However, there are still two major defects with existing methods. Firstly, the information capacity of hash code needs to be maximized. Secondly, the retrieval efficiency of partially repeated video is insufficient. In this paper, we propose a near-duplicate video retrieval method based on key frame hashing to improve retrieval performance. We design the semi-distributed hash layer to force the distribution of the continuous key frame hash code to approach the optimal distribution, i.e., the half-half distribution. By minimizing the semantic loss, quantization loss, and bit uncorrelated loss, we train our model to generate compact binary hash codes. To retrieve partially repeated videos, the proposed video subsequence matching method can accurately locate the near-duplicate fragments between the queried video and the target video. Experiments on two public datasets present that the mean average precision (MAP) of our hashing method is 0.63, which effectively improves the accuracy of video retrieval.\",\"PeriodicalId\":6838,\"journal\":{\"name\":\"2021 IEEE 24th International Conference on Computational Science and Engineering (CSE)\",\"volume\":\"117 1\",\"pages\":\"80-86\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 IEEE 24th International Conference on Computational Science and Engineering (CSE)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CSE53436.2021.00021\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 IEEE 24th International Conference on Computational Science and Engineering (CSE)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSE53436.2021.00021","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Near-duplicate Video Retrieval Based on Deep Unsupervised Key Frame Hashing
In recent years, original videos are often re-edited, modified and redistributed, which not only cause copyright problems, but also deteriorate users’ experience. Near-duplicate video retrieval based on learning to hash has been widely concerned by people. However, there are still two major defects with existing methods. Firstly, the information capacity of hash code needs to be maximized. Secondly, the retrieval efficiency of partially repeated video is insufficient. In this paper, we propose a near-duplicate video retrieval method based on key frame hashing to improve retrieval performance. We design the semi-distributed hash layer to force the distribution of the continuous key frame hash code to approach the optimal distribution, i.e., the half-half distribution. By minimizing the semantic loss, quantization loss, and bit uncorrelated loss, we train our model to generate compact binary hash codes. To retrieve partially repeated videos, the proposed video subsequence matching method can accurately locate the near-duplicate fragments between the queried video and the target video. Experiments on two public datasets present that the mean average precision (MAP) of our hashing method is 0.63, which effectively improves the accuracy of video retrieval.