{"title":"Review on Deep Adversarial Learning of Entity Resolution for Cross-Modal Data","authors":"Yizhuo Rao, Chengyuan Duan, Xiao Wei","doi":"10.1109/ITCA52113.2020.00128","DOIUrl":null,"url":null,"abstract":"With the repaid development of the Internet, multimedia data such as image, text, video, audio is increasing, which brings opportunities and challenges to the development of the economy and science. Cross-modal data entity resolution aims to find different objective descriptions of the semantically similar items from objects in different modalities. However, different modality data have the features with underlying heterogeneity and high-level semantic related. Starting from the problem of modality gap between cross-modal data, this paper introduces how to use the idea of confrontational learning to solve the cross-modal data entity resolution problem between images and text from the aspects of feature extraction and emotional state association.","PeriodicalId":103309,"journal":{"name":"2020 2nd International Conference on Information Technology and Computer Application (ITCA)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2020-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2020 2nd International Conference on Information Technology and Computer Application (ITCA)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ITCA52113.2020.00128","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
With the repaid development of the Internet, multimedia data such as image, text, video, audio is increasing, which brings opportunities and challenges to the development of the economy and science. Cross-modal data entity resolution aims to find different objective descriptions of the semantically similar items from objects in different modalities. However, different modality data have the features with underlying heterogeneity and high-level semantic related. Starting from the problem of modality gap between cross-modal data, this paper introduces how to use the idea of confrontational learning to solve the cross-modal data entity resolution problem between images and text from the aspects of feature extraction and emotional state association.