{"title":"Bag-of-Visual-Phrases via Local Contexts","authors":"E. Román-Rangel, S. Marchand-Maillet","doi":"10.1109/ACPR.2013.158","DOIUrl":null,"url":null,"abstract":"This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 2nd IAPR Asian Conference on Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACPR.2013.158","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3
Abstract
This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.