{"title":"通过本地语境的视觉短语袋","authors":"E. Román-Rangel, S. Marchand-Maillet","doi":"10.1109/ACPR.2013.158","DOIUrl":null,"url":null,"abstract":"This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.","PeriodicalId":365633,"journal":{"name":"2013 2nd IAPR Asian Conference on Pattern Recognition","volume":"48 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2013-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Bag-of-Visual-Phrases via Local Contexts\",\"authors\":\"E. Román-Rangel, S. Marchand-Maillet\",\"doi\":\"10.1109/ACPR.2013.158\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.\",\"PeriodicalId\":365633,\"journal\":{\"name\":\"2013 2nd IAPR Asian Conference on Pattern Recognition\",\"volume\":\"48 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2013-11-05\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2013 2nd IAPR Asian Conference on Pattern Recognition\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ACPR.2013.158\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2013 2nd IAPR Asian Conference on Pattern Recognition","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACPR.2013.158","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
This paper extends the bag-of-visual-words representations to a bag-of-visual-phrases model. The introduced bag-of-visual-phrases representation is constructed upon a proposed method for probabilistic description of co-occurring visual words, which is adapted for each reference word. This bag-of-visual-phrases representation implicitly encodes spatial relationships among visual words, thus being a richer representation while remaining as compact as the bag-of-visual-words model. We demonstrate the effectiveness of our method with a series of statistical analysis and retrieval experiments, and show that it largely outperforms previous methods for construction of bag representations. Furthermore, our method allows to query traditional bag-of-words vs the proposed bag-of-phrases. We conducted retrieval experiments on a dataset of complex shapes, whose instances correspond to hieroglyphs of the pre-Columbian Maya culture from the ancient Americas.