首页 > 最新文献

2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)最新文献

英文 中文
Continuous hierarchical exploration of multimedia collections 多媒体馆藏的持续分层探索
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153621
Tomás Grosup, Juraj Mosko, Premysl Cech
Preserving continuity between individual exploration steps in a process of multimedia exploration is a concept of natural intuition that sometimes decides if a particular exploration system is usable or not. One of ways how to emulate the continuity of the exploration process is adding some sort of granularity into this process. Then anyone who uses such system can explore particular areas in less or more details. In this paper we proposed new concept, hierarchical querying, which keeps consecutive steps of the exploration process more tight to each other. As a second concept, which directly supports the continuity of the exploration process, we proposed preservation of a user context between consecutive steps of the exploration process. In addition, we also presented an evaluation process and architecture design of our multimedia exploration system. For a validity confirmation of our ideas, we have implemented all proposed concepts in a web application that is accessible online.
在多媒体探索过程中保持各个探索步骤之间的连续性是一种自然直觉概念,它有时会决定特定探索系统是否可用。模拟探索过程连续性的方法之一是在这个过程中添加某种粒度。然后,任何使用这种系统的人都可以或多或少地探索特定领域的细节。在本文中,我们提出了分层查询的新概念,使连续的探索过程更加紧密。作为第二个直接支持探索过程连续性的概念,我们提出在探索过程的连续步骤之间保留用户上下文。此外,我们还提出了多媒体探索系统的评估过程和体系结构设计。为了验证我们的想法的有效性,我们在一个可在线访问的web应用程序中实现了所有提出的概念。
{"title":"Continuous hierarchical exploration of multimedia collections","authors":"Tomás Grosup, Juraj Mosko, Premysl Cech","doi":"10.1109/CBMI.2015.7153621","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153621","url":null,"abstract":"Preserving continuity between individual exploration steps in a process of multimedia exploration is a concept of natural intuition that sometimes decides if a particular exploration system is usable or not. One of ways how to emulate the continuity of the exploration process is adding some sort of granularity into this process. Then anyone who uses such system can explore particular areas in less or more details. In this paper we proposed new concept, hierarchical querying, which keeps consecutive steps of the exploration process more tight to each other. As a second concept, which directly supports the continuity of the exploration process, we proposed preservation of a user context between consecutive steps of the exploration process. In addition, we also presented an evaluation process and architecture design of our multimedia exploration system. For a validity confirmation of our ideas, we have implemented all proposed concepts in a web application that is accessible online.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122847139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Over-the-shoulder shot detection in art films 艺术电影中的过肩镜头检测
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153627
M. Svanera, Sergio Benini, N. Adami, R. Leonardi, A. Kovács
The ability to characterize a film, in terms of its narrative and style, is becoming a necessity especially for developing personal video recommendation systems to better deliver on-demand Internet streaming media. Among the set of identifiable stylistic features which play an important role in the film's emotional effects, the use of Over-the-shoulder (OtS) shots in movies is able to convey a big dramatic tension on the viewers. In this work we propose a methodology able to automatically detect this kind of shots by combining in a SVM learning scheme some state-of-the-art human presence detectors, with a set of saliency features based on colour and motion. In the experimental investigation, the comparison of obtained results with manual annotations made by cinema experts proves the validity of the framework. Experiments are conducted on two art films directed by Michelangelo Antonioni belonging to his famous “tetralogy on modernity and its discontent”, one in shades of gray (L'avventura, 1960), and the other in colour motion (Il deserto rosso, 1964).
从叙事和风格的角度来描述一部电影的特征,正变得越来越必要,尤其是对于开发个人视频推荐系统,以更好地提供按需的互联网流媒体来说。在一系列可识别的风格特征中,对电影的情感效果起着重要作用,电影中肩部镜头的使用能够给观众传达一种巨大的戏剧张力。在这项工作中,我们提出了一种能够自动检测这种镜头的方法,该方法将一些最先进的人类存在检测器与一组基于颜色和运动的显著性特征结合在一起。在实验研究中,将得到的结果与影院专家手工标注的结果进行了比较,证明了该框架的有效性。实验是对米开朗基罗·安东尼奥尼导演的两部艺术电影进行的,这两部电影属于他著名的“现代性及其不满四部曲”,一部是灰色阴影(《冒险》,1960),另一部是彩色运动(《红色沙漠》,1964)。
{"title":"Over-the-shoulder shot detection in art films","authors":"M. Svanera, Sergio Benini, N. Adami, R. Leonardi, A. Kovács","doi":"10.1109/CBMI.2015.7153627","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153627","url":null,"abstract":"The ability to characterize a film, in terms of its narrative and style, is becoming a necessity especially for developing personal video recommendation systems to better deliver on-demand Internet streaming media. Among the set of identifiable stylistic features which play an important role in the film's emotional effects, the use of Over-the-shoulder (OtS) shots in movies is able to convey a big dramatic tension on the viewers. In this work we propose a methodology able to automatically detect this kind of shots by combining in a SVM learning scheme some state-of-the-art human presence detectors, with a set of saliency features based on colour and motion. In the experimental investigation, the comparison of obtained results with manual annotations made by cinema experts proves the validity of the framework. Experiments are conducted on two art films directed by Michelangelo Antonioni belonging to his famous “tetralogy on modernity and its discontent”, one in shades of gray (L'avventura, 1960), and the other in colour motion (Il deserto rosso, 1964).","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116978870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A GPU-accelerated two stage visual matching pipeline for image and video retrieval 一种gpu加速的两阶段视觉匹配管道,用于图像和视频检索
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153620
Hannes Fassold, H. Stiegler, Jakub Rosner, M. Thaler, W. Bailer
We propose a two stage visual matching pipeline including a first step using VLAD signatures for filtering results, and a second step which reranks the top results using raw matching of SIFT descriptors. This enables adjusting the tradeoff between high computational cost of matching local descriptors and the insufficient accuracy of compact signatures in many application scenarios. We describe GPU accelerated extraction and matching algorithms for SIFT, which result in a speedup factor of at least 4. The VLAD filtering step reduces the number of images/frames for which the local descriptors need to be matched, thus speeding up retrieval by an additional factor of 9-10 without sacrificing mean average precision over full raw descriptor matching.
我们提出了一个两阶段的视觉匹配管道,包括第一步使用VLAD签名过滤结果,第二步使用SIFT描述符的原始匹配对顶级结果进行重新排序。在许多应用场景中,这可以调整匹配局部描述符的高计算成本和紧凑签名的不足准确性之间的权衡。我们描述了GPU加速的SIFT提取和匹配算法,其加速系数至少为4。VLAD过滤步骤减少了需要匹配局部描述符的图像/帧的数量,从而在不牺牲完整原始描述符匹配的平均精度的情况下,将检索速度提高了9-10倍。
{"title":"A GPU-accelerated two stage visual matching pipeline for image and video retrieval","authors":"Hannes Fassold, H. Stiegler, Jakub Rosner, M. Thaler, W. Bailer","doi":"10.1109/CBMI.2015.7153620","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153620","url":null,"abstract":"We propose a two stage visual matching pipeline including a first step using VLAD signatures for filtering results, and a second step which reranks the top results using raw matching of SIFT descriptors. This enables adjusting the tradeoff between high computational cost of matching local descriptors and the insufficient accuracy of compact signatures in many application scenarios. We describe GPU accelerated extraction and matching algorithms for SIFT, which result in a speedup factor of at least 4. The VLAD filtering step reduces the number of images/frames for which the local descriptors need to be matched, thus speeding up retrieval by an additional factor of 9-10 without sacrificing mean average precision over full raw descriptor matching.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132830175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
The Mex-Culture Multimedia platform: Preservation and dissemination of the Mexican Culture 墨西哥文化多媒体平台:墨西哥文化的保存与传播
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153624
A. Ramirez, J. Benois-Pineau, M. García-Vázquez, A. Stoian, M. Crucianu, M. Nakano-Miyatake, F. Garcia-Ugalde, Jean-Luc Rouas, H. Nicolas, J. Carrive
In this paper we present the Mex-Culture Multimedia platform, which is the first prototype of multimedia indexing and retrieval for a large-scale access to digitized Mexican cultural audio-visual content. The platform is designed as an open and extensible architecture of Web services. The different architectural layers and media services are presented, ensuring a rich set of scenarios. The latter comprises summarization of audio-visual content in cross-media description spaces, video queries by actions, key-frame and image queries by example and audio-analysis services. Specific attention is paid to the selection of data to be representative of Mexican cultural content. Scalability issues are addressed as well.
在本文中,我们提出了Mex-Culture多媒体平台,这是多媒体索引和检索的第一个原型,用于大规模访问数字化墨西哥文化视听内容。该平台被设计为一个开放的、可扩展的Web服务体系结构。展示了不同的体系结构层和媒体服务,确保了丰富的场景集。后者包括跨媒体描述空间的视听内容摘要、按动作进行视频查询、按示例进行关键帧和图像查询以及音频分析服务。特别注意数据的选择,以代表墨西哥的文化内容。可伸缩性问题也得到了解决。
{"title":"The Mex-Culture Multimedia platform: Preservation and dissemination of the Mexican Culture","authors":"A. Ramirez, J. Benois-Pineau, M. García-Vázquez, A. Stoian, M. Crucianu, M. Nakano-Miyatake, F. Garcia-Ugalde, Jean-Luc Rouas, H. Nicolas, J. Carrive","doi":"10.1109/CBMI.2015.7153624","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153624","url":null,"abstract":"In this paper we present the Mex-Culture Multimedia platform, which is the first prototype of multimedia indexing and retrieval for a large-scale access to digitized Mexican cultural audio-visual content. The platform is designed as an open and extensible architecture of Web services. The different architectural layers and media services are presented, ensuring a rich set of scenarios. The latter comprises summarization of audio-visual content in cross-media description spaces, video queries by actions, key-frame and image queries by example and audio-analysis services. Specific attention is paid to the selection of data to be representative of Mexican cultural content. Scalability issues are addressed as well.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128435467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Detection of ulcerative colitis severity in colonoscopy video frames 结肠镜检查录像中溃疡性结肠炎严重程度的检测
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153617
Ashok Dahal, Jung-Hwan Oh, Wallapak Tavanapong, J. Wong, P. C. Groen
Ulcerative colitis (UC) is a chronic inflammatory disease characterized by periods of relapses and remissions affecting more than 500,000 people in the United States. The therapeutic goals of UC are to first induce and then maintain disease remission. However, it is very difficult to evaluate the severity of UC objectively because of non-uniform nature of symptoms associated with UC, and large variations in their patterns. To address this, we objectively measure and classify the severity of UC presented in optical colonoscopy video frames based on the image textures. To extract distinct textures, we are using a hybrid approach in which a new proposed feature based on the accumulation of pixel value differences is combined with an existing feature such as LBP (Local Binary Pattern). The experimental results show the hybrid method can achieve more than 90% overall accuracy.
溃疡性结肠炎(UC)是一种以复发和缓解期为特征的慢性炎症性疾病,在美国影响了50多万人。UC的治疗目标是首先诱导并维持疾病缓解。然而,客观评估UC的严重程度是非常困难的,因为UC相关症状的不均匀性,以及其模式的巨大差异。为了解决这个问题,我们根据图像纹理客观地测量和分类光学结肠镜检查视频帧中UC的严重程度。为了提取不同的纹理,我们使用了一种混合方法,在这种方法中,基于像素值差异积累的新提出的特征与现有的特征(如LBP (Local Binary Pattern))相结合。实验结果表明,混合方法的总体准确率达到90%以上。
{"title":"Detection of ulcerative colitis severity in colonoscopy video frames","authors":"Ashok Dahal, Jung-Hwan Oh, Wallapak Tavanapong, J. Wong, P. C. Groen","doi":"10.1109/CBMI.2015.7153617","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153617","url":null,"abstract":"Ulcerative colitis (UC) is a chronic inflammatory disease characterized by periods of relapses and remissions affecting more than 500,000 people in the United States. The therapeutic goals of UC are to first induce and then maintain disease remission. However, it is very difficult to evaluate the severity of UC objectively because of non-uniform nature of symptoms associated with UC, and large variations in their patterns. To address this, we objectively measure and classify the severity of UC presented in optical colonoscopy video frames based on the image textures. To extract distinct textures, we are using a hybrid approach in which a new proposed feature based on the accumulation of pixel value differences is combined with an existing feature such as LBP (Local Binary Pattern). The experimental results show the hybrid method can achieve more than 90% overall accuracy.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133073961","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
VSD2014: A dataset for violent scenes detection in hollywood movies and web videos VSD2014:用于好莱坞电影和网络视频中暴力场景检测的数据集
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153604
M. Schedl, Mats Sjöberg, Ionut Mironica, B. Ionescu, Vu Lam Quang, Yu-Gang Jiang, C. Demarty
In this paper, we introduce a violent scenes and violence-related concept detection dataset named VSD2014. It contains annotations as well as auditory and visual features of Hollywood movies and user-generated footage shared on the web. The dataset is the result of a joint annotation endeavor of different research institutions and responds to the real-world use case of parental guidance in selecting appropriate content for children. The dataset has been validated during the Violent Scenes Detection (VSD) task at the MediaEval benchmarking initiative for multimedia evaluation.
本文介绍了一个名为VSD2014的暴力场景和暴力相关概念检测数据集。它包含注释以及好莱坞电影的听觉和视觉特征,以及在网络上共享的用户生成的镜头。该数据集是不同研究机构联合注释努力的结果,并响应父母指导为儿童选择适当内容的现实用例。该数据集已在MediaEval多媒体评估基准测试计划的暴力场景检测(VSD)任务中得到验证。
{"title":"VSD2014: A dataset for violent scenes detection in hollywood movies and web videos","authors":"M. Schedl, Mats Sjöberg, Ionut Mironica, B. Ionescu, Vu Lam Quang, Yu-Gang Jiang, C. Demarty","doi":"10.1109/CBMI.2015.7153604","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153604","url":null,"abstract":"In this paper, we introduce a violent scenes and violence-related concept detection dataset named VSD2014. It contains annotations as well as auditory and visual features of Hollywood movies and user-generated footage shared on the web. The dataset is the result of a joint annotation endeavor of different research institutions and responds to the real-world use case of parental guidance in selecting appropriate content for children. The dataset has been validated during the Violent Scenes Detection (VSD) task at the MediaEval benchmarking initiative for multimedia evaluation.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122640738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Event video retrieval using global and local descriptors in visual domain 基于全局和局部描述符的视觉域事件视频检索
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153622
Jennifer Roldan-Carlos, M. Lux, Xavier Giró-i-Nieto, P. Muñoz, N. Anagnostopoulos
With the advent of affordable multimedia smart phones, it has become common that people take videos when they are at events. The larger the event, the larger is the amount of videos taken there and also, the more videos get shared online. To search in this mass of videos is a challenging topic. In this paper we present and discuss a prototype software for searching in such videos. We focus only on visual information, and we report on experiments based on a research data set. With a small study we show that our prototype demonstrates promising results by identifying the same scene in different videos taken from different angles solely based on content based image retrieval.
随着多媒体智能手机的出现,人们在参加活动时拍摄视频已经变得很普遍。活动规模越大,拍摄的视频数量就越多,网上分享的视频也就越多。在这么多的视频中搜索是一个具有挑战性的话题。在本文中,我们提出并讨论了一个原型软件来搜索这些视频。我们只关注视觉信息,我们报告基于研究数据集的实验。通过一项小型研究,我们表明,我们的原型仅基于基于内容的图像检索,就可以识别从不同角度拍摄的不同视频中的相同场景,从而显示出有希望的结果。
{"title":"Event video retrieval using global and local descriptors in visual domain","authors":"Jennifer Roldan-Carlos, M. Lux, Xavier Giró-i-Nieto, P. Muñoz, N. Anagnostopoulos","doi":"10.1109/CBMI.2015.7153622","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153622","url":null,"abstract":"With the advent of affordable multimedia smart phones, it has become common that people take videos when they are at events. The larger the event, the larger is the amount of videos taken there and also, the more videos get shared online. To search in this mass of videos is a challenging topic. In this paper we present and discuss a prototype software for searching in such videos. We focus only on visual information, and we report on experiments based on a research data set. With a small study we show that our prototype demonstrates promising results by identifying the same scene in different videos taken from different angles solely based on content based image retrieval.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127902819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Instrument classification in laparoscopic videos 腹腔镜视频中的仪器分类
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153616
Manfred Jürgen Primus, Klaus Schöffmann, L. Böszörményi
In medical endoscopy more and more surgeons record videos of their interventions in a long-term storage archive for later retrieval. In order to allow content-based search in such endoscopic video archives, the video data needs to be indexed first. However, even the very basic step of content-based indexing, namely content segmentation, is already very challenging due to the special characteristics of such video data. Therefore, we propose to use instrument classification to enable semantic segmentation of laparoscopic videos. In this paper, we evaluate the performance of such an instrument classification approach. Our results show satisfying performance for all instruments used in our evaluation.
在医学内窥镜检查中,越来越多的外科医生将手术过程的视频记录在一个长期存储档案中,以便以后检索。为了在此类内窥镜视频档案中进行基于内容的搜索,首先需要对视频数据进行索引。然而,即使是基于内容的索引的最基本的步骤,即内容分割,由于这类视频数据的特殊特性,已经非常具有挑战性。因此,我们建议使用仪器分类来实现腹腔镜视频的语义分割。在本文中,我们评估了这种仪器分类方法的性能。我们的结果表明,在我们的评估中使用的所有仪器的性能都令人满意。
{"title":"Instrument classification in laparoscopic videos","authors":"Manfred Jürgen Primus, Klaus Schöffmann, L. Böszörményi","doi":"10.1109/CBMI.2015.7153616","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153616","url":null,"abstract":"In medical endoscopy more and more surgeons record videos of their interventions in a long-term storage archive for later retrieval. In order to allow content-based search in such endoscopic video archives, the video data needs to be indexed first. However, even the very basic step of content-based indexing, namely content segmentation, is already very challenging due to the special characteristics of such video data. Therefore, we propose to use instrument classification to enable semantic segmentation of laparoscopic videos. In this paper, we evaluate the performance of such an instrument classification approach. Our results show satisfying performance for all instruments used in our evaluation.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"88 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124374457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
An unsupervised approach for comparing styles of illustrations 一种用于比较插图风格的无监督方法
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153615
T. Furuya, Shigeru Kuriyama, Ryutarou Ohbuchi
In creating web pages, books, or presentation slides, consistent use of tasteful visual style(s) is quite important. In this paper, we consider the problem of style-based comparison and retrieval of illustrations. In their pioneering work, Garces et al. [2] proposed an algorithm for comparing illustrative style. The algorithm uses supervised learning that relied on stylistic labels present in a training dataset. In reality, obtaining such labels is quite difficult. In this paper, we propose an unsupervised approach to achieve accurate and efficient stylistic comparison among illustrations. The proposed algorithm combines heterogeneous local visual features extracted densely. These features are aggregated into a feature vector per illustration prior to be treated with distance metric learning based on unsupervised dimension reduction for saliency and compactness. Experimental evaluation of the proposed method by using multiple benchmark datasets indicates that the proposed method outperforms existing approaches.
在创建网页、书籍或演示幻灯片时,始终如一地使用有品位的视觉风格是非常重要的。在本文中,我们考虑了基于风格的插图比较和检索问题。Garces等人在他们的开创性工作中提出了一种比较插图风格的算法。该算法使用监督学习,依赖于训练数据集中存在的风格标签。实际上,获得这样的标签是相当困难的。在本文中,我们提出了一种无监督的方法来实现插图之间的准确和有效的风格比较。该算法结合了密集提取的异质局部视觉特征。在使用基于无监督降维的距离度量学习处理显著性和紧凑性之前,这些特征被聚合到每个插图的特征向量中。使用多个基准数据集对所提方法进行的实验评估表明,所提方法优于现有方法。
{"title":"An unsupervised approach for comparing styles of illustrations","authors":"T. Furuya, Shigeru Kuriyama, Ryutarou Ohbuchi","doi":"10.1109/CBMI.2015.7153615","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153615","url":null,"abstract":"In creating web pages, books, or presentation slides, consistent use of tasteful visual style(s) is quite important. In this paper, we consider the problem of style-based comparison and retrieval of illustrations. In their pioneering work, Garces et al. [2] proposed an algorithm for comparing illustrative style. The algorithm uses supervised learning that relied on stylistic labels present in a training dataset. In reality, obtaining such labels is quite difficult. In this paper, we propose an unsupervised approach to achieve accurate and efficient stylistic comparison among illustrations. The proposed algorithm combines heterogeneous local visual features extracted densely. These features are aggregated into a feature vector per illustration prior to be treated with distance metric learning based on unsupervised dimension reduction for saliency and compactness. Experimental evaluation of the proposed method by using multiple benchmark datasets indicates that the proposed method outperforms existing approaches.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126436927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Comparison of metric space browsing strategies for efficient image exploration 高效图像搜索的度量空间浏览策略比较
Pub Date : 2015-06-10 DOI: 10.1109/CBMI.2015.7153631
Premysl Cech, Tomás Grosup
In this paper, we compare eight different multimedia exploration methods. We describe each of them individually and evaluate their effectiveness in a user study focusing on different aspects of image exploration needs. We also created a testing scenario for the user study and defined several metrics to compare the exploration methods.
在本文中,我们比较了八种不同的多媒体探索方法。我们分别描述了它们中的每一个,并评估了它们在用户研究中的有效性,重点关注图像探索需求的不同方面。我们还为用户研究创建了一个测试场景,并定义了几个指标来比较勘探方法。
{"title":"Comparison of metric space browsing strategies for efficient image exploration","authors":"Premysl Cech, Tomás Grosup","doi":"10.1109/CBMI.2015.7153631","DOIUrl":"https://doi.org/10.1109/CBMI.2015.7153631","url":null,"abstract":"In this paper, we compare eight different multimedia exploration methods. We describe each of them individually and evaluate their effectiveness in a user study focusing on different aspects of image exploration needs. We also created a testing scenario for the user study and defined several metrics to compare the exploration methods.","PeriodicalId":387496,"journal":{"name":"2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-06-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130510792","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
2015 13th International Workshop on Content-Based Multimedia Indexing (CBMI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1