首页 > 最新文献

2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)最新文献

英文 中文
Animated movie genre detection using symbolic fusion of text and image descriptors 基于文本和图像描述符符号融合的动画电影类型检测
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269813
Grégory Païs, P. Lambert, Daniel Beauchêne, F. Deloule, B. Ionescu
This paper addresses the automatic movie genre classification in the specific case of animated movies. Two types of information are used. The first one are movie synopsis. For each genre, a symbolic representation of a thematic intensity is extracted from synopsis. Addressed visually, movie content is described with symbolic representations of different mid-level color and activity features. A fusion between the text and image descriptions is performed using a set of symbolic rules conveying human expertise. The approach is tested on a set of 107 animated movies in order to estimate their ”drama” character. It is observed that the text-image fusion achieves a precision up to 78% and a recall of 44%.
本文以动画电影为例,研究了电影类型的自动分类。使用了两种类型的信息。第一个是电影大纲。对于每种类型,从概要中提取出主题强度的符号表示。从视觉上讲,电影内容是用不同中级颜色和活动特征的符号表示来描述的。文本和图像描述之间的融合是使用一组传递人类专业知识的符号规则来完成的。该方法在107部动画电影中进行了测试,以评估它们的“戏剧性”特征。结果表明,文本图像融合的准确率高达78%,召回率为44%。
{"title":"Animated movie genre detection using symbolic fusion of text and image descriptors","authors":"Grégory Païs, P. Lambert, Daniel Beauchêne, F. Deloule, B. Ionescu","doi":"10.1109/CBMI.2012.6269813","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269813","url":null,"abstract":"This paper addresses the automatic movie genre classification in the specific case of animated movies. Two types of information are used. The first one are movie synopsis. For each genre, a symbolic representation of a thematic intensity is extracted from synopsis. Addressed visually, movie content is described with symbolic representations of different mid-level color and activity features. A fusion between the text and image descriptions is performed using a set of symbolic rules conveying human expertise. The approach is tested on a set of 107 animated movies in order to estimate their ”drama” character. It is observed that the text-image fusion achieves a precision up to 78% and a recall of 44%.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129814655","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Effective concept detection using Second order Co-occurence Flickr context similarity measure SOCFCS 基于二阶共现Flickr上下文相似度测度SOCFCS的有效概念检测
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269846
Amel Ksibi, A. Ammar, C. Amar
Automatic photo annotation task aims to describe the semantic content by detecting high level concepts. Most existing approaches are performed by training independent concept detectors omitting the interdependencies between concepts. The obtained annotations are often not so satisfactory. Therefore, a process of annotation refinement is mondatory to improve the imprecise annotation results. Recently, harnessing the contextual correlation between concepts is shown to be an important resource to improve concept detection. In this paper, we propose a new context based concept detection process. For this purpose, we define a new semantic measure called Second order Co-occurence Flickr context similarity (SOCFCS), which aggregates the FCS values of common Flickr related-tags of two target concepts in order to calculate their relative semantic context relatedness (SCR). Our proposed measure is applied to build a concept network as the context space. A Random Walk with Restart process is performed over this network to refine the annotation results by exploring the contextual correlation among concepts. Experimental studies are conducted on ImageCLEF 2011 Collection containing 99 concepts. The results demonstrate the effectiveness of our proposed approach.
照片自动标注任务旨在通过检测高层次的概念来描述语义内容。大多数现有的方法是通过训练独立的概念检测器来执行的,忽略了概念之间的相互依赖性。得到的注解往往不那么令人满意。因此,为了改善标注不精确的结果,需要对标注进行细化。近年来,利用概念之间的上下文相关性被证明是提高概念检测的重要资源。本文提出了一种新的基于上下文的概念检测方法。为此,我们定义了一种新的语义度量,称为二阶共现Flickr上下文相似性(SOCFCS),它将两个目标概念的常见Flickr相关标签的FCS值聚合在一起,以计算它们的相对语义上下文相关性(SCR)。我们提出的方法被应用于构建一个概念网络作为上下文空间。在该网络上执行随机行走(Random Walk with Restart)过程,通过探索概念之间的上下文相关性来改进注释结果。在包含99个概念的ImageCLEF 2011 Collection上进行实验研究。结果证明了我们所提出的方法的有效性。
{"title":"Effective concept detection using Second order Co-occurence Flickr context similarity measure SOCFCS","authors":"Amel Ksibi, A. Ammar, C. Amar","doi":"10.1109/CBMI.2012.6269846","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269846","url":null,"abstract":"Automatic photo annotation task aims to describe the semantic content by detecting high level concepts. Most existing approaches are performed by training independent concept detectors omitting the interdependencies between concepts. The obtained annotations are often not so satisfactory. Therefore, a process of annotation refinement is mondatory to improve the imprecise annotation results. Recently, harnessing the contextual correlation between concepts is shown to be an important resource to improve concept detection. In this paper, we propose a new context based concept detection process. For this purpose, we define a new semantic measure called Second order Co-occurence Flickr context similarity (SOCFCS), which aggregates the FCS values of common Flickr related-tags of two target concepts in order to calculate their relative semantic context relatedness (SCR). Our proposed measure is applied to build a concept network as the context space. A Random Walk with Restart process is performed over this network to refine the annotation results by exploring the contextual correlation among concepts. Experimental studies are conducted on ImageCLEF 2011 Collection containing 99 concepts. The results demonstrate the effectiveness of our proposed approach.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133482781","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Quaero-MSSE: A content based multimedia indexing prototype Quaero-MSSE:一个基于内容的多媒体索引原型
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269805
Jean-Philippe Cabanal
Quaero-MSSE is an applicative project of the Quaero collaborative program. This project develops a multimedia search and navigation demonstrator, which gives access to several types of contents (catch-up TV, archives, foreign videos, music) and which illustrates the benefits of advanced audio-video analysis technologies.
Quaero- msse是Quaero协作计划的一个应用项目。该项目开发了一个多媒体搜索和导航演示器,它可以访问几种类型的内容(流行电视、档案、外国视频、音乐),并说明了先进音视频分析技术的好处。
{"title":"Quaero-MSSE: A content based multimedia indexing prototype","authors":"Jean-Philippe Cabanal","doi":"10.1109/CBMI.2012.6269805","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269805","url":null,"abstract":"Quaero-MSSE is an applicative project of the Quaero collaborative program. This project develops a multimedia search and navigation demonstrator, which gives access to several types of contents (catch-up TV, archives, foreign videos, music) and which illustrates the benefits of advanced audio-video analysis technologies.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117331129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An empirical study of fusion operators for multimodal image retrieval 多模态图像检索融合算子的实证研究
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269843
G. Csurka, S. Clinchant
In this paper we propose an empirical study of late fusion operators for multimodal image retrieval. Therefore, we consider two experts, one based on textual and one on visual similarities between documents and study the possibilities to go beyond simple score averaging. The main idea is to exploit the correlation between the two experts by encoding explicitly or implicitly an "and" and an "or" operator in an efficient way. We show through several experiments that the operators that combine both of these two aspects generally outperform the ones that look only to one of them. Based on this observation we propose several generalized version of most classical fusion operators and compare them using ImageClef benchmark datasets both in an unsupervised and in a supervised framework.
本文提出了一种用于多模态图像检索的后期融合算子的实证研究。因此,我们考虑两个专家,一个基于文本,另一个基于文档之间的视觉相似性,并研究超越简单平均得分的可能性。其主要思想是以一种有效的方式显式或隐式地编码“和”和“或”操作符,从而利用两个专家之间的相关性。我们通过几个实验表明,将这两个方面结合起来的运营商通常比只关注其中一个方面的运营商表现更好。基于这一观察,我们提出了几种经典融合算子的广义版本,并使用ImageClef基准数据集在无监督和有监督框架下对它们进行了比较。
{"title":"An empirical study of fusion operators for multimodal image retrieval","authors":"G. Csurka, S. Clinchant","doi":"10.1109/CBMI.2012.6269843","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269843","url":null,"abstract":"In this paper we propose an empirical study of late fusion operators for multimodal image retrieval. Therefore, we consider two experts, one based on textual and one on visual similarities between documents and study the possibilities to go beyond simple score averaging. The main idea is to exploit the correlation between the two experts by encoding explicitly or implicitly an \"and\" and an \"or\" operator in an efficient way. We show through several experiments that the operators that combine both of these two aspects generally outperform the ones that look only to one of them. Based on this observation we propose several generalized version of most classical fusion operators and compare them using ImageClef benchmark datasets both in an unsupervised and in a supervised framework.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127655691","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
3D model retrieval using the 2D Poisson equation 使用二维泊松方程的三维模型检索
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269797
Fattah Alizadeh, Alistair Sutherland
3D Model Retrieval is one of the most popular topics in computer vision and huge efforts are dedicated to finding a way to improve retrieval accuracy. Defining a new efficient and effective way to describe 3D models plays a critical role in the retrieval process. In this paper we propose a view-based shape signature to search and retrieve 3D objects using the 2D Poisson equation. Our proposed method uses 60 different 2D silhouettes, which are automatically extracted from different view-angles of 3D models. Solving the Poisson equation for each Silhouette assigns a number to each pixel as the pixel's signature. Counting and accumulating these pixel signatures generates a histogram-based signature for each silhouette (Silhouette Poisson Histogram or simply SilPH). By doing some preprocessing steps one can see that the signature is insensitive to rotation, scaling and translation. The results show a high power of discrimination on the McGill dataset and demonstrate that the proposed method outperforms other existing methods.
三维模型检索是计算机视觉中最热门的课题之一,人们一直在努力寻找提高检索精度的方法。定义一种新的高效的描述三维模型的方法在检索过程中起着至关重要的作用。在本文中,我们提出了一种基于视图的形状签名,利用二维泊松方程来搜索和检索三维物体。我们提出的方法使用60种不同的2D轮廓,这些轮廓从3D模型的不同视角自动提取。求解每个轮廓的泊松方程为每个像素分配一个数字作为像素的签名。计数和累积这些像素签名为每个轮廓生成基于直方图的签名(轮廓泊松直方图或简称SilPH)。通过做一些预处理步骤,可以看到签名对旋转、缩放和平移不敏感。结果表明,该方法在McGill数据集上具有很高的分辨能力,并且优于其他现有方法。
{"title":"3D model retrieval using the 2D Poisson equation","authors":"Fattah Alizadeh, Alistair Sutherland","doi":"10.1109/CBMI.2012.6269797","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269797","url":null,"abstract":"3D Model Retrieval is one of the most popular topics in computer vision and huge efforts are dedicated to finding a way to improve retrieval accuracy. Defining a new efficient and effective way to describe 3D models plays a critical role in the retrieval process. In this paper we propose a view-based shape signature to search and retrieve 3D objects using the 2D Poisson equation. Our proposed method uses 60 different 2D silhouettes, which are automatically extracted from different view-angles of 3D models. Solving the Poisson equation for each Silhouette assigns a number to each pixel as the pixel's signature. Counting and accumulating these pixel signatures generates a histogram-based signature for each silhouette (Silhouette Poisson Histogram or simply SilPH). By doing some preprocessing steps one can see that the signature is insensitive to rotation, scaling and translation. The results show a high power of discrimination on the McGill dataset and demonstrate that the proposed method outperforms other existing methods.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121090437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Supervised models for multimodal image retrieval based on visual, semantic and geographic information 基于视觉、语义和地理信息的多模态图像检索监督模型
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269806
Duc-Tien Dang-Nguyen, G. Boato, Alessandro Moschitti, F. D. Natale
Nowadays, large-scale networked social media need better search technologies to achieve suitable performance. Multimodal approaches are promising technologies to improve image ranking. This is particularly true when metadata are not completely reliable, which is a rather common case as far as user annotation, time and location are concerned. In this paper, we propose to properly combine visual information with additional multi-faceted information, to define a novel multimodal similarity measure. More specifically, we combine visual features, which strongly relate to the image content, with semantic information represented by manually annotated concepts, and geo tagging, very often available in the form of object/subject location. Furthermore, we propose a supervised machine learning approach, based on Support Vector Machines (SVMs), to automatically learn optimized weights to combine the above features. The resulting models is used as a ranking function to sort the results of a multimodal query.
如今,大规模网络社交媒体需要更好的搜索技术来实现合适的性能。多模态方法是改进图像排名的有前途的技术。在元数据不完全可靠的情况下,尤其如此,就用户注释、时间和地点而言,这种情况相当普遍。在本文中,我们建议将视觉信息与额外的多方面信息适当结合起来,从而定义一种新的多模态相似度测量方法。更具体地说,我们将与图像内容密切相关的视觉特征与人工标注概念所代表的语义信息和地理标记(通常以对象/主体位置的形式提供)相结合。此外,我们还提出了一种基于支持向量机(SVM)的监督机器学习方法,以自动学习优化权重,从而将上述特征结合起来。由此产生的模型可用作排序函数,对多模态查询结果进行排序。
{"title":"Supervised models for multimodal image retrieval based on visual, semantic and geographic information","authors":"Duc-Tien Dang-Nguyen, G. Boato, Alessandro Moschitti, F. D. Natale","doi":"10.1109/CBMI.2012.6269806","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269806","url":null,"abstract":"Nowadays, large-scale networked social media need better search technologies to achieve suitable performance. Multimodal approaches are promising technologies to improve image ranking. This is particularly true when metadata are not completely reliable, which is a rather common case as far as user annotation, time and location are concerned. In this paper, we propose to properly combine visual information with additional multi-faceted information, to define a novel multimodal similarity measure. More specifically, we combine visual features, which strongly relate to the image content, with semantic information represented by manually annotated concepts, and geo tagging, very often available in the form of object/subject location. Furthermore, we propose a supervised machine learning approach, based on Support Vector Machines (SVMs), to automatically learn optimized weights to combine the above features. The resulting models is used as a ranking function to sort the results of a multimodal query.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126312327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A mobile visual search application for content based image retrieval in the fashion domain 一个移动视觉搜索应用程序,用于时尚领域中基于内容的图像检索
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269838
Angelo Nodari, Matteo Ghiringhelli, Alessandro Zamberletti, M. Vanetti, S. Albertini, I. Gallo
In this study we propose a mobile application which interfaces with a Content-Based Image Retrieval engine for online shopping in the fashion domain. Using this application it is possible to take a picture of a garment to retrieve its most similar products. The proposed method is firstly presented as an application in which the user manually select the name of the subject framed by the camera, before sending the request to the server. In the second part we propose an advanced approach which automatically classifies the object of interest, in this way it is possible to minimize the effort required by the user during the query process. In order to evaluate the performance of the proposed method, we have collected three datasets: the first contains clothing images of products taken from different online shops, whereas for the other datasets we have used images and video frames of clothes taken by Internet users. The results show the feasibility in the use of the proposed mobile application in a real scenario.
在这项研究中,我们提出了一个移动应用程序,该应用程序与基于内容的图像检索引擎接口,用于时尚领域的在线购物。使用这个应用程序,可以拍摄一件衣服的照片,以检索其最相似的产品。该方法首先以应用程序的形式呈现,用户在向服务器发送请求之前,手动选择由相机框成的主题的名称。在第二部分中,我们提出了一种自动分类感兴趣对象的高级方法,这样可以最大限度地减少用户在查询过程中所需的工作量。为了评估所提出的方法的性能,我们收集了三个数据集:第一个数据集包含从不同的在线商店拍摄的产品的服装图像,而对于其他数据集,我们使用了互联网用户拍摄的服装图像和视频帧。结果表明,所提出的移动应用程序在实际场景中使用是可行的。
{"title":"A mobile visual search application for content based image retrieval in the fashion domain","authors":"Angelo Nodari, Matteo Ghiringhelli, Alessandro Zamberletti, M. Vanetti, S. Albertini, I. Gallo","doi":"10.1109/CBMI.2012.6269838","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269838","url":null,"abstract":"In this study we propose a mobile application which interfaces with a Content-Based Image Retrieval engine for online shopping in the fashion domain. Using this application it is possible to take a picture of a garment to retrieve its most similar products. The proposed method is firstly presented as an application in which the user manually select the name of the subject framed by the camera, before sending the request to the server. In the second part we propose an advanced approach which automatically classifies the object of interest, in this way it is possible to minimize the effort required by the user during the query process. In order to evaluate the performance of the proposed method, we have collected three datasets: the first contains clothing images of products taken from different online shops, whereas for the other datasets we have used images and video frames of clothes taken by Internet users. The results show the feasibility in the use of the proposed mobile application in a real scenario.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125497898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Feasibility of the detection of choirs for ethnomusicologic music indexing 民族音乐学音乐索引中唱诗班检测的可行性
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269845
M. L. Coz, R. André-Obrecht, J. Pinquier
The music is commonly structured in terms of the three classical categories which are instrumental, singing or singing-instrumental parts. To refine this notion, the number of singers and/or instruments is searched. An important difficulty appears when a choir in unison is observed: several singers try to reach the same note at the same time and the classical pitch analysis fails. This paper presents a method to detect such situation in an a capella context (without instrument). The approach is based on a temporal segmentation followed by a frequency tracking inside located frequency bands; it exploits the apparent splitting of the high harmonics due to small difference between the singers. The first results obtained on ethnomusicological corpora are quite satisfying and offer interesting perspectives to our work.
音乐通常是根据三个古典类别组成的,即器乐,歌唱或歌唱-器乐部分。为了完善这个概念,搜索歌手和/或乐器的数量。当观察到唱诗班的一致时,一个重要的困难出现了:几个歌手试图同时达到同一个音符,而经典的音高分析失败了。本文提出了一种方法来检测这种情况在无伴奏的情况下(没有仪器)。该方法基于时间分割,然后在定位的频带内进行频率跟踪;它利用高谐波的明显分裂,由于歌手之间的小差异。在民族音乐学语料库上获得的初步结果令人满意,并为我们的工作提供了有趣的视角。
{"title":"Feasibility of the detection of choirs for ethnomusicologic music indexing","authors":"M. L. Coz, R. André-Obrecht, J. Pinquier","doi":"10.1109/CBMI.2012.6269845","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269845","url":null,"abstract":"The music is commonly structured in terms of the three classical categories which are instrumental, singing or singing-instrumental parts. To refine this notion, the number of singers and/or instruments is searched. An important difficulty appears when a choir in unison is observed: several singers try to reach the same note at the same time and the classical pitch analysis fails. This paper presents a method to detect such situation in an a capella context (without instrument). The approach is based on a temporal segmentation followed by a frequency tracking inside located frequency bands; it exploits the apparent splitting of the high harmonics due to small difference between the singers. The first results obtained on ethnomusicological corpora are quite satisfying and offer interesting perspectives to our work.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129172928","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Search of objects of interest in videos 搜索视频中感兴趣的对象
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269809
Boris Mansencal, J. Benois-Pineau, Rémi Vieux, J. Domenger
The paper addresses the problem of object search in video content. Both Query-By-Example paradigm and context search are explored. In QBE paradigm the object of interest is searched by matching of object signatures built from SURF descriptors with on-the-fly computed signatures in frames. The ”context” search is understood as a query on the whole frame with features extracted after a region-based segmentation. Both kinds of features are transcribed in Bag-Of-Words framework. The combination of Bag-of-Visual-Words and Bag-of-Region-Words gives promising results in TRECVID'2011 Instance Search Task.
本文研究了视频内容中的对象搜索问题。研究了按例查询范式和上下文搜索。在QBE范例中,通过将SURF描述符构建的对象签名与帧中动态计算的签名进行匹配来搜索感兴趣的对象。“上下文”搜索可以理解为对整个框架进行查询,并在基于区域的分割后提取特征。这两种特征都被转录到词袋框架中。在TRECVID的2011实例搜索任务中,视觉词袋和区域词袋的结合取得了令人满意的结果。
{"title":"Search of objects of interest in videos","authors":"Boris Mansencal, J. Benois-Pineau, Rémi Vieux, J. Domenger","doi":"10.1109/CBMI.2012.6269809","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269809","url":null,"abstract":"The paper addresses the problem of object search in video content. Both Query-By-Example paradigm and context search are explored. In QBE paradigm the object of interest is searched by matching of object signatures built from SURF descriptors with on-the-fly computed signatures in frames. The ”context” search is understood as a query on the whole frame with features extracted after a region-based segmentation. Both kinds of features are transcribed in Bag-Of-Words framework. The combination of Bag-of-Visual-Words and Bag-of-Region-Words gives promising results in TRECVID'2011 Instance Search Task.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127352532","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Distributed high-dimensional index creation using Hadoop, HDFS and C++ 使用Hadoop, HDFS和c++创建分布式高维索引
Pub Date : 2012-06-27 DOI: 10.1109/CBMI.2012.6269848
G. Gudmundsson, L. Amsaleg, B. Jónsson
This paper describes an initial study where the open-source Hadoop parallel and distributed run-time environment is used to speedup the construction phase of a large high-dimensional index. This paper first discusses the typical practical problems developers may run into when porting their code to Hadoop. It then presents early experimental results showing that the performance gains are substantial when indexing large data sets.
本文介绍了一种利用开源Hadoop并行分布式运行环境加速大型高维索引构建阶段的初步研究。本文首先讨论开发人员在将代码移植到Hadoop时可能遇到的典型实际问题。然后给出了早期的实验结果,表明在索引大型数据集时,性能获得了实质性的提高。
{"title":"Distributed high-dimensional index creation using Hadoop, HDFS and C++","authors":"G. Gudmundsson, L. Amsaleg, B. Jónsson","doi":"10.1109/CBMI.2012.6269848","DOIUrl":"https://doi.org/10.1109/CBMI.2012.6269848","url":null,"abstract":"This paper describes an initial study where the open-source Hadoop parallel and distributed run-time environment is used to speedup the construction phase of a large high-dimensional index. This paper first discusses the typical practical problems developers may run into when porting their code to Hadoop. It then presents early experimental results showing that the performance gains are substantial when indexing large data sets.","PeriodicalId":120769,"journal":{"name":"2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116741427","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
期刊
2012 10th International Workshop on Content-Based Multimedia Indexing (CBMI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1