首页 > 最新文献

2012 IEEE International Conference on Multimedia and Expo Workshops最新文献

英文 中文
Contextual Dominant Color Name Extraction for Web Image Search 上下文主色名称提取的网络图像搜索
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.61
Peng Wang, Dongqing Zhang, Gang Zeng, Jingdong Wang
This paper addresses the problem of extracting perceptually dominant color names of images. Our approach is motivated by the principle that the pixels corresponding to one dominant color name identified by human are often context dependent, spatially connected and form a perceptually meaningful region. Our algorithm first learns the probabilistic mapping from a RGB color to a color name. Then, a double-threshold approach is utilized to determine the color name of a RGB pixel in a specific image by considering its neighboring pixels. This scheme effectively deals with the pixels ambiguously belonging to several dominant color names. Last, the saliency information is combined to extract perceptually dominant colors. Experiments on our labeled image data set and the Ebay image set demonstrate the effectiveness of our approach.
本文研究了图像感知主色名称的提取问题。我们的方法是由这样一个原理驱动的,即与人类识别的一个主色名称对应的像素通常是上下文相关的,空间相连的,并形成一个感知上有意义的区域。我们的算法首先学习从RGB颜色到颜色名称的概率映射。然后,利用双阈值方法通过考虑相邻像素来确定特定图像中RGB像素的颜色名称。该方案有效地处理了属于多个主色名称的模糊像素。最后,结合显著性信息提取感知上的主色。在我们的标记图像数据集和Ebay图像集上的实验证明了我们的方法的有效性。
{"title":"Contextual Dominant Color Name Extraction for Web Image Search","authors":"Peng Wang, Dongqing Zhang, Gang Zeng, Jingdong Wang","doi":"10.1109/ICMEW.2012.61","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.61","url":null,"abstract":"This paper addresses the problem of extracting perceptually dominant color names of images. Our approach is motivated by the principle that the pixels corresponding to one dominant color name identified by human are often context dependent, spatially connected and form a perceptually meaningful region. Our algorithm first learns the probabilistic mapping from a RGB color to a color name. Then, a double-threshold approach is utilized to determine the color name of a RGB pixel in a specific image by considering its neighboring pixels. This scheme effectively deals with the pixels ambiguously belonging to several dominant color names. Last, the saliency information is combined to extract perceptually dominant colors. Experiments on our labeled image data set and the Ebay image set demonstrate the effectiveness of our approach.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127075219","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Improved Image Retargeting by Distinguishing between Faces in Focus and Out of Focus 通过区分对焦和失焦的人脸来改进图像重定位
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.32
J. Kiess, Rodrigo Garcia, S. Kopf, W. Effelsberg
The identification of relevant objects in an image is highly relevant in the context of image retargeting. Especially faces draw the attention of viewers. But the level of relevance may change between different faces depending on the size, the location, or whether a face is in focus or not. In this paper, we present a novel algorithm which distinguishes in-focus and out-of-focus faces. A face detector with multiple cascades is used first to locate initial face regions. We analyze the ratio of strong edges in each face region to classify out-of-focus faces. Finally, we use the Grab Cut algorithm to segment the faces and define binary face masks. These masks can then be used as an additional input to image retargeting algorithms.
在图像重定位的背景下,图像中相关对象的识别是高度相关的。尤其是脸能吸引观众的注意力。但是,不同的人脸之间的相关性水平可能会根据大小、位置或人脸是否被聚焦而有所不同。本文提出了一种新的人脸识别算法。首先使用具有多个级联的人脸检测器来定位初始人脸区域。我们通过分析每个人脸区域中强边缘的比例来对失焦人脸进行分类。最后,我们使用Grab Cut算法对人脸进行分割,并定义二值掩码。然后,这些掩码可以用作图像重定向算法的额外输入。
{"title":"Improved Image Retargeting by Distinguishing between Faces in Focus and Out of Focus","authors":"J. Kiess, Rodrigo Garcia, S. Kopf, W. Effelsberg","doi":"10.1109/ICMEW.2012.32","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.32","url":null,"abstract":"The identification of relevant objects in an image is highly relevant in the context of image retargeting. Especially faces draw the attention of viewers. But the level of relevance may change between different faces depending on the size, the location, or whether a face is in focus or not. In this paper, we present a novel algorithm which distinguishes in-focus and out-of-focus faces. A face detector with multiple cascades is used first to locate initial face regions. We analyze the ratio of strong edges in each face region to classify out-of-focus faces. Finally, we use the Grab Cut algorithm to segment the faces and define binary face masks. These masks can then be used as an additional input to image retargeting algorithms.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114465469","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Social Photo Tagging Recommendation Using Community-Based Group Associations 社会照片标签推荐使用社区为基础的团体协会
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.46
Chien-Li Chou, Yee-Choy Chean, Yi-Cheng Chen, Hua-Tsung Chen, Suh-Yin Lee
In the social network, living photos occupy a large portion of web contents. For sharing a photo with the people appearing in that, users have to manually tag the people with their names, and the social network system links the photo to the people immediately. However, tagging the photos manually is a time-consuming task while people take thousands of photos in their daily life. Therefore, more and more researchers put their eyes on how to recommend tags for a photo. In this paper, our goal is to recommend tags for a query photo with one tagged face. We fuse the results of face recognition and the user's relationships obtained from social contexts. In addition, the Community-Based Group Associations, called CBGA, is proposed to discover the group associations among users through the community detection. Finally, the experimental evaluations show that the performance of photo tagging recommendation is improved by combining the face recognition and social relationship. Furthermore, the proposed framework achieves the high quality for social photo tagging recommendation.
在社交网络中,生活照片占据了网页内容的很大一部分。为了与出现在其中的人分享照片,用户必须手动标记他们的名字,社交网络系统会立即将照片链接到这些人。然而,当人们在日常生活中拍摄成千上万的照片时,手动标记照片是一项耗时的任务。因此,如何为照片推荐标签成为越来越多研究者关注的问题。在本文中,我们的目标是为带有一个标记脸的查询照片推荐标签。我们融合了人脸识别的结果和从社会背景中获得的用户关系。此外,还提出了基于社区的组关联(CBGA),通过社区检测来发现用户之间的组关联。最后,实验结果表明,将人脸识别与社会关系相结合,提高了照片标签推荐的性能。此外,该框架还实现了高质量的社交照片标签推荐。
{"title":"Social Photo Tagging Recommendation Using Community-Based Group Associations","authors":"Chien-Li Chou, Yee-Choy Chean, Yi-Cheng Chen, Hua-Tsung Chen, Suh-Yin Lee","doi":"10.1109/ICMEW.2012.46","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.46","url":null,"abstract":"In the social network, living photos occupy a large portion of web contents. For sharing a photo with the people appearing in that, users have to manually tag the people with their names, and the social network system links the photo to the people immediately. However, tagging the photos manually is a time-consuming task while people take thousands of photos in their daily life. Therefore, more and more researchers put their eyes on how to recommend tags for a photo. In this paper, our goal is to recommend tags for a query photo with one tagged face. We fuse the results of face recognition and the user's relationships obtained from social contexts. In addition, the Community-Based Group Associations, called CBGA, is proposed to discover the group associations among users through the community detection. Finally, the experimental evaluations show that the performance of photo tagging recommendation is improved by combining the face recognition and social relationship. Furthermore, the proposed framework achieves the high quality for social photo tagging recommendation.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131093898","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
How Many Frames Does Facial Expression Recognition Require? 面部表情识别需要多少帧?
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.56
Kaimin Yu, Zhiyong Wang, Genliang Guan, Qiuxia Wu, Z. Chi, D. Feng
Facial expression analysis is essential to enable socially intelligent processing of multimedia video content. Most facial expression recognition algorithms generally analyze the whole image sequence of an expression to exploit its temporal characteristics. However, it is seldom studied whether it is necessary to utilize all the frames of a sequence, since human beings are able to capture the dynamics of facial expressions from very short sequences (even only one frame). In this paper, we investigate the impact of the number of frames in a facial expression sequence on facial expression recognition accuracy. In particular, we develop a key frame selection method through key point based frame representation. Experimental results on the popular CK facial expression dataset indicate that recognition accuracy achieved with half of the sequence frames is comparable to that of utilizing all the sequence frames. Our key frame selection method can further reduce the number of frames without clearly compromising recognition accuracy.
面部表情分析对于实现多媒体视频内容的社会化智能处理至关重要。大多数面部表情识别算法通常分析表情的整个图像序列,以利用其时间特征。然而,很少有人研究是否有必要利用一个序列的所有帧,因为人类能够从非常短的序列(甚至只有一帧)中捕捉面部表情的动态。在本文中,我们研究了面部表情序列帧数对面部表情识别精度的影响。特别地,我们开发了一种基于关键点的帧表示的关键帧选择方法。在流行的CK面部表情数据集上的实验结果表明,使用一半序列帧的识别精度与使用所有序列帧的识别精度相当。我们的关键帧选择方法可以在不明显影响识别精度的情况下进一步减少帧数。
{"title":"How Many Frames Does Facial Expression Recognition Require?","authors":"Kaimin Yu, Zhiyong Wang, Genliang Guan, Qiuxia Wu, Z. Chi, D. Feng","doi":"10.1109/ICMEW.2012.56","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.56","url":null,"abstract":"Facial expression analysis is essential to enable socially intelligent processing of multimedia video content. Most facial expression recognition algorithms generally analyze the whole image sequence of an expression to exploit its temporal characteristics. However, it is seldom studied whether it is necessary to utilize all the frames of a sequence, since human beings are able to capture the dynamics of facial expressions from very short sequences (even only one frame). In this paper, we investigate the impact of the number of frames in a facial expression sequence on facial expression recognition accuracy. In particular, we develop a key frame selection method through key point based frame representation. Experimental results on the popular CK facial expression dataset indicate that recognition accuracy achieved with half of the sequence frames is comparable to that of utilizing all the sequence frames. Our key frame selection method can further reduce the number of frames without clearly compromising recognition accuracy.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123529684","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Human Gesture Analysis Using Multimodal Features 基于多模态特征的人类手势分析
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.88
Dan Luo, H. K. Ekenel, J. Ohya
Human gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and PLS is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.
人体手势作为一种自然界面,在实现智能人机交互(HCI)中起着至关重要的作用。人类的手势包括视觉动作的不同组成部分,如手部动作、面部表情和躯干,以传达意思。到目前为止,在手势识别领域,以往的工作大多集中在手势的手动成分上。在本文中,我们提出了一个基于外观的多模态手势识别框架,该框架结合了从单个网络摄像机捕获的图像帧中提取的不同组特征,如面部表情特征和手部运动特征。我们参考了美国手语(ASL)中的12类人类面部表情手势,包括中性、消极和积极的含义。我们通过采用两种融合策略将特征组合在两个层次上。在特征层,通过对不同的特征组进行连接和加权来进行早期的特征组合,并使用PLS通过将特征投射到判别表达式空间来选择最具判别性的元素。第二种策略应用于决策层面。来自单一模式的加权决策在稍后阶段融合在一起。采用基于凝聚的分类算法。我们收集了三到七次记录的数据集,并用组合技术进行了实验。实验结果表明,人脸分析提高了手势识别能力,决策级融合优于特征级融合。
{"title":"Human Gesture Analysis Using Multimodal Features","authors":"Dan Luo, H. K. Ekenel, J. Ohya","doi":"10.1109/ICMEW.2012.88","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.88","url":null,"abstract":"Human gesture as a natural interface plays an utmost important role for achieving intelligent Human Computer Interaction (HCI). Human gestures include different components of visual actions such as motion of hands, facial expression, and torso, to convey meaning. So far, in the field of gesture recognition, most previous works have focused on the manual component of gestures. In this paper, we present an appearance-based multimodal gesture recognition framework, which combines the different groups of features such as facial expression features and hand motion features which are extracted from image frames captured by a single web camera. We refer 12 classes of human gestures with facial expression including neutral, negative and positive meanings from American Sign Languages (ASL). We combine the features in two levels by employing two fusion strategies. At the feature level, an early feature combination can be performed by concatenating and weighting different feature groups, and PLS is used to choose the most discriminative elements by projecting the feature on a discriminative expression space. The second strategy is applied on decision level. Weighted decisions from single modalities are fused in a later stage. A condensation-based algorithm is adopted for classification. We collected a data set with three to seven recording sessions and conducted experiments with the combination techniques. Experimental results showed that facial analysis improve hand gesture recognition, decision level fusion performs better than feature level fusion.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124609302","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Resource Allocation for Service Composition in Cloud-based Video Surveillance Platform 基于云的视频监控平台业务组合的资源分配
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.77
M. S. Hossain, M. Hassan, Muhammad Al-Qurishi, A. Alghamdi
Resource allocation play an important role in service composition for cloud-based video surveillance platform. In this platform, the utilization of computational resources is managed through accessing various services from Virtual Machine (VM) resources. A single service accessed from VMs running inside such a cloud platform may not cater the application demands of all surveillance users. Services require to be modeled as a value added composite service. In order to provide such a composite service to the customer, VM resources need to be utilized optimally so that QoS requirements is fulfilled. In order to optimize the VM resource allocation, we have used linear programming approach as well as heuristics. The simulation results show that our approach outperforms the existing VM allocation schemes in a cloud-based video surveillance environment, in terms of cost and response time.
资源分配在云视频监控平台的业务构成中起着重要的作用。在该平台中,通过访问虚拟机(VM)资源中的各种服务来管理计算资源的利用。从运行在这种云平台内的虚拟机访问的单个服务可能无法满足所有监控用户的应用需求。服务需要建模为增值组合服务。为了向客户提供这样的组合服务,需要优化VM资源,以满足QoS要求。为了优化虚拟机资源分配,我们使用了线性规划方法和启发式方法。仿真结果表明,在基于云的视频监控环境中,我们的方法在成本和响应时间方面优于现有的虚拟机分配方案。
{"title":"Resource Allocation for Service Composition in Cloud-based Video Surveillance Platform","authors":"M. S. Hossain, M. Hassan, Muhammad Al-Qurishi, A. Alghamdi","doi":"10.1109/ICMEW.2012.77","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.77","url":null,"abstract":"Resource allocation play an important role in service composition for cloud-based video surveillance platform. In this platform, the utilization of computational resources is managed through accessing various services from Virtual Machine (VM) resources. A single service accessed from VMs running inside such a cloud platform may not cater the application demands of all surveillance users. Services require to be modeled as a value added composite service. In order to provide such a composite service to the customer, VM resources need to be utilized optimally so that QoS requirements is fulfilled. In order to optimize the VM resource allocation, we have used linear programming approach as well as heuristics. The simulation results show that our approach outperforms the existing VM allocation schemes in a cloud-based video surveillance environment, in terms of cost and response time.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132606073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
Extracting Context Information from Microblog Based on Analysis of Online Reviews 基于在线评论分析的微博语境信息提取
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.49
T. Takehara, Shohei Miki, Naoko Nitta, N. Babaguchi
Recommender systems automatically determine suitable items for users. Although preferences or context of users have been widely utilized in order to evaluate the suitability of the items for users, the surrounding context have little been considered. Focusing on that many ordinary human beings voluntarily report their observations of the current situation of the world to microblogs, this paper proposes a recommender system which not only recommends suitable restaurants to users based on their preferences and context but also provides the surrounding context information reported to microblogs which will further affect the users' restaurant selection behaviors. In particular, considering that such influential surrounding context information in microblogs includes keywords related to restaurant assessment, we propose a method for automatically determining the keywords to extract the context information by analyzing online reviews, which have been gathered also from ordinary human beings over a long period of time. The experiments by using Twitter as microblogs and Tabelog, a popular online restaurant review site in Japan, to obtain online reviews, indicated that the influential context information can be extracted from Twitter with the highest recall of 93.3% by using the area-related keywords. Additionally using the restaurant-related keywords was effective in removing irrelevant information obtaining the precision of 15.9%.
推荐系统会自动为用户确定合适的项目。虽然用户的偏好或背景已被广泛利用,以评估项目对用户的适用性,但周围的背景很少被考虑。针对许多普通人自愿将自己对世界现状的观察报告到微博上的情况,本文提出了一种推荐系统,该系统不仅根据用户的偏好和语境为用户推荐合适的餐厅,还将周围的语境信息报告到微博上,从而进一步影响用户的餐厅选择行为。特别是,考虑到微博中这种有影响力的周边语境信息中包含了与餐厅评价相关的关键词,我们提出了一种通过分析在线评论来自动确定关键词提取语境信息的方法,这些评论也是长期从普通人那里收集来的。利用Twitter作为微博和日本著名的在线餐厅评论网站Tabelog获取在线评论的实验表明,利用与领域相关的关键词可以从Twitter中提取有影响力的上下文信息,召回率最高,达到93.3%。此外,使用与餐厅相关的关键词可以有效地去除无关信息,准确率达到15.9%。
{"title":"Extracting Context Information from Microblog Based on Analysis of Online Reviews","authors":"T. Takehara, Shohei Miki, Naoko Nitta, N. Babaguchi","doi":"10.1109/ICMEW.2012.49","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.49","url":null,"abstract":"Recommender systems automatically determine suitable items for users. Although preferences or context of users have been widely utilized in order to evaluate the suitability of the items for users, the surrounding context have little been considered. Focusing on that many ordinary human beings voluntarily report their observations of the current situation of the world to microblogs, this paper proposes a recommender system which not only recommends suitable restaurants to users based on their preferences and context but also provides the surrounding context information reported to microblogs which will further affect the users' restaurant selection behaviors. In particular, considering that such influential surrounding context information in microblogs includes keywords related to restaurant assessment, we propose a method for automatically determining the keywords to extract the context information by analyzing online reviews, which have been gathered also from ordinary human beings over a long period of time. The experiments by using Twitter as microblogs and Tabelog, a popular online restaurant review site in Japan, to obtain online reviews, indicated that the influential context information can be extracted from Twitter with the highest recall of 93.3% by using the area-related keywords. Additionally using the restaurant-related keywords was effective in removing irrelevant information obtaining the precision of 15.9%.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129016281","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Query by Humming by Using Locality Sensitive Hashing Based on Combination of Pitch and Note 基于音高和音符组合的局部敏感哈希算法的哼唱查询
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.58
Qiang Wang, Zhiyuan Guo, Gang Liu, Jun Guo, Yueming Lu
Query by humming (QBH) is a technique that is used for content-based music information retrieval. It is a challenging unsolved problem due to humming errors. In this paper a novel retrieval method called note-based locality sensitive hashing (NLSH) is presented and it is combined with pitch-based locality sensitive hashing (PLSH) to screen candidate fragments. The method extracts PLSH and NLSH vectors from the database to construct two indexes. In the phase of retrieval, it automatically extracts vectors similar to the index construction and searches the indexes to obtain a list of candidates. Then recursive alignment (RA) is executed on these surviving candidates. Experiments are conducted on a database of 5,000 MIDI files with the 2010 MIREX-QBH query corpus. The results show by using the combination approach the relatively improvements of mean reciprocal rank are 29.7% (humming from anywhere) and 23.8% (humming from beginning), respectively, compared with the current state-of-the-art method.
哼唱查询(QBH)是一种基于内容的音乐信息检索技术。由于嗡嗡声误差的存在,这是一个具有挑战性的未解决问题。本文提出了一种新的检索方法——基于笔记的局部敏感哈希(NLSH),并将其与基于音高的局部敏感哈希(PLSH)相结合来筛选候选片段。该方法从数据库中提取PLSH和NLSH向量,构建两个索引。在检索阶段,它自动提取与索引构造相似的向量,并搜索索引以获得候选列表。然后对这些幸存的候选对象执行递归对齐(RA)。利用2010年MIREX-QBH查询语料库在5000个MIDI文件数据库上进行了实验。结果表明,与现有方法相比,组合方法的平均倒数秩(从任何地方哼唱)和(从开始哼唱)分别提高了29.7%和23.8%。
{"title":"Query by Humming by Using Locality Sensitive Hashing Based on Combination of Pitch and Note","authors":"Qiang Wang, Zhiyuan Guo, Gang Liu, Jun Guo, Yueming Lu","doi":"10.1109/ICMEW.2012.58","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.58","url":null,"abstract":"Query by humming (QBH) is a technique that is used for content-based music information retrieval. It is a challenging unsolved problem due to humming errors. In this paper a novel retrieval method called note-based locality sensitive hashing (NLSH) is presented and it is combined with pitch-based locality sensitive hashing (PLSH) to screen candidate fragments. The method extracts PLSH and NLSH vectors from the database to construct two indexes. In the phase of retrieval, it automatically extracts vectors similar to the index construction and searches the indexes to obtain a list of candidates. Then recursive alignment (RA) is executed on these surviving candidates. Experiments are conducted on a database of 5,000 MIDI files with the 2010 MIREX-QBH query corpus. The results show by using the combination approach the relatively improvements of mean reciprocal rank are 29.7% (humming from anywhere) and 23.8% (humming from beginning), respectively, compared with the current state-of-the-art method.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125651824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Intelligent Vehicle Detection and Tracking for Highway Driving 面向高速公路行驶的智能车辆检测与跟踪
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.19
Wanxin Xu, Meikang Qiu, Zhi Chen, Hai Su
Due to the increment of vehicles, the traffic jamming in cities becomes a serious challenge and the safety of people is threatened. Intelligent transportation system (ITS) and intelligent vehicles are critical to the efficiency of city transportation. In the area related with ITS and intelligent vehicles, moving vehicle detection and tracking are the most challenging problems. In this paper, we propose a framework for vehicle detection and tracking and make an in-depth research in key algorithms and techniques. We also conduct a serial of experiments on the basis of the existing results. Experimental results show that our proposed approach is feasible and effective for vehicle detection and tracking.
随着机动车数量的增加,城市交通拥堵问题日益严重,人们的生命安全受到威胁。智能交通系统和智能车辆是提高城市交通效率的关键。在与ITS和智能车辆相关的领域中,移动车辆的检测和跟踪是最具挑战性的问题。本文提出了车辆检测与跟踪的框架,并对关键算法和技术进行了深入研究。我们还在已有结果的基础上进行了一系列的实验。实验结果表明,该方法对车辆的检测和跟踪是可行和有效的。
{"title":"Intelligent Vehicle Detection and Tracking for Highway Driving","authors":"Wanxin Xu, Meikang Qiu, Zhi Chen, Hai Su","doi":"10.1109/ICMEW.2012.19","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.19","url":null,"abstract":"Due to the increment of vehicles, the traffic jamming in cities becomes a serious challenge and the safety of people is threatened. Intelligent transportation system (ITS) and intelligent vehicles are critical to the efficiency of city transportation. In the area related with ITS and intelligent vehicles, moving vehicle detection and tracking are the most challenging problems. In this paper, we propose a framework for vehicle detection and tracking and make an in-depth research in key algorithms and techniques. We also conduct a serial of experiments on the basis of the existing results. Experimental results show that our proposed approach is feasible and effective for vehicle detection and tracking.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134329401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Research Design for Evaluating How to Engage Students with Urban Public Screens in Students' Neighbourhoods 评估如何让学生参与学生社区的城市公共屏幕的研究设计
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.68
A. Lugmayr, Yuan Fu
Public screens are spreading throughout urban residential environments - in busses, trains, shopping centers, or at bus stops. Currently they are mostly used for advertising purposes, however, within the scope of this publication we focus on a new non apparent application area: the application of public screens in student villages. However, with new emerging innovative technologies and the increasing demand of students to use the latest technologies, there is a need and desire to bridge residents and businesses in the local vicinity. In addition, social networks shall help to foster a deeper integration of the community and its services. We present a study of the usage of public screen environments in the student vicinity of Kelvin Grove, Brisbane, Australia and Hervanta, Tampere, Finland. The study had three goals: (1) interviews with business owners to evaluate their needs for content and services, (2) student questioner to gain insights into consumer desires and expectations, (3) development of a roadmap and service concepts for public screens in student vicinities.
公共屏幕正在遍布城市居住环境——公共汽车、火车、购物中心或公共汽车站。目前,它们主要用于广告目的,然而,在本出版物的范围内,我们关注一个新的不明显的应用领域:公共屏幕在学生村的应用。然而,随着新兴的创新技术和学生使用最新技术的需求不断增加,人们需要和渴望在当地附近的居民和企业之间架起一座桥梁。此外,社交网络将有助于促进社区及其服务更深层次的融合。我们对澳大利亚布里斯班Kelvin Grove和芬兰坦佩雷Hervanta附近的学生使用公共屏幕环境进行了研究。该研究有三个目标:(1)与企业主进行访谈,以评估他们对内容和服务的需求;(2)学生提问,以了解消费者的需求和期望;(3)为学生社区的公共屏幕制定路线图和服务概念。
{"title":"Research Design for Evaluating How to Engage Students with Urban Public Screens in Students' Neighbourhoods","authors":"A. Lugmayr, Yuan Fu","doi":"10.1109/ICMEW.2012.68","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.68","url":null,"abstract":"Public screens are spreading throughout urban residential environments - in busses, trains, shopping centers, or at bus stops. Currently they are mostly used for advertising purposes, however, within the scope of this publication we focus on a new non apparent application area: the application of public screens in student villages. However, with new emerging innovative technologies and the increasing demand of students to use the latest technologies, there is a need and desire to bridge residents and businesses in the local vicinity. In addition, social networks shall help to foster a deeper integration of the community and its services. We present a study of the usage of public screen environments in the student vicinity of Kelvin Grove, Brisbane, Australia and Hervanta, Tampere, Finland. The study had three goals: (1) interviews with business owners to evaluate their needs for content and services, (2) student questioner to gain insights into consumer desires and expectations, (3) development of a roadmap and service concepts for public screens in student vicinities.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114718867","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
2012 IEEE International Conference on Multimedia and Expo Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1