首页 > 最新文献

2009 Seventh International Workshop on Content-Based Multimedia Indexing最新文献

英文 中文
A Multi-faceted Browsing Interface for Digital Photo Collections 一个面向数字相册的多面浏览界面
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.23
Ilaria Bartolini
Managing photos by using visual features (e.g., color and texture) is known to be a powerful, yet imprecise, retrieval paradigm because of the semantic gap problem. The same is true if search relies only on keywords (or tags), derived from either the image context or user-provided annotations. In this paper we present a new multi-faceted image search and browsing system, named Scenique, that allows the user to manage her photo collections by using both visual features and tags, possibly organized into multiple dimensions (or facets). Each facet can be seen as a coordinate of a multidimensional space describing the image content (for example, the visual appearance, the content type, the geographic location, and so on). We present the basic principles of Scenique and provide evidence of the effectiveness of its visual tools. Feedback supplied by a set of real users indicates that the proposed interface is intuitive, easy-to-use, and that satisfies users' expectations in managing photo collections and quickly locating images of interest.
使用视觉特征(如颜色和纹理)管理照片是一种强大但不精确的检索范式,因为存在语义缺口问题。如果搜索仅依赖于关键字(或标记),则从图像上下文或用户提供的注释派生出来,情况也是如此。在本文中,我们提出了一个新的多面图像搜索和浏览系统,名为Scenique,它允许用户通过使用视觉特征和标签来管理她的照片收藏,可能组织成多个维度(或方面)。每个方面都可以看作是描述图像内容的多维空间的坐标(例如,视觉外观、内容类型、地理位置等等)。我们提出了Scenique的基本原则,并提供了其可视化工具有效性的证据。一组真实用户提供的反馈表明,所提出的界面直观、易于使用,满足了用户在管理图片集和快速定位感兴趣的图像方面的期望。
{"title":"A Multi-faceted Browsing Interface for Digital Photo Collections","authors":"Ilaria Bartolini","doi":"10.1109/CBMI.2009.23","DOIUrl":"https://doi.org/10.1109/CBMI.2009.23","url":null,"abstract":"Managing photos by using visual features (e.g., color and texture) is known to be a powerful, yet imprecise, retrieval paradigm because of the semantic gap problem. The same is true if search relies only on keywords (or tags), derived from either the image context or user-provided annotations. In this paper we present a new multi-faceted image search and browsing system, named Scenique, that allows the user to manage her photo collections by using both visual features and tags, possibly organized into multiple dimensions (or facets). Each facet can be seen as a coordinate of a multidimensional space describing the image content (for example, the visual appearance, the content type, the geographic location, and so on). We present the basic principles of Scenique and provide evidence of the effectiveness of its visual tools. Feedback supplied by a set of real users indicates that the proposed interface is intuitive, easy-to-use, and that satisfies users' expectations in managing photo collections and quickly locating images of interest.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126972701","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
RUSHES Retrieval of Multimedia Semantic Units for Enhanced Reusability 提高复用性的多媒体语义单元的rush检索
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.43
O. Schreer, I. Feldmann, Isabel Alonso Mediavilla, P. Concejero, A. Sadka, M. Swash
Multimedia analysis and reuse of raw un-edited audio visual content known as rushes is gaining acceptance by a large number of research labs and companies. A set of research projects are considering multimedia indexing, annotation, search and retrieval in the context of European funded research, but only the FP6 project RUSHES is focusing on automatic semantic annotation, indexing and retrieval of raw and un-edited audio-visual content. Even professional content creators and providers as well as home-users are dealing with this type of content and therefore novel technologies for semantic search and retrieval are required. In this paper, we present a summary of the most relevant achievements of the RUSHES project, focusing on specific approaches for automatic annotation as well as the main features of the final RUSHES search engine.
多媒体分析和重新使用原始的未编辑的视听内容被称为“rush”,它正在被大量的研究实验室和公司所接受。一组研究项目正在考虑在欧洲资助的研究背景下进行多媒体索引、注释、搜索和检索,但只有FP6项目“匆匆”专注于对原始和未编辑的视听内容进行自动语义注释、索引和检索。甚至专业的内容创建者和提供者以及家庭用户都在处理这种类型的内容,因此需要新的语义搜索和检索技术。在本文中,我们总结了rush项目最相关的成果,重点介绍了自动注释的具体方法以及最终的rush搜索引擎的主要特征。
{"title":"RUSHES Retrieval of Multimedia Semantic Units for Enhanced Reusability","authors":"O. Schreer, I. Feldmann, Isabel Alonso Mediavilla, P. Concejero, A. Sadka, M. Swash","doi":"10.1109/CBMI.2009.43","DOIUrl":"https://doi.org/10.1109/CBMI.2009.43","url":null,"abstract":"Multimedia analysis and reuse of raw un-edited audio visual content known as rushes is gaining acceptance by a large number of research labs and companies. A set of research projects are considering multimedia indexing, annotation, search and retrieval in the context of European funded research, but only the FP6 project RUSHES is focusing on automatic semantic annotation, indexing and retrieval of raw and un-edited audio-visual content. Even professional content creators and providers as well as home-users are dealing with this type of content and therefore novel technologies for semantic search and retrieval are required. In this paper, we present a summary of the most relevant achievements of the RUSHES project, focusing on specific approaches for automatic annotation as well as the main features of the final RUSHES search engine.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134017970","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Hierarchical Ontology-Based Robust Video Shots Indexation Using Global MPEG-7 Visual Descriptors 基于层次本体的基于全局MPEG-7视觉描述符的鲁棒视频片段索引
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.18
R. Benmokhtar, B. Huet
This paper proposes to improve our previous work on the concept-based video shot indexing, by considering an ontological concept construction in the TRECVid 2007 video retrieval, based on two steps. First, each single concept is modeled independently. Second, an ontology-based concept is introduced via the representation of the influence relations between concepts and the ontological readjustment of the confidence values. The main contribution of this paper is in the exploitation manner of the inter-concepts similarity in our indexing system, where three measures are represented: co-occurrence, visual similarity and LSCOM-lite ontology path length contribution. The experimental results report the efficiency and the significant improvement provided by the proposed scheme.
本文提出了基于概念的视频镜头索引的改进工作,在TRECVid 2007视频检索中考虑本体概念的构建,基于两个步骤。首先,每个概念都是独立建模的。其次,通过概念间影响关系的表征和置信度值的本体调整,引入了基于本体的概念;本文的主要贡献在于我们的索引系统中概念间相似度的开发方式,其中包括三个度量:共现性、视觉相似度和LSCOM-lite本体路径长度贡献。实验结果表明了该方案的有效性和显著的改进。
{"title":"Hierarchical Ontology-Based Robust Video Shots Indexation Using Global MPEG-7 Visual Descriptors","authors":"R. Benmokhtar, B. Huet","doi":"10.1109/CBMI.2009.18","DOIUrl":"https://doi.org/10.1109/CBMI.2009.18","url":null,"abstract":"This paper proposes to improve our previous work on the concept-based video shot indexing, by considering an ontological concept construction in the TRECVid 2007 video retrieval, based on two steps. First, each single concept is modeled independently. Second, an ontology-based concept is introduced via the representation of the influence relations between concepts and the ontological readjustment of the confidence values. The main contribution of this paper is in the exploitation manner of the inter-concepts similarity in our indexing system, where three measures are represented: co-occurrence, visual similarity and LSCOM-lite ontology path length contribution. The experimental results report the efficiency and the significant improvement provided by the proposed scheme.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124345485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Semantic High-Level Features for Automated Cross-Modal Slideshow Generation 用于自动跨模态幻灯片生成的语义高级特性
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.32
P. Dunker, C. Dittmar, André Begau, S. Nowak, M. Gruhne
This paper describes a technical solution for automated slideshow generation by extracting a set of high-level features from music, such as beat grid, mood and genre and intelligently combining this set with image high-level features, such as mood, daytime- and scene classification. An advantage of this high-level concept is to enable the user to incorporate his preferences regarding the semantic aspects of music and images. For example, the user might request the system to automatically create a slideshow, which plays soft music and shows pictures with sunsets from the last 10 years of his own photo collection.The high-level feature extraction on both, the audio and the visual information is based on the same underlying machine learning core, which processes different audio- and visual- low- and mid-level features. This paper describes the technical realization and evaluation of the algorithms with suitable test databases.
本文描述了一种自动生成幻灯片的技术方案,从音乐中提取一组高级特征,如节拍网格、情绪和类型,并将其与图像高级特征(如情绪、白天和场景分类)智能结合。这种高级概念的一个优点是使用户能够结合自己对音乐和图像语义方面的偏好。例如,用户可能要求系统自动创建一个幻灯片,播放柔和的音乐,并显示最近10年他自己的照片收藏中的日落图片。本文用合适的测试数据库描述了算法的技术实现和评估。
{"title":"Semantic High-Level Features for Automated Cross-Modal Slideshow Generation","authors":"P. Dunker, C. Dittmar, André Begau, S. Nowak, M. Gruhne","doi":"10.1109/CBMI.2009.32","DOIUrl":"https://doi.org/10.1109/CBMI.2009.32","url":null,"abstract":"This paper describes a technical solution for automated slideshow generation by extracting a set of high-level features from music, such as beat grid, mood and genre and intelligently combining this set with image high-level features, such as mood, daytime- and scene classification. An advantage of this high-level concept is to enable the user to incorporate his preferences regarding the semantic aspects of music and images. For example, the user might request the system to automatically create a slideshow, which plays soft music and shows pictures with sunsets from the last 10 years of his own photo collection.The high-level feature extraction on both, the audio and the visual information is based on the same underlying machine learning core, which processes different audio- and visual- low- and mid-level features. This paper describes the technical realization and evaluation of the algorithms with suitable test databases.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127476332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
ESPI Image Indexing and Similarity Search in Radon Transform Domain Radon变换域ESPI图像索引与相似度搜索
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.38
Rémi Vieux, J. Benois-Pineau, J. Domenger, A. Braquelaire
Content Based Image Retrieval is a topic which has received a lot of attention and increasing popularity due to a wide range of applications. In this paper, we present a similarity measure for CBIR in an industrial context, where the images of a vibration phenomenon are obtained by Electronic Speckle Pattern Interferometry. Images obtained have very poor visual characteristics, and traditional CBIR systems which rely on color or texture information could not be efficient. We propose a CBIR approach based on the 1-dimensional projections of the images obtained by the Radon transform. Experiments show that this signature is relevant and enables good retrieval performances compared to a baseline image correlation.
基于内容的图像检索由于其广泛的应用受到了广泛的关注和日益普及。在本文中,我们提出了工业背景下CBIR的相似性度量,其中振动现象的图像是通过电子散斑干涉法获得的。所获得的图像视觉特征很差,传统的基于颜色或纹理信息的CBIR系统无法达到有效的效果。我们提出了一种基于Radon变换得到的图像的一维投影的CBIR方法。实验表明,与基线图像相关相比,该特征是相关的,并且具有良好的检索性能。
{"title":"ESPI Image Indexing and Similarity Search in Radon Transform Domain","authors":"Rémi Vieux, J. Benois-Pineau, J. Domenger, A. Braquelaire","doi":"10.1109/CBMI.2009.38","DOIUrl":"https://doi.org/10.1109/CBMI.2009.38","url":null,"abstract":"Content Based Image Retrieval is a topic which has received a lot of attention and increasing popularity due to a wide range of applications. In this paper, we present a similarity measure for CBIR in an industrial context, where the images of a vibration phenomenon are obtained by Electronic Speckle Pattern Interferometry. Images obtained have very poor visual characteristics, and traditional CBIR systems which rely on color or texture information could not be efficient. We propose a CBIR approach based on the 1-dimensional projections of the images obtained by the Radon transform. Experiments show that this signature is relevant and enables good retrieval performances compared to a baseline image correlation.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"197 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132099993","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Subspace Tree 子树
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.14
A. Wichert
We are interested in designing a data structure for n objects of dimension d, with the following objectives: Space requirements should be  O(d * n) and the query time should be O(d * log(n)). Such a structure corresponds to subspace trees. A subspace tree divides the distances between the subspaces. It is realized by  the hierarchical linear subspace method. By doing so, the data is divided into disjoint entities. The asymptotic upper bound estimation of the maximum applicable number of subspaces is logarithmically constrained by the number of represented elements and their dimension.The search in such a tree starts at the subspace with the lowest dimension.  In this subspace, the set of all possible similar objects is determined. In the next subspace, additional metric information corresponding to a higher dimension is used to reduce this set.
我们有兴趣为d维的n个对象设计一个数据结构,其目标如下:空间需求应为O(d * n),查询时间应为O(d * log(n))。这种结构对应于子空间树。子空间树划分子空间之间的距离。该方法采用层次线性子空间方法实现。通过这样做,数据被划分为不相交的实体。子空间的最大可应用数目的渐近上界估计受表示元素的数目及其维数的对数约束。这种树的搜索从最低维的子空间开始。在这个子空间中,确定了所有可能的相似对象的集合。在下一个子空间中,使用与更高维度相对应的附加度量信息来约简该集合。
{"title":"Subspace Tree","authors":"A. Wichert","doi":"10.1109/CBMI.2009.14","DOIUrl":"https://doi.org/10.1109/CBMI.2009.14","url":null,"abstract":"We are interested in designing a data structure for n objects of dimension d, with the following objectives: Space requirements should be  O(d * n) and the query time should be O(d * log(n)). Such a structure corresponds to subspace trees. A subspace tree divides the distances between the subspaces. It is realized by  the hierarchical linear subspace method. By doing so, the data is divided into disjoint entities. The asymptotic upper bound estimation of the maximum applicable number of subspaces is logarithmically constrained by the number of represented elements and their dimension.The search in such a tree starts at the subspace with the lowest dimension.  In this subspace, the set of all possible similar objects is determined. In the next subspace, additional metric information corresponding to a higher dimension is used to reduce this set.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127895307","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Content-Aware Web Browsing and Visualization Tools for Cantillation and Chant Research 内容感知的网络浏览和可视化工具的吟唱和吟唱研究
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.46
S. Ness, G. Tzanetakis, D. Biró
Chant and cantillation research is particularly interesting as it explores the transition from oral to written transmission of music. The goal of this work to create web-based computational tools that can assist the study of how diverse recitation traditions, having their origin in primarily non-notated melodies, later became codified. One of the authors is a musicologist and music theorist who has guided the system design and development by providing manual annotations and participating in the design process. We describe novel content-based visualization and analysis algorithms that can be used for problem-seeking exploration of audio recordings of chant and recitations.
吟唱和吟唱研究特别有趣,因为它探索了音乐从口头到书面传播的过渡。这项工作的目标是创建基于网络的计算工具,以帮助研究不同的背诵传统,它们最初起源于无符号的旋律,后来如何被编纂。其中一位作者是音乐学家和音乐理论家,他通过提供手动注释和参与设计过程来指导系统的设计和开发。我们描述了新颖的基于内容的可视化和分析算法,可用于对吟诵和背诵的录音进行问题探索。
{"title":"Content-Aware Web Browsing and Visualization Tools for Cantillation and Chant Research","authors":"S. Ness, G. Tzanetakis, D. Biró","doi":"10.1109/CBMI.2009.46","DOIUrl":"https://doi.org/10.1109/CBMI.2009.46","url":null,"abstract":"Chant and cantillation research is particularly interesting as it explores the transition from oral to written transmission of music. The goal of this work to create web-based computational tools that can assist the study of how diverse recitation traditions, having their origin in primarily non-notated melodies, later became codified. One of the authors is a musicologist and music theorist who has guided the system design and development by providing manual annotations and participating in the design process. We describe novel content-based visualization and analysis algorithms that can be used for problem-seeking exploration of audio recordings of chant and recitations.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125546453","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Semantic Video Clustering in Ad Hoc Networks for Content-Based Retrieval 基于内容检索的Ad Hoc网络语义视频聚类
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.31
Bo Yang, M. Manohar
Traditional content-based retrieval approaches employ either centralized or flooding strategies in ad hoc networks, which may result in low fault tolerance and high search cost making them inefficient. To facilitate an efficient video retrieval, we propose a logic-based content summary framework that is able to represent semantic contents of video data using concise logic terms. In this method the video data is characterized by color and wavelet coefficients which will be converted into logical terms by using threshold operators. The logical terms are then summarized as node content descriptions. The nodes containing similar node descriptions are clustered into a virtual infrastructure according to the semantic content.
传统的基于内容的检索方法在自组织网络中采用集中式或泛洪式策略,导致容错性低,搜索成本高,效率低下。为了方便高效的视频检索,我们提出了一个基于逻辑的内容摘要框架,该框架能够使用简洁的逻辑术语表示视频数据的语义内容。在该方法中,视频数据由颜色和小波系数表征,并通过阈值算子转换为逻辑项。然后将逻辑术语总结为节点内容描述。将包含相似节点描述的节点根据语义内容聚类成虚拟基础结构。
{"title":"Semantic Video Clustering in Ad Hoc Networks for Content-Based Retrieval","authors":"Bo Yang, M. Manohar","doi":"10.1109/CBMI.2009.31","DOIUrl":"https://doi.org/10.1109/CBMI.2009.31","url":null,"abstract":"Traditional content-based retrieval approaches employ either centralized or flooding strategies in ad hoc networks, which may result in low fault tolerance and high search cost making them inefficient. To facilitate an efficient video retrieval, we propose a logic-based content summary framework that is able to represent semantic contents of video data using concise logic terms. In this method the video data is characterized by color and wavelet coefficients which will be converted into logical terms by using threshold operators. The logical terms are then summarized as node content descriptions. The nodes containing similar node descriptions are clustered into a virtual infrastructure according to the semantic content.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127792381","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Classification of Images Based on Hidden Markov Models 基于隐马尔可夫模型的图像分类
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.22
Marc Mouret, C. Solnon, Christian Wolf
We propose to use hidden Markov models (HMMs) to classify images. Images are modeled by extracting symbols corresponding to 3x3 binary neighborhoods of interest points, and by ordering these symbols by decreasing saliency order, thus obtaining strings of symbols. HMMs are learned from sets of strings modeling classes of images. The method has been tested on the SIMPLIcity database and shows an improvement over competing approaches based on interest points. We also evaluate these approaches for classifying thumbnail images, i.e., low resolution images.
我们提出使用隐马尔可夫模型(hmm)对图像进行分类。通过提取感兴趣点的3 × 3二进制邻域对应的符号,并通过降低显著性顺序对这些符号进行排序,从而获得符号串。hmm是从一组对图像进行建模的字符串中学习的。该方法已在simple数据库上进行了测试,并显示出基于兴趣点的竞争方法的改进。我们还评估了这些分类缩略图的方法,即低分辨率图像。
{"title":"Classification of Images Based on Hidden Markov Models","authors":"Marc Mouret, C. Solnon, Christian Wolf","doi":"10.1109/CBMI.2009.22","DOIUrl":"https://doi.org/10.1109/CBMI.2009.22","url":null,"abstract":"We propose to use hidden Markov models (HMMs) to classify images. Images are modeled by extracting symbols corresponding to 3x3 binary neighborhoods of interest points, and by ordering these symbols by decreasing saliency order, thus obtaining strings of symbols. HMMs are learned from sets of strings modeling classes of images. The method has been tested on the SIMPLIcity database and shows an improvement over competing approaches based on interest points. We also evaluate these approaches for classifying thumbnail images, i.e., low resolution images.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122524668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A Compact Multi-view Descriptor for 3D Object Retrieval 用于三维对象检索的紧凑多视图描述符
Pub Date : 2009-06-03 DOI: 10.1109/CBMI.2009.15
P. Daras, A. Axenopoulos
In this paper, a novel view-based approach for 3D object retrieval is introduced. A set of 2D images (multi-views) are automatically generated from a 3D object, by taking views from uniformly distributed viewpoints. For each image, a set of 2D rotation-invariant shape descriptors is extracted. The global shape similarity between two 3D models is achieved by applying a novel matching scheme, which effectively combines the information extracted from the multiview representation. The proposed approach can well serve as a unified framework, supporting multimodal queries (such as sketches, 2D images, 3D objects). The experimental results illustrate the superiority of the method over similar view-based approaches.
本文提出了一种基于视图的三维目标检索方法。通过从均匀分布的视点获取视图,从3D对象自动生成一组2D图像(多视图)。对于每张图像,提取一组二维旋转不变性形状描述子。采用一种新的匹配方案,有效地结合多视图表示中提取的信息,实现了两个三维模型之间的全局形状相似度。所提出的方法可以很好地作为一个统一的框架,支持多模式查询(如草图、2D图像、3D对象)。实验结果表明,该方法优于类似的基于视图的方法。
{"title":"A Compact Multi-view Descriptor for 3D Object Retrieval","authors":"P. Daras, A. Axenopoulos","doi":"10.1109/CBMI.2009.15","DOIUrl":"https://doi.org/10.1109/CBMI.2009.15","url":null,"abstract":"In this paper, a novel view-based approach for 3D object retrieval is introduced. A set of 2D images (multi-views) are automatically generated from a 3D object, by taking views from uniformly distributed viewpoints. For each image, a set of 2D rotation-invariant shape descriptors is extracted. The global shape similarity between two 3D models is achieved by applying a novel matching scheme, which effectively combines the information extracted from the multiview representation. The proposed approach can well serve as a unified framework, supporting multimodal queries (such as sketches, 2D images, 3D objects). The experimental results illustrate the superiority of the method over similar view-based approaches.","PeriodicalId":417012,"journal":{"name":"2009 Seventh International Workshop on Content-Based Multimedia Indexing","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123930586","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 59
期刊
2009 Seventh International Workshop on Content-Based Multimedia Indexing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1