首页 > 最新文献

2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)最新文献

英文 中文
A scattering transform combination with local binary pattern for texture classification 结合局部二值模式的散射变换纹理分类
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500238
Vu-Lam Nguyen, Ngoc-Son Vu, P. Gosselin
In this paper, we propose a combined feature approach which takes full advantages of local structure information and the more global one for improving texture image classification results. In this way, Local Binary Pattern is used for extracting local features, whilst the Scattering Transform feature plays the role of a global descriptor. Intensive experiments conducted on many texture benchmarks such as ALOT, CUReT, KTH-TIPS2-a, KTH-TIPS2b, and OUTEX show that the combined method outweigh each one which stands alone in term of classification accuracy. Also, our method outperforms many others, whilst it is comparable to state of the art on the experimented datasets.
本文提出了一种充分利用局部结构信息和全局结构信息的组合特征方法来提高纹理图像的分类效果。这样,局部二值模式被用来提取局部特征,而散射变换特征则扮演全局描述子的角色。在ALOT、CUReT、KTH-TIPS2-a、KTH-TIPS2b、OUTEX等多个纹理基准上进行的大量实验表明,组合方法在分类精度上优于单独使用的方法。此外,我们的方法优于许多其他方法,同时它可以与实验数据集上的最新技术相媲美。
{"title":"A scattering transform combination with local binary pattern for texture classification","authors":"Vu-Lam Nguyen, Ngoc-Son Vu, P. Gosselin","doi":"10.1109/CBMI.2016.7500238","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500238","url":null,"abstract":"In this paper, we propose a combined feature approach which takes full advantages of local structure information and the more global one for improving texture image classification results. In this way, Local Binary Pattern is used for extracting local features, whilst the Scattering Transform feature plays the role of a global descriptor. Intensive experiments conducted on many texture benchmarks such as ALOT, CUReT, KTH-TIPS2-a, KTH-TIPS2b, and OUTEX show that the combined method outweigh each one which stands alone in term of classification accuracy. Also, our method outperforms many others, whilst it is comparable to state of the art on the experimented datasets.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133344774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A probabilistic topic approach for context-aware visual attention modeling 上下文感知视觉注意建模的概率主题方法
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500272
M. Fernandez-Torres, I. González-Díaz, F. Díaz-de-María
The modeling of visual attention has gained much interest during the last few years since it allows to efficiently drive complex visual processes to particular areas of images or video frames. Although the literature concerning bottom-up saliency models is vast, we still lack of generic approaches modeling top-down task and context-driven visual attention. Indeed, many top-down models simply modulate the weights associated to low-level descriptors to learn more accurate representations of visual attention than those ones of the generic fusion schemes in bottom-up techniques. In this paper we propose a hierarchical generic probabilistic framework that decomposes the complex process of context-driven visual attention into a mixture of latent subtasks, each of them being in turn modeled as a combination of specific distributions of low-level descriptors. The inclusion of this intermediate level bridges the gap between low-level features and visual attention and enables more comprehensive representations of the later. Our experiments on a dataset in which videos are organized by genre demonstrate that, by learning specific distributions for each video category, we can notably enhance the system performance.
视觉注意的建模在过去几年中获得了很大的兴趣,因为它可以有效地将复杂的视觉过程驱动到图像或视频帧的特定区域。尽管关于自下而上显著性模型的文献很多,但我们仍然缺乏对自上而下任务和上下文驱动的视觉注意建模的通用方法。事实上,许多自顶向下的模型只是简单地调整与低级描述符相关的权重,以学习比自底向上技术中的通用融合方案更准确的视觉注意表示。在本文中,我们提出了一个分层的通用概率框架,该框架将上下文驱动的视觉注意的复杂过程分解为潜在子任务的混合物,每个子任务依次建模为低级描述符的特定分布的组合。这一中间水平的包含弥补了低水平特征和视觉注意之间的差距,并使后者能够更全面地表现出来。我们在一个视频按类型组织的数据集上的实验表明,通过学习每个视频类别的特定分布,我们可以显著提高系统性能。
{"title":"A probabilistic topic approach for context-aware visual attention modeling","authors":"M. Fernandez-Torres, I. González-Díaz, F. Díaz-de-María","doi":"10.1109/CBMI.2016.7500272","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500272","url":null,"abstract":"The modeling of visual attention has gained much interest during the last few years since it allows to efficiently drive complex visual processes to particular areas of images or video frames. Although the literature concerning bottom-up saliency models is vast, we still lack of generic approaches modeling top-down task and context-driven visual attention. Indeed, many top-down models simply modulate the weights associated to low-level descriptors to learn more accurate representations of visual attention than those ones of the generic fusion schemes in bottom-up techniques. In this paper we propose a hierarchical generic probabilistic framework that decomposes the complex process of context-driven visual attention into a mixture of latent subtasks, each of them being in turn modeled as a combination of specific distributions of low-level descriptors. The inclusion of this intermediate level bridges the gap between low-level features and visual attention and enables more comprehensive representations of the later. Our experiments on a dataset in which videos are organized by genre demonstrate that, by learning specific distributions for each video category, we can notably enhance the system performance.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"118 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116357670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Explorative hyperbolic-tree-based clustering tool for unsupervised knowledge discovery 无监督知识发现的探索性双曲树聚类工具
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500271
M. Riegler, Konstantin Pogorelov, M. Lux, P. Halvorsen, C. Griwodz, T. Lange, S. Eskeland
Exploring and annotating collections of images without meta-data is a laborious task. Visual analytics and information visualization can help users by providing interfaces for exploration and annotation. In this paper, we show a prototype application that allows users from the medical domain to use feature-based clustering to perform explorative browsing and annotation in an unsupervised manner. For this, we utilize global image feature extraction, different unsupervised clustering algorithms and hyperbolic tree representation. First, the prototype application extracts features from images or video frames, and then, one or multiple features at the same time can be used to perform clustering. The clusters are presented to the users as a hyperbolic tree for visual analysis and annotation.
在没有元数据的情况下探索和注释图像集合是一项费力的任务。可视化分析和信息可视化可以通过提供探索和注释的界面来帮助用户。在本文中,我们展示了一个原型应用程序,该应用程序允许医疗领域的用户使用基于特征的聚类以无监督的方式执行探索性浏览和注释。为此,我们利用了全局图像特征提取、不同的无监督聚类算法和双曲树表示。首先,原型应用程序从图像或视频帧中提取特征,然后可以同时使用一个或多个特征进行聚类。聚类以双曲树的形式呈现给用户,便于可视化分析和注释。
{"title":"Explorative hyperbolic-tree-based clustering tool for unsupervised knowledge discovery","authors":"M. Riegler, Konstantin Pogorelov, M. Lux, P. Halvorsen, C. Griwodz, T. Lange, S. Eskeland","doi":"10.1109/CBMI.2016.7500271","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500271","url":null,"abstract":"Exploring and annotating collections of images without meta-data is a laborious task. Visual analytics and information visualization can help users by providing interfaces for exploration and annotation. In this paper, we show a prototype application that allows users from the medical domain to use feature-based clustering to perform explorative browsing and annotation in an unsupervised manner. For this, we utilize global image feature extraction, different unsupervised clustering algorithms and hyperbolic tree representation. First, the prototype application extracts features from images or video frames, and then, one or multiple features at the same time can be used to perform clustering. The clusters are presented to the users as a hyperbolic tree for visual analysis and annotation.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132235715","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
DeepSketch 2: Deep convolutional neural networks for partial sketch recognition DeepSketch 2:用于部分草图识别的深度卷积神经网络
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500261
S. Dupont, Omar Seddati, S. Mahmoudi
Freehand sketches are a simple and powerful tool for communication. They are easily recognized across cultures and suitable for various applications. In this paper, we propose a new approach for partial sketch recognition. This could be used to design applications using real-time sketch recognition. We use deep convolutional neural networks (ConvNets), state-of-the-art in the field of sketch recognition. To the best of our knowledge, it is the first ConvNet for partial sketch classification. Our first aim is to build a ConvNet capable of recognizing partial sketches without compromising the accuracy reached for complete sketch recognition. Therefore, we evaluate different approaches and propose an efficient way for partial sketch recognition. Our second aim is improving complete sketch recognition using information about sketching progression. We obtained a ConvNet that outperforms state-of-the-art results in the TU-Berlin sketch benchmark. We reached an accuracy of 77.69%.
手绘草图是一种简单而强大的沟通工具。它们很容易在不同文化中被识别,并且适用于各种应用程序。本文提出了一种局部素描识别的新方法。这可以用于设计使用实时草图识别的应用程序。我们使用深度卷积神经网络(ConvNets),在草图识别领域的最新技术。据我们所知,这是第一个局部草图分类的卷积神经网络。我们的第一个目标是建立一个能够识别部分草图而不影响完整草图识别精度的卷积神经网络。因此,我们评估了不同的方法,并提出了一种有效的局部草图识别方法。我们的第二个目标是利用素描进程的信息来提高完整的素描识别。我们获得的卷积神经网络在TU-Berlin草图基准测试中表现优于最先进的结果。我们达到了77.69%的准确率。
{"title":"DeepSketch 2: Deep convolutional neural networks for partial sketch recognition","authors":"S. Dupont, Omar Seddati, S. Mahmoudi","doi":"10.1109/CBMI.2016.7500261","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500261","url":null,"abstract":"Freehand sketches are a simple and powerful tool for communication. They are easily recognized across cultures and suitable for various applications. In this paper, we propose a new approach for partial sketch recognition. This could be used to design applications using real-time sketch recognition. We use deep convolutional neural networks (ConvNets), state-of-the-art in the field of sketch recognition. To the best of our knowledge, it is the first ConvNet for partial sketch classification. Our first aim is to build a ConvNet capable of recognizing partial sketches without compromising the accuracy reached for complete sketch recognition. Therefore, we evaluate different approaches and propose an efficient way for partial sketch recognition. Our second aim is improving complete sketch recognition using information about sketching progression. We obtained a ConvNet that outperforms state-of-the-art results in the TU-Berlin sketch benchmark. We reached an accuracy of 77.69%.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132735277","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Spatial pyramids for boosting global features in content based image retrieval 在基于内容的图像检索中增强全局特征的空间金字塔
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500248
M. Lux, N. Anagnostopoulos, C. Iakovidou
Image retrieval deals with the problem of finding relevant images to satisfy a specific user need. Many methods for content based image retrieval have been developed over the years, ranging from global to local features and, lately, to convolutional neural networks. Each of the approaches has its own benefits and drawbacks, but they also have similarities. In this paper we investigate how a method initially developed for local features, pyramid matching, then employed on texture features, spatial pyramids, can enhance general global features. We apply a spatial pyramid based approach to add spatial information to well known and established global descriptors, and present the results of an extensive evaluation that shows that this combination is able to outperform the original versions of the global features.
图像检索处理的问题是找到相关的图像,以满足特定的用户需求。多年来,人们开发了许多基于内容的图像检索方法,从全局特征到局部特征,以及最近的卷积神经网络。每种方法都有自己的优点和缺点,但它们也有相似之处。在本文中,我们研究了一种最初用于局部特征,金字塔匹配的方法,然后用于纹理特征,空间金字塔,如何增强一般的全局特征。我们采用基于空间金字塔的方法将空间信息添加到已知和已建立的全局描述符中,并提出了广泛评估的结果,表明这种组合能够优于原始版本的全局特征。
{"title":"Spatial pyramids for boosting global features in content based image retrieval","authors":"M. Lux, N. Anagnostopoulos, C. Iakovidou","doi":"10.1109/CBMI.2016.7500248","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500248","url":null,"abstract":"Image retrieval deals with the problem of finding relevant images to satisfy a specific user need. Many methods for content based image retrieval have been developed over the years, ranging from global to local features and, lately, to convolutional neural networks. Each of the approaches has its own benefits and drawbacks, but they also have similarities. In this paper we investigate how a method initially developed for local features, pyramid matching, then employed on texture features, spatial pyramids, can enhance general global features. We apply a spatial pyramid based approach to add spatial information to well known and established global descriptors, and present the results of an extensive evaluation that shows that this combination is able to outperform the original versions of the global features.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126958832","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
EIR — Efficient computer aided diagnosis framework for gastrointestinal endoscopies 用于胃肠道内窥镜检查的高效计算机辅助诊断框架
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500257
M. Riegler, Konstantin Pogorelov, P. Halvorsen, T. Lange, C. Griwodz, P. Schmidt, S. Eskeland, Dag Johansen
Analysis of medical videos for detection of abnormalities like lesions and diseases requires both high precision and recall but also real-time processing for live feedback during standard colonoscopies and scalability for massive population based screening, which can be done using a capsular video endoscope. Existing related work in this field does not provide the necessary combination of detection accuracy and performance. In this paper, a multimedia system is presented where the aim is to tackle automatic analysis of videos from the human gastrointestinal (GI) tract. The system includes the whole pipeline from data collection, processing and analysis, to visualization. The system combines filters using machine learning, image recognition and extraction of global and local image features, and it is built in a modular way, so that it can easily be extended. At the same time, it is developed for efficient processing in order to provide real-time feedback to the doctor. Initial experiments show that our system has detection and localisation accuracy at least as good as existing systems, but it stands out in terms of real-time performance and low resource consumption for scalability.
对医学视频进行分析以检测病变和疾病等异常情况,不仅需要高精度和召回率,还需要在标准结肠镜检查期间实时处理实时反馈,并且可以使用荚膜视频内窥镜进行大规模人群筛查。该领域现有的相关工作没有提供检测精度和性能的必要结合。在本文中,提出了一个多媒体系统,其目的是解决自动分析视频从人体胃肠道(GI)。该系统包括从数据采集、处理、分析到可视化的整个流程。该系统结合了使用机器学习、图像识别和提取全局和局部图像特征的过滤器,并以模块化的方式构建,因此可以很容易地扩展。同时,为了向医生提供实时反馈,它被开发为高效处理。初步实验表明,我们的系统具有至少与现有系统一样好的检测和定位精度,但它在实时性能和低资源消耗方面具有突出的可扩展性。
{"title":"EIR — Efficient computer aided diagnosis framework for gastrointestinal endoscopies","authors":"M. Riegler, Konstantin Pogorelov, P. Halvorsen, T. Lange, C. Griwodz, P. Schmidt, S. Eskeland, Dag Johansen","doi":"10.1109/CBMI.2016.7500257","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500257","url":null,"abstract":"Analysis of medical videos for detection of abnormalities like lesions and diseases requires both high precision and recall but also real-time processing for live feedback during standard colonoscopies and scalability for massive population based screening, which can be done using a capsular video endoscope. Existing related work in this field does not provide the necessary combination of detection accuracy and performance. In this paper, a multimedia system is presented where the aim is to tackle automatic analysis of videos from the human gastrointestinal (GI) tract. The system includes the whole pipeline from data collection, processing and analysis, to visualization. The system combines filters using machine learning, image recognition and extraction of global and local image features, and it is built in a modular way, so that it can easily be extended. At the same time, it is developed for efficient processing in order to provide real-time feedback to the doctor. Initial experiments show that our system has detection and localisation accuracy at least as good as existing systems, but it stands out in terms of real-time performance and low resource consumption for scalability.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"362 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115953093","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
A user-study examining visualization of lifelogs 一个用户研究检查可视化的生活日志
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500236
Soumyadeb Chowdhury, M. Ferdous, J. Jose
With continuous advances in the pervasive sensing and lifelogging technologies for the quantified self, users now can record their daily life activities automatically and seamlessly. In the existing lifelogging research, visualization techniques for presenting the lifelogs and evaluating the effectiveness of such techniques from a lifelogger's perspective has not been adequately studied. In this paper, we investigate the effectiveness of four distinct visualization techniques for exploring the lifelogs, which were collected by 22 lifeloggers who volunteered to use a wearable camera and a GPS device simultaneously, for a period of 3 days. Based on a user study with these 22 lifeloggers, which required them to browse through their personal lifelogs, we seek to identify the most effective visualization technique. Our results suggest various ways to augment and improve the visualization of personal lifelogs to enrich the quality of user experience and making lifelogging tools more engaging. We also propose a new visualization feature-drill-down approach with details-on-demand, to make the lifelogging visualization process more meaningful and informative to the lifeloggers.
随着无处不在的传感和量化自我的生活记录技术的不断进步,用户现在可以自动无缝地记录他们的日常生活活动。在现有的生命记录研究中,对生命记录的可视化呈现技术以及从生命记录者的角度评价这些技术的有效性的研究还不够充分。在本文中,我们研究了四种不同的可视化技术对探索生活日志的有效性,这些日志由22名志愿同时使用可穿戴相机和GPS设备的生活日志收集,为期3天。基于对这22位生活记录者的用户研究,我们要求他们浏览他们的个人生活记录,试图找出最有效的可视化技术。我们的研究结果提出了各种增强和改进个人生活日志可视化的方法,以丰富用户体验的质量,使生活日志工具更具吸引力。我们还提出了一种新的可视化特征-按需细节钻取方法,使生活日志可视化过程对生活记录者更有意义和信息。
{"title":"A user-study examining visualization of lifelogs","authors":"Soumyadeb Chowdhury, M. Ferdous, J. Jose","doi":"10.1109/CBMI.2016.7500236","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500236","url":null,"abstract":"With continuous advances in the pervasive sensing and lifelogging technologies for the quantified self, users now can record their daily life activities automatically and seamlessly. In the existing lifelogging research, visualization techniques for presenting the lifelogs and evaluating the effectiveness of such techniques from a lifelogger's perspective has not been adequately studied. In this paper, we investigate the effectiveness of four distinct visualization techniques for exploring the lifelogs, which were collected by 22 lifeloggers who volunteered to use a wearable camera and a GPS device simultaneously, for a period of 3 days. Based on a user study with these 22 lifeloggers, which required them to browse through their personal lifelogs, we seek to identify the most effective visualization technique. Our results suggest various ways to augment and improve the visualization of personal lifelogs to enrich the quality of user experience and making lifelogging tools more engaging. We also propose a new visualization feature-drill-down approach with details-on-demand, to make the lifelogging visualization process more meaningful and informative to the lifeloggers.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116838707","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Particle physics model for content-based 3D exploration 基于内容的3D探索粒子物理模型
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500259
Miroslav Macík, Jakub Lokoč, Premysl Cech, T. Skopal
Recent studies show that 3D visualization and browsing interfaces in content-based exploration systems of unstructured data represent a promising alternative to classical 2D grids. In this paper, we study 3D visualization techniques based on the particle physics model with focus on efficient evaluation of presentation layouts. We show that the particle physics model is a versatile approach suitable for exploration systems, enabling generation of various types of layouts. We also show that the model is able to organize thousands of objects given only limited time which is crucial for content-based exploration.
最近的研究表明,基于内容的非结构化数据探索系统中的3D可视化和浏览界面代表了传统2D网格的一个有希望的替代方案。本文研究了基于粒子物理模型的三维可视化技术,重点研究了如何有效地评估演示布局。我们表明,粒子物理模型是一种适用于探索系统的通用方法,可以生成各种类型的布局。我们还表明,该模型能够在有限的时间内组织数千个对象,这对于基于内容的探索至关重要。
{"title":"Particle physics model for content-based 3D exploration","authors":"Miroslav Macík, Jakub Lokoč, Premysl Cech, T. Skopal","doi":"10.1109/CBMI.2016.7500259","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500259","url":null,"abstract":"Recent studies show that 3D visualization and browsing interfaces in content-based exploration systems of unstructured data represent a promising alternative to classical 2D grids. In this paper, we study 3D visualization techniques based on the particle physics model with focus on efficient evaluation of presentation layouts. We show that the particle physics model is a versatile approach suitable for exploration systems, enabling generation of various types of layouts. We also show that the model is able to organize thousands of objects given only limited time which is crucial for content-based exploration.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116617040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
UCS: Ultimate course search UCS:终极课程搜索
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500242
Sheetal Rajgure, Vincent Oria, Krithika Raghavan, Hardik Dasadia, Sai Shashank Devannagari, Reza Curtmola, J. Geller, P. Gouton, Edina Renfro-Michel, Soon Ae Chun
In this system prototype demonstration we present, Ultimate Course Search (UCS), a learning tool developed to provide students ways to efficiently search electronic educational materials. UCS integrates slides, lecture videos and textbooks into a single platform. The keywords extracted from the textbooks and the slides are the basis of the indexing scheme. For the videos, UCS relies on slide transitions and metadata to establish the correspondence between slides and video segments. The video segmentation is based on the slides being presented using the meta-data provided by the video recording software and image processing techniques.
在这个系统原型演示中,我们展示了终极课程搜索(UCS),这是一个为学生提供有效搜索电子教育材料的学习工具。UCS将幻灯片、讲座视频和教科书集成到一个平台中。从教科书和幻灯片中提取的关键词是索引方案的基础。对于视频,UCS依赖于幻灯片转换和元数据来建立幻灯片和视频段之间的对应关系。视频分割是基于使用视频录制软件和图像处理技术提供的元数据所呈现的幻灯片。
{"title":"UCS: Ultimate course search","authors":"Sheetal Rajgure, Vincent Oria, Krithika Raghavan, Hardik Dasadia, Sai Shashank Devannagari, Reza Curtmola, J. Geller, P. Gouton, Edina Renfro-Michel, Soon Ae Chun","doi":"10.1109/CBMI.2016.7500242","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500242","url":null,"abstract":"In this system prototype demonstration we present, Ultimate Course Search (UCS), a learning tool developed to provide students ways to efficiently search electronic educational materials. UCS integrates slides, lecture videos and textbooks into a single platform. The keywords extracted from the textbooks and the slides are the basis of the indexing scheme. For the videos, UCS relies on slide transitions and metadata to establish the correspondence between slides and video segments. The video segmentation is based on the slides being presented using the meta-data provided by the video recording software and image processing techniques.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123330472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
From textual queries to visual queries 从文本查询到视觉查询
Pub Date : 2016-06-15 DOI: 10.1109/CBMI.2016.7500270
N. Zikos, A. Delopoulos, Dafni Maria Vasilikari
In this paper we present a framework to transform textual queries into visual ones. The proposed method uses standard image retrieval techniques with textual queries and the Fast Geometric Consistency Test (FGCT) method. For every textual query a set of images is retrieved and for every image a set of descriptors is extracted. Extracted features are combined with respect to their similarity in their descriptors' space and afterwards with respect to their geometric consistency on the image plane. All pairs of images are tested for consistent geometric structures using the FGCT method. This procedure extracts the subset of images that have a persistent geometric formation in the descriptors' space. Descriptors that compose the persistent formation are extracted and used as the input in a visual query; those features constitute the visual context of the visual query. Afterwards we perform again the FGCT method, but this time using the set of extracted features of the persistent formation into the cloud of images that consists of images with out a priori textual knowledge. It is noteworthy that the proposed method is scale, rotation and translation invariant. Experimental results on the Microsoft's Clickture dataset which consist of 1 million images are presented to support these statements.
本文提出了一个将文本查询转换为视觉查询的框架。该方法采用标准的图像检索技术和快速几何一致性测试(FGCT)方法。对于每个文本查询,检索一组图像,并为每个图像提取一组描述符。提取的特征根据其在描述符空间中的相似性进行组合,然后根据其在图像平面上的几何一致性进行组合。使用FGCT方法测试所有对图像的一致几何结构。该过程提取在描述符空间中具有持久几何形状的图像子集。组成持久形式的描述符被提取并用作可视化查询中的输入;这些特征构成了可视化查询的可视化上下文。之后,我们再次执行FGCT方法,但这次使用的是一组提取的持久形成的特征到图像云中,该图像由没有先验文本知识的图像组成。值得注意的是,该方法具有尺度、旋转、平移不变性。在微软的Clickture数据集(包含100万张图像)上的实验结果支持了这些说法。
{"title":"From textual queries to visual queries","authors":"N. Zikos, A. Delopoulos, Dafni Maria Vasilikari","doi":"10.1109/CBMI.2016.7500270","DOIUrl":"https://doi.org/10.1109/CBMI.2016.7500270","url":null,"abstract":"In this paper we present a framework to transform textual queries into visual ones. The proposed method uses standard image retrieval techniques with textual queries and the Fast Geometric Consistency Test (FGCT) method. For every textual query a set of images is retrieved and for every image a set of descriptors is extracted. Extracted features are combined with respect to their similarity in their descriptors' space and afterwards with respect to their geometric consistency on the image plane. All pairs of images are tested for consistent geometric structures using the FGCT method. This procedure extracts the subset of images that have a persistent geometric formation in the descriptors' space. Descriptors that compose the persistent formation are extracted and used as the input in a visual query; those features constitute the visual context of the visual query. Afterwards we perform again the FGCT method, but this time using the set of extracted features of the persistent formation into the cloud of images that consists of images with out a priori textual knowledge. It is noteworthy that the proposed method is scale, rotation and translation invariant. Experimental results on the Microsoft's Clickture dataset which consist of 1 million images are presented to support these statements.","PeriodicalId":356608,"journal":{"name":"2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2016-06-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125537744","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2016 14th International Workshop on Content-Based Multimedia Indexing (CBMI)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1