首页 > 最新文献

2012 Ninth Conference on Computer and Robot Vision最新文献

英文 中文
Extracting Edge Voxels from 3D Volumetric Maps to Reduce Map Size and Accelerate Mapping Alignment 从三维体积地图中提取边缘体素以减小地图大小并加速地图对齐
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.50
J. Ryde, J. Delmerico
For effective mobile robots we need a concise yet adequately descriptive mechanism for representing their surroundings. Traditionally 2D occupancy grids have proven effective for task such as SLAM, path planning and obstacle avoidance. Applying this to 3D maps requires consideration due to the large memory requirements of the resulting dense arrays. Approaches to address this, such as octrees and occupied voxel lists, take advantage of the relative sparsity of occupied voxels. We enhance the occupied voxel list representation by filtering out those voxels that are on planar sections of the environment to leave edge-like voxels. To do this we apply a structure tensor operation to the voxel map followed by a classification of the eigen values to remove voxels that are part of flat regions such as floors, walls and ceilings. This leaves the voxels tracing the edges of the environment producing a wire-frame like model. Fewer edge voxels require less memory and enable faster alignment. We compare the performance of scan-to-map matching of extracted edge voxels with that of the corresponding full 3D scans. We show that alignment accuracy is preserved when using edge voxels, while achieving a five times speedup and reduced memory requirements, compared to matching with all occupied voxels. It is posited that these edge voxel maps could also be useful for appearance based localisation.
对于有效的移动机器人,我们需要一个简洁而充分的描述机制来表示它们的周围环境。传统上,二维占用网格已被证明对SLAM、路径规划和避障等任务是有效的。将此应用于3D地图需要考虑,因为生成的密集数组需要大量内存。解决这个问题的方法,如八叉树和已占用体素列表,利用了已占用体素的相对稀疏性。我们通过过滤掉那些在环境的平面部分上的体素以留下边缘样体素来增强占用体素列表的表示。为此,我们对体素图应用结构张量操作,然后对特征值进行分类,以去除作为平面区域(如地板、墙壁和天花板)一部分的体素。这使得体素跟踪环境的边缘,产生类似线框的模型。更少的边缘体素需要更少的内存并实现更快的对齐。我们比较了提取的边缘体素与相应的全3D扫描的扫描到映射匹配的性能。我们表明,与与所有占用的体素匹配相比,在使用边缘体素时保持对齐精度,同时实现五倍的加速和减少的内存需求。假设这些边缘体素地图也可以用于基于外观的定位。
{"title":"Extracting Edge Voxels from 3D Volumetric Maps to Reduce Map Size and Accelerate Mapping Alignment","authors":"J. Ryde, J. Delmerico","doi":"10.1109/CRV.2012.50","DOIUrl":"https://doi.org/10.1109/CRV.2012.50","url":null,"abstract":"For effective mobile robots we need a concise yet adequately descriptive mechanism for representing their surroundings. Traditionally 2D occupancy grids have proven effective for task such as SLAM, path planning and obstacle avoidance. Applying this to 3D maps requires consideration due to the large memory requirements of the resulting dense arrays. Approaches to address this, such as octrees and occupied voxel lists, take advantage of the relative sparsity of occupied voxels. We enhance the occupied voxel list representation by filtering out those voxels that are on planar sections of the environment to leave edge-like voxels. To do this we apply a structure tensor operation to the voxel map followed by a classification of the eigen values to remove voxels that are part of flat regions such as floors, walls and ceilings. This leaves the voxels tracing the edges of the environment producing a wire-frame like model. Fewer edge voxels require less memory and enable faster alignment. We compare the performance of scan-to-map matching of extracted edge voxels with that of the corresponding full 3D scans. We show that alignment accuracy is preserved when using edge voxels, while achieving a five times speedup and reduced memory requirements, compared to matching with all occupied voxels. It is posited that these edge voxel maps could also be useful for appearance based localisation.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"2013 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125668834","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Information Fusion in Visual-Task Inference 视觉任务推理中的信息融合
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.14
Amin Haji Abolhassani, James J. Clark
Eye movement is a rich modality that can provide us with a window into a person's mind. In a typical human-human interaction, we can get information about the behavioral state of the others by examining their eye movements. For instance, when a poker player looks into the eyes of his opponent, he looks for any indication of bluffing by verifying the dynamics of the eye movements. However, the information extracted from the eyes is not the only source of information we get in a human-human interaction and other modalities, such as speech or gesture, help us infer the behavioral state of the others. Most of the time this fusion of information refines our decisions and helps us better infer people's cognitive and behavioral activity based on their actions. In this paper, we develop a probabilistic framework to fuse different sources of information to infer the ongoing task in a visual search activity given the viewer's eye movement data. We propose to use a dynamic programming method called token passing in an eye-typing application to reveal what the subject is typing during a search process by observing his direction of gaze during the execution of the task. Token passing is a computationally simple technique that allows us to fuse higher order constraints in the inference process and build models dynamically so we can have unlimited number of hypotheses. In the experiments we examine the effect of higher order information, in the form of a lexicon dictionary, on the task recognition accuracy.
眼动是一种丰富的形态,它可以为我们提供一扇了解一个人思想的窗口。在典型的人际互动中,我们可以通过观察对方的眼球运动来了解对方的行为状态。例如,当一个扑克玩家看着对手的眼睛时,他会通过验证眼球运动的动态来寻找任何虚张声势的迹象。然而,从眼睛中提取的信息并不是我们在人际互动中获得的唯一信息来源,其他方式,如语言或手势,可以帮助我们推断他人的行为状态。大多数时候,这种信息融合可以改进我们的决策,帮助我们更好地根据人们的行为推断出他们的认知和行为活动。在本文中,我们开发了一个概率框架来融合不同来源的信息,以推断视觉搜索活动中正在进行的任务。我们建议在一个眼睛打字应用中使用一种动态规划方法,即标记传递,通过观察对象在执行任务过程中注视的方向来揭示对象在搜索过程中输入的内容。令牌传递是一种计算简单的技术,它允许我们在推理过程中融合高阶约束,并动态构建模型,这样我们就可以有无限数量的假设。在实验中,我们考察了词典词典形式的高阶信息对任务识别准确率的影响。
{"title":"Information Fusion in Visual-Task Inference","authors":"Amin Haji Abolhassani, James J. Clark","doi":"10.1109/CRV.2012.14","DOIUrl":"https://doi.org/10.1109/CRV.2012.14","url":null,"abstract":"Eye movement is a rich modality that can provide us with a window into a person's mind. In a typical human-human interaction, we can get information about the behavioral state of the others by examining their eye movements. For instance, when a poker player looks into the eyes of his opponent, he looks for any indication of bluffing by verifying the dynamics of the eye movements. However, the information extracted from the eyes is not the only source of information we get in a human-human interaction and other modalities, such as speech or gesture, help us infer the behavioral state of the others. Most of the time this fusion of information refines our decisions and helps us better infer people's cognitive and behavioral activity based on their actions. In this paper, we develop a probabilistic framework to fuse different sources of information to infer the ongoing task in a visual search activity given the viewer's eye movement data. We propose to use a dynamic programming method called token passing in an eye-typing application to reveal what the subject is typing during a search process by observing his direction of gaze during the execution of the task. Token passing is a computationally simple technique that allows us to fuse higher order constraints in the inference process and build models dynamically so we can have unlimited number of hypotheses. In the experiments we examine the effect of higher order information, in the form of a lexicon dictionary, on the task recognition accuracy.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129326359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Probabilistic Obstacle Detection Using 2 1/2 D Terrain Maps 基于2 1/2 D地形图的概率障碍物检测
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.10
G. Broten, David Mackay, J. Collier
Navigating unstructured environments requires reliable perception that generates an appropriate world representation. This representation must encompass all types of impediments to traversal, whether they be insurmountable obstacles, or mobility inhibitors such as soft soil. Traditionally, traversability and obstacle avoidance have represented separate capabilities with individual rangefinders dedicated to each task. This paper presents a statistical technique that, through the analysis of the underlying 21/2 D terrain map, determines the probability of an obstacle. This integrated approach eliminates the need for multiple data sources and is applicable to range data from various sources, including laser rangefinders and stereo vision. The proposed obstacle detection technique has been tested in simulated environments and under real world conditions, and these experiments revealed that it accurately identifies obstacles.
在非结构化环境中导航需要可靠的感知,从而产生适当的世界表示。这种表现必须包含所有类型的障碍,无论是不可逾越的障碍,还是像软土这样的移动抑制剂。传统上,可穿越性和避障都是单独的功能,每个任务都有单独的测距仪。本文提出了一种统计技术,通过分析底层的2.1 /2维地形图,确定障碍物的概率。这种集成方法消除了对多个数据源的需求,适用于来自各种来源的距离数据,包括激光测距仪和立体视觉。本文提出的障碍物检测技术已经在模拟环境和现实条件下进行了测试,实验结果表明,该技术能够准确地识别障碍物。
{"title":"Probabilistic Obstacle Detection Using 2 1/2 D Terrain Maps","authors":"G. Broten, David Mackay, J. Collier","doi":"10.1109/CRV.2012.10","DOIUrl":"https://doi.org/10.1109/CRV.2012.10","url":null,"abstract":"Navigating unstructured environments requires reliable perception that generates an appropriate world representation. This representation must encompass all types of impediments to traversal, whether they be insurmountable obstacles, or mobility inhibitors such as soft soil. Traditionally, traversability and obstacle avoidance have represented separate capabilities with individual rangefinders dedicated to each task. This paper presents a statistical technique that, through the analysis of the underlying 21/2 D terrain map, determines the probability of an obstacle. This integrated approach eliminates the need for multiple data sources and is applicable to range data from various sources, including laser rangefinders and stereo vision. The proposed obstacle detection technique has been tested in simulated environments and under real world conditions, and these experiments revealed that it accurately identifies obstacles.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114067989","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Multi-Scale Hierarchical Codebook Method for Human Action Recognition in Videos Using a Single Example 视频中人类动作识别的多尺度分层码本方法
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.32
M. J. Roshtkhari, M. Levine
This paper presents a novel action matching method based on a hierarchical codebook of local spatio-temporal video volumes (STVs). Given a single example of an activity as a query video, the proposed method finds similar videos to the query in a video dataset. It is based on the bag of video words (BOV) representation and does not require prior knowledge about actions, background subtraction, motion estimation or tracking. It is also robust to spatial and temporal scale changes, as well as some deformations. The hierarchical algorithm yields a compact subset of salient code words of STVs for the query video, and then the likelihood of similarity between the query video and all STVs in the target video is measured using a probabilistic inference mechanism. This hierarchy is achieved by initially constructing a codebook of STVs, while considering the uncertainty in the codebook construction, which is always ignored in current versions of the BOV approach. At the second level of the hierarchy, a large contextual region containing many STVs (Ensemble of STVs) is considered in order to construct a probabilistic model of STVs and their spatio-temporal compositions. At the third level of the hierarchy a codebook is formed for the ensembles of STVs based on their contextual similarities. The latter are the proposed labels (code words) for the actions being exhibited in the video. Finally, at the highest level of the hierarchy, the salient labels for the actions are selected by analyzing the high level code words assigned to each image pixel as a function of time. The algorithm was applied to three available video datasets for action recognition with different complexities (KTH, Weizmann, and MSR II) and the results were superior to other approaches, especially in the cases of a single training example and cross-dataset action recognition.
提出了一种基于局部时空视频卷分层码本的动作匹配方法。给定一个活动示例作为查询视频,该方法在视频数据集中找到与查询相似的视频。它基于视频词包(BOV)表示,不需要关于动作、背景减除、运动估计或跟踪的先验知识。它对空间和时间尺度的变化以及一些变形也具有鲁棒性。分层算法为查询视频生成stv显著码字的紧凑子集,然后使用概率推理机制测量查询视频与目标视频中所有stv之间的相似似然。这种层次结构是通过首先构造stv的码本来实现的,同时考虑到码本构造中的不确定性,这在当前版本的BOV方法中总是被忽略。在层次结构的第二层,考虑一个包含许多stv (stv集合)的大上下文区域,以构建stv及其时空组成的概率模型。在层次结构的第三层,基于上下文相似性为stv的集合形成一个码本。后者是建议的标签(码字),用于在视频中展示的动作。最后,在层次结构的最高层,通过分析分配给每个图像像素的高级码字作为时间的函数来选择动作的显著标签。将该算法应用于三个不同复杂度的视频数据集(KTH, Weizmann和MSR II)进行动作识别,结果优于其他方法,特别是在单个训练样例和跨数据集动作识别的情况下。
{"title":"A Multi-Scale Hierarchical Codebook Method for Human Action Recognition in Videos Using a Single Example","authors":"M. J. Roshtkhari, M. Levine","doi":"10.1109/CRV.2012.32","DOIUrl":"https://doi.org/10.1109/CRV.2012.32","url":null,"abstract":"This paper presents a novel action matching method based on a hierarchical codebook of local spatio-temporal video volumes (STVs). Given a single example of an activity as a query video, the proposed method finds similar videos to the query in a video dataset. It is based on the bag of video words (BOV) representation and does not require prior knowledge about actions, background subtraction, motion estimation or tracking. It is also robust to spatial and temporal scale changes, as well as some deformations. The hierarchical algorithm yields a compact subset of salient code words of STVs for the query video, and then the likelihood of similarity between the query video and all STVs in the target video is measured using a probabilistic inference mechanism. This hierarchy is achieved by initially constructing a codebook of STVs, while considering the uncertainty in the codebook construction, which is always ignored in current versions of the BOV approach. At the second level of the hierarchy, a large contextual region containing many STVs (Ensemble of STVs) is considered in order to construct a probabilistic model of STVs and their spatio-temporal compositions. At the third level of the hierarchy a codebook is formed for the ensembles of STVs based on their contextual similarities. The latter are the proposed labels (code words) for the actions being exhibited in the video. Finally, at the highest level of the hierarchy, the salient labels for the actions are selected by analyzing the high level code words assigned to each image pixel as a function of time. The algorithm was applied to three available video datasets for action recognition with different complexities (KTH, Weizmann, and MSR II) and the results were superior to other approaches, especially in the cases of a single training example and cross-dataset action recognition.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128040488","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Heat Kernels for Non-Rigid Shape Retrieval: Sparse Representation and Efficient Classification 非刚性形状检索的热核:稀疏表示和高效分类
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.28
M. Abdelrahman, M. El-Melegy, A. Farag
One of the major goals of computer vision and machine intelligence is the development of flexible and efficient methods for shape representation. This paper presents an approach for shape retrieval based on sparse representation of scale-invariant heat kernel. We use the Laplace-Beltrami eigen functions to detect a small number of critical points on the shape surface. Then a shape descriptor is formed based on the heat kernels at the detected critical points for different scales, combined with the normalized eigen values of the Lap lace-Beltrami operator. Sparse representation is used to reduce the dimensionality of the calculated descriptor. The proposed descriptor is used for classification via the collaborative representation-based classification with regularized least square algorithm. We compare our approach to two well-known approaches on two different data sets: the nonrigid world data set and the SHREC 2011. The results have indeed confirmed the improved performance of the proposed approach, yet reducing the time and space complicity of the shape retrieval problem.
计算机视觉和机器智能的主要目标之一是开发灵活有效的形状表示方法。提出了一种基于尺度不变热核稀疏表示的形状检索方法。我们使用拉普拉斯-贝尔特拉米特征函数来检测形状表面上的少量临界点。然后结合Lap lace-Beltrami算子的归一化特征值,根据检测到的不同尺度临界点处的热核形成形状描述子。稀疏表示用于降低计算描述符的维数。该描述符通过基于正则化最小二乘算法的协同表示分类进行分类。我们将我们的方法与两种众所周知的方法在两个不同的数据集上进行了比较:非刚性世界数据集和SHREC 2011。实验结果证实了该方法的有效性,同时降低了形状检索问题的时间复杂度和空间复杂度。
{"title":"Heat Kernels for Non-Rigid Shape Retrieval: Sparse Representation and Efficient Classification","authors":"M. Abdelrahman, M. El-Melegy, A. Farag","doi":"10.1109/CRV.2012.28","DOIUrl":"https://doi.org/10.1109/CRV.2012.28","url":null,"abstract":"One of the major goals of computer vision and machine intelligence is the development of flexible and efficient methods for shape representation. This paper presents an approach for shape retrieval based on sparse representation of scale-invariant heat kernel. We use the Laplace-Beltrami eigen functions to detect a small number of critical points on the shape surface. Then a shape descriptor is formed based on the heat kernels at the detected critical points for different scales, combined with the normalized eigen values of the Lap lace-Beltrami operator. Sparse representation is used to reduce the dimensionality of the calculated descriptor. The proposed descriptor is used for classification via the collaborative representation-based classification with regularized least square algorithm. We compare our approach to two well-known approaches on two different data sets: the nonrigid world data set and the SHREC 2011. The results have indeed confirmed the improved performance of the proposed approach, yet reducing the time and space complicity of the shape retrieval problem.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131159516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Evaluation of Local Spatio-temporal Salient Feature Detectors for Human Action Recognition 局部时空显著特征检测器对人体动作识别的评价
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.69
A. Shabani, David A Clausi, J. Zelek
Local spatio-temporal salient features are used for a sparse and compact representation of video contents in many computer vision tasks such as human action recognition. To localize these features (i.e., key point detection), existing methods perform either symmetric or asymmetric multi-resolution temporal filtering and use a structural or a motion saliency criteria. In a common discriminative framework for action classification, different saliency criteria of the structured-based detectors and different temporal filters of the motion-based detectors are compared. We have two main observations. (1) The motion-based detectors localize features which are more effective than those of structured-based detectors. (2) The salient motion features detected using an asymmetric temporal filtering performbetter than all other sparse salient detectors and dense sampling. Based on these two observations, we recommend the use of asymmetric motion features for effective sparse video content representation and action recognition.
在人类动作识别等计算机视觉任务中,局部时空显著特征用于视频内容的稀疏和紧凑表示。为了定位这些特征(即关键点检测),现有方法执行对称或非对称多分辨率时间滤波,并使用结构或运动显著性标准。在一个通用的动作分类判别框架中,比较了基于结构的检测器的不同显著性标准和基于运动的检测器的不同时间滤波器。我们有两个主要观察结果。(1)基于运动的检测器比基于结构的检测器更有效地定位特征。(2)使用非对称时间滤波检测显著性运动特征的性能优于所有其他稀疏显著性检测器和密集采样。基于这两个观察结果,我们建议使用非对称运动特征来进行有效的稀疏视频内容表示和动作识别。
{"title":"Evaluation of Local Spatio-temporal Salient Feature Detectors for Human Action Recognition","authors":"A. Shabani, David A Clausi, J. Zelek","doi":"10.1109/CRV.2012.69","DOIUrl":"https://doi.org/10.1109/CRV.2012.69","url":null,"abstract":"Local spatio-temporal salient features are used for a sparse and compact representation of video contents in many computer vision tasks such as human action recognition. To localize these features (i.e., key point detection), existing methods perform either symmetric or asymmetric multi-resolution temporal filtering and use a structural or a motion saliency criteria. In a common discriminative framework for action classification, different saliency criteria of the structured-based detectors and different temporal filters of the motion-based detectors are compared. We have two main observations. (1) The motion-based detectors localize features which are more effective than those of structured-based detectors. (2) The salient motion features detected using an asymmetric temporal filtering performbetter than all other sparse salient detectors and dense sampling. Based on these two observations, we recommend the use of asymmetric motion features for effective sparse video content representation and action recognition.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115727600","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Combination of Color and Binary Pattern Codification for an Error Correcting M-array Technique 一种纠错m阵列技术的彩色和二进制模式编码组合
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.26
Udaya Wijenayake, Sung-In Choi, Soon-Yong Park
Much research has been conducted so far to find a perfect structured light coding system. Among them, spatial neighborhood techniques which use a single pattern have become more popular as they can be used for dynamic scene capturing. But the difficulties of decoding the pattern when it loses few pattern symbols still remain as a problem. As a solution for this problem we introduce a new strategy which encodes two patterns into a single pattern image. In our particular experiment, we show that our decoding method can decode the pattern even it has some lost symbols.
为了找到一个完美的结构光编码系统,目前已经进行了大量的研究。其中,使用单一模式的空间邻域技术因其可用于动态场景的捕捉而越来越受欢迎。但是,当丢失少量模式符号时,解码模式的困难仍然是一个问题。为了解决这个问题,我们引入了一种新的策略,将两个模式编码成一个单一的模式图像。在我们的具体实验中,我们证明了我们的解码方法可以解码图案,即使它有一些丢失的符号。
{"title":"Combination of Color and Binary Pattern Codification for an Error Correcting M-array Technique","authors":"Udaya Wijenayake, Sung-In Choi, Soon-Yong Park","doi":"10.1109/CRV.2012.26","DOIUrl":"https://doi.org/10.1109/CRV.2012.26","url":null,"abstract":"Much research has been conducted so far to find a perfect structured light coding system. Among them, spatial neighborhood techniques which use a single pattern have become more popular as they can be used for dynamic scene capturing. But the difficulties of decoding the pattern when it loses few pattern symbols still remain as a problem. As a solution for this problem we introduce a new strategy which encodes two patterns into a single pattern image. In our particular experiment, we show that our decoding method can decode the pattern even it has some lost symbols.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"122 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122425977","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
What Can an Image of a Smooth Lambertian Surface Tell About Camera Nonlinearity? 光滑朗伯曲面的图像能告诉我们什么关于相机非线性的信息?
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.51
M. El-Melegy, A. Farag
This paper investigates the use of a single image of a smooth Lambertian surface to calibrate and remove some image nonlinearities due to the imaging device. To the best of our knowledge, this has not been addressed before in the literature. We show that this is possible, both theoretically and practically, taking advantage of some local shading measures that vary nonlinearly as a function of luminance and geometric nonlinearities (e.g., gamma correction and lens distortion). This can work as a basis for developing a simple method to estimate these nonlinearities from a single image. Several experiments are reported to validate the proposed method.
本文研究了使用光滑朗伯曲面的单幅图像来校准和消除由于成像装置引起的一些图像非线性。据我们所知,这在以前的文献中还没有提到过。我们表明,这是可能的,在理论上和实践中,利用一些局部遮光措施的非线性变化作为亮度和几何非线性的函数(例如,伽马校正和镜头畸变)。这可以作为开发从单个图像估计这些非线性的简单方法的基础。几个实验报告验证了所提出的方法。
{"title":"What Can an Image of a Smooth Lambertian Surface Tell About Camera Nonlinearity?","authors":"M. El-Melegy, A. Farag","doi":"10.1109/CRV.2012.51","DOIUrl":"https://doi.org/10.1109/CRV.2012.51","url":null,"abstract":"This paper investigates the use of a single image of a smooth Lambertian surface to calibrate and remove some image nonlinearities due to the imaging device. To the best of our knowledge, this has not been addressed before in the literature. We show that this is possible, both theoretically and practically, taking advantage of some local shading measures that vary nonlinearly as a function of luminance and geometric nonlinearities (e.g., gamma correction and lens distortion). This can work as a basis for developing a simple method to estimate these nonlinearities from a single image. Several experiments are reported to validate the proposed method.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121990246","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Shape-Based Registration of Kidneys Across Differently Contrasted CT Scans 不同对比CT扫描中基于形状的肾脏配准
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.39
F. Flores-Mangas, A. Jepson, M. Haider
We present a method to register kidneys from Computed Tomography (CT) scans with and without contrast enhancement. The method builds a patient-specific kidney shape model from the contrast enhanced image, and then matches it against automatically segmented candidate surfaces extracted from the pre-contrast image to find the alignment. Only the object of interest is used to drive the alignment, providing results that are robust to near-rigid relative motions of the kidney with respect to the surrounding tissues. Shape-based features are used, as opposed to intensity-based ones, and consequently the resulting registration is invariant to the inherent contrast variations. The contributions of this work are: a surface grouping and segmentation algorithm driven by smooth curvature constraints, and a framework to register image volumes under contrast variation, relative motion and local deformation with minimal user intervention. Encouraging experimental results with real patient images, all with various kinds and sizes of kidney lesions, validate the approach.
我们提出了一种方法,以登记肾脏从计算机断层扫描(CT)与没有对比增强。该方法从增强后的图像中建立患者肾脏形状模型,然后将其与从增强前图像中提取的自动分割的候选表面进行匹配,以找到匹配对象。只有感兴趣的对象被用来驱动对齐,提供的结果对肾脏相对于周围组织的近刚性相对运动是稳健的。使用基于形状的特征,而不是基于强度的特征,因此所得到的配准对固有的对比度变化是不变的。这项工作的贡献是:由光滑曲率约束驱动的表面分组和分割算法,以及在对比度变化,相对运动和局部变形下以最小用户干预进行图像体积配准的框架。令人鼓舞的实验结果与真实的病人图像,所有不同类型和大小的肾脏病变,验证了该方法。
{"title":"Shape-Based Registration of Kidneys Across Differently Contrasted CT Scans","authors":"F. Flores-Mangas, A. Jepson, M. Haider","doi":"10.1109/CRV.2012.39","DOIUrl":"https://doi.org/10.1109/CRV.2012.39","url":null,"abstract":"We present a method to register kidneys from Computed Tomography (CT) scans with and without contrast enhancement. The method builds a patient-specific kidney shape model from the contrast enhanced image, and then matches it against automatically segmented candidate surfaces extracted from the pre-contrast image to find the alignment. Only the object of interest is used to drive the alignment, providing results that are robust to near-rigid relative motions of the kidney with respect to the surrounding tissues. Shape-based features are used, as opposed to intensity-based ones, and consequently the resulting registration is invariant to the inherent contrast variations. The contributions of this work are: a surface grouping and segmentation algorithm driven by smooth curvature constraints, and a framework to register image volumes under contrast variation, relative motion and local deformation with minimal user intervention. Encouraging experimental results with real patient images, all with various kinds and sizes of kidney lesions, validate the approach.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132079411","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Design and Evaluation of a Flexible Interface for Spatial Navigation 空间导航灵活界面的设计与评价
Pub Date : 2012-05-28 DOI: 10.1109/CRV.2012.53
Emily Tsang, S. W. Ong, Joelle Pineau
The paper tackles the problem of designing intuitive graphical interfaces for selecting navigational targets for an autonomous robot. Our work focuses on the design and validation of such a flexible interface for an intelligent wheelchair navigating in a large indoor environment. We begin by describing the robot platform and interface design. We then present results from a user study in which participants were required to select navigational targets using a variety of input and filtering methods. We considered two types of input modalities (point-and-click and single-switch), to investigate the effect of constraints on the input mode. We take a particular look at the use of filtering methods to reduce the amount of information presented onscreen and thereby accelerate selection of the correct option.
研究了自主机器人导航目标选择的直观图形界面设计问题。我们的工作重点是在大型室内环境中为智能轮椅导航设计和验证这种灵活的界面。我们首先描述了机器人的平台和界面设计。然后,我们展示了一项用户研究的结果,该研究要求参与者使用各种输入和过滤方法选择导航目标。我们考虑了两种类型的输入模式(点击和单开关),以研究约束对输入模式的影响。我们特别关注过滤方法的使用,以减少屏幕上显示的信息量,从而加快正确选项的选择。
{"title":"Design and Evaluation of a Flexible Interface for Spatial Navigation","authors":"Emily Tsang, S. W. Ong, Joelle Pineau","doi":"10.1109/CRV.2012.53","DOIUrl":"https://doi.org/10.1109/CRV.2012.53","url":null,"abstract":"The paper tackles the problem of designing intuitive graphical interfaces for selecting navigational targets for an autonomous robot. Our work focuses on the design and validation of such a flexible interface for an intelligent wheelchair navigating in a large indoor environment. We begin by describing the robot platform and interface design. We then present results from a user study in which participants were required to select navigational targets using a variety of input and filtering methods. We considered two types of input modalities (point-and-click and single-switch), to investigate the effect of constraints on the input mode. We take a particular look at the use of filtering methods to reduce the amount of information presented onscreen and thereby accelerate selection of the correct option.","PeriodicalId":372951,"journal":{"name":"2012 Ninth Conference on Computer and Robot Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2012-05-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128425294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2012 Ninth Conference on Computer and Robot Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1