首页 > 最新文献

2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
A rapidly deployable virtual presence extended defense system 一个快速部署的虚拟存在扩展防御系统
M. W. Koch, C. Giron, Hung D. Nguyen
We have developed algorithms for a virtual presence and extended defense (VPED) system that automatically learns the detection map of a deployed sensor field without a-priori knowledge of the local terrain. The VPED system is a network of sensor pods, with each pod containing acoustic and seismic sensors. Each pod has a limited detection range, but a network of pods can form a virtual perimeter. The site's geography and soil conditions can affect the detection performance of the pods. Thus a network in the field may not have the same performance as a network designed in the lab. To solve this problem we automatically estimate a network's detection performance as it is being constructed. We demonstrate results using simulated and real data.
我们已经为虚拟存在和扩展防御(VPED)系统开发了算法,该系统可以自动学习部署的传感器场的检测地图,而无需先验地了解当地地形。VPED系统是一个传感器吊舱网络,每个吊舱包含声学和地震传感器。每个豆荚的探测范围有限,但豆荚网络可以形成一个虚拟的边界。场地的地理和土壤条件会影响吊舱的探测性能。因此,现场的网络可能与实验室设计的网络具有不同的性能。为了解决这个问题,我们在构建网络时自动估计网络的检测性能。我们用模拟数据和真实数据证明了结果。
{"title":"A rapidly deployable virtual presence extended defense system","authors":"M. W. Koch, C. Giron, Hung D. Nguyen","doi":"10.1109/CVPRW.2009.5204089","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204089","url":null,"abstract":"We have developed algorithms for a virtual presence and extended defense (VPED) system that automatically learns the detection map of a deployed sensor field without a-priori knowledge of the local terrain. The VPED system is a network of sensor pods, with each pod containing acoustic and seismic sensors. Each pod has a limited detection range, but a network of pods can form a virtual perimeter. The site's geography and soil conditions can affect the detection performance of the pods. Thus a network in the field may not have the same performance as a network designed in the lab. To solve this problem we automatically estimate a network's detection performance as it is being constructed. We demonstrate results using simulated and real data.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"34 7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125720149","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Groupwise morphometric analysis based on morphological appearance manifold 基于形态外观流形的群形态计量学分析
Naixiang Lian, C. Davatzikos
The field of computational anatomy has developed rigorous frameworks for analyzing anatomical shape, based on diffeomorphic transformations of a template. However, differences in algorithms used for template warping, in regularization parameters, and in the template itself, lead to different representations of the same anatomy. Variations of these parameters are considered as confounding factors as they give rise to non-unique representation. Recently, extensions of the conventional computational anatomy framework to account for such confounding variations have shown that learning the equivalence class derived from the multitude of representations can lead to improved and more stable morphological descriptors. Herein, we follow that approach, estimating the morphological appearance manifold obtained by varying parameters of the template warping procedure. Our approach parallels work in the computer vision field, in which variations lighting, pose and other parameters leads to image appearancemanifolds representing the exact same figure in different ways. The proposed framework is then used for groupwise registration and statistical analysis of biomedical images, by employing a minimum variance criterion on selected complete morphological descriptor to perform manifold-constrained optimization, i.e. to traverse each individual's morphological appearance manifold until group variance is minimal. Effectively, this process removes the aforementioned confounding effects and potentially leads to morphological representations reflecting purely biological variations, instead of variations introduced by modeling assumptions and parameter settings. The nonlinearity of a morphological appearancemanifold is treated via local approximations of the manifold via PCA.
计算解剖学领域已经开发出严格的框架来分析解剖形状,基于模板的微分同构变换。然而,用于模板翘曲的算法、正则化参数和模板本身的差异,导致相同解剖结构的不同表示。这些参数的变化被认为是混杂因素,因为它们会产生非唯一的表示。最近,对传统计算解剖学框架的扩展,以解释这种混淆的变化,表明学习从大量表示中派生的等价类可以导致改进和更稳定的形态描述符。在此,我们遵循该方法,估计由模板翘曲过程的不同参数获得的形态外观流形。我们的方法与计算机视觉领域的工作相似,在计算机视觉领域,光线、姿势和其他参数的变化会导致图像外观的变化,以不同的方式代表完全相同的人物。然后将该框架用于生物医学图像的分组配准和统计分析,通过对选定的完整形态学描述符采用最小方差准则进行流形约束优化,即遍历每个个体的形态学外观流形,直到组方差最小。有效地,这一过程消除了上述混淆效应,并可能导致形态表征反映纯粹的生物变异,而不是由建模假设和参数设置引入的变异。利用主成分分析法对流形进行局部近似,处理了形态外观流形的非线性。
{"title":"Groupwise morphometric analysis based on morphological appearance manifold","authors":"Naixiang Lian, C. Davatzikos","doi":"10.1109/CVPRW.2009.5204042","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204042","url":null,"abstract":"The field of computational anatomy has developed rigorous frameworks for analyzing anatomical shape, based on diffeomorphic transformations of a template. However, differences in algorithms used for template warping, in regularization parameters, and in the template itself, lead to different representations of the same anatomy. Variations of these parameters are considered as confounding factors as they give rise to non-unique representation. Recently, extensions of the conventional computational anatomy framework to account for such confounding variations have shown that learning the equivalence class derived from the multitude of representations can lead to improved and more stable morphological descriptors. Herein, we follow that approach, estimating the morphological appearance manifold obtained by varying parameters of the template warping procedure. Our approach parallels work in the computer vision field, in which variations lighting, pose and other parameters leads to image appearancemanifolds representing the exact same figure in different ways. The proposed framework is then used for groupwise registration and statistical analysis of biomedical images, by employing a minimum variance criterion on selected complete morphological descriptor to perform manifold-constrained optimization, i.e. to traverse each individual's morphological appearance manifold until group variance is minimal. Effectively, this process removes the aforementioned confounding effects and potentially leads to morphological representations reflecting purely biological variations, instead of variations introduced by modeling assumptions and parameter settings. The nonlinearity of a morphological appearancemanifold is treated via local approximations of the manifold via PCA.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"126 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128046984","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Use of Active Appearance Models for analysis and synthesis of naturally occurring behavior 使用活动外观模型分析和合成自然发生的行为
J. Cohn
Significant efforts have been made in the analysis and understanding of naturally occurring behavior. Active Appearance Models (AAM) are an especially exciting approach to this task for facial behavior. They may be used both to measure naturally occurring behavior and to synthesize photo-realistic real-time avatars with which to test hypotheses made possible by those measurements. We have used both of these capabilities, analysis and synthesis, to investigate the influence of depression on face-to-face interaction. With AAMs we have investigated large datasets of clinical interviews and successfully modeled and perturbed communicative behavior in a video conference paradigm to test causal hypotheses. These advances have lead to new understanding of the social functions of depression and dampened affect in dyadic interaction. Key challenges remain. These include automated detection and synthesis of subtle facial actions; hybrid methods that optimally integrate automated and manual processing; computational modeling of subjective states from multimodal input; and dynamic models of social and affective behavior.
在分析和理解自然发生的行为方面已经做出了重大努力。主动外观模型(AAM)是研究面部行为的一种特别令人兴奋的方法。它们既可以用来测量自然发生的行为,也可以用来合成逼真的实时化身,用来测试由这些测量产生的假设。我们使用了这两种能力,分析和综合,来调查抑郁症对面对面互动的影响。通过AAMs,我们调查了大量临床访谈数据集,并成功地在视频会议范式中模拟和干扰了沟通行为,以检验因果假设。这些进展使人们对抑郁症的社会功能和二元互动中的抑制情绪有了新的认识。主要挑战依然存在。其中包括自动检测和合成微妙的面部动作;混合方法,最佳地集成自动化和人工处理;基于多模态输入的主观状态计算建模以及社会和情感行为的动态模型。
{"title":"Use of Active Appearance Models for analysis and synthesis of naturally occurring behavior","authors":"J. Cohn","doi":"10.1109/CVPRW.2009.5204260","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204260","url":null,"abstract":"Significant efforts have been made in the analysis and understanding of naturally occurring behavior. Active Appearance Models (AAM) are an especially exciting approach to this task for facial behavior. They may be used both to measure naturally occurring behavior and to synthesize photo-realistic real-time avatars with which to test hypotheses made possible by those measurements. We have used both of these capabilities, analysis and synthesis, to investigate the influence of depression on face-to-face interaction. With AAMs we have investigated large datasets of clinical interviews and successfully modeled and perturbed communicative behavior in a video conference paradigm to test causal hypotheses. These advances have lead to new understanding of the social functions of depression and dampened affect in dyadic interaction. Key challenges remain. These include automated detection and synthesis of subtle facial actions; hybrid methods that optimally integrate automated and manual processing; computational modeling of subjective states from multimodal input; and dynamic models of social and affective behavior.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"17 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120888014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Geometric video projector auto-calibration 几何视频投影仪自动校准
Jamil Draréni, S. Roy, P. Sturm
In this paper we address the problem of geometric calibration of video projectors. Like in most previous methods we also use a camera that observes the projection on a planar surface. Contrary to those previous methods, we neither require the camera to be calibrated nor the presence of a calibration grid or other metric information about the scene. We thus speak of geometric auto-calibration of projectors (GAP). The fact that camera calibration is not needed increases the usability of the method and at the same time eliminates one potential source of inaccuracy, since errors in the camera calibration would otherwise inevitably propagate through to the projector calibration. Our method enjoys a good stability and gives good results when compared against existing methods as depicted by our experiments.
本文研究了视频投影仪的几何标定问题。像大多数以前的方法一样,我们也使用相机来观察平面上的投影。与以前的方法相反,我们既不需要校准相机,也不需要校准网格或其他关于场景的度量信息。因此,我们谈到投影仪的几何自动校准(GAP)。摄像机校准不需要的事实增加了该方法的可用性,同时消除了一个潜在的不准确来源,因为相机校准中的错误将不可避免地传播到投影仪校准。实验结果表明,与现有方法相比,该方法具有良好的稳定性和较好的结果。
{"title":"Geometric video projector auto-calibration","authors":"Jamil Draréni, S. Roy, P. Sturm","doi":"10.1109/CVPRW.2009.5204317","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204317","url":null,"abstract":"In this paper we address the problem of geometric calibration of video projectors. Like in most previous methods we also use a camera that observes the projection on a planar surface. Contrary to those previous methods, we neither require the camera to be calibrated nor the presence of a calibration grid or other metric information about the scene. We thus speak of geometric auto-calibration of projectors (GAP). The fact that camera calibration is not needed increases the usability of the method and at the same time eliminates one potential source of inaccuracy, since errors in the camera calibration would otherwise inevitably propagate through to the projector calibration. Our method enjoys a good stability and gives good results when compared against existing methods as depicted by our experiments.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123356447","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Cutout-search: Putting a name to the picture 搜索:给图片加上名字
Dhruv Batra, Adarsh Kowdle, Devi Parikh, Tsuhan Chen
We often come across photographs with content whose identity we can no longer recall. For instance, we may have a picture from a football game we went to, but do not remember the name of the team in the photograph. A natural instinct may be to query an image search engine with related general terms, such as `football' or `football teams' in this case. This would lead to many irrelevant retrievals, and the user would have to manually examine several pages of retrieval results before he can hope to find other images containing the same team players and look at the text associated with these images to identify the team. With the growing popularity of global image matching techniques, one may consider matching the query image to other images on the Web. However, this does not allow for ways to focus on the object-of-interest while matching, and may cause the background to overwhelm the matching results, especially when the object-of-interest is small and can occur in varying backgrounds, again, leading to irrelevant retrievals. We propose Cutout-Search, where a user employs an interactive segmentation tool to cut out the object-of-interest from the image, and use this Cutout-Query to retrieve images. As our experiments show, this leads to retrieval of more relevant images when compared to global image matching leading to more specific identification of the object-of-interest in the query image.
我们经常会遇到一些照片,其中的内容我们已经不记得是谁了。例如,我们可能有一张去看足球比赛的照片,但不记得照片中球队的名字。一种自然的本能可能是用相关的一般术语来查询图像搜索引擎,比如在这个例子中是“足球”或“足球队”。这将导致许多不相关的检索,并且用户必须手动检查检索结果的几个页面,然后才能希望找到包含相同团队球员的其他图像,并查看与这些图像相关的文本以识别团队。随着全局图像匹配技术的日益普及,可以考虑将查询图像与Web上的其他图像进行匹配。然而,这并不允许在匹配时关注感兴趣的对象,并且可能导致背景压倒匹配结果,特别是当感兴趣的对象很小并且可能出现在不同的背景中时,再次导致不相关的检索。我们提出了Cutout-Search,用户使用交互式分割工具从图像中剪切出感兴趣的对象,并使用该Cutout-Query来检索图像。正如我们的实验所示,与全局图像匹配相比,这导致检索更相关的图像,从而在查询图像中更具体地识别感兴趣的对象。
{"title":"Cutout-search: Putting a name to the picture","authors":"Dhruv Batra, Adarsh Kowdle, Devi Parikh, Tsuhan Chen","doi":"10.1109/CVPRW.2009.5204195","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204195","url":null,"abstract":"We often come across photographs with content whose identity we can no longer recall. For instance, we may have a picture from a football game we went to, but do not remember the name of the team in the photograph. A natural instinct may be to query an image search engine with related general terms, such as `football' or `football teams' in this case. This would lead to many irrelevant retrievals, and the user would have to manually examine several pages of retrieval results before he can hope to find other images containing the same team players and look at the text associated with these images to identify the team. With the growing popularity of global image matching techniques, one may consider matching the query image to other images on the Web. However, this does not allow for ways to focus on the object-of-interest while matching, and may cause the background to overwhelm the matching results, especially when the object-of-interest is small and can occur in varying backgrounds, again, leading to irrelevant retrievals. We propose Cutout-Search, where a user employs an interactive segmentation tool to cut out the object-of-interest from the image, and use this Cutout-Query to retrieve images. As our experiments show, this leads to retrieval of more relevant images when compared to global image matching leading to more specific identification of the object-of-interest in the query image.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"230 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123364018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Combining 2D and 3D hand geometry features for biometric verification 结合2D和3D手部几何特征进行生物识别验证
Vivek Kanhangad, Ajay Kumar, David Zhang
Traditional hand geometry based personal verification systems offer limited performance and therefore suitable only for small scale applications. This paper investigates a new approach to achieve performance improvement for hand geometry systems by simultaneously acquiring three dimensional features from the presented hands. The proposed system utilizes a laser based 3D digitizer to acquire registered intensity and range images of the presented hands in a completely contact-free manner, without using any hand position restricting mechanism. Two new representations that characterize the local features on the finger surface are extracted from the acquired range images and are matched using the proposed matching metrics. The proposed approach is evaluated on a database of 177 users, with 10 hand images for each user acquired in two sessions. Our experimental results suggest that the proposed 3D hand geometry features have significant discriminatory information to reliably authenticate individuals. Our experimental results also demonstrate that the combination of 3D hand geometry features with 2D geometry features can be employed to significantly improve the performance from 2D hand geometry features alone.
传统的基于手几何的个人验证系统提供有限的性能,因此只适用于小规模应用。本文研究了一种通过同时获取手的三维特征来提高手几何系统性能的新方法。该系统利用基于激光的三维数字化仪以完全无接触的方式获取所呈现手的注册强度和距离图像,而不使用任何手部位置限制机构。从获取的距离图像中提取两种新的表征手指表面局部特征的表示,并使用提出的匹配度量进行匹配。该方法在177个用户的数据库上进行了评估,每个用户在两个会话中获得10张手图像。实验结果表明,所提出的三维手部几何特征具有显著的判别信息,可以可靠地对个体进行身份验证。我们的实验结果还表明,3D手几何特征与2D手几何特征的结合可以显著提高单独使用2D手几何特征的性能。
{"title":"Combining 2D and 3D hand geometry features for biometric verification","authors":"Vivek Kanhangad, Ajay Kumar, David Zhang","doi":"10.1109/CVPRW.2009.5204306","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204306","url":null,"abstract":"Traditional hand geometry based personal verification systems offer limited performance and therefore suitable only for small scale applications. This paper investigates a new approach to achieve performance improvement for hand geometry systems by simultaneously acquiring three dimensional features from the presented hands. The proposed system utilizes a laser based 3D digitizer to acquire registered intensity and range images of the presented hands in a completely contact-free manner, without using any hand position restricting mechanism. Two new representations that characterize the local features on the finger surface are extracted from the acquired range images and are matched using the proposed matching metrics. The proposed approach is evaluated on a database of 177 users, with 10 hand images for each user acquired in two sessions. Our experimental results suggest that the proposed 3D hand geometry features have significant discriminatory information to reliably authenticate individuals. Our experimental results also demonstrate that the combination of 3D hand geometry features with 2D geometry features can be employed to significantly improve the performance from 2D hand geometry features alone.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124015391","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
EYEWATCHME—3D Hand and object tracking for inside out activity analysis EYEWATCHME-3D手和对象跟踪从内到外的活动分析
Li Sun, Ulrich Klank, M. Beetz
This paper investigates the “inside-out” recognition of everyday manipulation tasks using a gaze-directed camera, which is a camera that actively directs at the visual attention focus of the person wearing the camera. We present EYEWATCHME, an integrated vision and state estimation system that at the same time tracks the positions and the poses of the acting hands, the pose that the manipulated object, and the pose of the observing camera. Taken together, EYEWATCHME provides comprehensive data for learning predictive models of vision-guided manipulation that include the objects people are attending, the interaction of attention and reaching/grasping, and the segmentation of reaching and grasping using visual attention as evidence. Key technical contributions of this paper include an ego view hand tracking system that estimates 27 DOF hand poses. The hand tracking system is capable of detecting hands and estimating their poses despite substantial self-occlusion caused by the hand and occlusions caused by the manipulated object. EYEWATCHME can also cope with blurred images that are caused by rapid eye movements. The second key contribution is the of the integrated activity recognition system that simultaneously tracks the attention of the person, the hand poses, and the poses of the manipulated objects in terms of a global scene coordinates. We demonstrate the operation of EYEWATCHME in the context of kitchen tasks including filling a cup with water.
本文研究了使用视线导向相机对日常操作任务的“由内到外”识别,这是一种主动指向佩戴相机的人的视觉注意力焦点的相机。我们提出了EYEWATCHME,一个集成的视觉和状态估计系统,同时跟踪动作手的位置和姿势,被操纵物体的姿势,以及观察相机的姿势。综上所述,EYEWATCHME为学习视觉引导操作的预测模型提供了全面的数据,包括人们正在关注的对象,注意和伸手/抓握的相互作用,以及以视觉注意为证据的伸手和抓握的分割。本文的关键技术贡献包括一个自我视图手部跟踪系统,该系统可以估计27个DOF手部姿势。该手部跟踪系统能够检测手部并估计其姿势,尽管手部和被操纵物体造成了严重的自遮挡。EYEWATCHME还可以处理由快速眼球运动引起的图像模糊问题。第二个关键贡献是集成的活动识别系统,该系统可以根据全局场景坐标同时跟踪人的注意力、手的姿势和被操纵物体的姿势。我们在厨房任务的背景下演示EYEWATCHME的操作,包括将杯子装满水。
{"title":"EYEWATCHME—3D Hand and object tracking for inside out activity analysis","authors":"Li Sun, Ulrich Klank, M. Beetz","doi":"10.1109/CVPRW.2009.5204358","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204358","url":null,"abstract":"This paper investigates the “inside-out” recognition of everyday manipulation tasks using a gaze-directed camera, which is a camera that actively directs at the visual attention focus of the person wearing the camera. We present EYEWATCHME, an integrated vision and state estimation system that at the same time tracks the positions and the poses of the acting hands, the pose that the manipulated object, and the pose of the observing camera. Taken together, EYEWATCHME provides comprehensive data for learning predictive models of vision-guided manipulation that include the objects people are attending, the interaction of attention and reaching/grasping, and the segmentation of reaching and grasping using visual attention as evidence. Key technical contributions of this paper include an ego view hand tracking system that estimates 27 DOF hand poses. The hand tracking system is capable of detecting hands and estimating their poses despite substantial self-occlusion caused by the hand and occlusions caused by the manipulated object. EYEWATCHME can also cope with blurred images that are caused by rapid eye movements. The second key contribution is the of the integrated activity recognition system that simultaneously tracks the attention of the person, the hand poses, and the poses of the manipulated objects in terms of a global scene coordinates. We demonstrate the operation of EYEWATCHME in the context of kitchen tasks including filling a cup with water.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"994 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123096857","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 50
SUSurE: Speeded Up Surround Extrema feature detector and descriptor for realtime applications SUSurE:加速环绕极值特征检测器和描述符的实时应用
M. Ebrahimi, W. Mayol-Cuevas
There has been significant research into the development of visual feature detectors and descriptors that are robust to a number of image deformations. Some of these methods have emphasized the need to improve on computational speed and compact representations so that they can enable a range of real-time applications with reduced computational requirements. In this paper we present modified detectors and descriptors based on the recently introduced CenSurE [1], and show experimental results that aim to highlight the computational savings that can be made with limited reduction in performance. The developed methods are based on exploiting the concept of sparse sampling which may be of interest to a range of other existing approaches.
对于视觉特征检测器和描述符的开发已经有了重要的研究,这些特征检测器和描述符对许多图像变形都具有鲁棒性。其中一些方法强调需要提高计算速度和紧凑的表示,以便能够在减少计算需求的情况下实现一系列实时应用程序。在本文中,我们提出了基于最近引入的CenSurE[1]的改进检测器和描述符,并展示了旨在强调在有限的性能降低的情况下可以节省计算的实验结果。开发的方法是基于利用稀疏采样的概念,这可能对一系列其他现有方法感兴趣。
{"title":"SUSurE: Speeded Up Surround Extrema feature detector and descriptor for realtime applications","authors":"M. Ebrahimi, W. Mayol-Cuevas","doi":"10.1109/CVPRW.2009.5204313","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204313","url":null,"abstract":"There has been significant research into the development of visual feature detectors and descriptors that are robust to a number of image deformations. Some of these methods have emphasized the need to improve on computational speed and compact representations so that they can enable a range of real-time applications with reduced computational requirements. In this paper we present modified detectors and descriptors based on the recently introduced CenSurE [1], and show experimental results that aim to highlight the computational savings that can be made with limited reduction in performance. The developed methods are based on exploiting the concept of sparse sampling which may be of interest to a range of other existing approaches.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127344808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Head pose estimation using Spectral Regression Discriminant Analysis 基于光谱回归判别分析的头部姿态估计
Caifeng Shan, Wei Chen
In this paper, we investigate a recently proposed efficient subspace learning method, Spectral Regression Discriminant Analysis (SRDA), and its kernel version SRKDA for head pose estimation. One important unsolved issue of SRDA is how to automatically determine an appropriate regularization parameter. The parameter, which was empirically set in the existing work, has great impact on its performance. By formulating it as a constrained optimization problem, we present a method to estimate the optimal regularization parameter in SRDA and SRKDA. Our experiments on two databases illustrate that SRDA, especially SRKDA, is promising for head pose estimation. Moreover, our approach for estimating the regularization parameter is shown to be effective in head pose estimation and face recognition experiments.
在本文中,我们研究了最近提出的一种有效的子空间学习方法,光谱回归判别分析(SRDA)及其核版本SRKDA,用于头部姿态估计。SRDA尚未解决的一个重要问题是如何自动确定合适的正则化参数。该参数是现有工作中经验性设置的,对其性能影响较大。通过将其表述为约束优化问题,我们提出了一种估计SRDA和SRKDA中最优正则化参数的方法。我们在两个数据库上的实验表明,SRDA,尤其是SRKDA,在头姿估计方面是很有前景的。此外,我们的正则化参数估计方法在头部姿态估计和人脸识别实验中被证明是有效的。
{"title":"Head pose estimation using Spectral Regression Discriminant Analysis","authors":"Caifeng Shan, Wei Chen","doi":"10.1109/CVPRW.2009.5204261","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204261","url":null,"abstract":"In this paper, we investigate a recently proposed efficient subspace learning method, Spectral Regression Discriminant Analysis (SRDA), and its kernel version SRKDA for head pose estimation. One important unsolved issue of SRDA is how to automatically determine an appropriate regularization parameter. The parameter, which was empirically set in the existing work, has great impact on its performance. By formulating it as a constrained optimization problem, we present a method to estimate the optimal regularization parameter in SRDA and SRKDA. Our experiments on two databases illustrate that SRDA, especially SRKDA, is promising for head pose estimation. Moreover, our approach for estimating the regularization parameter is shown to be effective in head pose estimation and face recognition experiments.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124359175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Learning texton models for real-time scene context 学习实时场景上下文的文本模型
A. Flint, I. Reid, D. W. Murray
We present a new model for scene context based on the distribution of textons within images. Our approach provides continuous, consistent scene gist throughout a video sequence and is suitable for applications in which the camera regularly views uninformative parts of the scene. We show that our model outperforms the state-of-the-art for place recognition. We further show how to deduce the camera orientation from our scene gist and finally show how our system can be applied to active object search.
提出了一种基于图像内文本分布的场景上下文模型。我们的方法在整个视频序列中提供连续,一致的场景要点,适用于摄像机定期查看场景中无信息部分的应用。我们表明,我们的模型在位置识别方面优于最先进的技术。我们进一步展示了如何从场景要点推断相机方向,最后展示了我们的系统如何应用于活动对象搜索。
{"title":"Learning texton models for real-time scene context","authors":"A. Flint, I. Reid, D. W. Murray","doi":"10.1109/CVPRW.2009.5204356","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204356","url":null,"abstract":"We present a new model for scene context based on the distribution of textons within images. Our approach provides continuous, consistent scene gist throughout a video sequence and is suitable for applications in which the camera regularly views uninformative parts of the scene. We show that our model outperforms the state-of-the-art for place recognition. We further show how to deduce the camera orientation from our scene gist and finally show how our system can be applied to active object search.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115883505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1