首页 > 最新文献

2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
Head and gaze dynamics in visual attention and context learning 视觉注意和语境学习中的头部和凝视动态
A. Doshi, M. Trivedi
Future intelligent environments and systems may need to interact with humans while simultaneously analyzing events and critical situations. Assistive living, advanced driver assistance systems, and intelligent command-and-control centers are just a few of these cases where human interactions play a critical role in situation analysis. In particular, the behavior or body language of the human subject may be a strong indicator of the context of the situation. In this paper we demonstrate how the interaction of a human observer's head pose and eye gaze behaviors can provide significant insight into the context of the event. Such semantic data derived from human behaviors can be used to help interpret and recognize an ongoing event. We present examples from driving and intelligent meeting rooms to support these conclusions, and demonstrate how to use these techniques to improve contextual learning.
未来的智能环境和系统可能需要与人类互动,同时分析事件和关键情况。辅助生活、高级驾驶员辅助系统和智能指挥控制中心只是人类互动在情况分析中发挥关键作用的几个案例。特别是,人类主体的行为或肢体语言可能是情景背景的一个强有力的指标。在本文中,我们展示了人类观察者的头部姿势和眼睛注视行为的相互作用如何为事件的背景提供重要的见解。这种来自人类行为的语义数据可以用来帮助解释和识别正在进行的事件。我们提供了来自驾驶和智能会议室的例子来支持这些结论,并演示了如何使用这些技术来提高上下文学习。
{"title":"Head and gaze dynamics in visual attention and context learning","authors":"A. Doshi, M. Trivedi","doi":"10.1109/CVPRW.2009.5204215","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204215","url":null,"abstract":"Future intelligent environments and systems may need to interact with humans while simultaneously analyzing events and critical situations. Assistive living, advanced driver assistance systems, and intelligent command-and-control centers are just a few of these cases where human interactions play a critical role in situation analysis. In particular, the behavior or body language of the human subject may be a strong indicator of the context of the situation. In this paper we demonstrate how the interaction of a human observer's head pose and eye gaze behaviors can provide significant insight into the context of the event. Such semantic data derived from human behaviors can be used to help interpret and recognize an ongoing event. We present examples from driving and intelligent meeting rooms to support these conclusions, and demonstrate how to use these techniques to improve contextual learning.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129690818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Generative hierarchical models for image analysis 图像分析的生成层次模型
S. Geman
A probabilistic grammar for the groupings and labeling of parts and objects, when taken together with pose and part-dependent appearance models, constitutes a generative scene model and a Bayesian framework for image analysis. To the extent that the generative model generates features, as opposed to pixel intensities, the "inverse" or "posterior distribution" on interpretations given images is based on incomplete information; feature vectors are generally insufficient to recover the original intensities. I will argue for fully generative scene models, meaning models that in principle generate actual digital pictures. I will outline an approach to the construction of fully generative models through an extension of context-sensitive grammars and a re-formulation of the popular template models for image fragments. Mostly I will focus on the problem of constructing pixel-level appearance models. I will propose an approach based on image-fragment templates, as introduced by Ullman and others. However, rather than using a correlation between a template and a given image patch as an extracted feature.
用于零件和物体的分组和标记的概率语法,当与姿势和部分相关的外观模型一起使用时,构成了生成场景模型和用于图像分析的贝叶斯框架。在某种程度上,生成模型生成特征,而不是像素强度,对给定图像的解释的“逆”或“后验分布”是基于不完整的信息;特征向量通常不足以恢复原始强度。我将支持完全生成场景模型,即原则上生成实际数字图像的模型。我将概述一种通过扩展上下文敏感语法和重新制定流行的图像片段模板模型来构建完全生成模型的方法。我将主要关注构建像素级外观模型的问题。我将提出一种基于图像片段模板的方法,正如Ullman等人所介绍的那样。然而,而不是使用模板和给定图像补丁之间的相关性作为提取特征。
{"title":"Generative hierarchical models for image analysis","authors":"S. Geman","doi":"10.1109/CVPRW.2009.5204335","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204335","url":null,"abstract":"A probabilistic grammar for the groupings and labeling of parts and objects, when taken together with pose and part-dependent appearance models, constitutes a generative scene model and a Bayesian framework for image analysis. To the extent that the generative model generates features, as opposed to pixel intensities, the \"inverse\" or \"posterior distribution\" on interpretations given images is based on incomplete information; feature vectors are generally insufficient to recover the original intensities. I will argue for fully generative scene models, meaning models that in principle generate actual digital pictures. I will outline an approach to the construction of fully generative models through an extension of context-sensitive grammars and a re-formulation of the popular template models for image fragments. Mostly I will focus on the problem of constructing pixel-level appearance models. I will propose an approach based on image-fragment templates, as introduced by Ullman and others. However, rather than using a correlation between a template and a given image patch as an extracted feature.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132012316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Face recognition by fusion of local and global matching scores using DS theory: An evaluation with uni-classifier and multi-classifier paradigm 基于DS理论的局部与全局匹配分数融合人脸识别:单分类器与多分类器的评价
D. Kisku, M. Tistarelli, J. Sing, Phalguni Gupta
Faces are highly deformable objects which may easily change their appearance over time. Not all face areas are subject to the same variability. Therefore decoupling the information from independent areas of the face is of paramount importance to improve the robustness of any face recognition technique. This paper presents a robust face recognition technique based on the extraction and matching of SIFT features related to independent face areas. Both a global and local (as recognition from parts) matching strategy is proposed. The local strategy is based on matching individual salient facial SIFT features as connected to facial landmarks such as the eyes and the mouth. As for the global matching strategy, all SIFT features are combined together to form a single feature. In order to reduce the identification errors, the Dempster-Shafer decision theory is applied to fuse the two matching techniques. The proposed algorithms are evaluated with the ORL and the IITK face databases. The experimental results demonstrate the effectiveness and potential of the proposed face recognition techniques also in the case of partially occluded faces or with missing information.
脸是高度可变形的物体,随着时间的推移很容易改变外观。并不是所有的面部区域都有相同的可变性。因此,从人脸的独立区域解耦信息对于提高任何人脸识别技术的鲁棒性至关重要。提出了一种基于独立人脸区域相关SIFT特征提取与匹配的鲁棒人脸识别技术。提出了一种全局匹配策略和局部匹配策略。局部策略是基于匹配与面部标志(如眼睛和嘴巴)相连的单个显著面部SIFT特征。在全局匹配策略上,将SIFT的所有特征组合在一起形成单个特征。为了减小识别误差,采用Dempster-Shafer决策理论对两种匹配技术进行融合。用ORL和IITK人脸数据库对算法进行了评价。实验结果证明了所提出的人脸识别技术在部分遮挡或信息缺失情况下的有效性和潜力。
{"title":"Face recognition by fusion of local and global matching scores using DS theory: An evaluation with uni-classifier and multi-classifier paradigm","authors":"D. Kisku, M. Tistarelli, J. Sing, Phalguni Gupta","doi":"10.1109/CVPRW.2009.5204298","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204298","url":null,"abstract":"Faces are highly deformable objects which may easily change their appearance over time. Not all face areas are subject to the same variability. Therefore decoupling the information from independent areas of the face is of paramount importance to improve the robustness of any face recognition technique. This paper presents a robust face recognition technique based on the extraction and matching of SIFT features related to independent face areas. Both a global and local (as recognition from parts) matching strategy is proposed. The local strategy is based on matching individual salient facial SIFT features as connected to facial landmarks such as the eyes and the mouth. As for the global matching strategy, all SIFT features are combined together to form a single feature. In order to reduce the identification errors, the Dempster-Shafer decision theory is applied to fuse the two matching techniques. The proposed algorithms are evaluated with the ORL and the IITK face databases. The experimental results demonstrate the effectiveness and potential of the proposed face recognition techniques also in the case of partially occluded faces or with missing information.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130767528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 48
Modeling and exploiting the spatio-temporal facial action dependencies for robust spontaneous facial expression recognition 基于人脸动作时空依赖的鲁棒性自发面部表情识别建模与开发
Yan Tong, Jixu Chen, Q. Ji
Facial action provides various types of messages for human communications. Recognizing spontaneous facial actions, however, is very challenging due to subtle facial deformation, frequent head movements, and ambiguous and uncertain facial motion measurements. As a result, current research in facial action recognition is limited to posed facial actions and often in frontal view.Spontaneous facial action is characterized by rigid head movements and nonrigid facial muscular movements. More importantly, it is the spatiotemporal interactions among the rigid and nonrigid facial motions that produce a meaningful and natural facial display. Recognizing this fact, we introduce a probabilistic facial action model based on a dynamic Bayesian network (DBN) to simultaneously and coherently capture rigid and nonrigid facial motions, their spatiotemporal dependencies, and their image measurements. Advanced machine learning methods are introduced to learn the probabilistic facial action model based on both training data and prior knowledge. Facial action recognition is accomplished through probabilistic inference by systemically integrating measurements official motions with the facial action model. Experiments show that the proposed system yields significant improvements in recognizing spontaneous facial actions.
面部动作为人类交流提供了各种类型的信息。然而,由于细微的面部变形、频繁的头部运动以及模糊和不确定的面部运动测量,识别自发的面部动作非常具有挑战性。因此,目前对面部动作识别的研究仅限于摆姿势的面部动作,而且通常是在正面视图下。自发性面部动作的特点是刚性的头部运动和非刚性的面部肌肉运动。更重要的是,刚性和非刚性面部运动之间的时空相互作用产生了有意义和自然的面部表现。认识到这一事实,我们引入了一个基于动态贝叶斯网络(DBN)的概率面部动作模型,以同时连贯地捕捉刚性和非刚性面部运动、它们的时空依赖关系以及它们的图像测量。引入先进的机器学习方法来学习基于训练数据和先验知识的概率面部动作模型。面部动作识别是通过概率推理,系统地将测量、官方动作与面部动作模型相结合来实现的。实验表明,该系统在识别自发面部动作方面取得了显著的进步。
{"title":"Modeling and exploiting the spatio-temporal facial action dependencies for robust spontaneous facial expression recognition","authors":"Yan Tong, Jixu Chen, Q. Ji","doi":"10.1109/CVPRW.2009.5204263","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204263","url":null,"abstract":"Facial action provides various types of messages for human communications. Recognizing spontaneous facial actions, however, is very challenging due to subtle facial deformation, frequent head movements, and ambiguous and uncertain facial motion measurements. As a result, current research in facial action recognition is limited to posed facial actions and often in frontal view.Spontaneous facial action is characterized by rigid head movements and nonrigid facial muscular movements. More importantly, it is the spatiotemporal interactions among the rigid and nonrigid facial motions that produce a meaningful and natural facial display. Recognizing this fact, we introduce a probabilistic facial action model based on a dynamic Bayesian network (DBN) to simultaneously and coherently capture rigid and nonrigid facial motions, their spatiotemporal dependencies, and their image measurements. Advanced machine learning methods are introduced to learn the probabilistic facial action model based on both training data and prior knowledge. Facial action recognition is accomplished through probabilistic inference by systemically integrating measurements official motions with the facial action model. Experiments show that the proposed system yields significant improvements in recognizing spontaneous facial actions.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114296754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Geometric Sequence (GS) imaging with Bayesian smoothing for optical and capacitive imaging sensors 光学和电容成像传感器的贝叶斯平滑几何序列成像
K. Sengupta, F. Porikli
In this paper, we introduce a novel technique called geometric sequence (GS) imaging, specifically for the purpose of low power and light weight tracking in human computer interface design. The imaging sensor is programmed to capture the scene with a train of packets, where each packet constitutes a few images. The delay or the baseline associated with consecutive image pairs in a packet follows a fixed ratio, as in a geometric sequence. The image pair with shorter baseline or delay captures fast motion, while the image pair with larger baseline or delay captures slow motion. Given an image packet, the motion confidence maps computed from the slow and the fast image pairs are fused into a single map. Next, we use a Bayesian update scheme to compute the motion hypotheses probability map, given the information of prior packets. We estimate the motion from this probability map. The GS imaging system reliably tracks slow movements as well as fast movements, a feature that is important in realizing applications such as a touchpad type system. Compared to continuous imaging with short delay between consecutive pairs, the GS imaging technique enjoys several advantages. The overall power consumption and the CPU load are significantly low. We present results in the domain of optical camera based human computer interface (HCI) applications, as well as for capacitive fingerprint imaging sensor based touch pad systems.
在本文中,我们介绍了一种新的技术,称为几何序列(GS)成像,特别是为低功耗和轻重量的跟踪在人机界面设计的目的。成像传感器被编程为用一系列数据包捕获场景,其中每个数据包构成一些图像。与数据包中连续图像对相关联的延迟或基线遵循固定的比率,如在几何序列中。基线或延迟较短的图像对捕获快速运动,而基线或延迟较大的图像对捕获慢动作。给定一个图像包,从慢速和快速图像对计算得到的运动置信度映射融合成一个单一的映射。接下来,我们使用贝叶斯更新方案来计算运动假设概率映射,给定先前数据包的信息。我们从这个概率图中估计运动。GS成像系统可靠地跟踪慢速运动和快速运动,这是实现触摸板类型系统等应用的重要功能。与连续对之间的短延迟连续成像相比,GS成像技术具有以下优点。整体功耗和CPU负载都很低。我们介绍了基于光学相机的人机界面(HCI)应用领域的结果,以及基于电容式指纹成像传感器的触摸板系统。
{"title":"Geometric Sequence (GS) imaging with Bayesian smoothing for optical and capacitive imaging sensors","authors":"K. Sengupta, F. Porikli","doi":"10.1109/CVPRW.2009.5205205","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5205205","url":null,"abstract":"In this paper, we introduce a novel technique called geometric sequence (GS) imaging, specifically for the purpose of low power and light weight tracking in human computer interface design. The imaging sensor is programmed to capture the scene with a train of packets, where each packet constitutes a few images. The delay or the baseline associated with consecutive image pairs in a packet follows a fixed ratio, as in a geometric sequence. The image pair with shorter baseline or delay captures fast motion, while the image pair with larger baseline or delay captures slow motion. Given an image packet, the motion confidence maps computed from the slow and the fast image pairs are fused into a single map. Next, we use a Bayesian update scheme to compute the motion hypotheses probability map, given the information of prior packets. We estimate the motion from this probability map. The GS imaging system reliably tracks slow movements as well as fast movements, a feature that is important in realizing applications such as a touchpad type system. Compared to continuous imaging with short delay between consecutive pairs, the GS imaging technique enjoys several advantages. The overall power consumption and the CPU load are significantly low. We present results in the domain of optical camera based human computer interface (HCI) applications, as well as for capacitive fingerprint imaging sensor based touch pad systems.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124683947","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Face recognition at-a-distance based on sparse-stereo reconstruction 基于稀疏立体重建的远距离人脸识别
H. Rara, S. Elhabian, Asem M. Ali, Mike Miller, T. Starr, A. Farag
We describe a framework for face recognition at a distance based on sparse-stereo reconstruction. We develop a 3D acquisition system that consists of two CCD stereo cameras mounted on pan-tilt units with adjustable baseline. We first detect the facial region and extract its landmark points, which are used to initialize an AAM mesh fitting algorithm. The fitted mesh vertices provide point correspondences between the left and right images of a stereo pair; stereo-based reconstruction is then used to infer the 3D information of the mesh vertices. We perform experiments regarding the use of different features extracted from these vertices for face recognition. The cumulative rank curves (CMC), which are generated using the proposed framework, confirms the feasibility of the proposed work for long distance recognition of human faces with respect to the state-of-the-art.
我们描述了一种基于稀疏立体重建的远距离人脸识别框架。我们开发了一个三维采集系统,该系统由两个CCD立体摄像机组成,安装在具有可调基线的平移倾斜单元上。我们首先检测面部区域并提取其地标点,用于初始化AAM网格拟合算法。拟合的网格顶点在立体图像对的左右图像之间提供点对应;然后使用基于立体的重建来推断网格顶点的三维信息。我们进行了关于使用从这些顶点提取的不同特征进行人脸识别的实验。使用所建议的框架生成的累积等级曲线(CMC)证实了所建议的远距离人脸识别工作相对于最新技术的可行性。
{"title":"Face recognition at-a-distance based on sparse-stereo reconstruction","authors":"H. Rara, S. Elhabian, Asem M. Ali, Mike Miller, T. Starr, A. Farag","doi":"10.1109/CVPRW.2009.5204301","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204301","url":null,"abstract":"We describe a framework for face recognition at a distance based on sparse-stereo reconstruction. We develop a 3D acquisition system that consists of two CCD stereo cameras mounted on pan-tilt units with adjustable baseline. We first detect the facial region and extract its landmark points, which are used to initialize an AAM mesh fitting algorithm. The fitted mesh vertices provide point correspondences between the left and right images of a stereo pair; stereo-based reconstruction is then used to infer the 3D information of the mesh vertices. We perform experiments regarding the use of different features extracted from these vertices for face recognition. The cumulative rank curves (CMC), which are generated using the proposed framework, confirms the feasibility of the proposed work for long distance recognition of human faces with respect to the state-of-the-art.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130841135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Geo-location inference from image content and user tags 从图像内容和用户标签推断地理位置
Andrew C. Gallagher, D. Joshi, Jie Yu, Jiebo Luo
Associating image content with their geographic locations has been increasingly pursued in the computer vision community in recent years. In a recent work, large collections of geotagged images were found to be helpful in estimating geo-locations of query images by simple visual nearest-neighbors search. In this paper, we leverage user tags along with image content to infer the geo-location. Our model builds upon the fact that the visual content and user tags of pictures can provide significant hints about their geo-locations. Using a large collection of over a million geotagged photographs, we build location probability maps of user tags over the entire globe. These maps reflect the picture-taking and tagging behaviors of thousands of users from all over the world, and reveal interesting tag map patterns. Visual content matching is performed using multiple feature descriptors including tiny images, color histograms, GIST features, and bags of textons. The combination of visual content matching and local tag probability maps forms a strong geo-inference engine. Large-scale experiments have shown significant improvements over pure visual content-based geo-location inference.
将图像内容与其地理位置相关联是近年来计算机视觉界越来越关注的问题。在最近的一项工作中,通过简单的视觉近邻搜索,发现大量地理标记图像的集合有助于估计查询图像的地理位置。在本文中,我们利用用户标签和图像内容来推断地理位置。我们的模型建立在这样一个事实之上,即图片的视觉内容和用户标签可以提供有关其地理位置的重要提示。使用超过一百万张地理标记照片的大集合,我们构建了全球用户标签的位置概率图。这些地图反映了来自世界各地成千上万用户的拍照和标记行为,并揭示了有趣的标记地图模式。视觉内容匹配使用多个特征描述符执行,包括微小图像、颜色直方图、GIST特征和文本包。视觉内容匹配和局部标签概率图的结合形成了一个强大的地理推理引擎。大规模实验表明,与纯基于视觉内容的地理位置推断相比,有了显著的改进。
{"title":"Geo-location inference from image content and user tags","authors":"Andrew C. Gallagher, D. Joshi, Jie Yu, Jiebo Luo","doi":"10.1109/CVPRW.2009.5204168","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204168","url":null,"abstract":"Associating image content with their geographic locations has been increasingly pursued in the computer vision community in recent years. In a recent work, large collections of geotagged images were found to be helpful in estimating geo-locations of query images by simple visual nearest-neighbors search. In this paper, we leverage user tags along with image content to infer the geo-location. Our model builds upon the fact that the visual content and user tags of pictures can provide significant hints about their geo-locations. Using a large collection of over a million geotagged photographs, we build location probability maps of user tags over the entire globe. These maps reflect the picture-taking and tagging behaviors of thousands of users from all over the world, and reveal interesting tag map patterns. Visual content matching is performed using multiple feature descriptors including tiny images, color histograms, GIST features, and bags of textons. The combination of visual content matching and local tag probability maps forms a strong geo-inference engine. Large-scale experiments have shown significant improvements over pure visual content-based geo-location inference.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132684282","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 66
Cortical enhanced tissue segmentation of neonatal brain MR images acquired by a dedicated phased array coil 由专用相控阵线圈获得的新生儿脑MR图像的皮质增强组织分割
F. Shi, P. Yap, Yong Fan, Jie-Zhi Cheng, L. Wald, G. Gerig, Weili Lin, D. Shen
The acquisition of high quality MR images of neonatal brains is largely hampered by their characteristically small head size and low tissue contrast. As a result, subsequent image processing and analysis, especially for brain tissue segmentation, are often hindered. To overcome this problem, a dedicated phased array neonatal head coil is utilized to improve MR image quality by effectively combing images obtained from 8 coil elements without lengthening data acquisition time. In addition, a subject-specific atlas based tissue segmentation algorithm is specifically developed for the delineation of fine structures in the acquired neonatal brain MR images. The proposed tissue segmentation method first enhances the sheet-like cortical gray matter (GM) structures in neonatal images with a Hessian filter for generation of cortical GM prior. Then, the prior is combined with our neonatal population atlas to form a cortical enhanced hybrid atlas, which we refer to as the subject-specific atlas. Various experiments are conducted to compare the proposed method with manual segmentation results, as well as with additional two population atlas based segmentation methods. Results show that the proposed method is capable of segmenting the neonatal brain with the highest accuracy, compared to other two methods.
新生儿脑部的高质量磁共振图像的获取在很大程度上受到其特征性的小头部尺寸和低组织对比度的阻碍。因此,后续的图像处理和分析,特别是脑组织分割,往往受到阻碍。为了克服这一问题,利用专用的相控阵新生儿头部线圈,在不延长数据采集时间的情况下,有效地将8个线圈单元获得的图像进行组合,从而提高MR图像质量。此外,我们还开发了一种基于主题特异性图谱的组织分割算法,用于描绘获得性新生儿脑MR图像中的精细结构。提出的组织分割方法首先利用Hessian滤波器对新生儿图像中的片状皮质灰质(GM)结构进行增强。然后,先验与我们的新生儿种群图谱相结合,形成皮质增强杂交图谱,我们称之为主题特异性图谱。进行了各种实验,将该方法与人工分割结果以及另外两种基于种群图谱的分割方法进行了比较。结果表明,与其他两种方法相比,该方法能够以最高的准确率对新生儿大脑进行分割。
{"title":"Cortical enhanced tissue segmentation of neonatal brain MR images acquired by a dedicated phased array coil","authors":"F. Shi, P. Yap, Yong Fan, Jie-Zhi Cheng, L. Wald, G. Gerig, Weili Lin, D. Shen","doi":"10.1109/CVPRW.2009.5204348","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204348","url":null,"abstract":"The acquisition of high quality MR images of neonatal brains is largely hampered by their characteristically small head size and low tissue contrast. As a result, subsequent image processing and analysis, especially for brain tissue segmentation, are often hindered. To overcome this problem, a dedicated phased array neonatal head coil is utilized to improve MR image quality by effectively combing images obtained from 8 coil elements without lengthening data acquisition time. In addition, a subject-specific atlas based tissue segmentation algorithm is specifically developed for the delineation of fine structures in the acquired neonatal brain MR images. The proposed tissue segmentation method first enhances the sheet-like cortical gray matter (GM) structures in neonatal images with a Hessian filter for generation of cortical GM prior. Then, the prior is combined with our neonatal population atlas to form a cortical enhanced hybrid atlas, which we refer to as the subject-specific atlas. Various experiments are conducted to compare the proposed method with manual segmentation results, as well as with additional two population atlas based segmentation methods. Results show that the proposed method is capable of segmenting the neonatal brain with the highest accuracy, compared to other two methods.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129251040","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Lonely but attractive: Sparse color salient points for object retrieval and categorization 孤独但有吸引力:稀疏的颜色突出点用于对象检索和分类
Julian Stöttinger, A. Hanbury, T. Gevers, N. Sebe
Local image descriptors computed in areas around salient points in images are essential for many algorithms in computer vision. Recent work suggests using as many salient points as possible. While sophisticated classifiers have been proposed to cope with the resulting large number of descriptors, processing this large amount of data is computationally costly. In this paper, computational methods are proposed to compute salient points designed to allow a reduction in the number of salient points while maintaining state of the art performance in image retrieval and object recognition applications. To obtain a more sparse description, a color salient point and scale determination framework is proposed operating on color spaces that have useful perceptual and saliency properties. This allows for the necessary discriminative points to be located, allowing a significant reduction in the number of salient points and obtaining an invariant (repeatability) and discriminative (distinctiveness) image description. Experimental results on large image datasets show that the proposed method obtains state of the art results with the number of salient points reduced by half. This reduction in the number of points allows subsequent operations, such as feature extraction and clustering, to run more efficiently. It is shown that the method provides less ambiguous features, a more compact description of visual data, and therefore a faster classification of visual data.
在图像显著点周围的区域计算局部图像描述符是计算机视觉中许多算法的基础。最近的研究建议使用尽可能多的要点。虽然已经提出了复杂的分类器来处理由此产生的大量描述符,但处理如此大量的数据在计算上是昂贵的。本文提出了计算突出点的方法,旨在减少突出点的数量,同时在图像检索和对象识别应用中保持最先进的性能。为了获得更稀疏的描述,提出了一个颜色突出点和尺度确定框架,该框架在具有有用的感知和显著性的颜色空间上运行。这样就可以找到必要的判别点,从而显著减少显著点的数量,并获得不变(可重复性)和判别(独特性)的图像描述。在大型图像数据集上的实验结果表明,该方法得到了最先进的结果,显著点的数量减少了一半。点数量的减少使得后续操作,如特征提取和聚类,可以更有效地运行。结果表明,该方法提供了更少的模糊特征,更紧凑的视觉数据描述,从而更快地分类视觉数据。
{"title":"Lonely but attractive: Sparse color salient points for object retrieval and categorization","authors":"Julian Stöttinger, A. Hanbury, T. Gevers, N. Sebe","doi":"10.1109/CVPRW.2009.5204286","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204286","url":null,"abstract":"Local image descriptors computed in areas around salient points in images are essential for many algorithms in computer vision. Recent work suggests using as many salient points as possible. While sophisticated classifiers have been proposed to cope with the resulting large number of descriptors, processing this large amount of data is computationally costly. In this paper, computational methods are proposed to compute salient points designed to allow a reduction in the number of salient points while maintaining state of the art performance in image retrieval and object recognition applications. To obtain a more sparse description, a color salient point and scale determination framework is proposed operating on color spaces that have useful perceptual and saliency properties. This allows for the necessary discriminative points to be located, allowing a significant reduction in the number of salient points and obtaining an invariant (repeatability) and discriminative (distinctiveness) image description. Experimental results on large image datasets show that the proposed method obtains state of the art results with the number of salient points reduced by half. This reduction in the number of points allows subsequent operations, such as feature extraction and clustering, to run more efficiently. It is shown that the method provides less ambiguous features, a more compact description of visual data, and therefore a faster classification of visual data.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121774186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
Dominance detection in face-to-face conversations 面对面对话中的优势侦测
Sergio Escalera, R. M. Martinez, Jordi Vitrià, P. Radeva, M. Anguera
Dominance is referred to the level of influence a person has in a conversation. Dominance is an important research area in social psychology, but the problem of its automatic estimation is a very recent topic in the contexts of social and wearable computing. In this paper, we focus on dominance detection from visual cues. We estimate the correlation among observers by categorizing the dominant people in a set of face-to-face conversations. Different dominance indicators from gestural communication are defined, manually annotated, and compared to the observers opinion. Moreover, the considered indicators are automatically extracted from video sequences and learnt by using binary classifiers. Results from the three analysis shows a high correlation and allows the categorization of dominant people in public discussion video sequences.
支配力指的是一个人在谈话中的影响力。支配地位是社会心理学的一个重要研究领域,但支配地位的自动估计问题是最近在社交和可穿戴计算环境下才出现的一个话题。本文主要研究基于视觉线索的优势检测。我们通过对一组面对面对话中的主导人物进行分类来估计观察者之间的相关性。定义不同的优势指标,手工标注,并与观察者的意见进行比较。此外,从视频序列中自动提取考虑的指标,并使用二值分类器进行学习。这三种分析的结果显示出高度的相关性,并允许对公共讨论视频序列中的主导人物进行分类。
{"title":"Dominance detection in face-to-face conversations","authors":"Sergio Escalera, R. M. Martinez, Jordi Vitrià, P. Radeva, M. Anguera","doi":"10.1109/CVPRW.2009.5204267","DOIUrl":"https://doi.org/10.1109/CVPRW.2009.5204267","url":null,"abstract":"Dominance is referred to the level of influence a person has in a conversation. Dominance is an important research area in social psychology, but the problem of its automatic estimation is a very recent topic in the contexts of social and wearable computing. In this paper, we focus on dominance detection from visual cues. We estimate the correlation among observers by categorizing the dominant people in a set of face-to-face conversations. Different dominance indicators from gestural communication are defined, manually annotated, and compared to the observers opinion. Moreover, the considered indicators are automatically extracted from video sequences and learnt by using binary classifiers. Results from the three analysis shows a high correlation and allows the categorization of dominant people in public discussion video sequences.","PeriodicalId":431981,"journal":{"name":"2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122516710","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1