首页 > 最新文献

2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)最新文献

英文 中文
Recognition of Composite Human Activities through Context-Free Grammar Based Representation 基于上下文无关语法表示的复合人类活动识别
M. Ryoo, J. Aggarwal
This paper describes a general methodology for automated recognition of complex human activities. The methodology uses a context-free grammar (CFG) based representation scheme to represent composite actions and interactions. The CFG-based representation enables us to formally define complex human activities based on simple actions or movements. Human activities are classified into three categories: atomic action, composite action, and interaction. Our system is not only able to represent complex human activities formally, but also able to recognize represented actions and interactions with high accuracy. Image sequences are processed to extract poses and gestures. Based on gestures, the system detects actions and interactions occurring in a sequence of image frames. Our results show that the system is able to represent composite actions and interactions naturally. The system was tested to represent and recognize eight types of interactions: approach, depart, point, shake-hands, hug, punch, kick, and push. The experiments show that the system can recognize sequences of represented composite actions and interactions with a high recognition rate.
本文描述了复杂人类活动自动识别的一般方法。该方法使用基于上下文无关语法(CFG)的表示方案来表示复合动作和交互。基于cfg的表示使我们能够基于简单的动作或运动正式定义复杂的人类活动。人类活动分为三大类:原子作用、复合作用和相互作用。我们的系统不仅能够形式化地表示复杂的人类活动,而且能够高精度地识别所表示的动作和交互。处理图像序列以提取姿势和手势。基于手势,系统检测一系列图像帧中发生的动作和交互。结果表明,该系统能够自然地表示复合动作和交互。经过测试,该系统可以表示和识别八种类型的交互:接近、离开、指向、握手、拥抱、打拳、踢脚和推。实验结果表明,该系统能够以较高的识别率识别复合动作和交互序列。
{"title":"Recognition of Composite Human Activities through Context-Free Grammar Based Representation","authors":"M. Ryoo, J. Aggarwal","doi":"10.1109/CVPR.2006.242","DOIUrl":"https://doi.org/10.1109/CVPR.2006.242","url":null,"abstract":"This paper describes a general methodology for automated recognition of complex human activities. The methodology uses a context-free grammar (CFG) based representation scheme to represent composite actions and interactions. The CFG-based representation enables us to formally define complex human activities based on simple actions or movements. Human activities are classified into three categories: atomic action, composite action, and interaction. Our system is not only able to represent complex human activities formally, but also able to recognize represented actions and interactions with high accuracy. Image sequences are processed to extract poses and gestures. Based on gestures, the system detects actions and interactions occurring in a sequence of image frames. Our results show that the system is able to represent composite actions and interactions naturally. The system was tested to represent and recognize eight types of interactions: approach, depart, point, shake-hands, hug, punch, kick, and push. The experiments show that the system can recognize sequences of represented composite actions and interactions with a high recognition rate.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124417063","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 288
Robust AAM Fitting by Fusion of Images and Disparity Data 基于图像和视差数据融合的鲁棒AAM拟合
Joerg Liebelt, Jing Xiao, Jie Yang
Active Appearance Models (AAMs) have been popularly used to represent the appearance and shape variations of human faces. Fitting an AAM to images recovers the face pose as well as its deformable shape and varying appearance. Successful fitting requires that the AAM is sufficiently generic such that it covers all possible facial appearances and shapes in the images. Such a generic AAM is often difficult to be obtained in practice, especially when the image quality is low or when occlusion occurs. To achieve robust AAM fitting under such circumstances, this paper proposes to incorporate the disparity data obtained from a stereo camera with the image fitting process. We develop an iterative multi-level algorithm that combines efficient AAM fitting to 2D images and robust 3D shape alignment to disparity data. Experiments on tracking faces in low-resolution images captured from meeting scenarios show that the proposed method achieves better performance than the original 2D AAM fitting algorithm. We also demonstrate an application of the proposed method to a facial expression recognition task.
主动外观模型(aam)已被广泛用于表示人脸的外观和形状变化。将AAM拟合到图像中可以恢复人脸姿态及其可变形的形状和变化的外观。成功的拟合要求AAM具有足够的通用性,以覆盖图像中所有可能的面部外观和形状。这种通用AAM在实践中往往难以获得,特别是在图像质量较低或出现遮挡的情况下。为了在这种情况下实现鲁棒的AAM拟合,本文提出将立体相机获取的视差数据与图像拟合过程相结合。我们开发了一种迭代多级算法,该算法结合了对2D图像的高效AAM拟合和对视差数据的鲁棒3D形状对齐。对会议场景低分辨率图像的人脸跟踪实验表明,该方法比原有的二维AAM拟合算法具有更好的性能。我们还演示了该方法在面部表情识别任务中的应用。
{"title":"Robust AAM Fitting by Fusion of Images and Disparity Data","authors":"Joerg Liebelt, Jing Xiao, Jie Yang","doi":"10.1109/CVPR.2006.255","DOIUrl":"https://doi.org/10.1109/CVPR.2006.255","url":null,"abstract":"Active Appearance Models (AAMs) have been popularly used to represent the appearance and shape variations of human faces. Fitting an AAM to images recovers the face pose as well as its deformable shape and varying appearance. Successful fitting requires that the AAM is sufficiently generic such that it covers all possible facial appearances and shapes in the images. Such a generic AAM is often difficult to be obtained in practice, especially when the image quality is low or when occlusion occurs. To achieve robust AAM fitting under such circumstances, this paper proposes to incorporate the disparity data obtained from a stereo camera with the image fitting process. We develop an iterative multi-level algorithm that combines efficient AAM fitting to 2D images and robust 3D shape alignment to disparity data. Experiments on tracking faces in low-resolution images captured from meeting scenarios show that the proposed method achieves better performance than the original 2D AAM fitting algorithm. We also demonstrate an application of the proposed method to a facial expression recognition task.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114940186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Fast Variational Segmentation using Partial Extremal Initialization 使用部分极值初始化的快速变分分割
J. E. Solem, N. C. Overgaard, Markus Persson, A. Heyden
In this paper we consider region-based variational segmentation of two- and three-dimensional images by the minimization of functionals whose fidelity term is the quotient of two integrals. Users often refrain from quotient functionals, even when they seem to be the most natural choice, probably because the corresponding gradient descent PDEs are nonlocal and hence require the computation of global properties. Here it is shown how this problem may be overcome by employing the structure of the Euler-Lagrange equation of the fidelity term to construct a good initialization for the gradient descent PDE, which will then converge rapidly to the desired (local) minimum. The initializer is found by making a one-dimensional search among the level sets of a function related to the fidelity term, picking the level set which minimizes the segmentation functional. This partial extremal initialization is tested on a medical segmentation problem with velocity- and intensity data from MR images. In this particular application, the partial extremal initialization speeds up the segmentation by two orders of magnitude compared to straight forward gradient descent.
本文研究了一种基于区域的二维和三维图像的变分分割方法,其保真度项为两个积分的商。用户经常避免使用商函数,即使它们看起来是最自然的选择,这可能是因为相应的梯度下降偏微分方程是非局部的,因此需要计算全局属性。这里展示了如何通过使用保真度项的欧拉-拉格朗日方程的结构来构造梯度下降PDE的良好初始化来克服这个问题,然后该初始化将迅速收敛到所需的(局部)最小值。初始化器是通过在与保真度项相关的函数的水平集中进行一维搜索来找到的,选择最小化分割函数的水平集。在MR图像的速度和强度数据的医学分割问题上,对这种部分极值初始化进行了测试。在这个特殊的应用程序中,与直接梯度下降相比,部分极值初始化将分割速度提高了两个数量级。
{"title":"Fast Variational Segmentation using Partial Extremal Initialization","authors":"J. E. Solem, N. C. Overgaard, Markus Persson, A. Heyden","doi":"10.1109/CVPR.2006.120","DOIUrl":"https://doi.org/10.1109/CVPR.2006.120","url":null,"abstract":"In this paper we consider region-based variational segmentation of two- and three-dimensional images by the minimization of functionals whose fidelity term is the quotient of two integrals. Users often refrain from quotient functionals, even when they seem to be the most natural choice, probably because the corresponding gradient descent PDEs are nonlocal and hence require the computation of global properties. Here it is shown how this problem may be overcome by employing the structure of the Euler-Lagrange equation of the fidelity term to construct a good initialization for the gradient descent PDE, which will then converge rapidly to the desired (local) minimum. The initializer is found by making a one-dimensional search among the level sets of a function related to the fidelity term, picking the level set which minimizes the segmentation functional. This partial extremal initialization is tested on a medical segmentation problem with velocity- and intensity data from MR images. In this particular application, the partial extremal initialization speeds up the segmentation by two orders of magnitude compared to straight forward gradient descent.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131021735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Multiscale Nonlinear Diffusion and Shock Filter for Ultrasound Image Enhancement 超声图像增强的多尺度非线性扩散和冲击滤波
Fan Zhang, Y. Yoo, Yongmin Kim, Lichen Zhang, L. M. Koh
A new noise reduction and edge enhancement method, i.e., Laplacian pyramid-based nonlinear diffusion and shock filter (LPNDSF), is proposed for medical ultrasound imaging. In the proposed LPNDSF, a coupled nonlinear diffusion and shock filter process is applied in Laplacian pyramid domain of an image, to remove speckle and enhance edges simultaneously. The performance of the proposed method was evaluated on a phantom and a real ultrasound image. In the phantom study, we obtained an average gain of 0.55 and 1.11 in contrast-to-noise ratio compared to the speckle reducing anisotropic diffusion (SRAD) and nonlinear coherent diffusion (NCD), respectively. Also, the proposed LPNDSF showed clearer boundaries on the phantom and the real ultrasound image. These preliminary results indicate that the proposed LPNDSF can effectively reduce speckle noise while enhancing image edges for retaining subtle features.
提出了一种新的医学超声成像降噪和边缘增强方法,即基于拉普拉斯金字塔的非线性扩散和冲击滤波(LPNDSF)。在该算法中,在图像的拉普拉斯金字塔域采用非线性扩散和冲击耦合滤波处理,同时去除斑点和增强边缘。在仿真和真实超声图像上对该方法的性能进行了评价。在模体研究中,与散斑减少各向异性扩散(SRAD)和非线性相干扩散(NCD)相比,我们分别获得了0.55和1.11的平均增益。此外,所提出的LPNDSF在虚影和真实超声图像上的边界更清晰。这些初步结果表明,所提出的LPNDSF可以有效地降低散斑噪声,同时增强图像边缘以保留细微特征。
{"title":"Multiscale Nonlinear Diffusion and Shock Filter for Ultrasound Image Enhancement","authors":"Fan Zhang, Y. Yoo, Yongmin Kim, Lichen Zhang, L. M. Koh","doi":"10.1109/CVPR.2006.203","DOIUrl":"https://doi.org/10.1109/CVPR.2006.203","url":null,"abstract":"A new noise reduction and edge enhancement method, i.e., Laplacian pyramid-based nonlinear diffusion and shock filter (LPNDSF), is proposed for medical ultrasound imaging. In the proposed LPNDSF, a coupled nonlinear diffusion and shock filter process is applied in Laplacian pyramid domain of an image, to remove speckle and enhance edges simultaneously. The performance of the proposed method was evaluated on a phantom and a real ultrasound image. In the phantom study, we obtained an average gain of 0.55 and 1.11 in contrast-to-noise ratio compared to the speckle reducing anisotropic diffusion (SRAD) and nonlinear coherent diffusion (NCD), respectively. Also, the proposed LPNDSF showed clearer boundaries on the phantom and the real ultrasound image. These preliminary results indicate that the proposed LPNDSF can effectively reduce speckle noise while enhancing image edges for retaining subtle features.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128414186","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A General Framework and New Alignment Criterion for Dense Optical Flow 密集光流的一般框架和新的对准准则
Rami Ben-Ari, N. Sochen
The problem of dense optical flow computation is addressed from a variational viewpoint. A new geometric framework is introduced. It unifies previous art and yields new efficient methods. Along with the framework a new alignment criterion suggests itself. It is shown that the alignment between the gradients of the optical flow components and between the latter and the intensity gradients is an important measure of the flow’s quality. Adding this criterion as a requirement in the optimization process improves the resulting flow. This is demonstrated in synthetic and real sequences.
从变分的角度讨论了密集光流的计算问题。提出了一种新的几何框架。它统一了以前的技术,并产生了新的有效方法。与框架一起出现的是一个新的对齐标准。结果表明,光流分量梯度之间以及光流分量梯度与光流强度梯度之间的对准度是衡量光流质量的重要指标。在优化过程中将此标准作为需求添加,可以改进结果流。这在合成序列和真实序列中得到了证明。
{"title":"A General Framework and New Alignment Criterion for Dense Optical Flow","authors":"Rami Ben-Ari, N. Sochen","doi":"10.1109/CVPR.2006.25","DOIUrl":"https://doi.org/10.1109/CVPR.2006.25","url":null,"abstract":"The problem of dense optical flow computation is addressed from a variational viewpoint. A new geometric framework is introduced. It unifies previous art and yields new efficient methods. Along with the framework a new alignment criterion suggests itself. It is shown that the alignment between the gradients of the optical flow components and between the latter and the intensity gradients is an important measure of the flow’s quality. Adding this criterion as a requirement in the optimization process improves the resulting flow. This is demonstrated in synthetic and real sequences.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"110 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132161712","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Hierarchical Procrustes Matching for Shape Retrieval 用于形状检索的分层procruster匹配
Graham Mcneill, S. Vijayakumar
We introduce Hierarchical Procrustes Matching (HPM), a segment-based shape matching algorithm which avoids problems associated with purely global or local methods and performs well on benchmark shape retrieval tests. The simplicity of the shape representation leads to a powerful matching algorithm which incorporates intuitive ideas about the perceptual nature of shape while being computationally efficient. This includes the ability to match similar parts even when they occur at different scales or positions. While comparison of multiscale shape representations is typically based on specific features, HPM avoids the need to extract such features. The hierarchical structure of the algorithm captures the appealing notion that matching should proceed in a global to local direction.
本文介绍了一种基于分段的形状匹配算法——分层Procrustes匹配(HPM),该算法避免了纯全局或局部方法的相关问题,并在基准形状检索测试中表现良好。形状表示的简单性导致了一种强大的匹配算法,该算法结合了关于形状感知本质的直观想法,同时具有计算效率。这包括匹配相似部分的能力,即使它们出现在不同的尺度或位置。虽然多尺度形状表示的比较通常是基于特定的特征,但HPM避免了提取这些特征的需要。该算法的层次结构抓住了一个吸引人的概念,即匹配应该从全局到局部方向进行。
{"title":"Hierarchical Procrustes Matching for Shape Retrieval","authors":"Graham Mcneill, S. Vijayakumar","doi":"10.1109/CVPR.2006.133","DOIUrl":"https://doi.org/10.1109/CVPR.2006.133","url":null,"abstract":"We introduce Hierarchical Procrustes Matching (HPM), a segment-based shape matching algorithm which avoids problems associated with purely global or local methods and performs well on benchmark shape retrieval tests. The simplicity of the shape representation leads to a powerful matching algorithm which incorporates intuitive ideas about the perceptual nature of shape while being computationally efficient. This includes the ability to match similar parts even when they occur at different scales or positions. While comparison of multiscale shape representations is typically based on specific features, HPM avoids the need to extract such features. The hierarchical structure of the algorithm captures the appealing notion that matching should proceed in a global to local direction.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121415418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 146
An Adaptive Appearance Model Approach for Model-based Articulated Object Tracking 基于模型的关节目标跟踪的自适应外观模型方法
A. O. Balan, Michael J. Black
The detection and tracking of three-dimensional human body models has progressed rapidly but successful approaches typically rely on accurate foreground silhouettes obtained using background segmentation. There are many practical applications where such information is imprecise. Here we develop a new image likelihood function based on the visual appearance of the subject being tracked. We propose a robust, adaptive, appearance model based on the Wandering-Stable-Lost framework extended to the case of articulated body parts. The method models appearance using a mixture model that includes an adaptive template, frame-to-frame matching and an outlier process. We employ an annealed particle filtering algorithm for inference and take advantage of the 3D body model to predict selfocclusion and improve pose estimation accuracy. Quantitative tracking results are presented for a walking sequence with a 180 degree turn, captured with four synchronized and calibrated cameras and containing significant appearance changes and self-occlusion in each view.
三维人体模型的检测和跟踪进展迅速,但成功的方法通常依赖于通过背景分割获得准确的前景轮廓。在许多实际应用中,这些信息是不精确的。在这里,我们基于被跟踪对象的视觉外观开发了一个新的图像似然函数。我们提出了一个鲁棒的,自适应的,基于流浪-稳定-丢失框架的外观模型,扩展到铰接的身体部位。该方法使用混合模型对外观进行建模,该混合模型包括自适应模板、帧对帧匹配和离群值处理。我们采用退火粒子滤波算法进行推理,并利用三维身体模型来预测自聚焦,提高姿态估计精度。定量跟踪结果展示了一个180度转弯的行走序列,由四个同步和校准的相机捕获,每个视图中包含显著的外观变化和自遮挡。
{"title":"An Adaptive Appearance Model Approach for Model-based Articulated Object Tracking","authors":"A. O. Balan, Michael J. Black","doi":"10.1109/CVPR.2006.52","DOIUrl":"https://doi.org/10.1109/CVPR.2006.52","url":null,"abstract":"The detection and tracking of three-dimensional human body models has progressed rapidly but successful approaches typically rely on accurate foreground silhouettes obtained using background segmentation. There are many practical applications where such information is imprecise. Here we develop a new image likelihood function based on the visual appearance of the subject being tracked. We propose a robust, adaptive, appearance model based on the Wandering-Stable-Lost framework extended to the case of articulated body parts. The method models appearance using a mixture model that includes an adaptive template, frame-to-frame matching and an outlier process. We employ an annealed particle filtering algorithm for inference and take advantage of the 3D body model to predict selfocclusion and improve pose estimation accuracy. Quantitative tracking results are presented for a walking sequence with a 180 degree turn, captured with four synchronized and calibrated cameras and containing significant appearance changes and self-occlusion in each view.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128747896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 93
Scale Variant Image Pyramids 变型图像金字塔
J. Gluckman
Multi-scale representations are motivated by the scale invariant properties of natural images. While many low level statistical measures, such as the local mean and variance of intensity, behave in a scale invariant manner, there are many higher order deviations from scale invariance where zero-crossings merge and disappear. Such scale variant behavior is important information to represent because it is not easily predicted from lower resolution data. A scale variant image pyramid is a representation that separates this information from the more redundant and predictable scale invariant information.
多尺度表示的动机是自然图像的尺度不变性。虽然许多低水平的统计度量,如强度的局部平均值和方差,表现为尺度不变性,但有许多高阶偏离尺度不变性,其中零交叉合并并消失。这种尺度变化的行为是重要的信息,因为它不容易从低分辨率的数据预测。尺度变化图像金字塔是一种表示,它将这些信息与更冗余和可预测的尺度不变信息分开。
{"title":"Scale Variant Image Pyramids","authors":"J. Gluckman","doi":"10.1109/CVPR.2006.265","DOIUrl":"https://doi.org/10.1109/CVPR.2006.265","url":null,"abstract":"Multi-scale representations are motivated by the scale invariant properties of natural images. While many low level statistical measures, such as the local mean and variance of intensity, behave in a scale invariant manner, there are many higher order deviations from scale invariance where zero-crossings merge and disappear. Such scale variant behavior is important information to represent because it is not easily predicted from lower resolution data. A scale variant image pyramid is a representation that separates this information from the more redundant and predictable scale invariant information.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114287903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
An Integrated Segmentation and Classification Approach Applied to Multiple Sclerosis Analysis 综合分割分类方法在多发性硬化症分析中的应用
A. Akselrod-Ballin, M. Galun, R. Basri, A. Brandt, M. Gomori, M. Filippi, P. Valsasina
We present a novel multiscale approach that combines segmentation with classification to detect abnormal brain structures in medical imagery, and demonstrate its utility in detecting multiple sclerosis lesions in 3D MRI data. Our method uses segmentation to obtain a hierarchical decomposition of a multi-channel, anisotropic MRI scan. It then produces a rich set of features describing the segments in terms of intensity, shape, location, and neighborhood relations. These features are then fed into a decision tree-based classifier, trained with data labeled by experts, enabling the detection of lesions in all scales. Unlike common approaches that use voxel-by-voxel analysis, our system can utilize regional properties that are often important for characterizing abnormal brain structures. We provide experiments showing successful detections of lesions in both simulated and real MR images.
我们提出了一种新的多尺度方法,结合分割和分类来检测医学图像中的异常大脑结构,并展示了其在3D MRI数据中检测多发性硬化症病变的实用性。我们的方法使用分割来获得多通道各向异性MRI扫描的分层分解。然后根据强度、形状、位置和邻里关系,生成一组丰富的特征来描述这些片段。然后将这些特征输入到基于决策树的分类器中,使用专家标记的数据进行训练,从而能够在所有尺度上检测病变。与使用逐体素分析的常见方法不同,我们的系统可以利用通常对表征异常大脑结构很重要的区域属性。我们提供了在模拟和真实的MR图像中成功检测病变的实验。
{"title":"An Integrated Segmentation and Classification Approach Applied to Multiple Sclerosis Analysis","authors":"A. Akselrod-Ballin, M. Galun, R. Basri, A. Brandt, M. Gomori, M. Filippi, P. Valsasina","doi":"10.1109/CVPR.2006.55","DOIUrl":"https://doi.org/10.1109/CVPR.2006.55","url":null,"abstract":"We present a novel multiscale approach that combines segmentation with classification to detect abnormal brain structures in medical imagery, and demonstrate its utility in detecting multiple sclerosis lesions in 3D MRI data. Our method uses segmentation to obtain a hierarchical decomposition of a multi-channel, anisotropic MRI scan. It then produces a rich set of features describing the segments in terms of intensity, shape, location, and neighborhood relations. These features are then fed into a decision tree-based classifier, trained with data labeled by experts, enabling the detection of lesions in all scales. Unlike common approaches that use voxel-by-voxel analysis, our system can utilize regional properties that are often important for characterizing abnormal brain structures. We provide experiments showing successful detections of lesions in both simulated and real MR images.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114450607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Human Carrying Status in Visual Surveillance 视觉监测中的人体携带状态
D. Tao, Xuelong Li, S. Maybank, Xindong Wu
A person’s gait changes when he or she is carrying an object such as a bag, suitcase or rucksack. As a result, human identification and tracking are made more difficult because the averaged gait image is too simple to represent the carrying status. Therefore, in this paper we first introduce a set of Gabor based human gait appearance models, because Gabor functions are similar to the receptive field profiles in the mammalian cortical simple cells. The very high dimensionality of the feature space makes training difficult. In order to solve this problem we propose a general tensor discriminant analysis (GTDA), which seamlessly incorporates the object (Gabor based human gait appearance model) structure information as a natural constraint. GTDA differs from the previous tensor based discriminant analysis methods in that the training converges. Existing methods fail to converge in the training stage. This makes them unsuitable for practical tasks. Experiments are carried out on the USF baseline data set to recognize a human’s ID from the gait silhouette. The proposed Gabor gait incorporated with GTDA is demonstrated to significantly outperform the existing appearance-based methods.
当一个人背着包、手提箱或帆布背包等物品时,他或她的步态会发生变化。结果,由于平均步态图像过于简单而无法表示携带状态,给人体识别和跟踪增加了难度。因此,在本文中,我们首先引入了一套基于Gabor的人类步态外观模型,因为Gabor功能类似于哺乳动物皮层简单细胞的感受野轮廓。特征空间的高维使得训练变得困难。为了解决这一问题,我们提出了一种通用张量判别分析(GTDA),它无缝地将物体(基于Gabor的人类步态外观模型)的结构信息作为自然约束。GTDA与以往基于张量的判别分析方法的不同之处在于训练是收敛的。现有的方法在训练阶段无法收敛。这使得它们不适合执行实际任务。在USF基线数据集上进行实验,从步态轮廓中识别人的身份。结合GTDA的Gabor步态被证明明显优于现有的基于外观的方法。
{"title":"Human Carrying Status in Visual Surveillance","authors":"D. Tao, Xuelong Li, S. Maybank, Xindong Wu","doi":"10.1109/CVPR.2006.138","DOIUrl":"https://doi.org/10.1109/CVPR.2006.138","url":null,"abstract":"A person’s gait changes when he or she is carrying an object such as a bag, suitcase or rucksack. As a result, human identification and tracking are made more difficult because the averaged gait image is too simple to represent the carrying status. Therefore, in this paper we first introduce a set of Gabor based human gait appearance models, because Gabor functions are similar to the receptive field profiles in the mammalian cortical simple cells. The very high dimensionality of the feature space makes training difficult. In order to solve this problem we propose a general tensor discriminant analysis (GTDA), which seamlessly incorporates the object (Gabor based human gait appearance model) structure information as a natural constraint. GTDA differs from the previous tensor based discriminant analysis methods in that the training converges. Existing methods fail to converge in the training stage. This makes them unsuitable for practical tasks. Experiments are carried out on the USF baseline data set to recognize a human’s ID from the gait silhouette. The proposed Gabor gait incorporated with GTDA is demonstrated to significantly outperform the existing appearance-based methods.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125436402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 122
期刊
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1