首页 > 最新文献

2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)最新文献

英文 中文
Intelligent Collaborative Tracking by Mining Auxiliary Objects 基于辅助对象挖掘的智能协同跟踪
Ming Yang, Ying Wu, S. Lao
Many tracking methods face a fundamental dilemma in practice: tracking has to be computationally efficient but verifying if or not the tracker is following the true target tends to be demanding, especially when the background is cluttered and/or when occlusion occurs. Due to the lack of a good solution to this problem, many existing methods tend to be either computationally intensive with the use of sophisticated image observation models, or vulnerable to the false alarms. This greatly threatens long-duration robust tracking. This paper presents a novel solution to this dilemma by integrating into the tracking process a set of auxiliary objects that are automatically discovered in the video on the fly by data mining. Auxiliary objects have three properties at least in a short time interval: (1) persistent co-occurrence with the target; (2) consistent motion correlation with the target; and (3) easy to track. The collaborative tracking of these auxiliary objects leads to an efficient computation as well as a strong verification. Our extensive experiments have exhibited exciting performance in very challenging real-world testing cases.
许多跟踪方法在实践中面临着一个基本的困境:跟踪必须具有计算效率,但验证跟踪器是否跟踪真实目标往往是苛刻的,特别是当背景混乱和/或遮挡发生时。由于缺乏很好的解决方案,现有的许多方法要么使用复杂的图像观测模型计算量大,要么容易产生误报。这极大地威胁到长时间的稳健跟踪。本文提出了一种新颖的解决方案,即在跟踪过程中集成一组通过数据挖掘在视频中动态自动发现的辅助对象。辅助对象至少在短时间间隔内具有三个属性:(1)与目标持续共现;(2)与目标运动相关性一致;(3)易于跟踪。这些辅助目标的协同跟踪使得计算效率高,验证能力强。我们广泛的实验在非常具有挑战性的现实世界测试案例中展示了令人兴奋的性能。
{"title":"Intelligent Collaborative Tracking by Mining Auxiliary Objects","authors":"Ming Yang, Ying Wu, S. Lao","doi":"10.1109/CVPR.2006.157","DOIUrl":"https://doi.org/10.1109/CVPR.2006.157","url":null,"abstract":"Many tracking methods face a fundamental dilemma in practice: tracking has to be computationally efficient but verifying if or not the tracker is following the true target tends to be demanding, especially when the background is cluttered and/or when occlusion occurs. Due to the lack of a good solution to this problem, many existing methods tend to be either computationally intensive with the use of sophisticated image observation models, or vulnerable to the false alarms. This greatly threatens long-duration robust tracking. This paper presents a novel solution to this dilemma by integrating into the tracking process a set of auxiliary objects that are automatically discovered in the video on the fly by data mining. Auxiliary objects have three properties at least in a short time interval: (1) persistent co-occurrence with the target; (2) consistent motion correlation with the target; and (3) easy to track. The collaborative tracking of these auxiliary objects leads to an efficient computation as well as a strong verification. Our extensive experiments have exhibited exciting performance in very challenging real-world testing cases.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"251 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114246516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Structure and View Estimation for Tomographic Reconstruction: A Bayesian Approach 层析成像重建的结构和视图估计:贝叶斯方法
S. P. Mallick, Sameer Agarwal, D. Kriegman, Serge J. Belongie, B. Carragher, C. Potter
This paper addresses the problem of reconstructing the density of a scene from multiple projection images produced by modalities such as x-ray, electron microscopy, etc. where an image value is related to the integral of the scene density along a 3D line segment between a radiation source and a point on the image plane. While computed tomography (CT) addresses this problem when the absolute orientation of the image plane and radiation source directions are known, this paper addresses the problem when the orientations are unknown - it is akin to the structure-from-motion (SFM) problem when the extrinsic camera parameters are unknown. We study the problem within the context of reconstructing the density of protein macro-molecules in Cryogenic Electron Microscopy (cryo-EM), where images are very noisy and existing techniques use several thousands of images. In a non-degenerate configuration, the viewing planes corresponding to two projections, intersect in a line in 3D. Using the geometry of the imaging setup, it is possible to determine the projections of this 3D line on the two image planes. In turn, the problem can be formulated as a type of orthographic structure from motion from line correspondences where the line correspondences between two views are unreliable due to image noise. We formulate the task as the problem of denoising a correspondence matrix and present a Bayesian solution to it. Subsequently, the absolute orientation of each projection is determined followed by density reconstruction. We show results on cryo-EM images of proteins and compare our results to that of Electron Micrograph Analysis (EMAN) - a widely used reconstruction tool in cryo-EM.
本文解决了从x射线,电子显微镜等方式产生的多个投影图像中重建场景密度的问题,其中图像值与场景密度沿辐射源和图像平面上的点之间的三维线段的积分有关。虽然计算机断层扫描(CT)解决了当图像平面的绝对方向和辐射源方向已知时的问题,但本文解决了当方向未知时的问题-它类似于外部相机参数未知时的运动结构(SFM)问题。我们在低温电子显微镜(cryo-EM)中重建蛋白质大分子密度的背景下研究了这个问题,其中图像非常嘈杂,现有技术使用数千张图像。在非简并构型中,对应于两个投影的观察平面在三维中相交于一条直线。利用成像设置的几何形状,可以确定这条3D线在两个图像平面上的投影。反过来,这个问题可以被表述为一种来自直线对应的运动的正射影结构,其中两个视图之间的直线对应由于图像噪声而不可靠。我们将该任务表述为对对应矩阵去噪的问题,并给出了一个贝叶斯解。然后,确定每个投影的绝对方向,然后进行密度重建。我们展示了蛋白质的冷冻电镜图像的结果,并将我们的结果与电子显微图分析(EMAN)的结果进行了比较-电子显微图分析是冷冻电镜中广泛使用的重建工具。
{"title":"Structure and View Estimation for Tomographic Reconstruction: A Bayesian Approach","authors":"S. P. Mallick, Sameer Agarwal, D. Kriegman, Serge J. Belongie, B. Carragher, C. Potter","doi":"10.1109/CVPR.2006.295","DOIUrl":"https://doi.org/10.1109/CVPR.2006.295","url":null,"abstract":"This paper addresses the problem of reconstructing the density of a scene from multiple projection images produced by modalities such as x-ray, electron microscopy, etc. where an image value is related to the integral of the scene density along a 3D line segment between a radiation source and a point on the image plane. While computed tomography (CT) addresses this problem when the absolute orientation of the image plane and radiation source directions are known, this paper addresses the problem when the orientations are unknown - it is akin to the structure-from-motion (SFM) problem when the extrinsic camera parameters are unknown. We study the problem within the context of reconstructing the density of protein macro-molecules in Cryogenic Electron Microscopy (cryo-EM), where images are very noisy and existing techniques use several thousands of images. In a non-degenerate configuration, the viewing planes corresponding to two projections, intersect in a line in 3D. Using the geometry of the imaging setup, it is possible to determine the projections of this 3D line on the two image planes. In turn, the problem can be formulated as a type of orthographic structure from motion from line correspondences where the line correspondences between two views are unreliable due to image noise. We formulate the task as the problem of denoising a correspondence matrix and present a Bayesian solution to it. Subsequently, the absolute orientation of each projection is determined followed by density reconstruction. We show results on cryo-EM images of proteins and compare our results to that of Electron Micrograph Analysis (EMAN) - a widely used reconstruction tool in cryo-EM.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"94 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114601255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Classifying Human Dynamics Without Contact Forces 没有接触力的人体动力学分类
A. Bissacco, Stefano Soatto
We develop a classification algorithm for hybrid autoregressive models of human motion for the purpose of videobased analysis and recognition. We assume that some temporal statistics are extracted from the images, and we use them to infer a dynamical system that explicitly models contact forces. We then develop a distance between such models that explicitly factors out exogenous inputs that are not unique to an individual or her gait. We show that such a distance is more discriminative than the distance between simple linear systems, where most of the energy is devoted to modeling the dynamics of spurious nuisances such as contact forces.
我们开发了一种用于基于视频的分析和识别的混合自回归人体运动模型的分类算法。我们假设从图像中提取了一些时间统计数据,并使用它们来推断一个明确建模接触力的动力系统。然后,我们在这些模型之间建立了一个距离,这些模型明确地排除了外生输入,这些输入不是个体或其步态所独有的。我们表明,这样的距离比简单线性系统之间的距离更具判别性,在简单线性系统中,大部分能量用于模拟虚假干扰(如接触力)的动力学。
{"title":"Classifying Human Dynamics Without Contact Forces","authors":"A. Bissacco, Stefano Soatto","doi":"10.1109/CVPR.2006.75","DOIUrl":"https://doi.org/10.1109/CVPR.2006.75","url":null,"abstract":"We develop a classification algorithm for hybrid autoregressive models of human motion for the purpose of videobased analysis and recognition. We assume that some temporal statistics are extracted from the images, and we use them to infer a dynamical system that explicitly models contact forces. We then develop a distance between such models that explicitly factors out exogenous inputs that are not unique to an individual or her gait. We show that such a distance is more discriminative than the distance between simple linear systems, where most of the energy is devoted to modeling the dynamics of spurious nuisances such as contact forces.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122716220","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Accurate Tracking of Monotonically Advancing Fronts 单调推进锋面的精确跟踪
M. Hassouna, A. Farag
A wide range of computer vision applications such as distance field computation, shape from shading, and shape representation require an accurate solution of a particular Hamilton-Jacobi (HJ) equation, known as the Eikonal equation. Although the fast marching method (FMM) is the most stable and consistent method among existing techniques for solving such equation, it suffers from large numerical error along diagonal directions as well as its computational complexity is not optimal. In this paper, we propose an improved version of the FMMthat is both highly accurate and computationally efficient for Cartesian domains. The new method is called the multi-stencils fast marching (MSFM), which computes the solution at each grid point by solving the Eikonal equation along several stencils and then picks the solution that satisfies the fast marching causality relationship. The stencils are centered at each grid point x and cover its entire nearest neighbors. In 2D space, 2 stencils cover the 8-neighbors of x, while in 3D space, 6 stencils cover its 26-neighbors. For those stencils that are not aligned with the natural coordinate system, the Eikonal equation is derived using directional derivatives and then solved using a higher order finite difference scheme.
广泛的计算机视觉应用,如距离场计算,阴影形状和形状表示,需要一个特定的Hamilton-Jacobi (HJ)方程的精确解,称为Eikonal方程。快速推进法(FMM)是目前求解该类方程的最稳定、最一致的方法,但其在对角线方向上存在较大的数值误差,且计算复杂度不是最优的。在本文中,我们提出了一个改进的fmm版本,它在笛卡尔域上具有很高的精度和计算效率。这种新方法被称为多模板快速推进(MSFM),它通过沿多个模板求解Eikonal方程来计算每个网格点的解,然后选择满足快速推进因果关系的解。模板以每个网格点x为中心,并覆盖其整个最近的邻居。在二维空间中,2个模板覆盖x的8个邻居,而在三维空间中,6个模板覆盖x的26个邻居。对于不与自然坐标系对齐的模板,先用方向导数推导出Eikonal方程,然后用高阶有限差分格式求解。
{"title":"Accurate Tracking of Monotonically Advancing Fronts","authors":"M. Hassouna, A. Farag","doi":"10.1109/CVPR.2006.46","DOIUrl":"https://doi.org/10.1109/CVPR.2006.46","url":null,"abstract":"A wide range of computer vision applications such as distance field computation, shape from shading, and shape representation require an accurate solution of a particular Hamilton-Jacobi (HJ) equation, known as the Eikonal equation. Although the fast marching method (FMM) is the most stable and consistent method among existing techniques for solving such equation, it suffers from large numerical error along diagonal directions as well as its computational complexity is not optimal. In this paper, we propose an improved version of the FMMthat is both highly accurate and computationally efficient for Cartesian domains. The new method is called the multi-stencils fast marching (MSFM), which computes the solution at each grid point by solving the Eikonal equation along several stencils and then picks the solution that satisfies the fast marching causality relationship. The stencils are centered at each grid point x and cover its entire nearest neighbors. In 2D space, 2 stencils cover the 8-neighbors of x, while in 3D space, 6 stencils cover its 26-neighbors. For those stencils that are not aligned with the natural coordinate system, the Eikonal equation is derived using directional derivatives and then solved using a higher order finite difference scheme.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126214536","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Multi-Target Tracking - Linking Identities using Bayesian Network Inference 多目标跟踪-使用贝叶斯网络推理链接身份
Peter Nillius, Josephine Sullivan, S. Carlsson
Multi-target tracking requires locating the targets and labeling their identities. The latter is a challenge when many targets, with indistinct appearances, frequently occlude one another, as in football and surveillance tracking. We present an approach to solving this labeling problem. When isolated, a target can be tracked and its identity maintained. While, if targets interact this is not always the case. This paper assumes a track graph exists, denoting when targets are isolated and describing how they interact. Measures of similarity between isolated tracks are defined. The goal is to associate the identities of the isolated tracks, by exploiting the graph constraints and similarity measures. We formulate this as a Bayesian network inference problem, allowing us to use standard message propagation to find the most probable set of paths in an efficient way. The high complexity inevitable in large problems is gracefully reduced by removing dependency links between tracks. We apply the method to a 10 min sequence of an international football game and compare results to ground truth.
多目标跟踪需要对目标进行定位并标记其身份。后者是一个挑战,因为在足球和监视跟踪中,许多目标的外观模糊不清,经常相互遮挡。我们提出了一种解决这个标签问题的方法。当被隔离时,目标可以被跟踪并保持其身份。然而,如果目标相互作用,情况并非总是如此。本文假设存在一个轨迹图,表示目标何时被隔离并描述它们如何相互作用。定义了孤立轨道之间的相似性度量。目标是通过利用图形约束和相似性度量来关联孤立轨道的身份。我们将其表述为贝叶斯网络推理问题,允许我们使用标准消息传播以有效的方式找到最可能的路径集。在大型问题中不可避免的高复杂性可以通过移除轨道之间的依赖链接而优雅地降低。我们将该方法应用于一场国际足球比赛的10分钟序列,并将结果与地面事实进行比较。
{"title":"Multi-Target Tracking - Linking Identities using Bayesian Network Inference","authors":"Peter Nillius, Josephine Sullivan, S. Carlsson","doi":"10.1109/CVPR.2006.198","DOIUrl":"https://doi.org/10.1109/CVPR.2006.198","url":null,"abstract":"Multi-target tracking requires locating the targets and labeling their identities. The latter is a challenge when many targets, with indistinct appearances, frequently occlude one another, as in football and surveillance tracking. We present an approach to solving this labeling problem. When isolated, a target can be tracked and its identity maintained. While, if targets interact this is not always the case. This paper assumes a track graph exists, denoting when targets are isolated and describing how they interact. Measures of similarity between isolated tracks are defined. The goal is to associate the identities of the isolated tracks, by exploiting the graph constraints and similarity measures. We formulate this as a Bayesian network inference problem, allowing us to use standard message propagation to find the most probable set of paths in an efficient way. The high complexity inevitable in large problems is gracefully reduced by removing dependency links between tracks. We apply the method to a 10 min sequence of an international football game and compare results to ground truth.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126920485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 178
Hierarchical Procrustes Matching for Shape Retrieval 用于形状检索的分层procruster匹配
Graham Mcneill, S. Vijayakumar
We introduce Hierarchical Procrustes Matching (HPM), a segment-based shape matching algorithm which avoids problems associated with purely global or local methods and performs well on benchmark shape retrieval tests. The simplicity of the shape representation leads to a powerful matching algorithm which incorporates intuitive ideas about the perceptual nature of shape while being computationally efficient. This includes the ability to match similar parts even when they occur at different scales or positions. While comparison of multiscale shape representations is typically based on specific features, HPM avoids the need to extract such features. The hierarchical structure of the algorithm captures the appealing notion that matching should proceed in a global to local direction.
本文介绍了一种基于分段的形状匹配算法——分层Procrustes匹配(HPM),该算法避免了纯全局或局部方法的相关问题,并在基准形状检索测试中表现良好。形状表示的简单性导致了一种强大的匹配算法,该算法结合了关于形状感知本质的直观想法,同时具有计算效率。这包括匹配相似部分的能力,即使它们出现在不同的尺度或位置。虽然多尺度形状表示的比较通常是基于特定的特征,但HPM避免了提取这些特征的需要。该算法的层次结构抓住了一个吸引人的概念,即匹配应该从全局到局部方向进行。
{"title":"Hierarchical Procrustes Matching for Shape Retrieval","authors":"Graham Mcneill, S. Vijayakumar","doi":"10.1109/CVPR.2006.133","DOIUrl":"https://doi.org/10.1109/CVPR.2006.133","url":null,"abstract":"We introduce Hierarchical Procrustes Matching (HPM), a segment-based shape matching algorithm which avoids problems associated with purely global or local methods and performs well on benchmark shape retrieval tests. The simplicity of the shape representation leads to a powerful matching algorithm which incorporates intuitive ideas about the perceptual nature of shape while being computationally efficient. This includes the ability to match similar parts even when they occur at different scales or positions. While comparison of multiscale shape representations is typically based on specific features, HPM avoids the need to extract such features. The hierarchical structure of the algorithm captures the appealing notion that matching should proceed in a global to local direction.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121415418","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 146
An Adaptive Appearance Model Approach for Model-based Articulated Object Tracking 基于模型的关节目标跟踪的自适应外观模型方法
A. O. Balan, Michael J. Black
The detection and tracking of three-dimensional human body models has progressed rapidly but successful approaches typically rely on accurate foreground silhouettes obtained using background segmentation. There are many practical applications where such information is imprecise. Here we develop a new image likelihood function based on the visual appearance of the subject being tracked. We propose a robust, adaptive, appearance model based on the Wandering-Stable-Lost framework extended to the case of articulated body parts. The method models appearance using a mixture model that includes an adaptive template, frame-to-frame matching and an outlier process. We employ an annealed particle filtering algorithm for inference and take advantage of the 3D body model to predict selfocclusion and improve pose estimation accuracy. Quantitative tracking results are presented for a walking sequence with a 180 degree turn, captured with four synchronized and calibrated cameras and containing significant appearance changes and self-occlusion in each view.
三维人体模型的检测和跟踪进展迅速,但成功的方法通常依赖于通过背景分割获得准确的前景轮廓。在许多实际应用中,这些信息是不精确的。在这里,我们基于被跟踪对象的视觉外观开发了一个新的图像似然函数。我们提出了一个鲁棒的,自适应的,基于流浪-稳定-丢失框架的外观模型,扩展到铰接的身体部位。该方法使用混合模型对外观进行建模,该混合模型包括自适应模板、帧对帧匹配和离群值处理。我们采用退火粒子滤波算法进行推理,并利用三维身体模型来预测自聚焦,提高姿态估计精度。定量跟踪结果展示了一个180度转弯的行走序列,由四个同步和校准的相机捕获,每个视图中包含显著的外观变化和自遮挡。
{"title":"An Adaptive Appearance Model Approach for Model-based Articulated Object Tracking","authors":"A. O. Balan, Michael J. Black","doi":"10.1109/CVPR.2006.52","DOIUrl":"https://doi.org/10.1109/CVPR.2006.52","url":null,"abstract":"The detection and tracking of three-dimensional human body models has progressed rapidly but successful approaches typically rely on accurate foreground silhouettes obtained using background segmentation. There are many practical applications where such information is imprecise. Here we develop a new image likelihood function based on the visual appearance of the subject being tracked. We propose a robust, adaptive, appearance model based on the Wandering-Stable-Lost framework extended to the case of articulated body parts. The method models appearance using a mixture model that includes an adaptive template, frame-to-frame matching and an outlier process. We employ an annealed particle filtering algorithm for inference and take advantage of the 3D body model to predict selfocclusion and improve pose estimation accuracy. Quantitative tracking results are presented for a walking sequence with a 180 degree turn, captured with four synchronized and calibrated cameras and containing significant appearance changes and self-occlusion in each view.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128747896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 93
Scale Variant Image Pyramids 变型图像金字塔
J. Gluckman
Multi-scale representations are motivated by the scale invariant properties of natural images. While many low level statistical measures, such as the local mean and variance of intensity, behave in a scale invariant manner, there are many higher order deviations from scale invariance where zero-crossings merge and disappear. Such scale variant behavior is important information to represent because it is not easily predicted from lower resolution data. A scale variant image pyramid is a representation that separates this information from the more redundant and predictable scale invariant information.
多尺度表示的动机是自然图像的尺度不变性。虽然许多低水平的统计度量,如强度的局部平均值和方差,表现为尺度不变性,但有许多高阶偏离尺度不变性,其中零交叉合并并消失。这种尺度变化的行为是重要的信息,因为它不容易从低分辨率的数据预测。尺度变化图像金字塔是一种表示,它将这些信息与更冗余和可预测的尺度不变信息分开。
{"title":"Scale Variant Image Pyramids","authors":"J. Gluckman","doi":"10.1109/CVPR.2006.265","DOIUrl":"https://doi.org/10.1109/CVPR.2006.265","url":null,"abstract":"Multi-scale representations are motivated by the scale invariant properties of natural images. While many low level statistical measures, such as the local mean and variance of intensity, behave in a scale invariant manner, there are many higher order deviations from scale invariance where zero-crossings merge and disappear. Such scale variant behavior is important information to represent because it is not easily predicted from lower resolution data. A scale variant image pyramid is a representation that separates this information from the more redundant and predictable scale invariant information.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114287903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
An Integrated Segmentation and Classification Approach Applied to Multiple Sclerosis Analysis 综合分割分类方法在多发性硬化症分析中的应用
A. Akselrod-Ballin, M. Galun, R. Basri, A. Brandt, M. Gomori, M. Filippi, P. Valsasina
We present a novel multiscale approach that combines segmentation with classification to detect abnormal brain structures in medical imagery, and demonstrate its utility in detecting multiple sclerosis lesions in 3D MRI data. Our method uses segmentation to obtain a hierarchical decomposition of a multi-channel, anisotropic MRI scan. It then produces a rich set of features describing the segments in terms of intensity, shape, location, and neighborhood relations. These features are then fed into a decision tree-based classifier, trained with data labeled by experts, enabling the detection of lesions in all scales. Unlike common approaches that use voxel-by-voxel analysis, our system can utilize regional properties that are often important for characterizing abnormal brain structures. We provide experiments showing successful detections of lesions in both simulated and real MR images.
我们提出了一种新的多尺度方法,结合分割和分类来检测医学图像中的异常大脑结构,并展示了其在3D MRI数据中检测多发性硬化症病变的实用性。我们的方法使用分割来获得多通道各向异性MRI扫描的分层分解。然后根据强度、形状、位置和邻里关系,生成一组丰富的特征来描述这些片段。然后将这些特征输入到基于决策树的分类器中,使用专家标记的数据进行训练,从而能够在所有尺度上检测病变。与使用逐体素分析的常见方法不同,我们的系统可以利用通常对表征异常大脑结构很重要的区域属性。我们提供了在模拟和真实的MR图像中成功检测病变的实验。
{"title":"An Integrated Segmentation and Classification Approach Applied to Multiple Sclerosis Analysis","authors":"A. Akselrod-Ballin, M. Galun, R. Basri, A. Brandt, M. Gomori, M. Filippi, P. Valsasina","doi":"10.1109/CVPR.2006.55","DOIUrl":"https://doi.org/10.1109/CVPR.2006.55","url":null,"abstract":"We present a novel multiscale approach that combines segmentation with classification to detect abnormal brain structures in medical imagery, and demonstrate its utility in detecting multiple sclerosis lesions in 3D MRI data. Our method uses segmentation to obtain a hierarchical decomposition of a multi-channel, anisotropic MRI scan. It then produces a rich set of features describing the segments in terms of intensity, shape, location, and neighborhood relations. These features are then fed into a decision tree-based classifier, trained with data labeled by experts, enabling the detection of lesions in all scales. Unlike common approaches that use voxel-by-voxel analysis, our system can utilize regional properties that are often important for characterizing abnormal brain structures. We provide experiments showing successful detections of lesions in both simulated and real MR images.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114450607","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 45
Human Carrying Status in Visual Surveillance 视觉监测中的人体携带状态
D. Tao, Xuelong Li, S. Maybank, Xindong Wu
A person’s gait changes when he or she is carrying an object such as a bag, suitcase or rucksack. As a result, human identification and tracking are made more difficult because the averaged gait image is too simple to represent the carrying status. Therefore, in this paper we first introduce a set of Gabor based human gait appearance models, because Gabor functions are similar to the receptive field profiles in the mammalian cortical simple cells. The very high dimensionality of the feature space makes training difficult. In order to solve this problem we propose a general tensor discriminant analysis (GTDA), which seamlessly incorporates the object (Gabor based human gait appearance model) structure information as a natural constraint. GTDA differs from the previous tensor based discriminant analysis methods in that the training converges. Existing methods fail to converge in the training stage. This makes them unsuitable for practical tasks. Experiments are carried out on the USF baseline data set to recognize a human’s ID from the gait silhouette. The proposed Gabor gait incorporated with GTDA is demonstrated to significantly outperform the existing appearance-based methods.
当一个人背着包、手提箱或帆布背包等物品时,他或她的步态会发生变化。结果,由于平均步态图像过于简单而无法表示携带状态,给人体识别和跟踪增加了难度。因此,在本文中,我们首先引入了一套基于Gabor的人类步态外观模型,因为Gabor功能类似于哺乳动物皮层简单细胞的感受野轮廓。特征空间的高维使得训练变得困难。为了解决这一问题,我们提出了一种通用张量判别分析(GTDA),它无缝地将物体(基于Gabor的人类步态外观模型)的结构信息作为自然约束。GTDA与以往基于张量的判别分析方法的不同之处在于训练是收敛的。现有的方法在训练阶段无法收敛。这使得它们不适合执行实际任务。在USF基线数据集上进行实验,从步态轮廓中识别人的身份。结合GTDA的Gabor步态被证明明显优于现有的基于外观的方法。
{"title":"Human Carrying Status in Visual Surveillance","authors":"D. Tao, Xuelong Li, S. Maybank, Xindong Wu","doi":"10.1109/CVPR.2006.138","DOIUrl":"https://doi.org/10.1109/CVPR.2006.138","url":null,"abstract":"A person’s gait changes when he or she is carrying an object such as a bag, suitcase or rucksack. As a result, human identification and tracking are made more difficult because the averaged gait image is too simple to represent the carrying status. Therefore, in this paper we first introduce a set of Gabor based human gait appearance models, because Gabor functions are similar to the receptive field profiles in the mammalian cortical simple cells. The very high dimensionality of the feature space makes training difficult. In order to solve this problem we propose a general tensor discriminant analysis (GTDA), which seamlessly incorporates the object (Gabor based human gait appearance model) structure information as a natural constraint. GTDA differs from the previous tensor based discriminant analysis methods in that the training converges. Existing methods fail to converge in the training stage. This makes them unsuitable for practical tasks. Experiments are carried out on the USF baseline data set to recognize a human’s ID from the gait silhouette. The proposed Gabor gait incorporated with GTDA is demonstrated to significantly outperform the existing appearance-based methods.","PeriodicalId":421737,"journal":{"name":"2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2006-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125436402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 122
期刊
2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1