首页 > 最新文献

2009 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Motion pattern interpretation and detection for tracking moving vehicles in airborne video 机载视频中运动车辆跟踪的运动模式解释与检测
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206541
Qian Yu, G. Medioni
Detection and tracking of moving vehicles in airborne videos is a challenging problem. Many approaches have been proposed to improve motion segmentation on frame-by-frame and pixel-by-pixel bases, however, little attention has been paid to analyze the long-term motion pattern, which is a distinctive property for moving vehicles in airborne videos. In this paper, we provide a straightforward geometric interpretation of a general motion pattern in 4D space (x, y, vx, vy). We propose to use the tensor voting computational framework to detect and segment such motion patterns in 4D space. Specifically, in airborne videos, we analyze the essential difference in motion patterns caused by parallax and independent moving objects, which leads to a practical method for segmenting motion patterns (flows) created by moving vehicles in stabilized airborne videos. The flows are used in turn to facilitate detection and tracking of each individual object in the flow. Conceptually, this approach is similar to “track-before-detect” techniques, which involves temporal information in the process as early as possible. As shown in the experiments, many difficult cases in airborne videos, such as parallax, noisy background modeling and long term occlusions, can be addressed by our approach.
机载视频中运动飞行器的检测与跟踪是一个具有挑战性的问题。人们提出了许多方法来改进逐帧和逐像素的运动分割,然而,很少有人关注分析长期运动模式,这是机载视频中运动车辆的一个独特特性。在本文中,我们提供了四维空间(x, y, vx, vy)中一般运动模式的直接几何解释。我们建议使用张量投票计算框架来检测和分割四维空间中的这种运动模式。具体来说,在机载视频中,我们分析了视差和独立运动物体引起的运动模式的本质区别,从而得出了一种实用的方法来分割稳定机载视频中运动车辆产生的运动模式(流)。依次使用流来促进流中每个单独对象的检测和跟踪。从概念上讲,这种方法类似于“检测前跟踪”技术,它尽可能早地涉及到过程中的时间信息。实验表明,我们的方法可以解决机载视频中的许多困难情况,如视差、噪声背景建模和长期遮挡。
{"title":"Motion pattern interpretation and detection for tracking moving vehicles in airborne video","authors":"Qian Yu, G. Medioni","doi":"10.1109/CVPR.2009.5206541","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206541","url":null,"abstract":"Detection and tracking of moving vehicles in airborne videos is a challenging problem. Many approaches have been proposed to improve motion segmentation on frame-by-frame and pixel-by-pixel bases, however, little attention has been paid to analyze the long-term motion pattern, which is a distinctive property for moving vehicles in airborne videos. In this paper, we provide a straightforward geometric interpretation of a general motion pattern in 4D space (x, y, vx, vy). We propose to use the tensor voting computational framework to detect and segment such motion patterns in 4D space. Specifically, in airborne videos, we analyze the essential difference in motion patterns caused by parallax and independent moving objects, which leads to a practical method for segmenting motion patterns (flows) created by moving vehicles in stabilized airborne videos. The flows are used in turn to facilitate detection and tracking of each individual object in the flow. Conceptually, this approach is similar to “track-before-detect” techniques, which involves temporal information in the process as early as possible. As shown in the experiments, many difficult cases in airborne videos, such as parallax, noisy background modeling and long term occlusions, can be addressed by our approach.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116810879","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 58
Fast multiple shape correspondence by pre-organizing shape instances 通过预先组织形状实例,快速实现多个形状对应
Pub Date : 2009-06-20 DOI: 10.1109/cvpr.2009.5206611
B. Munsell, Andrew Temlyakov, Song Wang
Accurately identifying corresponded landmarks from a population of shape instances is the major challenge in constructing statistical shape models. In general, shape-correspondence methods can be grouped into one of two categories: global methods and pair-wise methods. In this paper, we develop a new method that attempts to address the limitations of both the global and pair-wise methods. In particular, we reorganize the input population into a tree structure that incorporates global information about the population of shape instances, where each node in the tree represents a shape instance and each edge connects two very similar shape instances. Using this organized tree, neighboring shape instances can be corresponded efficiently and accurately by a pair-wise method. In the experiments, we evaluate the proposed method and compare its performance to five available shape correspondence methods and show the proposed method achieves the accuracy of a global method with speed of a pair-wise method.
在统计形状模型的构建中,准确地从一群形状实例中识别出相应的标志是一个主要的挑战。一般来说,形状对应方法可以分为两类:全局方法和成对方法。在本文中,我们开发了一种新的方法,试图解决全局方法和成对方法的局限性。特别是,我们将输入填充重新组织成一个包含形状实例填充全局信息的树结构,其中树中的每个节点代表一个形状实例,每个边连接两个非常相似的形状实例。利用这种组织树,相邻的形状实例可以通过配对方法高效准确地对应。在实验中,我们对所提出的方法进行了评价,并将其性能与五种可用的形状对应方法进行了比较,结果表明所提出的方法达到了全局方法的精度和成对方法的速度。
{"title":"Fast multiple shape correspondence by pre-organizing shape instances","authors":"B. Munsell, Andrew Temlyakov, Song Wang","doi":"10.1109/cvpr.2009.5206611","DOIUrl":"https://doi.org/10.1109/cvpr.2009.5206611","url":null,"abstract":"Accurately identifying corresponded landmarks from a population of shape instances is the major challenge in constructing statistical shape models. In general, shape-correspondence methods can be grouped into one of two categories: global methods and pair-wise methods. In this paper, we develop a new method that attempts to address the limitations of both the global and pair-wise methods. In particular, we reorganize the input population into a tree structure that incorporates global information about the population of shape instances, where each node in the tree represents a shape instance and each edge connects two very similar shape instances. Using this organized tree, neighboring shape instances can be corresponded efficiently and accurately by a pair-wise method. In the experiments, we evaluate the proposed method and compare its performance to five available shape correspondence methods and show the proposed method achieves the accuracy of a global method with speed of a pair-wise method.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123982331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Towards a practical face recognition system: Robust registration and illumination by sparse representation 一个实用的人脸识别系统:稀疏表示的鲁棒配准和照明
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206654
Andrew Wagner, John Wright, Arvind Ganesh, Zihan Zhou, Yi Ma
Most contemporary face recognition algorithms work well under laboratory conditions but degrade when tested in less-controlled environments. This is mostly due to the difficulty of simultaneously handling variations in illumination, alignment, pose, and occlusion. In this paper, we propose a simple and practical face recognition system that achieves a high degree of robustness and stability to all these variations. We demonstrate how to use tools from sparse representation to align a test face image with a set of frontal training images in the presence of significant registration error and occlusion. We thoroughly characterize the region of attraction for our alignment algorithm on public face datasets such as Multi-PIE. We further study how to obtain a sufficient set of training illuminations for linearly interpolating practical lighting conditions. We have implemented a complete face recognition system, including a projector-based training acquisition system, in order to evaluate how our algorithms work under practical testing conditions. We show that our system can efficiently and effectively recognize faces under a variety of realistic conditions, using only frontal images under the proposed illuminations as training.
大多数当代人脸识别算法在实验室条件下工作良好,但在控制较少的环境中进行测试时就会下降。这主要是由于同时处理照明,对齐,姿势和遮挡变化的困难。在本文中,我们提出了一个简单实用的人脸识别系统,该系统对所有这些变化都具有高度的鲁棒性和稳定性。我们演示了如何使用稀疏表示的工具在存在显着配准错误和遮挡的情况下将测试人脸图像与一组正面训练图像对齐。我们在公共人脸数据集(如Multi-PIE)上对我们的对齐算法的吸引区域进行了彻底的表征。我们进一步研究了如何获得一组足够的训练照明来线性插值实际照明条件。我们已经实现了一个完整的人脸识别系统,包括一个基于投影仪的训练采集系统,以评估我们的算法在实际测试条件下的工作情况。我们证明了我们的系统可以在各种现实条件下高效地识别人脸,仅使用提出的照明下的正面图像作为训练。
{"title":"Towards a practical face recognition system: Robust registration and illumination by sparse representation","authors":"Andrew Wagner, John Wright, Arvind Ganesh, Zihan Zhou, Yi Ma","doi":"10.1109/CVPR.2009.5206654","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206654","url":null,"abstract":"Most contemporary face recognition algorithms work well under laboratory conditions but degrade when tested in less-controlled environments. This is mostly due to the difficulty of simultaneously handling variations in illumination, alignment, pose, and occlusion. In this paper, we propose a simple and practical face recognition system that achieves a high degree of robustness and stability to all these variations. We demonstrate how to use tools from sparse representation to align a test face image with a set of frontal training images in the presence of significant registration error and occlusion. We thoroughly characterize the region of attraction for our alignment algorithm on public face datasets such as Multi-PIE. We further study how to obtain a sufficient set of training illuminations for linearly interpolating practical lighting conditions. We have implemented a complete face recognition system, including a projector-based training acquisition system, in order to evaluate how our algorithms work under practical testing conditions. We show that our system can efficiently and effectively recognize faces under a variety of realistic conditions, using only frontal images under the proposed illuminations as training.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123988885","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 211
A nonparametric Riemannian framework for processing high angular resolution diffusion images (HARDI) 处理高角分辨率扩散图像的非参数黎曼框架
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206843
A. Goh, C. Lenglet, P. Thompson, R. Vidal
High angular resolution diffusion imaging has become an important magnetic resonance technique for in vivo imaging. Most current research in this field focuses on developing methods for computing the orientation distribution function (ODF), which is the probability distribution function of water molecule diffusion along any angle on the sphere. In this paper, we present a Riemannian framework to carry out computations on an ODF field. The proposed framework does not require that the ODFs be represented by any fixed parameterization, such as a mixture of von Mises-Fisher distributions or a spherical harmonic expansion. Instead, we use a non-parametric representation of the ODF, and exploit the fact that under the square-root re-parameterization, the space of ODFs forms a Riemannian manifold, namely the unit Hilbert sphere. Specifically, we use Riemannian operations to perform various geometric data processing algorithms, such as interpolation, convolution and linear and nonlinear filtering. We illustrate these concepts with numerical experiments on synthetic and real datasets.
高角分辨率扩散成像已成为磁共振体内成像的重要技术。目前该领域的研究主要集中在开发方向分布函数(ODF)的计算方法,方向分布函数是水分子沿球体上任意角度扩散的概率分布函数。在本文中,我们提出了一个黎曼框架来对ODF域进行计算。所提出的框架不要求odf用任何固定的参数化表示,例如von Mises-Fisher分布的混合或球谐展开。相反,我们使用ODF的非参数表示,并利用在平方根重新参数化下,ODF的空间形成黎曼流形,即单位希尔伯特球。具体来说,我们使用黎曼运算来执行各种几何数据处理算法,如插值,卷积以及线性和非线性滤波。我们用合成数据集和真实数据集的数值实验来说明这些概念。
{"title":"A nonparametric Riemannian framework for processing high angular resolution diffusion images (HARDI)","authors":"A. Goh, C. Lenglet, P. Thompson, R. Vidal","doi":"10.1109/CVPR.2009.5206843","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206843","url":null,"abstract":"High angular resolution diffusion imaging has become an important magnetic resonance technique for in vivo imaging. Most current research in this field focuses on developing methods for computing the orientation distribution function (ODF), which is the probability distribution function of water molecule diffusion along any angle on the sphere. In this paper, we present a Riemannian framework to carry out computations on an ODF field. The proposed framework does not require that the ODFs be represented by any fixed parameterization, such as a mixture of von Mises-Fisher distributions or a spherical harmonic expansion. Instead, we use a non-parametric representation of the ODF, and exploit the fact that under the square-root re-parameterization, the space of ODFs forms a Riemannian manifold, namely the unit Hilbert sphere. Specifically, we use Riemannian operations to perform various geometric data processing algorithms, such as interpolation, convolution and linear and nonlinear filtering. We illustrate these concepts with numerical experiments on synthetic and real datasets.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129725605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
Efficient planar graph cuts with applications in Computer Vision 高效平面图形切割及其在计算机视觉中的应用
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206863
Frank R. Schmidt, Eno Töppe, D. Cremers
We present a fast graph cut algorithm for planar graphs. It is based on the graph theoretical work and leads to an efficient method that we apply on shape matching and image segmentation. In contrast to currently used methods in computer vision, the presented approach provides an upper bound for its runtime behavior that is almost linear. In particular, we are able to match two different planar shapes of N points in O(N2 log N) and segment a given image of N pixels in O(N log N). We present two experimental benchmark studies which demonstrate that the presented method is also in practice faster than previously proposed graph cut methods: On planar shape matching and image segmentation we observe a speed-up of an order of magnitude, depending on resolution.
提出了一种用于平面图形的快速图切算法。它是在图理论工作的基础上提出的一种有效的形状匹配和图像分割方法。与目前使用的计算机视觉方法相比,该方法为其运行时行为提供了一个几乎线性的上界。特别是,我们能够在O(N2 log N)内匹配N个点的两个不同平面形状,并在O(N log N)内分割给定图像的N个像素。我们提出了两个实验基准研究,表明所提出的方法在实践中也比以前提出的图切方法更快:在平面形状匹配和图像分割上,我们观察到速度提高了一个数量级,这取决于分辨率。
{"title":"Efficient planar graph cuts with applications in Computer Vision","authors":"Frank R. Schmidt, Eno Töppe, D. Cremers","doi":"10.1109/CVPR.2009.5206863","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206863","url":null,"abstract":"We present a fast graph cut algorithm for planar graphs. It is based on the graph theoretical work and leads to an efficient method that we apply on shape matching and image segmentation. In contrast to currently used methods in computer vision, the presented approach provides an upper bound for its runtime behavior that is almost linear. In particular, we are able to match two different planar shapes of N points in O(N2 log N) and segment a given image of N pixels in O(N log N). We present two experimental benchmark studies which demonstrate that the presented method is also in practice faster than previously proposed graph cut methods: On planar shape matching and image segmentation we observe a speed-up of an order of magnitude, depending on resolution.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127013163","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
A robust parametric method for bias field estimation and segmentation of MR images 一种鲁棒的磁共振图像偏置场估计和分割方法
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206553
Chunming Li, Chris Gatenby, Li Wang, J. Gore
This paper proposes a new energy minimization framework for simultaneous estimation of the bias field and segmentation of tissues for magnetic resonance images. The bias field is modeled as a linear combination of a set of basis functions, and thereby parameterized by the coefficients of the basis functions. We define an energy that depends on the coefficients of the basis functions, the membership functions of the tissues in the image, and the constants approximating the true signal from the corresponding tissues. This energy is convex in each of its variables. Bias field estimation and image segmentation are simultaneously achieved as the result of minimizing this energy. We provide an efficient iterative algorithm for energy minimization, which converges to the optimal solution at a fast rate. A salient advantage of our method is that its result is independent of initialization, which allows robust and fully automated application. The proposed method has been successfully applied to 3-Tesla MR images with desirable results. Comparisons with other approaches demonstrate the superior performance of this algorithm.
提出了一种新的能量最小化框架,用于同时估计磁共振图像的偏置场和组织分割。将偏置场建模为一组基函数的线性组合,从而用基函数的系数来参数化偏置场。我们定义了一个能量,它取决于基函数的系数,图像中组织的隶属函数,以及接近相应组织的真实信号的常数。这个能量在它的每个变量中都是凸的。由于该能量最小,因此可以同时实现偏置场估计和图像分割。给出了一种高效的能量最小化迭代算法,该算法快速收敛到最优解。我们的方法的一个显著优点是它的结果是独立于初始化的,这允许健壮和完全自动化的应用程序。该方法已成功应用于3-特斯拉MR图像,效果良好。与其他方法的比较表明,该算法具有较好的性能。
{"title":"A robust parametric method for bias field estimation and segmentation of MR images","authors":"Chunming Li, Chris Gatenby, Li Wang, J. Gore","doi":"10.1109/CVPR.2009.5206553","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206553","url":null,"abstract":"This paper proposes a new energy minimization framework for simultaneous estimation of the bias field and segmentation of tissues for magnetic resonance images. The bias field is modeled as a linear combination of a set of basis functions, and thereby parameterized by the coefficients of the basis functions. We define an energy that depends on the coefficients of the basis functions, the membership functions of the tissues in the image, and the constants approximating the true signal from the corresponding tissues. This energy is convex in each of its variables. Bias field estimation and image segmentation are simultaneously achieved as the result of minimizing this energy. We provide an efficient iterative algorithm for energy minimization, which converges to the optimal solution at a fast rate. A salient advantage of our method is that its result is independent of initialization, which allows robust and fully automated application. The proposed method has been successfully applied to 3-Tesla MR images with desirable results. Comparisons with other approaches demonstrate the superior performance of this algorithm.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127544805","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 90
Automatic fetal face detection from ultrasound volumes via learning 3D and 2D information 通过学习3D和2D信息,从超声体积中自动检测胎儿面部
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206527
Shaolei Feng, S. Zhou, Sara Good, D. Comaniciu
3D ultrasound imaging has been increasingly used in clinics for fetal examination. However, manually searching for the optimal view of the fetal face in 3D ultrasound volumes is cumbersome and time-consuming even for expert physicians and sonographers. In this paper we propose a learning-based approach which combines both 3D and 2D information for automatic and fast fetal face detection from 3D ultrasound volumes. Our approach applies a new technique - constrained marginal space learning - for 3D face mesh detection, and combines a boosting-based 2D profile detection to refine 3D face pose. To enhance the rendering of the fetal face, an automatic carving algorithm is proposed to remove all obstructions in front of the face based on the detected face mesh. Experiments are performed on a challenging 3D ultrasound data set containing 1010 fetal volumes. The results show that our system not only achieves excellent detection accuracy but also runs very fast - it can detect the fetal face from the 3D data in 1 second on a dual-core 2.0 GHz computer.
三维超声成像已越来越多地用于临床胎儿检查。然而,即使对专家医生和超声医师来说,在3D超声体积中手动搜索胎儿面部的最佳视图既麻烦又耗时。在本文中,我们提出了一种基于学习的方法,该方法结合了3D和2D信息,用于从3D超声体积中自动快速检测胎儿面部。我们的方法应用了一种新技术-约束边缘空间学习-用于3D人脸网格检测,并结合了基于增强的2D轮廓检测来优化3D人脸姿态。为了增强胎儿面部的渲染效果,提出了一种基于检测到的面部网格,去除面部前方所有障碍物的自动雕刻算法。实验是在包含1010个胎儿体积的具有挑战性的3D超声数据集上进行的。结果表明,该系统不仅具有良好的检测精度,而且运行速度非常快,在双核2.0 GHz计算机上,可以在1秒内从3D数据中检测出胎儿面部。
{"title":"Automatic fetal face detection from ultrasound volumes via learning 3D and 2D information","authors":"Shaolei Feng, S. Zhou, Sara Good, D. Comaniciu","doi":"10.1109/CVPR.2009.5206527","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206527","url":null,"abstract":"3D ultrasound imaging has been increasingly used in clinics for fetal examination. However, manually searching for the optimal view of the fetal face in 3D ultrasound volumes is cumbersome and time-consuming even for expert physicians and sonographers. In this paper we propose a learning-based approach which combines both 3D and 2D information for automatic and fast fetal face detection from 3D ultrasound volumes. Our approach applies a new technique - constrained marginal space learning - for 3D face mesh detection, and combines a boosting-based 2D profile detection to refine 3D face pose. To enhance the rendering of the fetal face, an automatic carving algorithm is proposed to remove all obstructions in front of the face based on the detected face mesh. Experiments are performed on a challenging 3D ultrasound data set containing 1010 fetal volumes. The results show that our system not only achieves excellent detection accuracy but also runs very fast - it can detect the fetal face from the 3D data in 1 second on a dual-core 2.0 GHz computer.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1969 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129973997","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Learning semantic scene models by object classification and trajectory clustering 通过对象分类和轨迹聚类学习语义场景模型
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206809
Tianzhu Zhang, Hanqing Lu, S. Li
Activity analysis is a basic task in video surveillance and has become an active research area. However, due to the diversity of moving objects category and their motion patterns, developing robust semantic scene models for activity analysis remains a challenging problem in traffic scenarios. This paper proposes a novel framework to learn semantic scene models. In this framework, the detected moving objects are first classified as pedestrians or vehicles via a co-trained classifier which takes advantage of the multiview information of objects. As a result, the framework can automatically learn motion patterns respectively for pedestrians and vehicles. Then, a graph is proposed to learn and cluster the motion patterns. To this end, trajectory is parameterized and the image is cut into multiple blocks which are taken as the nodes in the graph. Based on the parameters of trajectories, the primary motion patterns in each node (block) are extracted via Gaussian mixture model (GMM), and supplied to this graph. The graph cut algorithm is finally employed to group the motion patterns together, and trajectories are clustered to learn semantic scene models. Experimental results and applications to real world scenes show the validity of our proposed method.
活动分析是视频监控的一项基本任务,已成为一个活跃的研究领域。然而,由于运动物体类别及其运动模式的多样性,开发鲁棒的语义场景模型用于交通场景的活动分析仍然是一个具有挑战性的问题。本文提出了一种新的语义场景模型学习框架。在该框架中,首先利用物体的多视图信息,通过共同训练的分类器将检测到的运动物体分类为行人或车辆。因此,该框架可以自动学习行人和车辆的运动模式。然后,提出了一个图来学习和聚类运动模式。为此,将轨迹参数化,并将图像切割成多个块作为图中的节点。基于轨迹参数,通过高斯混合模型(GMM)提取每个节点(块)的主要运动模式,并提供给该图。最后利用图切算法对运动模式进行分组,并对轨迹进行聚类学习语义场景模型。实验结果和对真实场景的应用表明了该方法的有效性。
{"title":"Learning semantic scene models by object classification and trajectory clustering","authors":"Tianzhu Zhang, Hanqing Lu, S. Li","doi":"10.1109/CVPR.2009.5206809","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206809","url":null,"abstract":"Activity analysis is a basic task in video surveillance and has become an active research area. However, due to the diversity of moving objects category and their motion patterns, developing robust semantic scene models for activity analysis remains a challenging problem in traffic scenarios. This paper proposes a novel framework to learn semantic scene models. In this framework, the detected moving objects are first classified as pedestrians or vehicles via a co-trained classifier which takes advantage of the multiview information of objects. As a result, the framework can automatically learn motion patterns respectively for pedestrians and vehicles. Then, a graph is proposed to learn and cluster the motion patterns. To this end, trajectory is parameterized and the image is cut into multiple blocks which are taken as the nodes in the graph. Based on the parameters of trajectories, the primary motion patterns in each node (block) are extracted via Gaussian mixture model (GMM), and supplied to this graph. The graph cut algorithm is finally employed to group the motion patterns together, and trajectories are clustered to learn semantic scene models. Experimental results and applications to real world scenes show the validity of our proposed method.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129134467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 117
Mutual information-based stereo matching combined with SIFT descriptor in log-chromaticity color space 对数色度空间中基于互信息的立体匹配与SIFT描述子相结合
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206507
Y. S. Heo, Kyoung Mu Lee, Sang Uk Lee
Radiometric variations between input images can seriously degrade the performance of stereo matching algorithms. In this situation, mutual information is a very popular and powerful measure which can find any global relationship of intensities between two input images taken from unknown sources. The mutual information-based method, however, is still ambiguous or erroneous as regards local radiometric variations, since it only accounts for global variation between images, and does not contain spatial information properly. In this paper, we present a new method based on mutual information combined with SIFT descriptor to find correspondence for images which undergo local as well as global radiometric variations. We transform the input color images to log-chromaticity color space from which a linear relationship can be established. To incorporate spatial information in mutual information, we utilize the SIFT descriptor which includes near pixel gradient histogram to construct a joint probability in log-chromaticity color space. By combining the mutual information as an appearance measure and the SIFT descriptor as a geometric measure, we devise a robust and accurate stereo system. Experimental results show that our method is superior to the state-of-the art algorithms including conventional mutual information-based methods and window correlation methods under various radiometric changes.
输入图像之间的辐射变化会严重降低立体匹配算法的性能。在这种情况下,互信息是一种非常流行和强大的度量,它可以找到来自未知来源的两个输入图像之间的任何全局强度关系。然而,基于互信息的方法在局部辐射变化方面仍然是模糊或错误的,因为它只考虑图像之间的全局变化,而没有适当地包含空间信息。本文提出了一种基于互信息与SIFT描述子相结合的方法,用于寻找局部和全局辐射变化图像的对应关系。我们将输入的彩色图像转换为对数色度色彩空间,从对数色度色彩空间可以建立线性关系。为了将空间信息融合到互信息中,我们利用包含近像素梯度直方图的SIFT描述子在对数色度色彩空间中构造联合概率。通过结合互信息作为外观度量和SIFT描述子作为几何度量,我们设计了一个鲁棒和精确的立体系统。实验结果表明,在各种辐射变化情况下,该方法优于传统的互信息方法和窗口相关方法。
{"title":"Mutual information-based stereo matching combined with SIFT descriptor in log-chromaticity color space","authors":"Y. S. Heo, Kyoung Mu Lee, Sang Uk Lee","doi":"10.1109/CVPR.2009.5206507","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206507","url":null,"abstract":"Radiometric variations between input images can seriously degrade the performance of stereo matching algorithms. In this situation, mutual information is a very popular and powerful measure which can find any global relationship of intensities between two input images taken from unknown sources. The mutual information-based method, however, is still ambiguous or erroneous as regards local radiometric variations, since it only accounts for global variation between images, and does not contain spatial information properly. In this paper, we present a new method based on mutual information combined with SIFT descriptor to find correspondence for images which undergo local as well as global radiometric variations. We transform the input color images to log-chromaticity color space from which a linear relationship can be established. To incorporate spatial information in mutual information, we utilize the SIFT descriptor which includes near pixel gradient histogram to construct a joint probability in log-chromaticity color space. By combining the mutual information as an appearance measure and the SIFT descriptor as a geometric measure, we devise a robust and accurate stereo system. Experimental results show that our method is superior to the state-of-the art algorithms including conventional mutual information-based methods and window correlation methods under various radiometric changes.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132861938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Adaptive image and video retargeting technique based on Fourier analysis 基于傅里叶分析的自适应图像和视频重定向技术
Pub Date : 2009-06-20 DOI: 10.1109/CVPR.2009.5206666
Jun-Seong Kim, Jin-Hwan Kim, Chang-Su Kim
An adaptive image and video retargeting algorithm based on Fourier analysis is proposed in this work. We first divide an input image into several strips using the gradient information so that each strip consists of textures of similar complexities. Then, we scale each strip adaptively according to its importance measure. More specifically, the distortions, generated by the scaling procedure, are formulated in the frequency domain using the Fourier transform. Then, the objective is to determine the sizes of scaled strips to minimize the sum of distortions, subject to the constraint that the sum of their sizes should equal the size of the target output image. We solve this constrained optimization problem using the Lagrangian multiplier technique. Moreover, we extend the approach to the retargeting of video sequences. Simulation results demonstrate that the proposed algorithm provides reliable retargeting performance efficiently.
本文提出了一种基于傅里叶分析的自适应图像和视频重定向算法。我们首先使用梯度信息将输入图像分成若干条,使每个条由相似复杂性的纹理组成。然后,我们根据其重要性度量自适应缩放每个条带。更具体地说,由缩放过程产生的畸变在频域中使用傅里叶变换表示。然后,目标是确定缩放条带的大小,以最小化扭曲的总和,并受到其大小之和应等于目标输出图像大小的约束。我们用拉格朗日乘子技术解决了这个约束优化问题。此外,我们将该方法扩展到视频序列的重定向。仿真结果表明,该算法具有可靠的重定向性能。
{"title":"Adaptive image and video retargeting technique based on Fourier analysis","authors":"Jun-Seong Kim, Jin-Hwan Kim, Chang-Su Kim","doi":"10.1109/CVPR.2009.5206666","DOIUrl":"https://doi.org/10.1109/CVPR.2009.5206666","url":null,"abstract":"An adaptive image and video retargeting algorithm based on Fourier analysis is proposed in this work. We first divide an input image into several strips using the gradient information so that each strip consists of textures of similar complexities. Then, we scale each strip adaptively according to its importance measure. More specifically, the distortions, generated by the scaling procedure, are formulated in the frequency domain using the Fourier transform. Then, the objective is to determine the sizes of scaled strips to minimize the sum of distortions, subject to the constraint that the sum of their sizes should equal the size of the target output image. We solve this constrained optimization problem using the Lagrangian multiplier technique. Moreover, we extend the approach to the retargeting of video sequences. Simulation results demonstrate that the proposed algorithm provides reliable retargeting performance efficiently.","PeriodicalId":386532,"journal":{"name":"2009 IEEE Conference on Computer Vision and Pattern Recognition","volume":"89 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2009-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126228766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 77
期刊
2009 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1