首页 > 最新文献

2007 IEEE Conference on Computer Vision and Pattern Recognition最新文献

英文 中文
Connecting the Out-of-Sample and Pre-Image Problems in Kernel Methods 结合核方法中的样本外和预像问题
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383038
P. Arias, G. Randall, G. Sapiro
Kernel methods have been widely studied in the field of pattern recognition. These methods implicitly map, "the kernel trick," the data into a space which is more appropriate for analysis. Many manifold learning and dimensionality reduction techniques are simply kernel methods for which the mapping is explicitly computed. In such cases, two problems related with the mapping arise: The out-of-sample extension and the pre-image computation. In this paper we propose a new pre-image method based on the Nystrom formulation for the out-of-sample extension, showing the connections between both problems. We also address the importance of normalization in the feature space, which has been ignored by standard pre-image algorithms. As an example, we apply these ideas to the Gaussian kernel, and relate our approach to other popular pre-image methods. Finally, we show the application of these techniques in the study of dynamic shapes.
核方法在模式识别领域得到了广泛的研究。这些方法隐式地将“核技巧”数据映射到更适合分析的空间中。许多流形学习和降维技术都是简单的核方法,其映射是显式计算的。在这种情况下,与映射相关的两个问题出现了:样本外扩展和图像前计算。在本文中,我们提出了一种新的基于Nystrom公式的样本外扩展的预像方法,展示了这两个问题之间的联系。我们还讨论了特征空间中归一化的重要性,这一点被标准的预图像算法所忽略。作为一个例子,我们将这些思想应用于高斯核,并将我们的方法与其他流行的预图像方法联系起来。最后,我们展示了这些技术在动态形状研究中的应用。
{"title":"Connecting the Out-of-Sample and Pre-Image Problems in Kernel Methods","authors":"P. Arias, G. Randall, G. Sapiro","doi":"10.1109/CVPR.2007.383038","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383038","url":null,"abstract":"Kernel methods have been widely studied in the field of pattern recognition. These methods implicitly map, \"the kernel trick,\" the data into a space which is more appropriate for analysis. Many manifold learning and dimensionality reduction techniques are simply kernel methods for which the mapping is explicitly computed. In such cases, two problems related with the mapping arise: The out-of-sample extension and the pre-image computation. In this paper we propose a new pre-image method based on the Nystrom formulation for the out-of-sample extension, showing the connections between both problems. We also address the importance of normalization in the feature space, which has been ignored by standard pre-image algorithms. As an example, we apply these ideas to the Gaussian kernel, and relate our approach to other popular pre-image methods. Finally, we show the application of these techniques in the study of dynamic shapes.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115715975","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Integration of Motion Cues in Optical and Sonar Videos for 3-D Positioning 运动线索在光学和声纳视频中的集成用于三维定位
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383354
S. Negahdaripour, H. Pirsiavash, H. Sekkati
Target-based positioning and 3-D target reconstruction are critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. While optical cameras provide high-resolution and target details, they are constrained by limited visibility range. In highly turbid waters, target at up to distances of 10 s of meters can be recorded by high-frequency (MHz) 2-D sonar imaging systems that have become introduced to the commercial market in years. Because of lower resolution and SNR level and inferior target details compared to optical camera in favorable visibility conditions, the integration of both sensing modalities can enable operation in a wider range of conditions with generally better performance compared to deploying either system alone. In this paper, estimate of the 3-D motion of the integrated system and the 3-D reconstruction of scene features are addressed. We do not require establishing matches between optical and sonar features, referred to as opti-acoustic correspondences, but rather matches in either the sonar or optical motion sequences. In addition to improving the motion estimation accuracy, advantages of the system comprise overcoming certain inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity, and dual interpretation of planar scenes. We discuss how the proposed solution provides an effective strategy to address the rather complex opti-acoustic stereo matching problem. Experiment with real data demonstrate our technical contribution.
基于目标的定位和三维目标重建是部署潜水平台进行一系列水下应用的关键能力,例如搜索和检查任务。虽然光学相机提供高分辨率和目标细节,但它们受到有限的可见范围的限制。在高度浑浊的水域中,高频(MHz)二维声纳成像系统可以记录距离达10米的目标,该系统多年来已被引入商业市场。由于在有利的能见度条件下,与光学相机相比,分辨率和信噪比较低,目标细节较差,因此与单独部署任一系统相比,两种传感模式的集成可以在更广泛的条件下运行,并且通常具有更好的性能。本文主要研究了集成系统的三维运动估计和场景特征的三维重建。我们不需要在光学和声纳特征之间建立匹配,称为光声对应,而是在声纳或光学运动序列中建立匹配。除了提高运动估计精度外,该系统的优点还包括克服了单目视觉固有的某些模糊性,例如比例因子模糊性和平面场景的双重解释。我们讨论了所提出的解决方案如何提供一个有效的策略来解决相当复杂的光声立体匹配问题。实际数据实验证明了我们的技术贡献。
{"title":"Integration of Motion Cues in Optical and Sonar Videos for 3-D Positioning","authors":"S. Negahdaripour, H. Pirsiavash, H. Sekkati","doi":"10.1109/CVPR.2007.383354","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383354","url":null,"abstract":"Target-based positioning and 3-D target reconstruction are critical capabilities in deploying submersible platforms for a range of underwater applications, e.g., search and inspection missions. While optical cameras provide high-resolution and target details, they are constrained by limited visibility range. In highly turbid waters, target at up to distances of 10 s of meters can be recorded by high-frequency (MHz) 2-D sonar imaging systems that have become introduced to the commercial market in years. Because of lower resolution and SNR level and inferior target details compared to optical camera in favorable visibility conditions, the integration of both sensing modalities can enable operation in a wider range of conditions with generally better performance compared to deploying either system alone. In this paper, estimate of the 3-D motion of the integrated system and the 3-D reconstruction of scene features are addressed. We do not require establishing matches between optical and sonar features, referred to as opti-acoustic correspondences, but rather matches in either the sonar or optical motion sequences. In addition to improving the motion estimation accuracy, advantages of the system comprise overcoming certain inherent ambiguities of monocular vision, e.g., the scale-factor ambiguity, and dual interpretation of planar scenes. We discuss how the proposed solution provides an effective strategy to address the rather complex opti-acoustic stereo matching problem. Experiment with real data demonstrate our technical contribution.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124367543","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Opti-Acoustic Stereo Imaging, System Calibration and 3-D Reconstruction 光声立体成像,系统校准和三维重建
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383361
S. Negahdaripour, H. Sekkati, H. Pirsiavash
Utilization of an acoustic camera for range measurements is a key advantage for 3-D shape recovery of underwater targets by opti-acoustic stereo imaging, where the associated epipolar geometry of optical and acoustic image correspondences can be described in terms of conic sections. In this paper, we propose methods for system calibration and 3-D scene reconstruction by maximum likelihood estimation from noisy image measurements. The recursive 3-D reconstruction method utilized as initial condition a closed-form solution that integrates the advantages of so-called range and azimuth solutions. Synthetic data tests are given to provide insight into the merits of the new target imaging and 3-D reconstruction paradigm, while experiments with real data confirm the findings based on computer simulations, and demonstrate the merits of this novel 3-D reconstruction paradigm.
利用声相机进行距离测量是通过光声立体成像恢复水下目标三维形状的关键优势,其中光学和声学图像对应的相关极极几何可以用圆锥截面来描述。在本文中,我们提出了基于噪声图像测量的极大似然估计的系统校准和三维场景重建方法。递归三维重建方法采用了一个封闭解作为初始条件,该解综合了所谓距离解和方位角解的优点。通过综合数据测试,深入了解了新的目标成像和三维重建范式的优点,而基于计算机模拟的真实数据实验证实了这些发现,并证明了这种新的三维重建范式的优点。
{"title":"Opti-Acoustic Stereo Imaging, System Calibration and 3-D Reconstruction","authors":"S. Negahdaripour, H. Sekkati, H. Pirsiavash","doi":"10.1109/CVPR.2007.383361","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383361","url":null,"abstract":"Utilization of an acoustic camera for range measurements is a key advantage for 3-D shape recovery of underwater targets by opti-acoustic stereo imaging, where the associated epipolar geometry of optical and acoustic image correspondences can be described in terms of conic sections. In this paper, we propose methods for system calibration and 3-D scene reconstruction by maximum likelihood estimation from noisy image measurements. The recursive 3-D reconstruction method utilized as initial condition a closed-form solution that integrates the advantages of so-called range and azimuth solutions. Synthetic data tests are given to provide insight into the merits of the new target imaging and 3-D reconstruction paradigm, while experiments with real data confirm the findings based on computer simulations, and demonstrate the merits of this novel 3-D reconstruction paradigm.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"68 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124393283","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Shape from Shading Based on Lax-Friedrichs Fast Sweeping and Regularization Techniques With Applications to Document Image Restoration 基于拉克斯-弗里德里希快速扫描和正则化技术的阴影形状及其在文档图像恢复中的应用
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383287
Li Zhang, A. Yip, C. Tan
In this paper, we describe a 2-pass iterative scheme to solve the general partial differential equation (PDE) related to the Shape-from-Shading (SFS) problem under both distant and close point light sources. In particular, we discuss its applications in restoring warped document images that often appear in the daily snapshots. The proposed method consists of two steps. First the image irradiance equation is formulated as a static Hamilton-Jacobi (HJ) equation and solved using a fast sweeping strategy with Lax-Friedrichs Hamiltonian. However, abrupt errors may arise when applying to real document images due to noises in the approximated shading image. To reduce the noise sensitivity, a minimization method thus follows to smooth out the abrupt ridges in the initial result and produce a better reconstruction. Experiments on synthetic surfaces show promising results comparing to the ground truth data. Moreover, a general framework is developed, which demonstrates that the SFS method can help to remove both geometric and photometric distortions in warped document images for better visual appearance and higher recognition rate.
在本文中,我们描述了一种2次迭代方案来解决与远距离和近距离点光源下形状-阴影(SFS)问题相关的一般偏微分方程(PDE)。特别地,我们讨论了它在恢复日常快照中经常出现的扭曲文档图像中的应用。该方法分为两个步骤。首先将图像辐照度方程表示为静态哈密顿-雅可比(HJ)方程,并采用拉克斯-弗里德里希斯哈密顿量的快速扫描策略求解。然而,在应用于实际文档图像时,由于逼近的阴影图像中存在噪声,可能会产生突发性误差。为了降低噪声敏感性,采用最小化方法来平滑初始结果中的突变脊,从而获得更好的重建结果。与地面真值数据相比,在合成表面上的实验结果令人满意。此外,还开发了一个通用框架,证明了SFS方法可以帮助去除扭曲文档图像中的几何和光度畸变,从而获得更好的视觉外观和更高的识别率。
{"title":"Shape from Shading Based on Lax-Friedrichs Fast Sweeping and Regularization Techniques With Applications to Document Image Restoration","authors":"Li Zhang, A. Yip, C. Tan","doi":"10.1109/CVPR.2007.383287","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383287","url":null,"abstract":"In this paper, we describe a 2-pass iterative scheme to solve the general partial differential equation (PDE) related to the Shape-from-Shading (SFS) problem under both distant and close point light sources. In particular, we discuss its applications in restoring warped document images that often appear in the daily snapshots. The proposed method consists of two steps. First the image irradiance equation is formulated as a static Hamilton-Jacobi (HJ) equation and solved using a fast sweeping strategy with Lax-Friedrichs Hamiltonian. However, abrupt errors may arise when applying to real document images due to noises in the approximated shading image. To reduce the noise sensitivity, a minimization method thus follows to smooth out the abrupt ridges in the initial result and produce a better reconstruction. Experiments on synthetic surfaces show promising results comparing to the ground truth data. Moreover, a general framework is developed, which demonstrates that the SFS method can help to remove both geometric and photometric distortions in warped document images for better visual appearance and higher recognition rate.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124535731","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Kinematics from Lines in a Single Rolling Shutter Image 单张滚动快门图像中线条的运动学
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383119
Omar Ait-Aider, A. Bartoli, N. Andreff
Recent work shows that recovering pose and velocity from a single view of a moving rigid object is possible with a rolling shutter camera, based on feature point correspondences. We extend this method to line correspondences. Owing to the combined effect of rolling shutter and object motion, straight lines are distorted to curves as they get imaged with a rolling shutter camera. Lines thus capture more information than points, which is not the case with standard projection models for which both points and lines give two constraints. We extend the standard line reprojection error, and propose a nonlinear method for retrieving a solution to the pose and velocity computation problem. A careful inspection of the design matrix in the normal equations reveals that it is highly sparse and patterned. We propose a blockwise solution procedure based on bundle-adjustment-like sparse inversion. This makes nonlinear optimization fast and numerically stable. The method is validated using real data.
最近的研究表明,基于特征点对应,使用卷帘式相机从移动的刚性物体的单一视图中恢复姿态和速度是可能的。我们将这种方法推广到行通信。由于卷帘门和物体运动的共同作用,直线在使用卷帘门相机拍摄时被扭曲成曲线。因此,线比点捕获更多的信息,这与标准投影模型的情况不同,因为点和线都有两个约束。我们扩展了标准直线重投影误差,并提出了一种非线性方法来检索姿态和速度计算问题的解。仔细检查正常方程中的设计矩阵,可以发现它是高度稀疏和图案化的。提出了一种基于类束调整稀疏反演的分块求解方法。这使得非线性优化快速且数值稳定。用实际数据对该方法进行了验证。
{"title":"Kinematics from Lines in a Single Rolling Shutter Image","authors":"Omar Ait-Aider, A. Bartoli, N. Andreff","doi":"10.1109/CVPR.2007.383119","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383119","url":null,"abstract":"Recent work shows that recovering pose and velocity from a single view of a moving rigid object is possible with a rolling shutter camera, based on feature point correspondences. We extend this method to line correspondences. Owing to the combined effect of rolling shutter and object motion, straight lines are distorted to curves as they get imaged with a rolling shutter camera. Lines thus capture more information than points, which is not the case with standard projection models for which both points and lines give two constraints. We extend the standard line reprojection error, and propose a nonlinear method for retrieving a solution to the pose and velocity computation problem. A careful inspection of the design matrix in the normal equations reveals that it is highly sparse and patterned. We propose a blockwise solution procedure based on bundle-adjustment-like sparse inversion. This makes nonlinear optimization fast and numerically stable. The method is validated using real data.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"204 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114753974","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 60
Layered Depth Panoramas 分层深度全景
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383295
K. Zheng, S. B. Kang, Michael F. Cohen, R. Szeliski
Representations for interactive photorealistic visualization of scenes range from compact 2D panoramas to data-intensive 4D light fields. In this paper, we propose a technique for creating a layered representation from a sparse set of images taken with a hand-held camera. This representation, which we call a layered depth panorama (LDP), allows the user to experience 3D by off-axis panning. It combines the compelling experience of panoramas with limited 3D navigation. Our choice of representation is motivated by ease of capture and compactness. We formulate the problem of constructing the LDP as the recovery of color and geometry in a multi-perspective cylindrical disparity space. We leverage a graph cut approach to sequentially determine the disparity and color of each layer using multi-view stereo. Geometry visible through the cracks at depth discontinuities in a frontmost layer is determined and assigned to layers behind the frontmost layer. All layers are then used to render novel panoramic views with parallax. We demonstrate our approach on a variety of complex outdoor and indoor scenes.
交互式逼真场景可视化的表示范围从紧凑的2D全景图到数据密集型的4D光场。在本文中,我们提出了一种从手持相机拍摄的稀疏图像集创建分层表示的技术。这种表示,我们称之为分层深度全景(LDP),允许用户通过离轴平移来体验3D。它结合了令人信服的全景体验和有限的3D导航。我们选择表示法的动机是易于捕获和紧凑性。我们将LDP的构造问题表述为多视角圆柱视差空间中颜色和几何的恢复问题。我们利用图形切割方法来确定使用多视图立体的每一层的视差和颜色。在最前端层的深度不连续处通过裂缝可见的几何形状被确定并分配给最前端层后面的层。所有的图层都用来渲染新颖的视差全景。我们在各种复杂的室外和室内场景中展示了我们的方法。
{"title":"Layered Depth Panoramas","authors":"K. Zheng, S. B. Kang, Michael F. Cohen, R. Szeliski","doi":"10.1109/CVPR.2007.383295","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383295","url":null,"abstract":"Representations for interactive photorealistic visualization of scenes range from compact 2D panoramas to data-intensive 4D light fields. In this paper, we propose a technique for creating a layered representation from a sparse set of images taken with a hand-held camera. This representation, which we call a layered depth panorama (LDP), allows the user to experience 3D by off-axis panning. It combines the compelling experience of panoramas with limited 3D navigation. Our choice of representation is motivated by ease of capture and compactness. We formulate the problem of constructing the LDP as the recovery of color and geometry in a multi-perspective cylindrical disparity space. We leverage a graph cut approach to sequentially determine the disparity and color of each layer using multi-view stereo. Geometry visible through the cracks at depth discontinuities in a frontmost layer is determined and assigned to layers behind the frontmost layer. All layers are then used to render novel panoramic views with parallax. We demonstrate our approach on a variety of complex outdoor and indoor scenes.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"147 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116027198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Improved Video Registration using Non-Distinctive Local Image Features 改进视频配准使用非显著的局部图像特征
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.382989
Robin Hess, Alan Fern
The task of registering video frames with a static model is a common problem in many computer vision domains. The standard approach to registration involves finding point correspondences between the video and the model and using those correspondences to numerically determine registration transforms. Current methods locate video-to-model point correspondences by assembling a set of reference images to represent the model and then detecting and matching invariant local image features between the video frames and the set of reference images. These methods work well when all video frames can be guaranteed to contain a sufficient number of distinctive visual features. However, as we demonstrate, these methods are prone to severe misregistration errors in domains where many video frames lack distinctive image features. To overcome these errors, we introduce a concept of local distinctiveness which allows us to find model matches for nearly all video features, regardless of their distinctiveness on a global scale. We present results from the American football domain-where many video frames lack distinctive image features-which show a drastic improvement in registration accuracy over current methods. In addition, we introduce a simple, empirical stability test that allows our method to be fully automated. Finally, we present a registration dataset from the American football domain we hope can be used as a benchmarking tool for registration methods.
在许多计算机视觉领域中,用静态模型对视频帧进行配准是一个常见问题。配准的标准方法包括找到视频和模型之间的点对应关系,并使用这些对应关系在数值上确定配准变换。目前的方法是通过组装一组代表模型的参考图像,然后在视频帧和参考图像集之间检测和匹配不变的局部图像特征来定位视频与模型点的对应关系。当所有视频帧都能保证包含足够数量的不同的视觉特征时,这些方法就能很好地工作。然而,正如我们所证明的,这些方法在许多视频帧缺乏显著图像特征的域中容易出现严重的配错错误。为了克服这些错误,我们引入了一个局部显著性的概念,使我们能够找到几乎所有视频特征的模型匹配,而不管它们在全球范围内的显著性如何。我们展示了来自美式橄榄球领域的结果,其中许多视频帧缺乏鲜明的图像特征,这表明与当前方法相比,配准精度有了极大的提高。此外,我们引入了一个简单的,经验性的稳定性测试,使我们的方法完全自动化。最后,我们提供了一个来自美式橄榄球领域的注册数据集,我们希望可以用作注册方法的基准测试工具。
{"title":"Improved Video Registration using Non-Distinctive Local Image Features","authors":"Robin Hess, Alan Fern","doi":"10.1109/CVPR.2007.382989","DOIUrl":"https://doi.org/10.1109/CVPR.2007.382989","url":null,"abstract":"The task of registering video frames with a static model is a common problem in many computer vision domains. The standard approach to registration involves finding point correspondences between the video and the model and using those correspondences to numerically determine registration transforms. Current methods locate video-to-model point correspondences by assembling a set of reference images to represent the model and then detecting and matching invariant local image features between the video frames and the set of reference images. These methods work well when all video frames can be guaranteed to contain a sufficient number of distinctive visual features. However, as we demonstrate, these methods are prone to severe misregistration errors in domains where many video frames lack distinctive image features. To overcome these errors, we introduce a concept of local distinctiveness which allows us to find model matches for nearly all video features, regardless of their distinctiveness on a global scale. We present results from the American football domain-where many video frames lack distinctive image features-which show a drastic improvement in registration accuracy over current methods. In addition, we introduce a simple, empirical stability test that allows our method to be fully automated. Finally, we present a registration dataset from the American football domain we hope can be used as a benchmarking tool for registration methods.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"210 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116364786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 81
Bottom-up Recognition and Parsing of the Human Body 人体自底向上的识别与解析
Pub Date : 2007-06-17 DOI: 10.1007/978-3-540-74198-5_13
Praveen Srinivasan, Jianbo Shi
{"title":"Bottom-up Recognition and Parsing of the Human Body","authors":"Praveen Srinivasan, Jianbo Shi","doi":"10.1007/978-3-540-74198-5_13","DOIUrl":"https://doi.org/10.1007/978-3-540-74198-5_13","url":null,"abstract":"","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"107 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123507852","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 97
Quantifying Facial Expression Abnormality in Schizophrenia by Combining 2D and 3D Features 结合二维和三维特征量化精神分裂症患者面部表情异常
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383061
Peng Wang, Christiane Köhler, Fred Barrett, R. Gur, R. Gur, R. Verma
Most of current computer-based facial expression analysis methods focus on the recognition of perfectly posed expressions, and hence are incapable of handling the individuals with expression impairments. In particular, patients with schizophrenia usually have impaired expressions in the form of "flat" or "inappropriate" affects, which make the quantification of their facial expressions a challenging problem. This paper presents methods to quantify the group differences between patients with schizophrenia and healthy controls, by extracting specialized features and analyzing group differences on a feature manifold. The features include 2D and 3D geometric features, and the moment invariants combining both 3D geometry and 2D textures. Facial expression recognition experiments on actors demonstrate that our combined features can better characterize facial expressions than either 2D geometric or texture features. The features are then embedded into an ISOMAP manifold to quantify the group differences between controls and patients. Experiments show that our results are strongly supported by the human rating results and clinical findings, thus providing a framework that is able to quantify the abnormality in patients with schizophrenia.
目前大多数基于计算机的面部表情分析方法侧重于识别完美姿势的表情,因此无法处理有表情障碍的个体。特别是,精神分裂症患者通常会出现“平淡”或“不恰当”的表情受损,这使得对他们面部表情的量化成为一个具有挑战性的问题。本文提出了一种量化精神分裂症患者与健康对照组之间群体差异的方法,该方法通过提取专门的特征并分析特征流形上的群体差异。特征包括二维和三维几何特征,以及结合三维几何和二维纹理的矩不变量。演员面部表情识别实验表明,我们的组合特征比二维几何特征或纹理特征更能表征面部表情。然后将这些特征嵌入到ISOMAP歧管中,以量化对照组和患者之间的组差异。实验表明,我们的结果得到了人类评分结果和临床结果的有力支持,从而提供了一个能够量化精神分裂症患者异常的框架。
{"title":"Quantifying Facial Expression Abnormality in Schizophrenia by Combining 2D and 3D Features","authors":"Peng Wang, Christiane Köhler, Fred Barrett, R. Gur, R. Gur, R. Verma","doi":"10.1109/CVPR.2007.383061","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383061","url":null,"abstract":"Most of current computer-based facial expression analysis methods focus on the recognition of perfectly posed expressions, and hence are incapable of handling the individuals with expression impairments. In particular, patients with schizophrenia usually have impaired expressions in the form of \"flat\" or \"inappropriate\" affects, which make the quantification of their facial expressions a challenging problem. This paper presents methods to quantify the group differences between patients with schizophrenia and healthy controls, by extracting specialized features and analyzing group differences on a feature manifold. The features include 2D and 3D geometric features, and the moment invariants combining both 3D geometry and 2D textures. Facial expression recognition experiments on actors demonstrate that our combined features can better characterize facial expressions than either 2D geometric or texture features. The features are then embedded into an ISOMAP manifold to quantify the group differences between controls and patients. Experiments show that our results are strongly supported by the human rating results and clinical findings, thus providing a framework that is able to quantify the abnormality in patients with schizophrenia.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"363 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123557271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Fusion of Face and Palmprint for Personal Identification Based on Ordinal Features 基于顺序特征的人脸与掌纹融合身份识别
Pub Date : 2007-06-17 DOI: 10.1109/CVPR.2007.383522
R. Chu, Shengcai Liao, Yufei Han, Zhenan Sun, S. Li, T. Tan
In this paper, we present a face and palmprint multimodal biometric identification method and system to improve the identification performance. Effective classifiers based on ordinal features are constructed for faces and palmprints, respectively. Then, the matching scores from the two classifiers are combined using several fusion strategies. Experimental results on a middle-scale data set have demonstrated the effectiveness of the proposed system.
本文提出了一种人脸与掌纹的多模态生物特征识别方法和系统,以提高识别性能。分别对人脸和掌纹构造了基于有序特征的有效分类器。然后,使用多种融合策略对两个分类器的匹配分数进行组合。在中等规模数据集上的实验结果证明了该系统的有效性。
{"title":"Fusion of Face and Palmprint for Personal Identification Based on Ordinal Features","authors":"R. Chu, Shengcai Liao, Yufei Han, Zhenan Sun, S. Li, T. Tan","doi":"10.1109/CVPR.2007.383522","DOIUrl":"https://doi.org/10.1109/CVPR.2007.383522","url":null,"abstract":"In this paper, we present a face and palmprint multimodal biometric identification method and system to improve the identification performance. Effective classifiers based on ordinal features are constructed for faces and palmprints, respectively. Then, the matching scores from the two classifiers are combined using several fusion strategies. Experimental results on a middle-scale data set have demonstrated the effectiveness of the proposed system.","PeriodicalId":351008,"journal":{"name":"2007 IEEE Conference on Computer Vision and Pattern Recognition","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-06-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123617676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
期刊
2007 IEEE Conference on Computer Vision and Pattern Recognition
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1