首页 > 最新文献

2011 International Conference on Computer Vision最新文献

英文 中文
Simultaneous correspondence and non-rigid 3D reconstruction of the coronary tree from single X-ray images 同时对应和非刚性三维重建从单一的x射线图像冠状树
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126325
Eduard Serradell, Adriana Romero, R. Leta, C. Gatta, F. Moreno-Noguer
We present a novel approach to simultaneously reconstruct the 3D structure of a non-rigid coronary tree and estimate point correspondences between an input X-ray image and a reference 3D shape. At the core of our approach lies an optimization scheme that iteratively fits a generative 3D model of increasing complexity and guides the matching process. As a result, and in contrast to existing approaches that assume rigidity or quasi-rigidity of the structure, our method is able to retrieve large non-linear deformations even when the input data is corrupted by the presence of noise and partial occlusions. We extensively evaluate our approach under synthetic and real data and demonstrate a remarkable improvement compared to state-of-the-art.
我们提出了一种新的方法来同时重建非刚性冠状树的三维结构,并估计输入x射线图像和参考三维形状之间的点对应关系。我们方法的核心是一个优化方案,该方案迭代地适合日益复杂的生成3D模型,并指导匹配过程。因此,与假设结构刚性或准刚性的现有方法相反,即使输入数据被噪声和部分遮挡破坏,我们的方法也能够检索大的非线性变形。我们在综合和真实数据下广泛评估了我们的方法,并证明了与最先进的方法相比有了显着的改进。
{"title":"Simultaneous correspondence and non-rigid 3D reconstruction of the coronary tree from single X-ray images","authors":"Eduard Serradell, Adriana Romero, R. Leta, C. Gatta, F. Moreno-Noguer","doi":"10.1109/ICCV.2011.6126325","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126325","url":null,"abstract":"We present a novel approach to simultaneously reconstruct the 3D structure of a non-rigid coronary tree and estimate point correspondences between an input X-ray image and a reference 3D shape. At the core of our approach lies an optimization scheme that iteratively fits a generative 3D model of increasing complexity and guides the matching process. As a result, and in contrast to existing approaches that assume rigidity or quasi-rigidity of the structure, our method is able to retrieve large non-linear deformations even when the input data is corrupted by the presence of noise and partial occlusions. We extensively evaluate our approach under synthetic and real data and demonstrate a remarkable improvement compared to state-of-the-art.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74275456","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Tracking multiple people under global appearance constraints 在全局外观约束下跟踪多个人
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126235
Horesh Ben Shitrit, J. Berclaz, F. Fleuret, P. Fua
In this paper, we show that tracking multiple people whose paths may intersect can be formulated as a convex global optimization problem. Our proposed framework is designed to exploit image appearance cues to prevent identity switches. Our method is effective even when such cues are only available at distant time intervals. This is unlike many current approaches that depend on appearance being exploitable from frame to frame. We validate our approach on three multi-camera sport and pedestrian datasets that contain long and complex sequences. Our algorithm perseveres identities better than state-of-the-art algorithms while keeping similar MOTA scores.
在本文中,我们证明了跟踪路径可能相交的多个人可以表述为一个凸全局优化问题。我们提出的框架旨在利用图像外观线索来防止身份转换。我们的方法是有效的,即使这些线索只在遥远的时间间隔可用。这不同于当前许多依赖于从一帧到另一帧可利用的外观的方法。我们在三个包含长而复杂序列的多相机运动和行人数据集上验证了我们的方法。我们的算法比最先进的算法更好地保留身份,同时保持相似的MOTA分数。
{"title":"Tracking multiple people under global appearance constraints","authors":"Horesh Ben Shitrit, J. Berclaz, F. Fleuret, P. Fua","doi":"10.1109/ICCV.2011.6126235","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126235","url":null,"abstract":"In this paper, we show that tracking multiple people whose paths may intersect can be formulated as a convex global optimization problem. Our proposed framework is designed to exploit image appearance cues to prevent identity switches. Our method is effective even when such cues are only available at distant time intervals. This is unlike many current approaches that depend on appearance being exploitable from frame to frame. We validate our approach on three multi-camera sport and pedestrian datasets that contain long and complex sequences. Our algorithm perseveres identities better than state-of-the-art algorithms while keeping similar MOTA scores.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75385237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 258
Assessing the aesthetic quality of photographs using generic image descriptors 使用通用图像描述符评估照片的美学质量
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126444
L. Marchesotti, F. Perronnin, Diane Larlus, G. Csurka
In this paper, we automatically assess the aesthetic properties of images. In the past, this problem has been addressed by hand-crafting features which would correlate with best photographic practices (e.g. “Does this image respect the rule of thirds?”) or with photographic techniques (e.g. “Is this image a macro?”). We depart from this line of research and propose to use generic image descriptors to assess aesthetic quality. We experimentally show that the descriptors we use, which aggregate statistics computed from low-level local features, implicitly encode the aesthetic properties explicitly used by state-of-the-art methods and outperform them by a significant margin.
在本文中,我们自动评估图像的美学属性。在过去,这个问题已经通过与最佳摄影实践相关的手工制作功能(例如“这张照片是否遵守三分法?”)或摄影技术(例如“这张照片是微距的吗?”)来解决。我们从这条研究路线出发,建议使用通用图像描述符来评估美学质量。我们通过实验表明,我们使用的描述符(汇总了从低级局部特征计算的统计数据)隐式地编码了最先进方法显式使用的美学属性,并且在很大程度上优于它们。
{"title":"Assessing the aesthetic quality of photographs using generic image descriptors","authors":"L. Marchesotti, F. Perronnin, Diane Larlus, G. Csurka","doi":"10.1109/ICCV.2011.6126444","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126444","url":null,"abstract":"In this paper, we automatically assess the aesthetic properties of images. In the past, this problem has been addressed by hand-crafting features which would correlate with best photographic practices (e.g. “Does this image respect the rule of thirds?”) or with photographic techniques (e.g. “Is this image a macro?”). We depart from this line of research and propose to use generic image descriptors to assess aesthetic quality. We experimentally show that the descriptors we use, which aggregate statistics computed from low-level local features, implicitly encode the aesthetic properties explicitly used by state-of-the-art methods and outperform them by a significant margin.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75471472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 388
Gaussian process regression flow for analysis of motion trajectories 高斯过程回归流分析运动轨迹
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126365
Kihwan Kim, Dongryeol Lee, Irfan Essa
Recognition of motions and activities of objects in videos requires effective representations for analysis and matching of motion trajectories. In this paper, we introduce a new representation specifically aimed at matching motion trajectories. We model a trajectory as a continuous dense flow field from a sparse set of vector sequences using Gaussian Process Regression. Furthermore, we introduce a random sampling strategy for learning stable classes of motions from limited data. Our representation allows for incrementally predicting possible paths and detecting anomalous events from online trajectories. This representation also supports matching of complex motions with acceleration changes and pauses or stops within a trajectory. We use the proposed approach for classifying and predicting motion trajectories in traffic monitoring domains and test on several data sets. We show that our approach works well on various types of complete and incomplete trajectories from a variety of video data sets with different frame rates.
识别视频中物体的运动和活动需要有效的表示来分析和匹配运动轨迹。在本文中,我们引入了一种专门用于匹配运动轨迹的新表示。我们用高斯过程回归从稀疏的向量序列集将轨迹建模为连续的密集流场。此外,我们还引入了一种随机采样策略,用于从有限的数据中学习稳定的运动类别。我们的表示允许增量预测可能的路径,并从在线轨迹检测异常事件。这种表示还支持在轨迹中匹配具有加速度变化和暂停或停止的复杂运动。我们使用该方法对交通监控域中的运动轨迹进行分类和预测,并在多个数据集上进行了测试。我们表明,我们的方法在来自不同帧率的各种视频数据集的各种类型的完整和不完整轨迹上都能很好地工作。
{"title":"Gaussian process regression flow for analysis of motion trajectories","authors":"Kihwan Kim, Dongryeol Lee, Irfan Essa","doi":"10.1109/ICCV.2011.6126365","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126365","url":null,"abstract":"Recognition of motions and activities of objects in videos requires effective representations for analysis and matching of motion trajectories. In this paper, we introduce a new representation specifically aimed at matching motion trajectories. We model a trajectory as a continuous dense flow field from a sparse set of vector sequences using Gaussian Process Regression. Furthermore, we introduce a random sampling strategy for learning stable classes of motions from limited data. Our representation allows for incrementally predicting possible paths and detecting anomalous events from online trajectories. This representation also supports matching of complex motions with acceleration changes and pauses or stops within a trajectory. We use the proposed approach for classifying and predicting motion trajectories in traffic monitoring domains and test on several data sets. We show that our approach works well on various types of complete and incomplete trajectories from a variety of video data sets with different frame rates.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74557672","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 183
Unsupervised and semi-supervised learning via ℓ1-norm graph 基于1-范数图的无监督和半监督学习
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126506
F. Nie, Hua Wang, Heng Huang, C. Ding
In this paper, we propose a novel ℓ1-norm graph model to perform unsupervised and semi-supervised learning methods. Instead of minimizing the ℓ2-norm of spectral embedding as traditional graph based learning methods, our new graph learning model minimizes the ℓ1-norm of spectral embedding with well motivation. The sparsity produced by the ℓ1-norm minimization results in the solutions with much clearer cluster structures, which are suitable for both image clustering and classification tasks. We introduce a new efficient iterative algorithm to solve the ℓ1-norm of spectral embedding minimization problem, and prove the convergence of the algorithm. More specifically, our algorithm adaptively re-weight the original weights of graph to discover clearer cluster structure. Experimental results on both toy data and real image data sets show the effectiveness and advantages of our proposed method.
在本文中,我们提出了一种新的1-范数图模型来执行无监督和半监督学习方法。与传统的基于图的学习方法最小化谱嵌入的l2范数不同,我们的新图学习模型在动机良好的情况下最小化谱嵌入的l2范数。由1范数最小化产生的稀疏性使得解具有更清晰的聚类结构,适合于图像聚类和分类任务。提出了一种新的求解谱嵌入最小化问题的高效迭代算法,并证明了该算法的收敛性。更具体地说,我们的算法自适应地对图的原始权值进行重新加权,以发现更清晰的聚类结构。在玩具数据和真实图像数据集上的实验结果表明了该方法的有效性和优越性。
{"title":"Unsupervised and semi-supervised learning via ℓ1-norm graph","authors":"F. Nie, Hua Wang, Heng Huang, C. Ding","doi":"10.1109/ICCV.2011.6126506","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126506","url":null,"abstract":"In this paper, we propose a novel ℓ1-norm graph model to perform unsupervised and semi-supervised learning methods. Instead of minimizing the ℓ2-norm of spectral embedding as traditional graph based learning methods, our new graph learning model minimizes the ℓ1-norm of spectral embedding with well motivation. The sparsity produced by the ℓ1-norm minimization results in the solutions with much clearer cluster structures, which are suitable for both image clustering and classification tasks. We introduce a new efficient iterative algorithm to solve the ℓ1-norm of spectral embedding minimization problem, and prove the convergence of the algorithm. More specifically, our algorithm adaptively re-weight the original weights of graph to discover clearer cluster structure. Experimental results on both toy data and real image data sets show the effectiveness and advantages of our proposed method.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74597016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 76
Color photometric stereo for multicolored surfaces 用于多色表面的彩色光度立体
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126495
Robert Anderson, B. Stenger, R. Cipolla
We present a multispectral photometric stereo method for capturing geometry of deforming surfaces. A novel photometric calibration technique allows calibration of scenes containing multiple piecewise constant chromaticities. This method estimates per-pixel photometric properties, then uses a RANSAC-based approach to estimate the dominant chromaticities in the scene. A likelihood term is developed linking surface normal, image intensity and photometric properties, which allows estimating the number of chromaticities present in a scene to be framed as a model estimation problem. The Bayesian Information Criterion is applied to automatically estimate the number of chromaticities present during calibration. A two-camera stereo system provides low resolution geometry, allowing the likelihood term to be used in segmenting new images into regions of constant chromaticity. This segmentation is carried out in a Markov Random Field framework and allows the correct photometric properties to be used at each pixel to estimate a dense normal map. Results are shown on several challenging real-world sequences, demonstrating state-of-the-art results using only two cameras and three light sources. Quantitative evaluation is provided against synthetic ground truth data.
我们提出了一种多光谱光度立体方法来捕捉变形表面的几何形状。一种新的光度校准技术允许校准包含多个分段恒定色度的场景。该方法估计每像素的光度属性,然后使用基于ransac的方法估计场景中的主色度。开发了连接表面法线,图像强度和光度属性的似然项,它允许估计场景中存在的色度数量,并将其框架为模型估计问题。应用贝叶斯信息准则自动估计校准过程中存在的色度数。双摄像头立体系统提供低分辨率几何,允许在分割新图像到恒定色度的区域使用的可能性项。这种分割是在马尔科夫随机场框架中进行的,并允许在每个像素上使用正确的光度属性来估计密集的法线贴图。结果显示了几个具有挑战性的现实世界序列,展示了仅使用两个摄像头和三个光源的最先进的结果。对合成的地面真值数据进行定量评价。
{"title":"Color photometric stereo for multicolored surfaces","authors":"Robert Anderson, B. Stenger, R. Cipolla","doi":"10.1109/ICCV.2011.6126495","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126495","url":null,"abstract":"We present a multispectral photometric stereo method for capturing geometry of deforming surfaces. A novel photometric calibration technique allows calibration of scenes containing multiple piecewise constant chromaticities. This method estimates per-pixel photometric properties, then uses a RANSAC-based approach to estimate the dominant chromaticities in the scene. A likelihood term is developed linking surface normal, image intensity and photometric properties, which allows estimating the number of chromaticities present in a scene to be framed as a model estimation problem. The Bayesian Information Criterion is applied to automatically estimate the number of chromaticities present during calibration. A two-camera stereo system provides low resolution geometry, allowing the likelihood term to be used in segmenting new images into regions of constant chromaticity. This segmentation is carried out in a Markov Random Field framework and allows the correct photometric properties to be used at each pixel to estimate a dense normal map. Results are shown on several challenging real-world sequences, demonstrating state-of-the-art results using only two cameras and three light sources. Quantitative evaluation is provided against synthetic ground truth data.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74741606","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Self-calibrating depth from refraction 自校准深度从折射
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126298
Zhihu Chen, Kwan-Yee Kenneth Wong, Y. Matsushita, Xiaolong Zhu, Miaomiao Liu
In this paper, we introduce a novel method for depth acquisition based on refraction of light. A scene is captured twice by a fixed perspective camera, with the first image captured directly by the camera and the second by placing a transparent medium between the scene and the camera. A depth map of the scene is then recovered from the displacements of scene points in the images. Unlike other existing depth from refraction methods, our method does not require the knowledge of the pose and refractive index of the transparent medium, but can recover them directly from the input images. We hence call our method self-calibrating depth from refraction. Experimental results on both synthetic and real-world data are presented, which demonstrate the effectiveness of the proposed method.
本文介绍了一种基于光折射的深度采集新方法。一个场景被固定透视相机捕获两次,第一张图像直接被相机捕获,第二次通过在场景和相机之间放置透明介质捕获。然后从图像中场景点的位移恢复场景的深度图。与其他现有的深度折射方法不同,我们的方法不需要了解透明介质的姿态和折射率,而是可以直接从输入图像中恢复它们。因此,我们称这种方法为自校准折射深度。在合成数据和实际数据上的实验结果证明了该方法的有效性。
{"title":"Self-calibrating depth from refraction","authors":"Zhihu Chen, Kwan-Yee Kenneth Wong, Y. Matsushita, Xiaolong Zhu, Miaomiao Liu","doi":"10.1109/ICCV.2011.6126298","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126298","url":null,"abstract":"In this paper, we introduce a novel method for depth acquisition based on refraction of light. A scene is captured twice by a fixed perspective camera, with the first image captured directly by the camera and the second by placing a transparent medium between the scene and the camera. A depth map of the scene is then recovered from the displacements of scene points in the images. Unlike other existing depth from refraction methods, our method does not require the knowledge of the pose and refractive index of the transparent medium, but can recover them directly from the input images. We hence call our method self-calibrating depth from refraction. Experimental results on both synthetic and real-world data are presented, which demonstrate the effectiveness of the proposed method.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73112717","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Diffusion runs low on persistence fast 扩散在持久性上运行得很快
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126271
Chao Chen, H. Edelsbrunner
Interpreting an image as a function on a compact subset of the Euclidean plane, we get its scale-space by diffusion, spreading the image over the entire plane. This generates a 1-parameter family of functions alternatively defined as convolutions with a progressively wider Gaussian kernel. We prove that the corresponding 1-parameter family of persistence diagrams have norms that go rapidly to zero as time goes to infinity. This result rationalizes experimental observations about scale-space. We hope this will lead to targeted improvements of related computer vision methods.
将图像解释为欧几里得平面紧子集上的函数,我们通过扩散得到它的尺度空间,将图像扩展到整个平面上。这产生了一个1参数的函数族,或者定义为具有逐渐变宽的高斯核的卷积。我们证明了相应的1参数持久性图族具有随着时间趋于无穷而迅速趋近于零的范数。这一结果为尺度空间的实验观察提供了理论依据。我们希望这将导致相关计算机视觉方法的有针对性的改进。
{"title":"Diffusion runs low on persistence fast","authors":"Chao Chen, H. Edelsbrunner","doi":"10.1109/ICCV.2011.6126271","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126271","url":null,"abstract":"Interpreting an image as a function on a compact subset of the Euclidean plane, we get its scale-space by diffusion, spreading the image over the entire plane. This generates a 1-parameter family of functions alternatively defined as convolutions with a progressively wider Gaussian kernel. We prove that the corresponding 1-parameter family of persistence diagrams have norms that go rapidly to zero as time goes to infinity. This result rationalizes experimental observations about scale-space. We hope this will lead to targeted improvements of related computer vision methods.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76107376","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Fast articulated motion tracking using a sums of Gaussians body model 基于高斯和体模型的快速关节运动跟踪
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126338
Carsten Stoll, N. Hasler, Juergen Gall, H. Seidel, C. Theobalt
We present an approach for modeling the human body by Sums of spatial Gaussians (SoG), allowing us to perform fast and high-quality markerless motion capture from multi-view video sequences. The SoG model is equipped with a color model to represent the shape and appearance of the human and can be reconstructed from a sparse set of images. Similar to the human body, we also represent the image domain as SoG that models color consistent image blobs. Based on the SoG models of the image and the human body, we introduce a novel continuous and differentiable model-to-image similarity measure that can be used to estimate the skeletal motion of a human at 5–15 frames per second even for many camera views. In our experiments, we show that our method, which does not rely on silhouettes or training data, offers an good balance between accuracy and computational cost.
我们提出了一种通过空间高斯和(SoG)对人体建模的方法,使我们能够从多视图视频序列中执行快速和高质量的无标记运动捕获。SoG模型配备了一个颜色模型来表示人的形状和外观,可以从稀疏的图像集合中重建。与人体相似,我们也将图像域表示为SoG,该SoG对颜色一致的图像斑点进行建模。基于图像和人体的SoG模型,我们引入了一种新的连续和可微的模型-图像相似性度量,即使在许多摄像机视图下,该度量也可用于估计人体骨骼每秒5-15帧的运动。在我们的实验中,我们表明我们的方法不依赖于轮廓或训练数据,在精度和计算成本之间提供了很好的平衡。
{"title":"Fast articulated motion tracking using a sums of Gaussians body model","authors":"Carsten Stoll, N. Hasler, Juergen Gall, H. Seidel, C. Theobalt","doi":"10.1109/ICCV.2011.6126338","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126338","url":null,"abstract":"We present an approach for modeling the human body by Sums of spatial Gaussians (SoG), allowing us to perform fast and high-quality markerless motion capture from multi-view video sequences. The SoG model is equipped with a color model to represent the shape and appearance of the human and can be reconstructed from a sparse set of images. Similar to the human body, we also represent the image domain as SoG that models color consistent image blobs. Based on the SoG models of the image and the human body, we introduce a novel continuous and differentiable model-to-image similarity measure that can be used to estimate the skeletal motion of a human at 5–15 frames per second even for many camera views. In our experiments, we show that our method, which does not rely on silhouettes or training data, offers an good balance between accuracy and computational cost.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80105962","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 220
The medial feature detector: Stable regions from image boundaries 中间特征检测器:来自图像边界的稳定区域
Pub Date : 2011-11-06 DOI: 10.1109/ICCV.2011.6126436
Yannis Avrithis, Konstantinos Rapantzikos
We present a local feature detector that is able to detect regions of arbitrary scale and shape, without scale space construction. We compute a weighted distance map on image gradient, using our exact linear-time algorithm, a variant of group marching for Euclidean space. We find the weighted medial axis by extending residues, typically used in Voronoi skeletons. We decompose the medial axis into a graph representing image structure in terms of peaks and saddle points. A duality property enables reconstruction of regions using the same marching method. We greedily group regions taking both contrast and shape into account. On the way, we select regions according to our shape fragmentation factor, favoring those well enclosed by boundaries—even incomplete. We achieve state of the art performance in matching and retrieval experiments with reduced memory and computational requirements.
我们提出了一种局部特征检测器,它能够检测任意尺度和形状的区域,而不需要构建尺度空间。我们使用我们的精确线性时间算法计算图像梯度上的加权距离图,这是欧几里得空间的一种变体。我们通过扩展残基找到加权的内侧轴,通常用于Voronoi骨架。我们将中间轴分解成一个以峰和鞍点表示图像结构的图。对偶属性允许使用相同的行进方法重建区域。我们贪婪地对区域进行分组,同时考虑到对比度和形状。在这个过程中,我们根据我们的形状碎片因子来选择区域,偏爱那些被边界包围得很好的区域——甚至是不完整的区域。我们在匹配和检索实验中实现了最先进的性能,减少了内存和计算需求。
{"title":"The medial feature detector: Stable regions from image boundaries","authors":"Yannis Avrithis, Konstantinos Rapantzikos","doi":"10.1109/ICCV.2011.6126436","DOIUrl":"https://doi.org/10.1109/ICCV.2011.6126436","url":null,"abstract":"We present a local feature detector that is able to detect regions of arbitrary scale and shape, without scale space construction. We compute a weighted distance map on image gradient, using our exact linear-time algorithm, a variant of group marching for Euclidean space. We find the weighted medial axis by extending residues, typically used in Voronoi skeletons. We decompose the medial axis into a graph representing image structure in terms of peaks and saddle points. A duality property enables reconstruction of regions using the same marching method. We greedily group regions taking both contrast and shape into account. On the way, we select regions according to our shape fragmentation factor, favoring those well enclosed by boundaries—even incomplete. We achieve state of the art performance in matching and retrieval experiments with reduced memory and computational requirements.","PeriodicalId":6391,"journal":{"name":"2011 International Conference on Computer Vision","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2011-11-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86549554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
期刊
2011 International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1