首页 > 最新文献

Proceedings Ninth IEEE International Conference on Computer Vision最新文献

英文 中文
Background modeling and subtraction of dynamic scenes 动态场景的背景建模和减法
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238641
Antoine Monnet, Anurag Mittal, N. Paragios, Visvanathan Ramesh
Background modeling and subtraction is a core component in motion analysis. The central idea behind such module is to create a probabilistic representation of the static scene that is compared with the current input to perform subtraction. Such approach is efficient when the scene to be modeled refers to a static structure with limited perturbation. In this paper, we address the problem of modeling dynamic scenes where the assumption of a static background is not valid. Waving trees, beaches, escalators, natural scenes with rain or snow are examples. Inspired by the work proposed by Doretto et al. (2003), we propose an on-line auto-regressive model to capture and predict the behavior of such scenes. Towards detection of events we introduce a new metric that is based on a state-driven comparison between the prediction and the actual frame. Promising results demonstrate the potentials of the proposed framework.
背景建模和减法是运动分析的核心组成部分。该模块背后的核心思想是创建静态场景的概率表示,将其与当前输入进行比较以执行减法。当要建模的场景是摄动有限的静态结构时,这种方法是有效的。在本文中,我们解决了动态场景建模的问题,其中静态背景的假设是无效的。摇曳的树木、海滩、自动扶梯、下雨或下雪的自然景观都是例子。受Doretto等人(2003)提出的工作的启发,我们提出了一个在线自回归模型来捕获和预测这些场景的行为。为了检测事件,我们引入了一种新的度量,该度量基于预测和实际帧之间的状态驱动比较。令人鼓舞的结果证明了所提出框架的潜力。
{"title":"Background modeling and subtraction of dynamic scenes","authors":"Antoine Monnet, Anurag Mittal, N. Paragios, Visvanathan Ramesh","doi":"10.1109/ICCV.2003.1238641","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238641","url":null,"abstract":"Background modeling and subtraction is a core component in motion analysis. The central idea behind such module is to create a probabilistic representation of the static scene that is compared with the current input to perform subtraction. Such approach is efficient when the scene to be modeled refers to a static structure with limited perturbation. In this paper, we address the problem of modeling dynamic scenes where the assumption of a static background is not valid. Waving trees, beaches, escalators, natural scenes with rain or snow are examples. Inspired by the work proposed by Doretto et al. (2003), we propose an on-line auto-regressive model to capture and predict the behavior of such scenes. Towards detection of events we introduce a new metric that is based on a state-driven comparison between the prediction and the actual frame. Promising results demonstrate the potentials of the proposed framework.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"244 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123311335","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 481
Recognition of group activities using dynamic probabilistic networks 基于动态概率网络的群体活动识别
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238423
S. Gong, T. Xiang
Dynamic Probabilistic Networks (DPNs) are exploited for modeling the temporal relationships among a set of different object temporal events in the scene for a coherent and robust scene-level behaviour interpretation. In particular, we develop a Dynamically Multi-Linked Hidden Markov Model (DML-HMM) to interpret group activities involving multiple objects captured in an outdoor scene. The model is based on the discovery of salient dynamic interlinks among multiple temporal events using DPNs. Object temporal events are detected and labeled using Gaussian Mixture Models with automatic model order selection. A DML-HMM is built using Schwarz's Bayesian Information Criterion based factorisation resulting in its topology being intrinsically determined by the underlying causality and temporal order among different object events. Our experiments demonstrate that its performance on modelling group activities in a noisy outdoor scene is superior compared to that of a Multi-Observation Hidden Markov Model (MOHMM), a Parallel Hidden Markov Model (PaHMM) and a Coupled Hidden Markov Model (CHMM).
动态概率网络(dpn)用于建模场景中一组不同对象时间事件之间的时间关系,以实现连贯和鲁棒的场景级行为解释。特别是,我们开发了一个动态多链接隐马尔可夫模型(DML-HMM)来解释在户外场景中捕获的涉及多个对象的群体活动。该模型基于使用DPNs发现多个时间事件之间显著的动态相互联系。使用具有自动模型顺序选择的高斯混合模型对目标时间事件进行检测和标记。DML-HMM是基于Schwarz的贝叶斯信息准则构建的,其拓扑结构本质上是由不同对象事件之间的潜在因果关系和时间顺序决定的。实验表明,该模型在模拟嘈杂室外环境下的群体活动方面优于多观测隐马尔可夫模型(MOHMM)、并行隐马尔可夫模型(PaHMM)和耦合隐马尔可夫模型(CHMM)。
{"title":"Recognition of group activities using dynamic probabilistic networks","authors":"S. Gong, T. Xiang","doi":"10.1109/ICCV.2003.1238423","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238423","url":null,"abstract":"Dynamic Probabilistic Networks (DPNs) are exploited for modeling the temporal relationships among a set of different object temporal events in the scene for a coherent and robust scene-level behaviour interpretation. In particular, we develop a Dynamically Multi-Linked Hidden Markov Model (DML-HMM) to interpret group activities involving multiple objects captured in an outdoor scene. The model is based on the discovery of salient dynamic interlinks among multiple temporal events using DPNs. Object temporal events are detected and labeled using Gaussian Mixture Models with automatic model order selection. A DML-HMM is built using Schwarz's Bayesian Information Criterion based factorisation resulting in its topology being intrinsically determined by the underlying causality and temporal order among different object events. Our experiments demonstrate that its performance on modelling group activities in a noisy outdoor scene is superior compared to that of a Multi-Observation Hidden Markov Model (MOHMM), a Parallel Hidden Markov Model (PaHMM) and a Coupled Hidden Markov Model (CHMM).","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125939096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 303
Combinatorial constraints on multiple projections of set points 集点的多个投影的组合约束
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238459
Tomáš Werner
Multiple projections of a scene cannot be arbitrary, the allowed configurations being given by matching constraints. This paper presents new matching constraints on multiple projections of a rigid point set by uncalibrated cameras, obtained by formulation in the oriented projective rather than projective geometry. They follow from consistency of orientations of camera rays and from the fact that the scene is the affine rather that projective space. For their non-parametric nature, we call them combinatorial. The constraints are derived in a unified theoretical framework using the theory of oriented matroids. For example, we present constraints on 4 point correspondences for 2D camera resectioning, on 3 correspondences in two 1D cameras, and on 4 correspondences in two 2D cameras.
一个场景的多个投影不能是任意的,允许的配置由匹配约束给出。本文提出了一种新的匹配约束,该约束是由定向射影几何而非射影几何公式得到的。它们遵循相机光线方向的一致性以及场景是仿射空间而不是投影空间的事实。由于它们的非参数性质,我们称它们为组合的。利用定向拟阵理论,在统一的理论框架下导出了约束条件。例如,我们提出了2D相机切片的4点对应约束,两台1D相机的3点对应约束,以及两台2D相机的4点对应约束。
{"title":"Combinatorial constraints on multiple projections of set points","authors":"Tomáš Werner","doi":"10.1109/ICCV.2003.1238459","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238459","url":null,"abstract":"Multiple projections of a scene cannot be arbitrary, the allowed configurations being given by matching constraints. This paper presents new matching constraints on multiple projections of a rigid point set by uncalibrated cameras, obtained by formulation in the oriented projective rather than projective geometry. They follow from consistency of orientations of camera rays and from the fact that the scene is the affine rather that projective space. For their non-parametric nature, we call them combinatorial. The constraints are derived in a unified theoretical framework using the theory of oriented matroids. For example, we present constraints on 4 point correspondences for 2D camera resectioning, on 3 correspondences in two 1D cameras, and on 4 correspondences in two 2D cameras.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124803512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Phenomenological eigenfunctions for image irradiance 图像辐照度的现象学特征函数
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238398
Peter Nillius, J. Eklundh
We present a framework for calculating low-dimensional bases to represent image irradiance from surfaces with isotropic reflectance under arbitrary illumination. By representing the illumination and the bidirectional reflectance distribution function (BRDF) in frequency space, a model for the image irradiance is derived. This model is then reduced in dimensionality by analytically constructing the principal component basis for all images given the variations in both the illumination and the surface material. The principal component basis are constructed in such a way that all the symmetries (Helmholtz reciprocity and isotropy) of the BRDF are preserved in the basis functions. Using the framework we calculate a basis using a database of natural illumination and the CURET database containing BRDFs of real world surface materials.
我们提出了一种计算低维基底的框架,以表示任意照明下具有各向同性反射率的表面的图像辐照度。通过在频率空间中表示光照和双向反射分布函数(BRDF),推导了图像辐照度的模型。然后,在给定光照和表面材料变化的情况下,通过分析构造所有图像的主成分基,该模型的维数被降低。主成分基的构造方式使得BRDF的所有对称性(亥姆霍兹互易性和各向同性)都保留在基函数中。使用该框架,我们使用自然光照数据库和包含真实世界表面材料brdf的CURET数据库来计算基础。
{"title":"Phenomenological eigenfunctions for image irradiance","authors":"Peter Nillius, J. Eklundh","doi":"10.1109/ICCV.2003.1238398","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238398","url":null,"abstract":"We present a framework for calculating low-dimensional bases to represent image irradiance from surfaces with isotropic reflectance under arbitrary illumination. By representing the illumination and the bidirectional reflectance distribution function (BRDF) in frequency space, a model for the image irradiance is derived. This model is then reduced in dimensionality by analytically constructing the principal component basis for all images given the variations in both the illumination and the surface material. The principal component basis are constructed in such a way that all the symmetries (Helmholtz reciprocity and isotropy) of the BRDF are preserved in the basis functions. Using the framework we calculate a basis using a database of natural illumination and the CURET database containing BRDFs of real world surface materials.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127286419","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Spectral partitioning for structure from motion 从运动中划分结构的光谱
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238457
Drew Steedly, Irfan Essa, F. Dellaert
We propose a spectral partitioning approach for large-scale optimization problems, specifically structure from motion. In structure from motion, partitioning methods reduce the problem into smaller and better conditioned subproblems which can be efficiently optimized. Our partitioning method uses only the Hessian of the reprojection error and its eigenvector. We show that partitioned systems that preserve the eigenvectors corresponding to small eigenvalues result in lower residual error when optimized. We create partitions by clustering the entries of the eigenvectors of the Hessian corresponding to small eigenvalues. This is a more general technique than relying on domain knowledge and heuristics such as bottom-up structure from motion approaches. Simultaneously, it takes advantage of more information than generic matrix partitioning algorithms.
我们提出了一种用于大规模优化问题的谱划分方法,特别是运动结构优化问题。在从运动到结构的问题中,划分方法将问题分解为更小、条件更好的子问题,可以有效地进行优化。我们的划分方法只使用重投影误差及其特征向量的Hessian。我们证明了保留小特征值对应的特征向量的分割系统在优化时可以得到较小的残差。我们通过聚类Hessian特征向量对应于小特征值的条目来创建分区。这是一种比依赖领域知识和启发式(如运动方法的自下而上结构)更通用的技术。同时,它比一般的矩阵划分算法利用了更多的信息。
{"title":"Spectral partitioning for structure from motion","authors":"Drew Steedly, Irfan Essa, F. Dellaert","doi":"10.1109/ICCV.2003.1238457","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238457","url":null,"abstract":"We propose a spectral partitioning approach for large-scale optimization problems, specifically structure from motion. In structure from motion, partitioning methods reduce the problem into smaller and better conditioned subproblems which can be efficiently optimized. Our partitioning method uses only the Hessian of the reprojection error and its eigenvector. We show that partitioned systems that preserve the eigenvectors corresponding to small eigenvalues result in lower residual error when optimized. We create partitions by clustering the entries of the eigenvectors of the Hessian corresponding to small eigenvalues. This is a more general technique than relying on domain knowledge and heuristics such as bottom-up structure from motion approaches. Simultaneously, it takes advantage of more information than generic matrix partitioning algorithms.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126907404","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
On the use of marginal statistics of subband images 子带图像边缘统计的应用
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238381
J. Gluckman
A commonly used representation of a visual pattern is the set of marginal probability distributions of the output of a bank of filters (Gaussian, Laplacian, Gabor etc.). This representation has been used effectively for a variety of vision tasks including texture classification, texture synthesis, object detection and image retrieval. We examine the ability of this representation to discriminate between an arbitrary pair of visual stimuli. Examples of patterns are derived that provably possess the same marginal statistical properties, yet are "visually distinct." These results suggest the need for either employing a large and diverse filter bank or incorporating joint statistics in order to represent a large class of visual patterns.
视觉模式的常用表示是一组滤波器(高斯、拉普拉斯、Gabor等)输出的边际概率分布的集合。该表示已被有效地用于多种视觉任务,包括纹理分类、纹理合成、目标检测和图像检索。我们检验这种表征区分任意一对视觉刺激的能力。模式的例子可以证明具有相同的边际统计特性,但在“视觉上是不同的”。这些结果表明,需要采用一个大的和多样化的过滤器组或合并联合统计,以表示一个大类的视觉模式。
{"title":"On the use of marginal statistics of subband images","authors":"J. Gluckman","doi":"10.1109/ICCV.2003.1238381","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238381","url":null,"abstract":"A commonly used representation of a visual pattern is the set of marginal probability distributions of the output of a bank of filters (Gaussian, Laplacian, Gabor etc.). This representation has been used effectively for a variety of vision tasks including texture classification, texture synthesis, object detection and image retrieval. We examine the ability of this representation to discriminate between an arbitrary pair of visual stimuli. Examples of patterns are derived that provably possess the same marginal statistical properties, yet are \"visually distinct.\" These results suggest the need for either employing a large and diverse filter bank or incorporating joint statistics in order to represent a large class of visual patterns.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126740356","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Video input driven animation (VIDA) 视频输入驱动动画(VIDA)
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238319
Mengqin Sun, A. Jepson, E. Fiume
There are many challenges associated with the integration of synthetic and real imagery. One particularly difficult problem is the automatic extraction of salient parameters of natural phenomena in real video footage for subsequent application to synthetic objects. We can ensure that the hair and clothing of a synthetic actor placed in a meadow of swaying grass will move consistently with the wind that moved that grass. The video footage can be seen as a controller for the motion of synthetic features, a concept we call video input driven animation (VIDA). We propose a schema that analyzes an input video sequence, extracts parameters from the motion of objects in the video, and uses this information to drive the motion of synthetic objects. To validate the principles of VIDA, we approximate the inverse problem to harmonic oscillation, which we use to extract parameters of wind and of regular water waves. We observe the effect of wind on a tree in a video, estimate wind speed parameters from its motion, and then use this to make synthetic objects move. We also extract water elevation parameters from the observed motion of boats and apply the resulting water waves to synthetic boats.
合成图像和真实图像的整合存在许多挑战。一个特别困难的问题是自动提取真实视频片段中自然现象的显著参数,以便随后应用于合成对象。我们可以确保将一个合成演员的头发和衣服放置在一片摇摆的草地上,并随着风移动草而移动。视频片段可以看作是合成特征运动的控制器,我们称之为视频输入驱动动画(VIDA)。我们提出了一种分析输入视频序列的模式,从视频中对象的运动中提取参数,并使用这些信息来驱动合成对象的运动。为了验证VIDA的原理,我们将反问题近似为谐波振荡,并利用它来提取风和规则水波的参数。我们在视频中观察风对树的影响,从它的运动中估计风速参数,然后用它来让合成物体移动。我们还从观测到的船只运动中提取水面高度参数,并将所得水波应用于合成船只。
{"title":"Video input driven animation (VIDA)","authors":"Mengqin Sun, A. Jepson, E. Fiume","doi":"10.1109/ICCV.2003.1238319","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238319","url":null,"abstract":"There are many challenges associated with the integration of synthetic and real imagery. One particularly difficult problem is the automatic extraction of salient parameters of natural phenomena in real video footage for subsequent application to synthetic objects. We can ensure that the hair and clothing of a synthetic actor placed in a meadow of swaying grass will move consistently with the wind that moved that grass. The video footage can be seen as a controller for the motion of synthetic features, a concept we call video input driven animation (VIDA). We propose a schema that analyzes an input video sequence, extracts parameters from the motion of objects in the video, and uses this information to drive the motion of synthetic objects. To validate the principles of VIDA, we approximate the inverse problem to harmonic oscillation, which we use to extract parameters of wind and of regular water waves. We observe the effect of wind on a tree in a video, estimate wind speed parameters from its motion, and then use this to make synthetic objects move. We also extract water elevation parameters from the observed motion of boats and apply the resulting water waves to synthetic boats.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126255258","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Catadioptric camera calibration using geometric invariants 反射相机几何不变量标定
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238647
Xianghua Ying, Zhanyi Hu
Central catadioptric cameras are imaging devices that use mirrors to enhance the field of view while preserving a single effective viewpoint. In this paper, we propose a novel method for the calibration of central catadioptric cameras using geometric invariants. Lines in space are projected into conics in the catadioptric image plane as well as spheres in space. We proved that the projection of a line can provide three invariants whereas the projection of a sphere can provide two. From these invariants, constraint equations for the intrinsic parameters of catadioptric camera are derived. Therefore, there are two variants of this novel method. The first one uses the projections of lines and the second one uses the projections of spheres. In general, the projections of two lines or three spheres are sufficient to achieve the catadioptric camera calibration. One important observation in this paper is that the method based on the projections of spheres is more robust and has higher accuracy than that using the projections of lines. The performances of our method are demonstrated by the results of simulations and experiments with real images.
中央反射相机是一种成像设备,它使用反射镜来增强视野,同时保持单一有效的视点。本文提出了一种利用几何不变量标定中央反射相机的新方法。空间中的直线在反射成像平面上被投影成圆锥,在空间中被投影成球体。我们证明了直线的投影可以提供三个不变量,而球面的投影可以提供两个不变量。从这些不变量出发,导出了反射相机固有参数的约束方程。因此,这种新方法有两种变体。第一个用直线的投影,第二个用球体的投影。一般情况下,两条直线或三个球面的投影就足以实现反射相机的标定。本文的一个重要观察结果是,基于球面投影的方法比基于直线投影的方法具有更高的鲁棒性和精度。仿真和真实图像的实验结果证明了该方法的有效性。
{"title":"Catadioptric camera calibration using geometric invariants","authors":"Xianghua Ying, Zhanyi Hu","doi":"10.1109/ICCV.2003.1238647","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238647","url":null,"abstract":"Central catadioptric cameras are imaging devices that use mirrors to enhance the field of view while preserving a single effective viewpoint. In this paper, we propose a novel method for the calibration of central catadioptric cameras using geometric invariants. Lines in space are projected into conics in the catadioptric image plane as well as spheres in space. We proved that the projection of a line can provide three invariants whereas the projection of a sphere can provide two. From these invariants, constraint equations for the intrinsic parameters of catadioptric camera are derived. Therefore, there are two variants of this novel method. The first one uses the projections of lines and the second one uses the projections of spheres. In general, the projections of two lines or three spheres are sufficient to achieve the catadioptric camera calibration. One important observation in this paper is that the method based on the projections of spheres is more robust and has higher accuracy than that using the projections of lines. The performances of our method are demonstrated by the results of simulations and experiments with real images.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"1938 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128778825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 172
Calibration of a hybrid camera network 混合摄像机网络的标定
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238330
Xilin Chen, Jie Yang, A. Waibel
Visual surveillance using a camera network has imposed new challenges to camera calibration. An essential problem is that a large number of cameras may not have a common field of view or even be synchronized well. We propose to use a hybrid camera network that consists of catadioptric and perspective cameras for a visual surveillance task. The relations between multiple views of a scene captured from different cameras can be then calibrated under the catadioptric camera's coordinate system. We address the important issue of how to calibrate the hybrid camera network. We calibrate the hybrid camera network in three steps. First, we calibrate the catadioptric camera using only the vanishing points. In order to reduce computational complexity, we calibrate the camera without the mirror first and then calibrate the catadioptric camera system. Second, we determine 3D positions of some points using as few as two spatial parallel lines and some equidistance points. Finally, we calibrate other perspective cameras based on these known spatial points.
基于摄像机网络的视觉监控对摄像机标定提出了新的挑战。一个重要的问题是,大量的相机可能没有一个共同的视野,甚至不能很好地同步。我们建议使用混合摄像机网络,由反射和透视摄像机组成的视觉监控任务。从不同相机拍摄的场景的多个视图之间的关系可以在反射相机的坐标系下进行校准。我们解决了如何校准混合摄像机网络的重要问题。我们分三步校准混合摄像机网络。首先,我们只使用消失点校准反射相机。为了降低计算复杂度,我们先对无反射镜的摄像机进行标定,然后再对反射镜摄像机系统进行标定。其次,我们使用少至两条空间平行线和一些等距点来确定一些点的三维位置。最后,我们根据这些已知的空间点校准其他透视相机。
{"title":"Calibration of a hybrid camera network","authors":"Xilin Chen, Jie Yang, A. Waibel","doi":"10.1109/ICCV.2003.1238330","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238330","url":null,"abstract":"Visual surveillance using a camera network has imposed new challenges to camera calibration. An essential problem is that a large number of cameras may not have a common field of view or even be synchronized well. We propose to use a hybrid camera network that consists of catadioptric and perspective cameras for a visual surveillance task. The relations between multiple views of a scene captured from different cameras can be then calibrated under the catadioptric camera's coordinate system. We address the important issue of how to calibrate the hybrid camera network. We calibrate the hybrid camera network in three steps. First, we calibrate the catadioptric camera using only the vanishing points. In order to reduce computational complexity, we calibrate the camera without the mirror first and then calibrate the catadioptric camera system. Second, we determine 3D positions of some points using as few as two spatial parallel lines and some equidistance points. Finally, we calibrate other perspective cameras based on these known spatial points.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129036718","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
Markov-based failure prediction for human motion analysis 基于马尔可夫的人体运动分析失效预测
Pub Date : 2003-10-13 DOI: 10.1109/ICCV.2003.1238638
S. Dockstader, Nikita S. Imennov, A. Tekalp
This paper presents a new method of detecting and predicting motion tracking failures with applications in human motion and gait analysis. We define a tracking failure as an event and describe its temporal characteristics using a hidden Markov model (HMM). This stochastic model is trained using previous examples of tracking failures. We derive vector observations for the HMM using the noise covariance matrices characterizing a tracked, 3D structural model of the human body. We show a causal relationship between the conditional output probability of the HMM, as transformed using a logarithmic mapping function, and impending tracking failures. Results are illustrated on several multi-view sequences of complex human motion.
本文提出了一种检测和预测运动跟踪故障的新方法,并在人体运动和步态分析中应用。我们将跟踪故障定义为事件,并使用隐马尔可夫模型(HMM)描述其时间特征。这个随机模型是用之前跟踪失败的例子来训练的。我们使用表征跟踪的人体三维结构模型的噪声协方差矩阵来推导HMM的矢量观测值。我们展示了HMM的条件输出概率(使用对数映射函数转换)与即将发生的跟踪故障之间的因果关系。结果说明了几个多视图序列的复杂人体运动。
{"title":"Markov-based failure prediction for human motion analysis","authors":"S. Dockstader, Nikita S. Imennov, A. Tekalp","doi":"10.1109/ICCV.2003.1238638","DOIUrl":"https://doi.org/10.1109/ICCV.2003.1238638","url":null,"abstract":"This paper presents a new method of detecting and predicting motion tracking failures with applications in human motion and gait analysis. We define a tracking failure as an event and describe its temporal characteristics using a hidden Markov model (HMM). This stochastic model is trained using previous examples of tracking failures. We derive vector observations for the HMM using the noise covariance matrices characterizing a tracked, 3D structural model of the human body. We show a causal relationship between the conditional output probability of the HMM, as transformed using a logarithmic mapping function, and impending tracking failures. Results are illustrated on several multi-view sequences of complex human motion.","PeriodicalId":131580,"journal":{"name":"Proceedings Ninth IEEE International Conference on Computer Vision","volume":"56 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-10-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116576303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
Proceedings Ninth IEEE International Conference on Computer Vision
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1