首页 > 最新文献

2003 Conference on Computer Vision and Pattern Recognition Workshop最新文献

英文 中文
MDL-based Genetic Programming for Object Detection 基于mdl的目标检测遗传规划
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10062
Yingqiang Lin, B. Bhanu
In this paper, genetic programming (GP) is applied to synthesize composite operators from primitive operators and primitive features for object detection. To improve the efficiency of GP, smart crossover, smart mutation and a public library are proposed to identify and keep the effective components of composite operators. To prevent code bloat and avoid severe restriction on the GP search, a MDL-based fitness function is designed to incorporate the size of composite operator into the fitness evaluation process. The experiments with real synthetic aperture radar (SAR) images show that compared to normal GP, GP algorithm proposed here finds effective composite operators more quickly.
本文将遗传规划(GP)应用于原语算子和原语特征合成复合算子进行目标检测。为了提高遗传算法的效率,提出了智能交叉、智能突变和公共库来识别和保留复合算子的有效成分。为了防止代码膨胀和避免对GP搜索的严重限制,设计了一个基于mdl的适应度函数,将复合算子的大小纳入适应度评估过程。在真实合成孔径雷达(SAR)图像上的实验表明,与常规GP算法相比,本文提出的GP算法能更快地找到有效的复合算子。
{"title":"MDL-based Genetic Programming for Object Detection","authors":"Yingqiang Lin, B. Bhanu","doi":"10.1109/CVPRW.2003.10062","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10062","url":null,"abstract":"In this paper, genetic programming (GP) is applied to synthesize composite operators from primitive operators and primitive features for object detection. To improve the efficiency of GP, smart crossover, smart mutation and a public library are proposed to identify and keep the effective components of composite operators. To prevent code bloat and avoid severe restriction on the GP search, a MDL-based fitness function is designed to incorporate the size of composite operator into the fitness evaluation process. The experiments with real synthetic aperture radar (SAR) images show that compared to normal GP, GP algorithm proposed here finds effective composite operators more quickly.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131088682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Time-dependent HMMs for visual intrusion detection 基于时间的hmm视觉入侵检测
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10035
Vera M. Kettnaker
We propose a new Hidden Markov Model with time-dependent states. Estimation of this model is shown to be as fast and easy as the estimation of regular HMMs. We demonstrate the usefulness and feasibility of such time-dependent HMMs with an application in which illegitimate access to personnel-only rooms in airports etc. can be distinguished from access by legitimate personnel, based on differences in the time of access or differences in the motion trajectories.
提出了一种具有时间依赖状态的隐马尔可夫模型。该模型的估计与常规hmm的估计一样快速简便。我们通过一个应用程序证明了这种时间依赖hmm的有效性和可行性,在该应用程序中,根据进入时间的差异或运动轨迹的差异,可以将非法进入机场等人员专用房间与合法人员的进入区分开来。
{"title":"Time-dependent HMMs for visual intrusion detection","authors":"Vera M. Kettnaker","doi":"10.1109/CVPRW.2003.10035","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10035","url":null,"abstract":"We propose a new Hidden Markov Model with time-dependent states. Estimation of this model is shown to be as fast and easy as the estimation of regular HMMs. We demonstrate the usefulness and feasibility of such time-dependent HMMs with an application in which illegitimate access to personnel-only rooms in airports etc. can be distinguished from access by legitimate personnel, based on differences in the time of access or differences in the motion trajectories.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115404108","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Omnidirectional Distributed Vision System for a Team of Heterogeneous Robots 面向异构机器人团队的全向分布式视觉系统
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10068
E. Menegatti, A. Scarpa, Dario Massarin, Enrico Ros, E. Pagello
This paper presents a system designed to cooperatively track and share the information about moving objects using a multi-robot team. Every robot of the team is fitted with a different omnidirectional vision system running at different frame rates. The information gathered from every robot is broadcast to all the other robots and every robot fuses its own measurements with the information received from the teammates, building its own "vision of the world". The cooperation of the vision sensors enhances the capabilities of the single vision sensor. This work was implemented in the RoboCup domain, using our team of heterogeneous robot, but the approach is very general and can be used in any application where a team of robot has to track multiple objects. The system is designed to work with heterogeneous vision systems both in the camera design and in computational resources. Experiments in real game scenarios are presented.
本文提出了一个多机器人团队协同跟踪和共享运动目标信息的系统。团队的每个机器人都配备了不同的全方位视觉系统,以不同的帧速率运行。从每个机器人收集到的信息被广播给所有其他机器人,每个机器人将自己的测量结果与从队友那里接收到的信息融合在一起,构建自己的“世界视野”。视觉传感器的协同工作提高了单个视觉传感器的性能。这项工作是在RoboCup领域实现的,使用我们的异构机器人团队,但该方法非常通用,可以用于机器人团队必须跟踪多个对象的任何应用程序。该系统在相机设计和计算资源两方面都能与异构视觉系统一起工作。给出了在真实游戏场景下的实验。
{"title":"Omnidirectional Distributed Vision System for a Team of Heterogeneous Robots","authors":"E. Menegatti, A. Scarpa, Dario Massarin, Enrico Ros, E. Pagello","doi":"10.1109/CVPRW.2003.10068","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10068","url":null,"abstract":"This paper presents a system designed to cooperatively track and share the information about moving objects using a multi-robot team. Every robot of the team is fitted with a different omnidirectional vision system running at different frame rates. The information gathered from every robot is broadcast to all the other robots and every robot fuses its own measurements with the information received from the teammates, building its own \"vision of the world\". The cooperation of the vision sensors enhances the capabilities of the single vision sensor. This work was implemented in the RoboCup domain, using our team of heterogeneous robot, but the approach is very general and can be used in any application where a team of robot has to track multiple objects. The system is designed to work with heterogeneous vision systems both in the camera design and in computational resources. Experiments in real game scenarios are presented.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116074727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Low-Overlap Range Image Registration for Archaeological Applications 用于考古应用的低重叠范围图像配准
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10011
Luciano Silva, O. Bellon, K. Boyer, P. Gotardo
In digital Archaeology, the 3D modeling of physical objects from range views is an important issue. Generally, the applications demand a great number of views to create a precise 3D model through a registration process. Most range image registration techniques are based on variants of the ICP (Iterative Closest Point) algorithm. The ICP algorithm has two main drawbacks: the possibility of convergence to a local minimum, and the need to prealign the images. Genetic Algorithms (GAs) were recently applied to range image registration providing good convergence results without the constraints observed in the ICP approaches. To improve range image registration, we explore the use of GAs and develop a novel approach that combines a GA with hillclimbing heuristics (GH). The experimental results show that our method is effective in aligning low overlap views and yield more accurate registration results than either ICP or standard GA approaches. Our method is highly advantageous in archaeological applications, where it is necessary to reduce the number of views to be aligned because data acquisition is expensive and also to minimize error accumulation in the 3D model. We also present a new measure of surface interpenetration with which to evaluate the registration and prove its utility with experimental results.
在数字考古学中,从距离视图对物理对象进行三维建模是一个重要的问题。通常,应用程序需要大量的视图来通过注册过程创建精确的3D模型。大多数距离图像配准技术是基于ICP(迭代最近点)算法的变体。ICP算法有两个主要缺点:收敛到局部最小值的可能性,以及需要对图像进行预对齐。遗传算法(GAs)最近被应用于距离图像配准,提供了良好的收敛结果,而没有在ICP方法中观察到的约束。为了改善距离图像配准,我们探索了遗传算法的使用,并开发了一种将遗传算法与爬坡启发式(GH)相结合的新方法。实验结果表明,该方法可以有效地对准低重叠视图,并且比ICP或标准GA方法获得更准确的配准结果。我们的方法在考古应用中非常有优势,因为数据采集非常昂贵,需要减少要对齐的视图数量,并且还可以最大限度地减少3D模型中的误差积累。我们还提出了一种新的表面互穿测量方法来评价配准,并用实验结果证明了它的实用性。
{"title":"Low-Overlap Range Image Registration for Archaeological Applications","authors":"Luciano Silva, O. Bellon, K. Boyer, P. Gotardo","doi":"10.1109/CVPRW.2003.10011","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10011","url":null,"abstract":"In digital Archaeology, the 3D modeling of physical objects from range views is an important issue. Generally, the applications demand a great number of views to create a precise 3D model through a registration process. Most range image registration techniques are based on variants of the ICP (Iterative Closest Point) algorithm. The ICP algorithm has two main drawbacks: the possibility of convergence to a local minimum, and the need to prealign the images. Genetic Algorithms (GAs) were recently applied to range image registration providing good convergence results without the constraints observed in the ICP approaches. To improve range image registration, we explore the use of GAs and develop a novel approach that combines a GA with hillclimbing heuristics (GH). The experimental results show that our method is effective in aligning low overlap views and yield more accurate registration results than either ICP or standard GA approaches. Our method is highly advantageous in archaeological applications, where it is necessary to reduce the number of views to be aligned because data acquisition is expensive and also to minimize error accumulation in the 3D model. We also present a new measure of surface interpenetration with which to evaluate the registration and prove its utility with experimental results.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122807873","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Using Corner Feature Correspondences to Rank Word Images by Similarity 利用角点特征对应对词图像进行相似度排序
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10021
Jamie L. Rothfeder, Shaolei Feng, T. Rath
Libraries contain enormous amounts of handwritten historical documents which cannot be made available on-line because they do not have a searchable index. The wordspotting idea has previously been proposed as a solution to creating indexes for such documents and collections by matching word images. In this paper we present an algorithm which compares whole word-images based on their appearance. This algorithm recovers correspondences of points of interest in two images, and then uses these correspondences to construct a similarity measure. This similarity measure can then be used to rank word-images in order of their closeness to a querying image. We achieved an average precision of 62.57% on a set of 2372 images of reasonable quality and an average precision of 15.49% on a set of 3262 images from documents of poor quality that are even hard to read for humans.
图书馆中有大量手写的历史文献,由于没有可搜索的索引,这些文献不能在网上提供。单词点出的想法以前曾被提出作为一种解决方案,通过匹配单词图像为此类文档和集合创建索引。本文提出了一种基于外观的全字图像比较算法。该算法恢复两幅图像中感兴趣点的对应关系,然后利用这些对应关系构建相似度度量。然后,这种相似度度量可以用来对单词图像按照它们与查询图像的接近程度进行排序。我们在一组2372张质量合理的图像上实现了62.57%的平均精度,在一组3262张质量较差的图像上实现了15.49%的平均精度,这些图像甚至很难被人类阅读。
{"title":"Using Corner Feature Correspondences to Rank Word Images by Similarity","authors":"Jamie L. Rothfeder, Shaolei Feng, T. Rath","doi":"10.1109/CVPRW.2003.10021","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10021","url":null,"abstract":"Libraries contain enormous amounts of handwritten historical documents which cannot be made available on-line because they do not have a searchable index. The wordspotting idea has previously been proposed as a solution to creating indexes for such documents and collections by matching word images. In this paper we present an algorithm which compares whole word-images based on their appearance. This algorithm recovers correspondences of points of interest in two images, and then uses these correspondences to construct a similarity measure. This similarity measure can then be used to rank word-images in order of their closeness to a querying image. We achieved an average precision of 62.57% on a set of 2372 images of reasonable quality and an average precision of 15.49% on a set of 3262 images from documents of poor quality that are even hard to read for humans.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"369 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122774397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 93
A Study on Bayes Feature Fusion for Image Classification 基于贝叶斯特征融合的图像分类研究
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10090
Xiaojin Shi, R. Manduchi
We consider here the problem of image classification when more than one visual feature are available. In these cases, Bayes fusion offers an attractive solution by combining the results of different classifiers (one classifier per feature). This is a general form of the so-called "naive Bayes" approach. Analyzing the performance of Bayes fusion with respect to a Bayesian classifier over the joint feature distribution, however, is tricky. On the one hand, it is well-known that the latter has lower bias than the former, unless the features are conditionally independent, in which case the two coincide. On the other hand, as noted by Friedman, the low variance associated with naive Bayes estimation may dramatically mitigate the effect of its bias. In this paper, we attempt to assess the tradeoff between these two factors by means of experimental tests on two image data sets using color and texture features. Our results suggest that (1) the difference between the correct classification rates using Bayes fusion and using the joint feature distribution is a function of the conditional dependence of the features (measured in terms of mutual information), however: (2) for small training data size, Bayes fusion performs almost as well as the classifier on the joint distribution.
我们在这里考虑了当有多个视觉特征可用时的图像分类问题。在这些情况下,贝叶斯融合通过组合不同分类器的结果(每个特征一个分类器)提供了一个有吸引力的解决方案。这是所谓的“朴素贝叶斯”方法的一般形式。然而,在联合特征分布上分析贝叶斯分类器的贝叶斯融合性能是很棘手的。一方面,众所周知,后者比前者具有更低的偏差,除非特征是条件独立的,在这种情况下,两者是重合的。另一方面,正如弗里德曼所指出的,与朴素贝叶斯估计相关的低方差可能会显著减轻其偏差的影响。在本文中,我们试图通过使用颜色和纹理特征对两个图像数据集进行实验测试来评估这两个因素之间的权衡。我们的研究结果表明:(1)使用贝叶斯融合和使用联合特征分布的正确分类率之间的差异是特征的条件依赖性的函数(以互信息来衡量),然而:(2)对于较小的训练数据规模,贝叶斯融合的表现几乎与分类器在联合分布上的表现一样好。
{"title":"A Study on Bayes Feature Fusion for Image Classification","authors":"Xiaojin Shi, R. Manduchi","doi":"10.1109/CVPRW.2003.10090","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10090","url":null,"abstract":"We consider here the problem of image classification when more than one visual feature are available. In these cases, Bayes fusion offers an attractive solution by combining the results of different classifiers (one classifier per feature). This is a general form of the so-called \"naive Bayes\" approach. Analyzing the performance of Bayes fusion with respect to a Bayesian classifier over the joint feature distribution, however, is tricky. On the one hand, it is well-known that the latter has lower bias than the former, unless the features are conditionally independent, in which case the two coincide. On the other hand, as noted by Friedman, the low variance associated with naive Bayes estimation may dramatically mitigate the effect of its bias. In this paper, we attempt to assess the tradeoff between these two factors by means of experimental tests on two image data sets using color and texture features. Our results suggest that (1) the difference between the correct classification rates using Bayes fusion and using the joint feature distribution is a function of the conditional dependence of the features (measured in terms of mutual information), however: (2) for small training data size, Bayes fusion performs almost as well as the classifier on the joint distribution.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116943711","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 57
Parametric Subpixel Matchpoint Recovery with Uncertainty Estimation: A Statistical Approach 不确定估计的参数亚像素匹配点恢复:一种统计方法
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10091
R. M. Steele, C. Jaynes
We present a novel matchpoint acquisition method capable of producing accurate correspondences at subpixel precision. Given the known representation of the point to be matched, such as a projected fiducial in a structured light system, the method estimates the fiducial location and its expected uncertainty. Improved matchpoint precision has application in a number of calibration tasks, and uncertainty estimates can be used to significantly improve overall calibration results. A simple parametric model captures the relationship between the known fiducial and its corresponding position, shape, and intensity on the image plane. For each match-point pair, these unknown model parameters are recovered using maximum likelihood estimation to determine a sub-pixel center for the fiducial. The uncertainty of the match-point center is estimated by performing forward error analysis on the expected image noise. Uncertainty estimates used in conjunction with the accurate matchpoints can improve calibration accuracy for multi-view systems.
我们提出了一种新的匹配点获取方法,能够在亚像素精度下产生准确的对应。给定要匹配点的已知表示,例如结构光系统中的投影基准,该方法估计基准位置及其预期的不确定性。改进的匹配点精度已应用于许多校准任务,不确定度估计可用于显着改善整体校准结果。一个简单的参数化模型捕获了已知基准与其在图像平面上相应的位置、形状和强度之间的关系。对于每个匹配点对,使用最大似然估计来恢复这些未知模型参数,以确定基准的亚像素中心。通过对期望图像噪声进行前向误差分析来估计匹配点中心的不确定性。结合精确匹配点的不确定度估计可以提高多视点系统的校准精度。
{"title":"Parametric Subpixel Matchpoint Recovery with Uncertainty Estimation: A Statistical Approach","authors":"R. M. Steele, C. Jaynes","doi":"10.1109/CVPRW.2003.10091","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10091","url":null,"abstract":"We present a novel matchpoint acquisition method capable of producing accurate correspondences at subpixel precision. Given the known representation of the point to be matched, such as a projected fiducial in a structured light system, the method estimates the fiducial location and its expected uncertainty. Improved matchpoint precision has application in a number of calibration tasks, and uncertainty estimates can be used to significantly improve overall calibration results. A simple parametric model captures the relationship between the known fiducial and its corresponding position, shape, and intensity on the image plane. For each match-point pair, these unknown model parameters are recovered using maximum likelihood estimation to determine a sub-pixel center for the fiducial. The uncertainty of the match-point center is estimated by performing forward error analysis on the expected image noise. Uncertainty estimates used in conjunction with the accurate matchpoints can improve calibration accuracy for multi-view systems.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128141634","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Statistical Error Propagation in 3D Modeling From Monocular Video 单目视频三维建模中的统计误差传播
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10092
A. Roy-Chowdhury, R. Chellappa
A significant portion of recent research in computer vision has focused on issues related to sensitivity and robustness of existing techniques. In this paper, we study the classical structure from motion problem and analyze how the statistics representing the quality of the input video propagates through the reconstruction algorithm and affects the quality of the output reconstruction. Specifically, we show that it is possible to derive analytical expressions of the first and second order statistics (bias and error covariance) of the solution as a function of the statistics of the input. We concentrate on the case of reconstruction from a monocular video, where the small baseline makes any algorithm very susceptible to noise in the motion estimates from the video sequence. We derive an expression relating the error covariance of the reconstruction to the error covariance of the feature tracks in the input video. This is done using the implicit function theorem of real analysis and does not require strong statistical assumptions. Next, we prove that the 3D reconstruction is statistically biased, derive an expression for it and show that it is numerically significant. Combining these two results, we also establish a new bound on the minimum error in the depth reconstruction. We present the numerical significance of these analytical results on real video data.
近年来,计算机视觉的研究主要集中在现有技术的灵敏度和鲁棒性方面。本文从运动问题出发研究经典结构,分析了代表输入视频质量的统计量如何通过重构算法传播并影响输出重构的质量。具体地说,我们表明有可能推导出解的一阶和二阶统计量(偏差和误差协方差)作为输入统计量的函数的解析表达式。我们专注于单目视频重建的情况,其中小基线使得任何算法都很容易受到视频序列运动估计中的噪声的影响。我们推导了重构误差协方差与输入视频中特征轨迹误差协方差的关系式。这是使用实分析的隐函数定理完成的,不需要很强的统计假设。接下来,我们证明了三维重建在统计上是有偏差的,推导了它的表达式,并证明了它在数字上是显著的。结合这两个结果,我们还建立了深度重建最小误差的新界限。我们给出了这些分析结果在实际视频数据上的数值意义。
{"title":"Statistical Error Propagation in 3D Modeling From Monocular Video","authors":"A. Roy-Chowdhury, R. Chellappa","doi":"10.1109/CVPRW.2003.10092","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10092","url":null,"abstract":"A significant portion of recent research in computer vision has focused on issues related to sensitivity and robustness of existing techniques. In this paper, we study the classical structure from motion problem and analyze how the statistics representing the quality of the input video propagates through the reconstruction algorithm and affects the quality of the output reconstruction. Specifically, we show that it is possible to derive analytical expressions of the first and second order statistics (bias and error covariance) of the solution as a function of the statistics of the input. We concentrate on the case of reconstruction from a monocular video, where the small baseline makes any algorithm very susceptible to noise in the motion estimates from the video sequence. We derive an expression relating the error covariance of the reconstruction to the error covariance of the feature tracks in the input video. This is done using the implicit function theorem of real analysis and does not require strong statistical assumptions. Next, we prove that the 3D reconstruction is statistically biased, derive an expression for it and show that it is numerically significant. Combining these two results, we also establish a new bound on the minimum error in the depth reconstruction. We present the numerical significance of these analytical results on real video data.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114539811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Statistical Models for Skin Detection 皮肤检测的统计模型
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10094
B. Jedynak, Huicheng Zheng, M. Daoudi
We consider a sequence of three models for skin detection built from a large collection of labelled images. Each model is a maximum entropy model with respect to constraints concerning marginal distributions. Our models are nested. The first model is well known from practitioners. Pixels are considered as independent. The second model is a Hidden Markov Model. It includes constraints that force smoothness of the solution. The third model is a first order model. The full color gradient is included. Parameter estimation as well as optimization cannot be tackled without approximations. We use thoroughly Bethe tree approximation of the pixel lattice. Within it , parameter estimation is eradicated and the belief propagation algorithm permits to obtain exact and fast solution for skin probability at pixel locations. We then assess the performance on the Compaq database.
我们考虑从大量标记图像中构建的三个皮肤检测模型序列。每个模型都是关于边际分布约束的最大熵模型。我们的模型是嵌套的。第一个模型为实践者所熟知。像素被认为是独立的。第二个模型是隐马尔可夫模型。它包括强制解决方案平滑的约束。第三个模型是一阶模型。包括完整的颜色渐变。参数估计和优化不能没有近似处理。我们彻底使用贝特树近似像素点阵。在该算法中,消除了参数估计,并采用信念传播算法,可以在像素位置精确、快速地求解皮肤概率。然后我们在康柏数据库上评估性能。
{"title":"Statistical Models for Skin Detection","authors":"B. Jedynak, Huicheng Zheng, M. Daoudi","doi":"10.1109/CVPRW.2003.10094","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10094","url":null,"abstract":"We consider a sequence of three models for skin detection built from a large collection of labelled images. Each model is a maximum entropy model with respect to constraints concerning marginal distributions. Our models are nested. The first model is well known from practitioners. Pixels are considered as independent. The second model is a Hidden Markov Model. It includes constraints that force smoothness of the solution. The third model is a first order model. The full color gradient is included. Parameter estimation as well as optimization cannot be tackled without approximations. We use thoroughly Bethe tree approximation of the pixel lattice. Within it , parameter estimation is eradicated and the belief propagation algorithm permits to obtain exact and fast solution for skin probability at pixel locations. We then assess the performance on the Compaq database.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"386 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123092958","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Archaeological Fragment Reconstruction Using Curve-Matching 基于曲线匹配的考古碎片重建
Pub Date : 2003-06-16 DOI: 10.1109/CVPRW.2003.10008
J. McBride, B. Kimia
We present a novel approach to the problem of puzzle solving as it relates to archaeological fragment reconstruction. We begin with a set of broken fragments. In the first stage, we compare every pair of fragments and use partial curve matching to find similar portions of their respective boundaries. Partial curve matching is typically a very difficult problem because the specification of the partial curves are highly unconstrained and curve matching is computationally expensive. To address the first problem, we only consider matches which begin at fragment corners and then use curve-matching with normalized energy to determine how far the match extends. We also reduce computational cost by employing a multi-scale approach. This allows us to quickly generate many possible matches at a coarse scale and only keep the best ones to be matched again at a finer scale. In the second stage, we take a rank-ordered list of pairwise matches to search for a globally optimal arrangement. The search is based on a best-first strategy which adds fragments with the highest pairwise affinity first, but then evaluates their confidence as part of the global solution by rewarding the formation of triple junctions which are dominant in archaeological puzzles. To prevent failure due to the inclusion of spurious matches, we employ a standard beam-search to simultaneously expand on multiple solutions. Results on several cases are demonstrated.
我们提出了一种新的方法来解决难题,因为它涉及到考古碎片重建。我们从一组破碎的碎片开始。在第一阶段,我们比较每对片段,并使用部分曲线匹配来找到它们各自边界的相似部分。部分曲线匹配通常是一个非常困难的问题,因为部分曲线的规格是高度不受约束的,曲线匹配的计算成本很高。为了解决第一个问题,我们只考虑从碎片角开始的匹配,然后使用归一化能量的曲线匹配来确定匹配扩展的距离。我们还通过采用多尺度方法降低了计算成本。这使我们能够在粗略的尺度上快速生成许多可能的匹配,并只保留最好的匹配,以便在更精细的尺度上再次匹配。在第二阶段,我们采用一个成对匹配的排序列表来搜索全局最优排列。搜索基于最佳优先策略,即首先添加具有最高成对亲和力的碎片,然后通过奖励在考古谜题中占主导地位的三重连接的形成来评估它们作为全局解决方案的一部分的信心。为了防止由于包含虚假匹配而导致的失败,我们采用标准波束搜索来同时扩展多个解。对几个实例的结果进行了验证。
{"title":"Archaeological Fragment Reconstruction Using Curve-Matching","authors":"J. McBride, B. Kimia","doi":"10.1109/CVPRW.2003.10008","DOIUrl":"https://doi.org/10.1109/CVPRW.2003.10008","url":null,"abstract":"We present a novel approach to the problem of puzzle solving as it relates to archaeological fragment reconstruction. We begin with a set of broken fragments. In the first stage, we compare every pair of fragments and use partial curve matching to find similar portions of their respective boundaries. Partial curve matching is typically a very difficult problem because the specification of the partial curves are highly unconstrained and curve matching is computationally expensive. To address the first problem, we only consider matches which begin at fragment corners and then use curve-matching with normalized energy to determine how far the match extends. We also reduce computational cost by employing a multi-scale approach. This allows us to quickly generate many possible matches at a coarse scale and only keep the best ones to be matched again at a finer scale. In the second stage, we take a rank-ordered list of pairwise matches to search for a globally optimal arrangement. The search is based on a best-first strategy which adds fragments with the highest pairwise affinity first, but then evaluates their confidence as part of the global solution by rewarding the formation of triple junctions which are dominant in archaeological puzzles. To prevent failure due to the inclusion of spurious matches, we employ a standard beam-search to simultaneously expand on multiple solutions. Results on several cases are demonstrated.","PeriodicalId":121249,"journal":{"name":"2003 Conference on Computer Vision and Pattern Recognition Workshop","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2003-06-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131529503","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 88
期刊
2003 Conference on Computer Vision and Pattern Recognition Workshop
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1