首页 > 最新文献

2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops最新文献

英文 中文
A dual-layer estimator architecture for long-term localization 用于长期定位的双层估计器体系结构
Anastasios I. Mourikis, S. Roumeliotis
In this paper, we present a localization algorithm for estimating the 3D position and orientation (pose) of a moving vehicle based on visual and inertial measurements. The main advantage of the proposed method is that it provides precise pose estimates at low computational cost. This is achieved by introducing a two-layer estimation architecture that processes measurements based on their information content. Inertial measurements and feature tracks between consecutive images are processed locally in the first layer (multi-state-constraint Kalman filter) providing estimates for the motion of the vehicle at a high rate. The second layer comprises a bundle adjustment iterative estimator that operates intermittently so as to (i) reduce the effect of the linearization errors, and (ii) update the state estimates every time an area is re-visited and features are re-detected (loop closure). Through this process reliable state estimates are available continuously, while the estimation errors remain bounded during long-term operation. The performance of the developed system is demonstrated in large-scale experiments, involving a vehicle localizing within an urban area.
在本文中,我们提出了一种基于视觉和惯性测量估计移动车辆的三维位置和方向(姿态)的定位算法。该方法的主要优点是能够以较低的计算成本提供精确的姿态估计。这是通过引入基于信息内容处理度量的两层评估体系结构来实现的。惯性测量和连续图像之间的特征轨迹在第一层(多状态约束卡尔曼滤波)进行局部处理,以高速估计车辆的运动。第二层包括一个间歇运行的束调整迭代估计器,以便(i)减少线性化误差的影响,(ii)每次重新访问一个区域和重新检测特征时更新状态估计(环路关闭)。通过该过程,可以连续获得可靠的状态估计,而在长期运行过程中,估计误差保持有界。该系统的性能在大规模实验中得到了验证,其中包括在城市区域内进行车辆定位。
{"title":"A dual-layer estimator architecture for long-term localization","authors":"Anastasios I. Mourikis, S. Roumeliotis","doi":"10.1109/CVPRW.2008.4563131","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563131","url":null,"abstract":"In this paper, we present a localization algorithm for estimating the 3D position and orientation (pose) of a moving vehicle based on visual and inertial measurements. The main advantage of the proposed method is that it provides precise pose estimates at low computational cost. This is achieved by introducing a two-layer estimation architecture that processes measurements based on their information content. Inertial measurements and feature tracks between consecutive images are processed locally in the first layer (multi-state-constraint Kalman filter) providing estimates for the motion of the vehicle at a high rate. The second layer comprises a bundle adjustment iterative estimator that operates intermittently so as to (i) reduce the effect of the linearization errors, and (ii) update the state estimates every time an area is re-visited and features are re-detected (loop closure). Through this process reliable state estimates are available continuously, while the estimation errors remain bounded during long-term operation. The performance of the developed system is demonstrated in large-scale experiments, involving a vehicle localizing within an urban area.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124425174","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Mutual information computation and maximization using GPU 基于GPU的互信息计算与最大化
Yuping Lin, G. Medioni
We present a GPU implementation to compute both mutual information and its derivatives. Mutual information computation is a highly demanding process due to the enormous number of exponential computations. It is therefore the bottleneck in many image registration applications. However, we show that these computations are fully parallizable and can be efficiently ported onto the GPU architecture. Compared with the same CPU implementation running on a workstation level CPU, we reached a factor of 170 in computing mutual information, and a factor of 400 in computing its derivatives.
我们提出了一种计算互信息及其导数的GPU实现。互信息计算是一个要求很高的过程,因为它需要大量的指数计算。因此,它是许多图像配准应用中的瓶颈。然而,我们证明了这些计算是完全可并行的,可以有效地移植到GPU架构上。与在工作站级CPU上运行的相同CPU实现相比,我们在计算互信息方面达到了170倍,在计算其导数方面达到了400倍。
{"title":"Mutual information computation and maximization using GPU","authors":"Yuping Lin, G. Medioni","doi":"10.1109/CVPRW.2008.4563101","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563101","url":null,"abstract":"We present a GPU implementation to compute both mutual information and its derivatives. Mutual information computation is a highly demanding process due to the enormous number of exponential computations. It is therefore the bottleneck in many image registration applications. However, we show that these computations are fully parallizable and can be efficiently ported onto the GPU architecture. Compared with the same CPU implementation running on a workstation level CPU, we reached a factor of 170 in computing mutual information, and a factor of 400 in computing its derivatives.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124645359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 39
Implementation of Advanced Encryption Standard for encryption and decryption of images and text on a GPU 高级加密标准在GPU上对图像和文本进行加密和解密的实现
Manoj Seshadrinathan, K. Dempski
In this paper, we propose a system for the complete implementation of the advanced encryption standard (AES) for encryption and decryption of images and text on a graphics processing unit. The GPU acts as a valuable co-processor that relieves the load off the CPU. In the decryption stage, we use a novel technique to display the decrypted images and text on the screen without bringing it onto CPU memory. We also present a system for encryption and decryption of hybrid map tiles generated from GIS data sets.
在本文中,我们提出了一种在图形处理单元上完全实现用于图像和文本加密和解密的高级加密标准(AES)的系统。GPU作为一个有价值的协处理器,减轻了CPU的负载。在解密阶段,我们使用一种新颖的技术在屏幕上显示解密后的图像和文本,而无需将其放入CPU内存。我们还提出了一个系统,用于加密和解密由GIS数据集生成的混合地图块。
{"title":"Implementation of Advanced Encryption Standard for encryption and decryption of images and text on a GPU","authors":"Manoj Seshadrinathan, K. Dempski","doi":"10.1109/CVPRW.2008.4563094","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563094","url":null,"abstract":"In this paper, we propose a system for the complete implementation of the advanced encryption standard (AES) for encryption and decryption of images and text on a graphics processing unit. The GPU acts as a valuable co-processor that relieves the load off the CPU. In the decryption stage, we use a novel technique to display the decrypted images and text on the screen without bringing it onto CPU memory. We also present a system for encryption and decryption of hybrid map tiles generated from GIS data sets.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134185389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Variational registration of tensor-valued images 张量值图像的变分配准
S. Barbieri, M. Welk, J. Weickert
We present a variational framework for the registration of tensor-valued images. It is based on an energy functional with four terms: a data term based on a diffusion tensor constancy constraint, a compatibility term encoding the physical model linking domain deformations and tensor reorientation, and smoothness terms for deformation and tensor reorientation. Although the tensor deformation model employed here is designed with regard to diffusion tensor MRI data, the separation of data and compatibility term allows to adapt the model easily to different tensor deformation models. We minimise the energy functional with respect to both transformation fields by a multiscale gradient descent. Experiments demonstrate the viability and potential of this approach in the registration of tensor-valued images.
我们提出了一个张量值图像配准的变分框架。它基于一个包含四项的能量泛函:一个基于扩散张量常数约束的数据项,一个编码连接域变形和张量重定向的物理模型的兼容性项,以及一个用于变形和张量重定向的平滑项。虽然本文采用的张量变形模型是针对弥散张量MRI数据设计的,但由于数据与相容项的分离,使得模型可以很容易地适应不同的张量变形模型。我们通过多尺度梯度下降最小化了两个变换场的能量泛函。实验证明了该方法在张量值图像配准中的可行性和潜力。
{"title":"Variational registration of tensor-valued images","authors":"S. Barbieri, M. Welk, J. Weickert","doi":"10.1109/CVPRW.2008.4562964","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4562964","url":null,"abstract":"We present a variational framework for the registration of tensor-valued images. It is based on an energy functional with four terms: a data term based on a diffusion tensor constancy constraint, a compatibility term encoding the physical model linking domain deformations and tensor reorientation, and smoothness terms for deformation and tensor reorientation. Although the tensor deformation model employed here is designed with regard to diffusion tensor MRI data, the separation of data and compatibility term allows to adapt the model easily to different tensor deformation models. We minimise the energy functional with respect to both transformation fields by a multiscale gradient descent. Experiments demonstrate the viability and potential of this approach in the registration of tensor-valued images.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133617911","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Multiple cue integration in transductive confidence machines for head pose classification 基于多线索集成的头部姿态分类换能型置信度机器
V. Balasubramanian, S. Panchanathan, Shayok Chakraborty
An important facet of learning in an online setting is the confidence associated with a prediction on a given test data point. In an online learning scenario, it would be expected that the system can increase its confidence of prediction as training data increases. We present a statistical approach in this work to associate a confidence value with a predicted class label in an online learning scenario. Our work is based on the existing work on transductive confidence machines (TCM) [1], which provided a methodology to define a heuristic confidence measure. We applied this approach to the problem of head pose classification from face images, and extended the framework to compute a confidence value when multiple cues are extracted from images to perform classification. Our approach is based on combining the results of multiple hypotheses and obtaining an integrated p-value to validate a single test hypothesis. From our experiments on the widely accepted FERET database, we obtained results which corroborated the significance of confidence measures - particularly, in online learning approaches. We could infer from our results with transductive learning that using confidence measures in online learning could yield significant boosts in the prediction accuracy, which would be very useful in critical pattern recognition applications.
在线学习的一个重要方面是与给定测试数据点的预测相关的置信度。在在线学习场景中,可以期望系统随着训练数据的增加而增加其预测的置信度。在这项工作中,我们提出了一种统计方法,将置信度值与在线学习场景中的预测类标签相关联。我们的工作是基于现有的关于传导置信机(TCM)的工作[1],它提供了一种定义启发式置信度度量的方法。我们将该方法应用于人脸图像的头部姿态分类问题,并扩展了该框架,在从图像中提取多个线索进行分类时计算置信值。我们的方法是基于组合多个假设的结果并获得一个集成的p值来验证单个检验假设。从我们在广泛接受的FERET数据库上的实验中,我们获得的结果证实了信心措施的重要性,特别是在在线学习方法中。我们可以从转换学习的结果中推断,在在线学习中使用置信度度量可以显著提高预测精度,这在关键的模式识别应用中非常有用。
{"title":"Multiple cue integration in transductive confidence machines for head pose classification","authors":"V. Balasubramanian, S. Panchanathan, Shayok Chakraborty","doi":"10.1109/CVPRW.2008.4563070","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563070","url":null,"abstract":"An important facet of learning in an online setting is the confidence associated with a prediction on a given test data point. In an online learning scenario, it would be expected that the system can increase its confidence of prediction as training data increases. We present a statistical approach in this work to associate a confidence value with a predicted class label in an online learning scenario. Our work is based on the existing work on transductive confidence machines (TCM) [1], which provided a methodology to define a heuristic confidence measure. We applied this approach to the problem of head pose classification from face images, and extended the framework to compute a confidence value when multiple cues are extracted from images to perform classification. Our approach is based on combining the results of multiple hypotheses and obtaining an integrated p-value to validate a single test hypothesis. From our experiments on the widely accepted FERET database, we obtained results which corroborated the significance of confidence measures - particularly, in online learning approaches. We could infer from our results with transductive learning that using confidence measures in online learning could yield significant boosts in the prediction accuracy, which would be very useful in critical pattern recognition applications.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116951436","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Efficient scan-window based object detection using GPGPU 使用GPGPU高效的基于扫描窗口的目标检测
Li Zhang, R. Nevatia
We describe an efficient design for scan-window based object detectors using a general purpose graphics hardware computing (GPGPU) framework. While the design is particularly applied to built a pedestrian detector that uses histogram of oriented gradient (HOG) features and the support vector machine (SVM) classifiers, the methodology we use is generic and can be applied to other objects, using different features and classifiers. The GPGPU paradigm is utilized for feature extraction and classification, so that the scan windows can be processed in parallel. We further propose to precompute and cache all the histograms in advance, instead of using integral images, which greatly lowers the computation cost. A multi-scale reduce strategy is employed to save expensive CPU-GPU data transfers. Experimental results show that our implementation achieves a more-than-ten-times speed up with no loss on detection rates.
我们描述了一种基于扫描窗口的目标检测器的高效设计,使用通用图形硬件计算(GPGPU)框架。虽然该设计特别适用于构建一个使用定向梯度直方图(HOG)特征和支持向量机(SVM)分类器的行人检测器,但我们使用的方法是通用的,可以应用于其他对象,使用不同的特征和分类器。利用GPGPU范式进行特征提取和分类,使扫描窗口可以并行处理。我们进一步提出预先计算和缓存所有的直方图,而不是使用积分图像,这大大降低了计算成本。采用多尺度缩减策略,节省了昂贵的CPU-GPU数据传输。实验结果表明,我们的实现在没有检测率损失的情况下实现了十倍以上的提速。
{"title":"Efficient scan-window based object detection using GPGPU","authors":"Li Zhang, R. Nevatia","doi":"10.1109/CVPRW.2008.4563097","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563097","url":null,"abstract":"We describe an efficient design for scan-window based object detectors using a general purpose graphics hardware computing (GPGPU) framework. While the design is particularly applied to built a pedestrian detector that uses histogram of oriented gradient (HOG) features and the support vector machine (SVM) classifiers, the methodology we use is generic and can be applied to other objects, using different features and classifiers. The GPGPU paradigm is utilized for feature extraction and classification, so that the scan windows can be processed in parallel. We further propose to precompute and cache all the histograms in advance, instead of using integral images, which greatly lowers the computation cost. A multi-scale reduce strategy is employed to save expensive CPU-GPU data transfers. Experimental results show that our implementation achieves a more-than-ten-times speed up with no loss on detection rates.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114999172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
Exploiting spatio-temporal information for view recognition in cardiac echo videos 利用时空信息进行心脏回波视频的视点识别
D. Beymer, T. Syeda-Mahmood, Fei Wang
2D Echocardiography is an important diagnostic aid for morphological and functional assessment of the heart. The transducer position is varied during an echo exam to elicit important information about the heart function and its anatomy. The knowledge of the transducer viewpoint is important in automatic cardiac echo interpretation to understand the regions being depicted as well as in the quantification of their attributes. In this paper, we address the problem of inferring the transducer viewpoint from the spatio-temporal information in cardiac echo videos. Unlike previous approaches, we exploit motion of the heart within a cardiac cycle in addition to spatial information to discriminate between viewpoints. Specifically, we use an active shape model (ASM) to model shape and texture information in an echo frame. The motion information derived by tracking ASMs through a heart cycle is then projected into the eigen-motion feature space of the viewpoint class for matching. We report comparison with a re-implementation of state-of-the-art view recognition methods in echos on a large database of patients with various cardiac diseases.
二维超声心动图是评价心脏形态和功能的重要诊断手段。在超声检查中,换能器的位置会发生变化,以获得有关心脏功能及其解剖结构的重要信息。换能器观点的知识在自动心脏回波解释中很重要,可以理解被描述的区域以及它们属性的量化。本文研究了从心脏回波视频的时空信息推断换能器视角的问题。与以前的方法不同,我们利用心脏在心脏周期内的运动以及空间信息来区分视点。具体来说,我们使用主动形状模型(ASM)对回波帧中的形状和纹理信息进行建模。通过心脏周期跟踪asm得到的运动信息被投影到视点类的本征运动特征空间中进行匹配。我们报告了与回声中最先进的视图识别方法在各种心脏病患者的大型数据库中的重新实现的比较。
{"title":"Exploiting spatio-temporal information for view recognition in cardiac echo videos","authors":"D. Beymer, T. Syeda-Mahmood, Fei Wang","doi":"10.1109/CVPRW.2008.4563008","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563008","url":null,"abstract":"2D Echocardiography is an important diagnostic aid for morphological and functional assessment of the heart. The transducer position is varied during an echo exam to elicit important information about the heart function and its anatomy. The knowledge of the transducer viewpoint is important in automatic cardiac echo interpretation to understand the regions being depicted as well as in the quantification of their attributes. In this paper, we address the problem of inferring the transducer viewpoint from the spatio-temporal information in cardiac echo videos. Unlike previous approaches, we exploit motion of the heart within a cardiac cycle in addition to spatial information to discriminate between viewpoints. Specifically, we use an active shape model (ASM) to model shape and texture information in an echo frame. The motion information derived by tracking ASMs through a heart cycle is then projected into the eigen-motion feature space of the viewpoint class for matching. We report comparison with a re-implementation of state-of-the-art view recognition methods in echos on a large database of patients with various cardiac diseases.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"146 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123917349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Fast gain-adaptive KLT tracking on the GPU GPU上的快速增益自适应KLT跟踪
C. Zach, D. Gallup, Jan-Michael Frahm
High-performance feature tracking from video input is a valuable tool in many computer vision techniques and mixed reality applications. This work presents a refined and substantially accelerated approach to KLT feature tracking performed on the GPU. Additionally, a global gain ratio between successive frames is estimated to compensate for changes in the camera exposure. The proposed approach achieves more than 200 frames per second on state-of-the art consumer GPUs for PAL (720 times 576) resolution data, and delivers real-time performance even on low-end mobile graphics processors.
视频输入的高性能特征跟踪在许多计算机视觉技术和混合现实应用中是一个有价值的工具。这项工作提出了一种在GPU上执行KLT特征跟踪的改进和实质上加速的方法。此外,估计连续帧之间的全局增益比以补偿相机曝光的变化。所提出的方法可以在最先进的消费级gpu上实现每秒200帧以上的PAL (720 × 576)分辨率数据,并且即使在低端移动图形处理器上也能提供实时性能。
{"title":"Fast gain-adaptive KLT tracking on the GPU","authors":"C. Zach, D. Gallup, Jan-Michael Frahm","doi":"10.1109/CVPRW.2008.4563089","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563089","url":null,"abstract":"High-performance feature tracking from video input is a valuable tool in many computer vision techniques and mixed reality applications. This work presents a refined and substantially accelerated approach to KLT feature tracking performed on the GPU. Additionally, a global gain ratio between successive frames is estimated to compensate for changes in the camera exposure. The proposed approach achieves more than 200 frames per second on state-of-the art consumer GPUs for PAL (720 times 576) resolution data, and delivers real-time performance even on low-end mobile graphics processors.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127970508","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 83
Incident light related distance error study and calibration of the PMD-range imaging camera pmd距离成像相机的入射光相关距离误差研究与标定
Jochen Radmer, Pol Moser Fuste, H. Schmidt, J. Krüger
For various applications, such as object recognition or tracking and especially when the object is partly occluded or articulated, 3D information is crucial for the robustness of the application. A recently developed sensor to acquire distance information is based on the Photo Mixer Device (PMD)for which a distance error based on different causes can be observed. This article presents an improved distance calibration approach for PMD-based distance sensoring which handles objects with different Lambertian reflectance properties. Within this scope the relation of the sources of distance errors were investigated. Where applicable they were isolated for relational studies with the actuating variables, i.e. integration time, amplitude and measured distance, as these are the only parameters available for the calibration. The calibration results of the proposed method excel the results of all other known methods. In particular with objects with unknown reflectance properties a significant reduction of the error is achieved.
对于各种应用程序,例如对象识别或跟踪,特别是当对象部分遮挡或铰接时,3D信息对于应用程序的鲁棒性至关重要。最近开发的一种用于获取距离信息的传感器是基于光混合装置(PMD)的,它可以观察到基于不同原因的距离误差。本文提出了一种改进的基于pmd的距离校准方法,用于处理具有不同朗伯反射率的物体。在此范围内,研究了距离误差源之间的关系。在适用的情况下,它们被隔离,用于与驱动变量的关系研究,即积分时间,振幅和测量距离,因为这些是唯一可用于校准的参数。该方法的校准结果优于所有其他已知方法的校准结果。特别是对于具有未知反射特性的物体,可以显著减少误差。
{"title":"Incident light related distance error study and calibration of the PMD-range imaging camera","authors":"Jochen Radmer, Pol Moser Fuste, H. Schmidt, J. Krüger","doi":"10.1109/CVPRW.2008.4563168","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563168","url":null,"abstract":"For various applications, such as object recognition or tracking and especially when the object is partly occluded or articulated, 3D information is crucial for the robustness of the application. A recently developed sensor to acquire distance information is based on the Photo Mixer Device (PMD)for which a distance error based on different causes can be observed. This article presents an improved distance calibration approach for PMD-based distance sensoring which handles objects with different Lambertian reflectance properties. Within this scope the relation of the sources of distance errors were investigated. Where applicable they were isolated for relational studies with the actuating variables, i.e. integration time, amplitude and measured distance, as these are the only parameters available for the calibration. The calibration results of the proposed method excel the results of all other known methods. In particular with objects with unknown reflectance properties a significant reduction of the error is achieved.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127983507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Camera localization and building reconstruction from single monocular images 单目图像的摄像机定位与建筑物重建
Ruisheng Wang, F. Ferrie
This paper presents a new method for reconstructing rectilinear buildings from single images under the assumption of flat terrain. An intuition of the method is that, given an image composed of rectilinear buildings, the 3D buildings can be geometrically reconstructed by using the image only. The recovery algorithm is formulated in terms of two objective functions which are based on the equivalence between the vector normal to the interpretation plane in the image space and the vector normal to the rotated interpretation plane in the object space. These objective functions are minimized with respect to the camera pose, the building dimensions, locations and orientations to obtain estimates for the structure of the scene. The method potentially provides a solution for large-scale urban modelling using aerial images, and can be easily extended to deal with piecewise planar objects in a more general situation.
本文提出了一种在平坦地形条件下从单幅图像重建直线建筑物的新方法。该方法的直观效果是,给定由直线建筑组成的图像,仅使用该图像就可以对三维建筑进行几何重构。基于图像空间中解释平面的法向量和物体空间中旋转解释平面的法向量的等价性,用两个目标函数来表述恢复算法。这些目标函数根据相机姿态、建筑尺寸、位置和方向进行最小化,以获得场景结构的估计。该方法潜在地为使用航空图像的大规模城市建模提供了解决方案,并且可以很容易地扩展到在更一般的情况下处理分段平面物体。
{"title":"Camera localization and building reconstruction from single monocular images","authors":"Ruisheng Wang, F. Ferrie","doi":"10.1109/CVPRW.2008.4563132","DOIUrl":"https://doi.org/10.1109/CVPRW.2008.4563132","url":null,"abstract":"This paper presents a new method for reconstructing rectilinear buildings from single images under the assumption of flat terrain. An intuition of the method is that, given an image composed of rectilinear buildings, the 3D buildings can be geometrically reconstructed by using the image only. The recovery algorithm is formulated in terms of two objective functions which are based on the equivalence between the vector normal to the interpretation plane in the image space and the vector normal to the rotated interpretation plane in the object space. These objective functions are minimized with respect to the camera pose, the building dimensions, locations and orientations to obtain estimates for the structure of the scene. The method potentially provides a solution for large-scale urban modelling using aerial images, and can be easily extended to deal with piecewise planar objects in a more general situation.","PeriodicalId":102206,"journal":{"name":"2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops","volume":"79 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125523584","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1