首页 > 最新文献

2011 IEEE Workshop on Applications of Computer Vision (WACV)最新文献

英文 中文
Historical comparison of vehicles using scanned x-ray images 使用扫描x射线图像对车辆进行历史比较
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711516
W. Ahmed, Ming Zhang, O. Al-Kofahi
X-ray scanners are increasingly used for scanning vehicles crossing international borders or entering critical infrastructure installations. The ability to penetrate through steel and other opaque materials and the nondestructive nature of x-ray radiation make them ideal for finding drugs, explosives and other contraband. In many situations, the same vehicles cross the checkpoint repeatedly, such as the employee vehicles entering a high-risk facility or cargo vehicles crossing international borders back and forth. Manual analysis of these images puts extra burden on the operator and results in slow throughput. In this paper we report an integrated and fully automated system to solve this problem. In the first stage of the algorithm, a model-based segmentation approach is used to find the vehicle outline. It proceeds by first using background subtraction to find the overall body of the vehicle. Next, we find the outlines of tires by using rotating edge detection kernels. The lower outline of the vehicle is found using active contours. We then use a deformable registration approach to align the vehicles which is specifically designed for the requirements of this problem. An intensity normalization step is then performed to account for the intensity variations between the scans at two time points. We use a histogram-based approach that scales and shifts the histogram of one image to match that of the other. The differences between the two inspection results are computed next. We then apply knowledge-based rules to remove false alarms such as lights and driver's body. The system is specifically designed for back-scatter x-ray imaging which is a powerful modality for detecting organic materials such as drugs and explosives. We have applied this system to images scanned by a deployed x-ray scanner and have achieved satisfactory results.
x射线扫描仪越来越多地用于扫描穿越国际边界或进入关键基础设施设施的车辆。穿透钢铁和其他不透明材料的能力,以及x射线辐射的非破坏性,使它们成为寻找毒品、爆炸物和其他违禁品的理想选择。在许多情况下,相同的车辆反复通过检查站,例如进入高风险设施的雇员车辆或往返穿越国际边界的货运车辆。手动分析这些图像会给操作人员带来额外的负担,并导致较慢的吞吐量。在本文中,我们报告了一个集成的全自动化系统来解决这个问题。在算法的第一阶段,采用基于模型的分割方法寻找车辆轮廓。它首先使用背景减法来找到车辆的整体。接下来,我们使用旋转边缘检测核找到轮胎的轮廓。车辆的下轮廓是使用活动轮廓来找到的。然后,我们使用可变形的注册方法来对齐车辆,这是专门为这个问题的要求而设计的。然后执行强度归一化步骤,以解释两个时间点扫描之间的强度变化。我们使用基于直方图的方法,缩放和移动一个图像的直方图以匹配另一个图像。接下来计算两个检测结果之间的差异。然后,我们应用基于知识的规则来消除假警报,如灯光和司机的身体。该系统是专门为后向散射x射线成像设计的,这是一种检测有机材料(如毒品和爆炸物)的强大模式。我们已将该系统应用于部署的x射线扫描仪扫描的图像,并取得了令人满意的结果。
{"title":"Historical comparison of vehicles using scanned x-ray images","authors":"W. Ahmed, Ming Zhang, O. Al-Kofahi","doi":"10.1109/WACV.2011.5711516","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711516","url":null,"abstract":"X-ray scanners are increasingly used for scanning vehicles crossing international borders or entering critical infrastructure installations. The ability to penetrate through steel and other opaque materials and the nondestructive nature of x-ray radiation make them ideal for finding drugs, explosives and other contraband. In many situations, the same vehicles cross the checkpoint repeatedly, such as the employee vehicles entering a high-risk facility or cargo vehicles crossing international borders back and forth. Manual analysis of these images puts extra burden on the operator and results in slow throughput. In this paper we report an integrated and fully automated system to solve this problem. In the first stage of the algorithm, a model-based segmentation approach is used to find the vehicle outline. It proceeds by first using background subtraction to find the overall body of the vehicle. Next, we find the outlines of tires by using rotating edge detection kernels. The lower outline of the vehicle is found using active contours. We then use a deformable registration approach to align the vehicles which is specifically designed for the requirements of this problem. An intensity normalization step is then performed to account for the intensity variations between the scans at two time points. We use a histogram-based approach that scales and shifts the histogram of one image to match that of the other. The differences between the two inspection results are computed next. We then apply knowledge-based rules to remove false alarms such as lights and driver's body. The system is specifically designed for back-scatter x-ray imaging which is a powerful modality for detecting organic materials such as drugs and explosives. We have applied this system to images scanned by a deployed x-ray scanner and have achieved satisfactory results.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123686467","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Robust alignment of wide baseline terrestrial laser scans via 3D viewpoint normalization 基于三维视点归一化的宽基线地面激光扫描鲁棒对准
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711539
Yanpeng Cao, M. Yang, J. McDonald
The complexity of natural scenes and the amount of information acquired by terrestrial laser scanners turn the registration among scans into a complex problem. This problem becomes even more challenging when two individual scans captured at significantly changed viewpoints (wide baseline). Since laser-scanning instruments nowadays are often equipped with an additional image sensor, it stands to reason making use of the image content to improve the registration process of 3D scanning data. In this paper, we present a novel improvement to the existing feature techniques to enable automatic alignment between two widely separated 3D scans. The key idea consists of extracting dominant planar structures from 3D point clouds and then utilizing the recovered 3D geometry to improve the performance of 2D image feature extraction and matching. The resulting features are very discriminative and robust to perspective distortions and viewpoint changes due to exploiting the underlying 3D structure. Using this novel viewpoint invariant feature, the corresponding 3D points are automatically linked in terms of wide baseline image matching. Initial experiments with real data demonstrate the potential of the proposed method for the challenging wide baseline 3D scanning data alignment tasks.
自然场景的复杂性和地面激光扫描仪获取的信息量使得扫描间的配准成为一个复杂的问题。当两个单独的扫描在显著改变的视点(宽基线)上捕获时,这个问题变得更加具有挑战性。由于现在的激光扫描仪器通常配备额外的图像传感器,因此利用图像内容来改进3D扫描数据的配准过程是理所当然的。在本文中,我们提出了一种新的改进现有的特征技术,以实现两个广泛分离的3D扫描之间的自动对齐。其核心思想是从三维点云中提取优势平面结构,然后利用恢复的三维几何形状来提高二维图像特征提取和匹配的性能。由于利用了底层3D结构,所得到的特征对透视扭曲和视点变化具有很强的鉴别性和鲁棒性。利用这种新颖的视点不变性特征,在宽基线图像匹配方面自动链接相应的三维点。实际数据的初步实验证明了该方法在具有挑战性的宽基线3D扫描数据对齐任务中的潜力。
{"title":"Robust alignment of wide baseline terrestrial laser scans via 3D viewpoint normalization","authors":"Yanpeng Cao, M. Yang, J. McDonald","doi":"10.1109/WACV.2011.5711539","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711539","url":null,"abstract":"The complexity of natural scenes and the amount of information acquired by terrestrial laser scanners turn the registration among scans into a complex problem. This problem becomes even more challenging when two individual scans captured at significantly changed viewpoints (wide baseline). Since laser-scanning instruments nowadays are often equipped with an additional image sensor, it stands to reason making use of the image content to improve the registration process of 3D scanning data. In this paper, we present a novel improvement to the existing feature techniques to enable automatic alignment between two widely separated 3D scans. The key idea consists of extracting dominant planar structures from 3D point clouds and then utilizing the recovered 3D geometry to improve the performance of 2D image feature extraction and matching. The resulting features are very discriminative and robust to perspective distortions and viewpoint changes due to exploiting the underlying 3D structure. Using this novel viewpoint invariant feature, the corresponding 3D points are automatically linked in terms of wide baseline image matching. Initial experiments with real data demonstrate the potential of the proposed method for the challenging wide baseline 3D scanning data alignment tasks.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"141 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125833705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Evolving improved transforms for reconstruction of quantized ultrasound images 基于改进变换的量化超声图像重建
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711511
Chris Miller, B. Babb, F. Moore, M. R. Peterson
State-of-the-art lossy compression schemes for medical imagery utilize the 9/7 wavelet. Recent research has established a methodology for using evolutionary computation (EC) to evolve wavelet and scaling numbers describing novel reconstruction transforms that outperform the 9/7 under lossy conditions. This paper describes an investigation into whether evolved transforms could automatically compensate for the detrimental effects of quantization for ultrasound (US) images. Results for 16:1, 32:1, and 64:1 quantization consistently demonstrate superior performance of evolved transforms in comparison to the 9/7 wavelet; in general, this advantage increases in proportion to the selected quantization level.
医学图像的最先进的有损压缩方案利用9/7小波。最近的研究已经建立了一种使用进化计算(EC)来进化小波和描述在有损条件下优于9/7的新型重构变换的尺度数的方法。本文描述了一项关于进化变换是否能自动补偿量化对超声图像的有害影响的研究。16:1、32:1和64:1量化的结果一致表明进化变换的性能优于9/7小波;一般来说,这种优势与所选择的量化水平成比例地增加。
{"title":"Evolving improved transforms for reconstruction of quantized ultrasound images","authors":"Chris Miller, B. Babb, F. Moore, M. R. Peterson","doi":"10.1109/WACV.2011.5711511","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711511","url":null,"abstract":"State-of-the-art lossy compression schemes for medical imagery utilize the 9/7 wavelet. Recent research has established a methodology for using evolutionary computation (EC) to evolve wavelet and scaling numbers describing novel reconstruction transforms that outperform the 9/7 under lossy conditions. This paper describes an investigation into whether evolved transforms could automatically compensate for the detrimental effects of quantization for ultrasound (US) images. Results for 16:1, 32:1, and 64:1 quantization consistently demonstrate superior performance of evolved transforms in comparison to the 9/7 wavelet; in general, this advantage increases in proportion to the selected quantization level.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128482862","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Classification of traffic video based on a spatiotemporal orientation analysis 基于时空方向分析的交通视频分类
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711560
K. Derpanis, Richard P. Wildes
This paper describes a system for classifying traffic congestion videos based on their observed visual dynamics. Central to the proposed system is treating traffic flow identification as an instance of dynamic texture classification. More specifically, a recent discriminative model of dynamic textures is adapted for the special case of traffic flows. This approach avoids the need for segmentation, tracking and motion estimation that typify extant approaches. Classification is based on matching distributions (or histograms) of spacetime orientation structure. Empirical evaluation on a publicly available data set shows high classification performance and robustness to typical environmental conditions (e.g., variable lighting).
本文描述了一种基于观察到的视觉动态对交通拥堵视频进行分类的系统。该系统的核心是将交通流识别作为动态纹理分类的一个实例。更具体地说,最近的动态纹理判别模型适用于交通流的特殊情况。这种方法避免了对分割、跟踪和运动估计的需要,这些都是现有方法的特点。分类是基于时空方向结构的匹配分布(或直方图)。对公开可用数据集的经验评估显示出高分类性能和对典型环境条件(例如,可变照明)的鲁棒性。
{"title":"Classification of traffic video based on a spatiotemporal orientation analysis","authors":"K. Derpanis, Richard P. Wildes","doi":"10.1109/WACV.2011.5711560","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711560","url":null,"abstract":"This paper describes a system for classifying traffic congestion videos based on their observed visual dynamics. Central to the proposed system is treating traffic flow identification as an instance of dynamic texture classification. More specifically, a recent discriminative model of dynamic textures is adapted for the special case of traffic flows. This approach avoids the need for segmentation, tracking and motion estimation that typify extant approaches. Classification is based on matching distributions (or histograms) of spacetime orientation structure. Empirical evaluation on a publicly available data set shows high classification performance and robustness to typical environmental conditions (e.g., variable lighting).","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"33 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128500386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Detection of static objects for the task of video surveillance 检测静态物体为视频监控的任务
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711550
Rubén Heras Evangelio, T. Senst, T. Sikora
Detecting static objects in video sequences has a high relevance in many surveillance scenarios like airports and railwaystations. In this paper we propose a system for the detection of static objects in crowded scenes that, based on the detection of two background models learning at different rates, classifies pixels with the help of a finite-state machine. The background is modelled by two mixtures of Gaussians with identical parameters except for the learning rate. The state machine provides the meaning for the interpretation of the results obtained from background subtraction and can be used to incorporate additional information cues, obtaining thus a flexible system specially suitable for real-life applications. The system was built in our surveillance application and successfully validated with several public datasets.
在机场和火车站等监控场景中,检测视频序列中的静态物体具有很高的相关性。在本文中,我们提出了一种用于拥挤场景中静态物体检测的系统,该系统基于以不同速率学习的两个背景模型的检测,在有限状态机的帮助下对像素进行分类。除学习率外,背景由两个具有相同参数的高斯混合模型建模。状态机为解释从背景减法中获得的结果提供了意义,并可用于合并额外的信息线索,从而获得特别适合实际应用的灵活系统。该系统建立在我们的监控应用程序中,并成功地通过几个公共数据集进行了验证。
{"title":"Detection of static objects for the task of video surveillance","authors":"Rubén Heras Evangelio, T. Senst, T. Sikora","doi":"10.1109/WACV.2011.5711550","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711550","url":null,"abstract":"Detecting static objects in video sequences has a high relevance in many surveillance scenarios like airports and railwaystations. In this paper we propose a system for the detection of static objects in crowded scenes that, based on the detection of two background models learning at different rates, classifies pixels with the help of a finite-state machine. The background is modelled by two mixtures of Gaussians with identical parameters except for the learning rate. The state machine provides the meaning for the interpretation of the results obtained from background subtraction and can be used to incorporate additional information cues, obtaining thus a flexible system specially suitable for real-life applications. The system was built in our surveillance application and successfully validated with several public datasets.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122253106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
On the use of multispectral conjunctival vasculature as a soft biometric 利用多光谱结膜血管作为软生物识别
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711504
S. Crihalmeanu, A. Ross
Ocular biometrics has made significant progress over the past decade primarily due to advances in iris recognition. Initial research in the field of iris recognition focused on the acquisition and processing of frontal irides which may require considerable subject cooperation. However, when the iris is off-angle with respect to the acquisition device, the sclera (the white part of the eye) is exposed. The sclera is covered by a thin transparent layer called conjunctiva. Both the episclera and conjunctiva contain blood vessels that are observable from the outside. In this work, these blood vessels are referred to as conjunctival vasculature. Iris patterns are better observed in the near infrared spectrum while conjunctival vasculature is better seen in the visible spectrum. Therefore, multispectral (i.e., color-infrared) images of the eye are acquired to allow for the combination of the iris biometric with the conjunctival vasculature. The paper focuses on conjunctival vasculature enhancement, registration and matching. Initial results are promising and suggest the need for further investigation of this biometric in a bimodal configuration with iris.
眼部生物识别技术在过去十年中取得了重大进展,主要是由于虹膜识别技术的进步。虹膜识别领域的初步研究主要集中在额叶虹膜的获取和处理,这可能需要大量的受试者合作。然而,当虹膜相对于采集设备偏离角度时,巩膜(眼睛的白色部分)就会暴露出来。巩膜上覆盖着一层薄薄的透明膜,叫做结膜。外膜和结膜都含有从外部可见的血管。在这项工作中,这些血管被称为结膜血管。在近红外光谱中可以更好地观察到虹膜图案,而在可见光谱中可以更好地看到结膜血管。因此,获得眼睛的多光谱(即彩色红外)图像,以便将虹膜生物特征与结膜血管系统相结合。本文的重点是结膜血管增强、配准和匹配。初步结果是有希望的,并表明需要进一步研究这种生物特征在虹膜双峰配置。
{"title":"On the use of multispectral conjunctival vasculature as a soft biometric","authors":"S. Crihalmeanu, A. Ross","doi":"10.1109/WACV.2011.5711504","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711504","url":null,"abstract":"Ocular biometrics has made significant progress over the past decade primarily due to advances in iris recognition. Initial research in the field of iris recognition focused on the acquisition and processing of frontal irides which may require considerable subject cooperation. However, when the iris is off-angle with respect to the acquisition device, the sclera (the white part of the eye) is exposed. The sclera is covered by a thin transparent layer called conjunctiva. Both the episclera and conjunctiva contain blood vessels that are observable from the outside. In this work, these blood vessels are referred to as conjunctival vasculature. Iris patterns are better observed in the near infrared spectrum while conjunctival vasculature is better seen in the visible spectrum. Therefore, multispectral (i.e., color-infrared) images of the eye are acquired to allow for the combination of the iris biometric with the conjunctival vasculature. The paper focuses on conjunctival vasculature enhancement, registration and matching. Initial results are promising and suggest the need for further investigation of this biometric in a bimodal configuration with iris.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121050385","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Tracking gaze direction from far-field surveillance cameras 通过远场监控摄像头追踪注视方向
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711548
K. Sankaranarayanan, Ming-Ching Chang, N. Krahnstoever
We present a real-time approach to estimating the gaze direction of multiple individuals using a network of far-field surveillance cameras. This work is part of a larger surveillance system that utilizes a network of fixed cameras as well as PTZ cameras to perform site-wide tracking of individuals. Based on the tracking information, one or more PTZ cameras are cooperatively controlled to obtain close-up facial images of individuals. Within these close-up shots, face detection and head pose estimation are performed and the results are provided back to the tracking system to track the individual gazes. A new cost metric based on location and gaze orientation is proposed to robustly associate head observations with tracker states. The tracking system can thus leverage the newly obtained gaze information for two purposes: (i) improve the localization of individuals in crowded settings, and (ii) aid high-level surveillance tasks such as understanding gesturing, interactions between individuals, and finding the object-of-interest that people are looking at. In security application, our system can detect if a subject is looking at the security cameras or guard posts.
我们提出了一种利用远场监控摄像机网络实时估计多个个体注视方向的方法。这项工作是一个更大的监视系统的一部分,该系统利用固定摄像机网络和PTZ摄像机对个人进行全场址跟踪。基于跟踪信息,协同控制一个或多个PTZ摄像机获取个体的近距离面部图像。在这些特写镜头中,人脸检测和头部姿势估计被执行,并将结果提供给跟踪系统以跟踪单个凝视。提出了一种新的基于位置和凝视方向的代价度量,以鲁棒地将头部观察与跟踪器状态关联起来。因此,跟踪系统可以将新获得的凝视信息用于两个目的:(i)提高拥挤环境中个体的定位;(ii)帮助执行高级监视任务,如理解手势、个体之间的互动,以及找到人们正在看的感兴趣的对象。在安防应用中,我们的系统可以检测到目标是否在看监控摄像头或岗哨。
{"title":"Tracking gaze direction from far-field surveillance cameras","authors":"K. Sankaranarayanan, Ming-Ching Chang, N. Krahnstoever","doi":"10.1109/WACV.2011.5711548","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711548","url":null,"abstract":"We present a real-time approach to estimating the gaze direction of multiple individuals using a network of far-field surveillance cameras. This work is part of a larger surveillance system that utilizes a network of fixed cameras as well as PTZ cameras to perform site-wide tracking of individuals. Based on the tracking information, one or more PTZ cameras are cooperatively controlled to obtain close-up facial images of individuals. Within these close-up shots, face detection and head pose estimation are performed and the results are provided back to the tracking system to track the individual gazes. A new cost metric based on location and gaze orientation is proposed to robustly associate head observations with tracker states. The tracking system can thus leverage the newly obtained gaze information for two purposes: (i) improve the localization of individuals in crowded settings, and (ii) aid high-level surveillance tasks such as understanding gesturing, interactions between individuals, and finding the object-of-interest that people are looking at. In security application, our system can detect if a subject is looking at the security cameras or guard posts.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116755650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Window detection from mobile LiDAR data 从移动激光雷达数据进行窗口检测
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711484
Ruisheng Wang, Jeff Bach, F. Ferrie
We present an automatic approach to window and façade detection from LiDAR (Light Detection And Ranging) data collected from a moving vehicle along streets in urban environments. The proposed method combines bottom-up with top-down strategies to extract façade planes from noisy LiDAR point clouds. The window detection is achieved through a two-step approach: potential window point detection and window localization. The facade pattern is automatically inferred to enhance the robustness of the window detection. Experimental results on six datasets result in 71.2% and 88.9% in the first two datasets, 100% for the rest four datasets in terms of completeness rate, and 100% correctness rate for all the tested datasets, which demonstrate the effectiveness of the proposed solution. The application potential includes generation of building facade models with street-level details and texture synthesis for producing realistic occlusion-free façade texture.
我们提出了一种从城市环境中沿着街道行驶的车辆收集的激光雷达(光探测和测距)数据中自动检测窗口和前方的方法。该方法结合自底向上和自顶向下两种策略,从噪声激光雷达点云中提取近场面。窗口检测通过两个步骤实现:潜在窗口点检测和窗口定位。自动推断立面模式以增强窗口检测的鲁棒性。在6个数据集上的实验结果显示,前2个数据集的完备率分别为71.2%和88.9%,其余4个数据集的完备率均为100%,所有被测数据集的完备率均为100%,验证了所提方案的有效性。应用潜力包括生成具有街道级细节的建筑立面模型和纹理合成,以产生逼真的无遮挡的立面纹理。
{"title":"Window detection from mobile LiDAR data","authors":"Ruisheng Wang, Jeff Bach, F. Ferrie","doi":"10.1109/WACV.2011.5711484","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711484","url":null,"abstract":"We present an automatic approach to window and façade detection from LiDAR (Light Detection And Ranging) data collected from a moving vehicle along streets in urban environments. The proposed method combines bottom-up with top-down strategies to extract façade planes from noisy LiDAR point clouds. The window detection is achieved through a two-step approach: potential window point detection and window localization. The facade pattern is automatically inferred to enhance the robustness of the window detection. Experimental results on six datasets result in 71.2% and 88.9% in the first two datasets, 100% for the rest four datasets in terms of completeness rate, and 100% correctness rate for all the tested datasets, which demonstrate the effectiveness of the proposed solution. The application potential includes generation of building facade models with street-level details and texture synthesis for producing realistic occlusion-free façade texture.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"41 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131489099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
An assisted photography method for street scenes 一种街景辅助摄影方法
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711488
Marynel Vázquez, Aaron Steinfeld
We present an interactive, computational approach for assisting users with visual impairments during photographic documentation of transit problems. Our technique can be described as a method to improve picture composition, while retaining visual information that is expected to be most relevant. Our system considers the position of the estimated region of interest (ROI) of a photo, and camera orientation. Saliency maps and Gestalt theory are used for guiding the user towards a more balanced picture. Our current implementation for mobile phones uses optic flow to update the internal knowledge of the position of the ROI and tilt sensor readings to correct non horizontal or vertical camera orientations. Using ground truth labels, we confirmed our method proposes valid strategies for improving image composition. Future work includes an optimized implementation and user studies.
我们提出了一种交互式的计算方法,以帮助在交通问题的摄影记录中有视觉障碍的用户。我们的技术可以被描述为一种改善图片构图的方法,同时保留预期最相关的视觉信息。我们的系统考虑了照片估计感兴趣区域(ROI)的位置和相机方向。显著性图和格式塔理论被用来引导用户走向一个更平衡的画面。我们目前在手机上的实现使用光流来更新ROI位置的内部知识和倾斜传感器读数,以纠正非水平或垂直相机方向。使用地面真值标签,我们证实了我们的方法为改善图像组成提出了有效的策略。未来的工作包括优化实现和用户研究。
{"title":"An assisted photography method for street scenes","authors":"Marynel Vázquez, Aaron Steinfeld","doi":"10.1109/WACV.2011.5711488","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711488","url":null,"abstract":"We present an interactive, computational approach for assisting users with visual impairments during photographic documentation of transit problems. Our technique can be described as a method to improve picture composition, while retaining visual information that is expected to be most relevant. Our system considers the position of the estimated region of interest (ROI) of a photo, and camera orientation. Saliency maps and Gestalt theory are used for guiding the user towards a more balanced picture. Our current implementation for mobile phones uses optic flow to update the internal knowledge of the position of the ROI and tilt sensor readings to correct non horizontal or vertical camera orientations. Using ground truth labels, we confirmed our method proposes valid strategies for improving image composition. Future work includes an optimized implementation and user studies.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117320500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
SLAM combining ToF and high-resolution cameras SLAM结合ToF和高分辨率相机
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711569
V. Castañeda, D. Mateus, Nassir Navab
This paper describes an extension to the Monocular Simultaneous Localization and Mapping (MonoSLAM) method that relies on the images provided by a combined high resolution Time of Flight (HR-ToF) sensor. In its standard formulation MonoSLAM estimates the depth of each tracked feature as the camera moves. This depth estimation depends both on the quality of the feature tracking and the previous camera position estimates. Additionally, MonoSLAM requires a set of known features to initialize the scale of the map and the world coordinate system. We propose to use the combined high resolution ToF sensor to incorporate depth measures into the MonoSLAM framework while keeping the accuracy of the feature detection. In practice, we use a ToF (Time of Flight) and a high-resolution (HR) camera in a calibrated and synchronized set-up and modify the measurement model and observation updates of MonoSLAM. The proposed method does not require known features to initialize a map. Experiments show first, that the depth measurements in our method improve the results of camera localization when compared to the MonoSLAM approach using HR images alone; and second, that HR images are required for reliable tracking.
本文描述了单目同时定位和映射(MonoSLAM)方法的扩展,该方法依赖于组合高分辨率飞行时间(HR-ToF)传感器提供的图像。在其标准公式中,MonoSLAM在相机移动时估计每个跟踪特征的深度。这种深度估计既取决于特征跟踪的质量,也取决于之前的相机位置估计。此外,MonoSLAM需要一组已知的特征来初始化地图的比例和世界坐标系统。我们建议使用组合的高分辨率ToF传感器将深度测量纳入MonoSLAM框架,同时保持特征检测的准确性。在实际应用中,我们使用飞行时间(ToF)和高分辨率(HR)相机进行校准和同步设置,并修改MonoSLAM的测量模型和观测更新。所提出的方法不需要已知的特征来初始化映射。实验表明,与仅使用HR图像的MonoSLAM方法相比,我们的方法中的深度测量改善了相机定位结果;其次,需要HR图像进行可靠的跟踪。
{"title":"SLAM combining ToF and high-resolution cameras","authors":"V. Castañeda, D. Mateus, Nassir Navab","doi":"10.1109/WACV.2011.5711569","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711569","url":null,"abstract":"This paper describes an extension to the Monocular Simultaneous Localization and Mapping (MonoSLAM) method that relies on the images provided by a combined high resolution Time of Flight (HR-ToF) sensor. In its standard formulation MonoSLAM estimates the depth of each tracked feature as the camera moves. This depth estimation depends both on the quality of the feature tracking and the previous camera position estimates. Additionally, MonoSLAM requires a set of known features to initialize the scale of the map and the world coordinate system. We propose to use the combined high resolution ToF sensor to incorporate depth measures into the MonoSLAM framework while keeping the accuracy of the feature detection. In practice, we use a ToF (Time of Flight) and a high-resolution (HR) camera in a calibrated and synchronized set-up and modify the measurement model and observation updates of MonoSLAM. The proposed method does not require known features to initialize a map. Experiments show first, that the depth measurements in our method improve the results of camera localization when compared to the MonoSLAM approach using HR images alone; and second, that HR images are required for reliable tracking.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122176448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
期刊
2011 IEEE Workshop on Applications of Computer Vision (WACV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1