首页 > 最新文献

2011 IEEE Workshop on Applications of Computer Vision (WACV)最新文献

英文 中文
Image matching with distinctive visual vocabulary 具有鲜明视觉语汇的形象匹配
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711532
Hongwen Kang, M. Hebert, T. Kanade
In this paper we propose an image indexing and matching algorithm that relies on selecting distinctive high dimensional features. In contrast with conventional techniques that treated all features equally, we claim that one can benefit significantly from focusing on distinctive features. We propose a bag-of-words algorithm that combines the feature distinctiveness in visual vocabulary generation. Our approach compares favorably with the state of the art in image matching tasks on the University of Kentucky Recognition Benchmark dataset and on an indoor localization dataset. We also show that our approach scales up more gracefully on a large scale Flickr dataset.
本文提出了一种基于高维特征选取的图像索引与匹配算法。与同等对待所有特征的传统技术相比,我们声称人们可以从专注于独特的特征中获益良多。我们提出了一种结合视觉词汇生成中特征显著性的词袋算法。我们的方法与肯塔基大学识别基准数据集和室内定位数据集上的图像匹配任务的最新状态相比具有优势。我们还展示了我们的方法在大规模的Flickr数据集上更优雅地扩展。
{"title":"Image matching with distinctive visual vocabulary","authors":"Hongwen Kang, M. Hebert, T. Kanade","doi":"10.1109/WACV.2011.5711532","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711532","url":null,"abstract":"In this paper we propose an image indexing and matching algorithm that relies on selecting distinctive high dimensional features. In contrast with conventional techniques that treated all features equally, we claim that one can benefit significantly from focusing on distinctive features. We propose a bag-of-words algorithm that combines the feature distinctiveness in visual vocabulary generation. Our approach compares favorably with the state of the art in image matching tasks on the University of Kentucky Recognition Benchmark dataset and on an indoor localization dataset. We also show that our approach scales up more gracefully on a large scale Flickr dataset.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122253876","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Stacked spatial-pyramid kernel: An object-class recognition method to combine scores from random trees 堆叠空间金字塔核:一种结合随机树得分的对象类识别方法
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711522
N. Larios, Junyuan Lin, Mengzi Zhang, D. Lytle, A. Moldenke, L. Shapiro, Thomas G. Dietterich
The combination of local features, complementary feature types, and relative position information has been successfully applied to many object-class recognition tasks. Stacking is a common classification approach that combines the results from multiple classifiers, having the added benefit of allowing each classifier to handle a different feature space. However, the standard stacking method by its own nature discards any spatial information contained in the features, because only the combination of raw classification scores are input to the final classifier. The object-class recognition method proposed in this paper combines different feature types in a new stacking framework that efficiently quantizes input data and boosts classification accuracy, while allowing the use of spatial information. This classification method is applied to the task of automated insect-species identification for biomonitoring purposes. The test data set for this work contains 4722 images with 29 insect species, belonging to the three most common orders used to measure stream water quality, several of which are closely related and very difficult to distinguish. The specimens are in different 3D positions, different orientations, and different developmental and degradation stages with wide intra-class variation. On this very challenging data set, our new algorithm outperforms other classifiers, showing the benefits of using spatial information in the stacking framework with multiple dissimilar feature types.
局部特征、互补特征类型和相对位置信息的组合已成功应用于许多目标类识别任务。堆叠是一种常见的分类方法,它将来自多个分类器的结果组合在一起,还具有允许每个分类器处理不同特征空间的额外好处。然而,标准的叠加方法由于其本身的性质,丢弃了特征中包含的任何空间信息,因为只有原始分类分数的组合被输入到最终的分类器中。本文提出的目标类识别方法将不同的特征类型结合在一个新的叠加框架中,在允许使用空间信息的同时,有效地量化了输入数据,提高了分类精度。这种分类方法适用于生物监测目的的昆虫物种自动鉴定任务。这项工作的测试数据集包含4722张图像,包含29种昆虫,属于用于测量溪流水质的三个最常见的目,其中一些是密切相关的,很难区分。不同的三维位置、不同的方位、不同的发育和退化阶段,类内差异较大。在这个非常具有挑战性的数据集上,我们的新算法优于其他分类器,显示了在具有多个不同特征类型的堆叠框架中使用空间信息的好处。
{"title":"Stacked spatial-pyramid kernel: An object-class recognition method to combine scores from random trees","authors":"N. Larios, Junyuan Lin, Mengzi Zhang, D. Lytle, A. Moldenke, L. Shapiro, Thomas G. Dietterich","doi":"10.1109/WACV.2011.5711522","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711522","url":null,"abstract":"The combination of local features, complementary feature types, and relative position information has been successfully applied to many object-class recognition tasks. Stacking is a common classification approach that combines the results from multiple classifiers, having the added benefit of allowing each classifier to handle a different feature space. However, the standard stacking method by its own nature discards any spatial information contained in the features, because only the combination of raw classification scores are input to the final classifier. The object-class recognition method proposed in this paper combines different feature types in a new stacking framework that efficiently quantizes input data and boosts classification accuracy, while allowing the use of spatial information. This classification method is applied to the task of automated insect-species identification for biomonitoring purposes. The test data set for this work contains 4722 images with 29 insect species, belonging to the three most common orders used to measure stream water quality, several of which are closely related and very difficult to distinguish. The specimens are in different 3D positions, different orientations, and different developmental and degradation stages with wide intra-class variation. On this very challenging data set, our new algorithm outperforms other classifiers, showing the benefits of using spatial information in the stacking framework with multiple dissimilar feature types.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123788029","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
A study on recognizing non-artistic face sketches 非艺术人脸素描识别的研究
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711509
Hossein Nejati, T. Sim
Face sketches are being used in eyewitness testimonies for about a century. These sketches are crucial in finding suspects when no photo is available, but a mental image in the eyewitness's mind. However, research shows that current procedures used for eyewitness testimonies have two main problems. First, they can significantly disturb the memories of the eyewitness. Second, in many cases, these procedures result in face images far from their target faces. These two problems are related to the plasticity of the human visual system and the differences between face perception in humans (holistic) and current methods of sketch production (piecemeal). In this paper, we present some insights for more realistic sketch to photo matching. We describe how to retrieve identity specific information from crude sketches, directly drawn by the non-artistic eyewitnesses. The sketches we used merely contain facial component outlines and facial marks (e.g. wrinkles and moles). We compare results of automatically matching two types sketches (trace-over and user-provided, 25 each) to four types of faces (original, locally exaggerated, configurally exaggerated, and globally exaggerated, 249 each), using two methods (PDM distance comparison and PCA classification). Based on our results, we argue that for automatic non-artistic sketch to photo matching, the algorithms should compare the user-provided sketches with globally exaggerated faces, with a soft constraint on facial marks, to achieve the best matching rates. This is because the user-provided sketch from the user's mental image, seems to be caricatured both locally and configurally.
人脸素描被用于目击者证词已经有一个世纪的历史了。在没有照片的情况下,这些素描在寻找嫌疑人时至关重要,但在目击者的脑海中形成了一个形象。然而,研究表明,目前用于目击者证词的程序存在两个主要问题。首先,它们会严重干扰目击者的记忆。其次,在很多情况下,这些程序会导致人脸图像与目标人脸相差很远。这两个问题与人类视觉系统的可塑性以及人类面部感知(整体)与当前素描制作方法(碎片)之间的差异有关。在本文中,我们提出了一些更真实的素描到照片匹配的见解。我们描述了如何从由非艺术目击者直接绘制的粗糙草图中检索身份特定信息。我们使用的草图仅包含面部成分轮廓和面部痕迹(例如皱纹和痣)。我们使用两种方法(PDM距离比较和PCA分类)将两种类型的草图(trace-over和用户提供的,各25张)与四种类型的人脸(原始、局部夸张、配置夸张和全局夸张,各249张)的自动匹配结果进行了比较。基于我们的研究结果,我们认为对于非艺术素描与照片的自动匹配,算法应该将用户提供的草图与全局夸张的人脸进行比较,并对面部标记进行软约束,以达到最佳匹配率。这是因为用户提供的草图来自用户的心理形象,似乎在局部和配置上都被漫画化了。
{"title":"A study on recognizing non-artistic face sketches","authors":"Hossein Nejati, T. Sim","doi":"10.1109/WACV.2011.5711509","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711509","url":null,"abstract":"Face sketches are being used in eyewitness testimonies for about a century. These sketches are crucial in finding suspects when no photo is available, but a mental image in the eyewitness's mind. However, research shows that current procedures used for eyewitness testimonies have two main problems. First, they can significantly disturb the memories of the eyewitness. Second, in many cases, these procedures result in face images far from their target faces. These two problems are related to the plasticity of the human visual system and the differences between face perception in humans (holistic) and current methods of sketch production (piecemeal). In this paper, we present some insights for more realistic sketch to photo matching. We describe how to retrieve identity specific information from crude sketches, directly drawn by the non-artistic eyewitnesses. The sketches we used merely contain facial component outlines and facial marks (e.g. wrinkles and moles). We compare results of automatically matching two types sketches (trace-over and user-provided, 25 each) to four types of faces (original, locally exaggerated, configurally exaggerated, and globally exaggerated, 249 each), using two methods (PDM distance comparison and PCA classification). Based on our results, we argue that for automatic non-artistic sketch to photo matching, the algorithms should compare the user-provided sketches with globally exaggerated faces, with a soft constraint on facial marks, to achieve the best matching rates. This is because the user-provided sketch from the user's mental image, seems to be caricatured both locally and configurally.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128124705","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Augmented transit maps 增强型交通地图
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711543
Matei Stroila, J. Mays, Bill Gale, Jeff Bach
We introduce a new class of mobile augmented reality navigation applications that allow people to interact with transit maps in public transit stations and vehicles. Our system consists of a database of coded transit maps, a vision engine for recognizing and tracking planar objects, and a graphics engine to overlay relevant real-time navigation information, such as the user's current location and the time to destination. We demonstrate this system with a prototype application built from open source components only. The application runs on a Nokia N900 mobile phone equipped with Maemo, a Debian Linux-based operating system. We use the OpenCV library and the new Frankencamera API for the vision engine. The application is written using the LGPL licensed Qt C++ Framework.
我们介绍了一类新的移动增强现实导航应用程序,它允许人们与公共交通站点和车辆中的交通地图进行交互。我们的系统包括一个编码交通地图数据库,一个用于识别和跟踪平面物体的视觉引擎,以及一个用于覆盖相关实时导航信息的图形引擎,例如用户当前位置和到达目的地的时间。我们用一个仅由开源组件构建的原型应用程序来演示这个系统。这款应用运行在搭载Maemo(一款基于Debian linux的操作系统)的诺基亚N900手机上。我们使用OpenCV库和新的franencamera API作为视觉引擎。该应用程序是使用LGPL许可的Qt c++框架编写的。
{"title":"Augmented transit maps","authors":"Matei Stroila, J. Mays, Bill Gale, Jeff Bach","doi":"10.1109/WACV.2011.5711543","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711543","url":null,"abstract":"We introduce a new class of mobile augmented reality navigation applications that allow people to interact with transit maps in public transit stations and vehicles. Our system consists of a database of coded transit maps, a vision engine for recognizing and tracking planar objects, and a graphics engine to overlay relevant real-time navigation information, such as the user's current location and the time to destination. We demonstrate this system with a prototype application built from open source components only. The application runs on a Nokia N900 mobile phone equipped with Maemo, a Debian Linux-based operating system. We use the OpenCV library and the new Frankencamera API for the vision engine. The application is written using the LGPL licensed Qt C++ Framework.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128373935","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A random center surround bottom up visual attention model useful for salient region detection 一个随机中心环绕自下而上的视觉注意模型,用于显著区域检测
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711499
T. Vikram, M. Tscherepanow, B. Wrede
In this article, we propose a bottom-up saliency model which works on capturing the contrast between random pixels in an image. The model is explained on the basis of the stimulus bias between two given stimuli (pixel intensity values) in an image and has a minimal set of tunable parameters. The methodology does not require any training bases or priors. We followed an established experimental setting and obtained state-of-the-art-results for salient region detection on the MSR dataset. Further experiments demonstrate that our method is robust to noise and has, in comparison to six other state-of-the-art models, a consistent performance in terms of recall, precision and F-measure.
在本文中,我们提出了一种自下而上的显著性模型,用于捕获图像中随机像素之间的对比度。该模型是基于图像中两个给定刺激(像素强度值)之间的刺激偏差来解释的,并且具有最小的可调参数集。该方法不需要任何培训基础或经验。我们遵循既定的实验设置,并获得了在MSR数据集上显著区域检测的最先进结果。进一步的实验表明,我们的方法对噪声具有鲁棒性,并且与其他六种最先进的模型相比,在召回率、精度和F-measure方面具有一致的性能。
{"title":"A random center surround bottom up visual attention model useful for salient region detection","authors":"T. Vikram, M. Tscherepanow, B. Wrede","doi":"10.1109/WACV.2011.5711499","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711499","url":null,"abstract":"In this article, we propose a bottom-up saliency model which works on capturing the contrast between random pixels in an image. The model is explained on the basis of the stimulus bias between two given stimuli (pixel intensity values) in an image and has a minimal set of tunable parameters. The methodology does not require any training bases or priors. We followed an established experimental setting and obtained state-of-the-art-results for salient region detection on the MSR dataset. Further experiments demonstrate that our method is robust to noise and has, in comparison to six other state-of-the-art models, a consistent performance in terms of recall, precision and F-measure.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134205774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Detecting questionable observers using face track clustering 利用人脸轨迹聚类检测可疑观察者
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711501
Jeremiah R. Barr, K. Bowyer, P. Flynn
We introduce the questionable observer detection problem: Given a collection of videos of crowds, determine which individuals appear unusually often across the set of videos. The algorithm proposed here detects these individuals by clustering sequences of face images. To provide robustness to sensor noise, facial expression and resolution variations, blur, and intermittent occlusions, we merge similar face image sequences from the same video and discard outlying face patterns prior to clustering. We present experiments on a challenging video dataset. The results show that the proposed method can surpass the performance of a clustering algorithm based on the VeriLook face recognition software by Neurotechnology both in terms of the detection rate and the false detection frequency.
我们引入了一个可疑的观察者检测问题:给定一组人群的视频,确定哪些个体在视频集中出现得异常频繁。本文提出的算法通过对人脸图像序列进行聚类来检测这些个体。为了提供对传感器噪声、面部表情和分辨率变化、模糊和间歇性遮挡的鲁棒性,我们合并了来自同一视频的相似面部图像序列,并在聚类之前丢弃了边缘的面部模式。我们在一个具有挑战性的视频数据集上进行了实验。结果表明,该方法在检测率和误检频率方面均优于基于VeriLook人脸识别软件的聚类算法。
{"title":"Detecting questionable observers using face track clustering","authors":"Jeremiah R. Barr, K. Bowyer, P. Flynn","doi":"10.1109/WACV.2011.5711501","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711501","url":null,"abstract":"We introduce the questionable observer detection problem: Given a collection of videos of crowds, determine which individuals appear unusually often across the set of videos. The algorithm proposed here detects these individuals by clustering sequences of face images. To provide robustness to sensor noise, facial expression and resolution variations, blur, and intermittent occlusions, we merge similar face image sequences from the same video and discard outlying face patterns prior to clustering. We present experiments on a challenging video dataset. The results show that the proposed method can surpass the performance of a clustering algorithm based on the VeriLook face recognition software by Neurotechnology both in terms of the detection rate and the false detection frequency.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122921507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Localized support vector machines using Parzen window for incomplete sets of categories 局部支持向量机使用Parzen窗口的不完全类别集
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711538
Kevin L. Veon, M. Mahoor
This paper describes a novel approach to pattern classification that combines Parzen window and support vector machines. Pattern classification is usually performed in universes where all possible categories are defined. Most of the current supervised learning classification techniques do not account for undefined categories. In a universe that is only partially defined, there may be objects that do not fall into the known set of categories. It would be a mistake to always classify these objects as a known category. We propose a Parzen window-based approach which is capable of classifying an object as not belonging to a known class. In our approach we use a Parzen window to identify local neighbors of a test point and train a localized support vector machine on the identified neighbors. Visual category recognition experiments are performed to compare the results of our approach, localized support vector machines using a k-nearest neighbors approach, and global support vector machines. Our experiments show that our Parzen window approach has superior results when testing with incomplete sets, and comparable results when testing with complete sets.
本文提出了一种结合Parzen窗口和支持向量机的模式分类方法。模式分类通常在定义了所有可能类别的宇宙中进行。目前大多数监督学习分类技术都没有考虑到未定义的类别。在一个只有部分定义的宇宙中,可能存在不属于已知类别集合的物体。总是把这些物体归为已知的一类是错误的。我们提出了一种基于Parzen窗口的方法,该方法能够将对象分类为不属于已知类。在我们的方法中,我们使用Parzen窗口来识别测试点的局部邻居,并在识别的邻居上训练局部支持向量机。进行了视觉类别识别实验,以比较我们的方法、使用k近邻方法的局部支持向量机和全局支持向量机的结果。我们的实验表明,我们的Parzen窗口方法在不完备集测试时具有优越的结果,在完备集测试时具有相当的结果。
{"title":"Localized support vector machines using Parzen window for incomplete sets of categories","authors":"Kevin L. Veon, M. Mahoor","doi":"10.1109/WACV.2011.5711538","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711538","url":null,"abstract":"This paper describes a novel approach to pattern classification that combines Parzen window and support vector machines. Pattern classification is usually performed in universes where all possible categories are defined. Most of the current supervised learning classification techniques do not account for undefined categories. In a universe that is only partially defined, there may be objects that do not fall into the known set of categories. It would be a mistake to always classify these objects as a known category. We propose a Parzen window-based approach which is capable of classifying an object as not belonging to a known class. In our approach we use a Parzen window to identify local neighbors of a test point and train a localized support vector machine on the identified neighbors. Visual category recognition experiments are performed to compare the results of our approach, localized support vector machines using a k-nearest neighbors approach, and global support vector machines. Our experiments show that our Parzen window approach has superior results when testing with incomplete sets, and comparable results when testing with complete sets.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127042196","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Supervised particle filter for tracking 2D human pose in monocular video 用于单目视频中二维人体姿态跟踪的监督粒子滤波
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711527
S. Sedai, D. Huynh, Bennamoun
In this paper, we propose a hybrid method that combines supervised learning and particle filtering to track the 2D pose of a human subject in monocular video sequences. Our approach, which we call a supervised particle filter method, consists of two steps: the training step and the tracking step. In the training step, we use a supervised learning method to train the regressors that take the silhouette descriptors as input and produce the 2D poses as output. In the tracking step, the output pose estimated from the regressors is combined with the particle filter to track the 2D pose in each video frame. Unlike the particle filter, our method does not require any manual initialization. We have tested our approach using the HumanEva video datasets and compared it with the standard particle filter and 2D pose estimation on individual frames. Our experimental results show that our approach can successfully track the pose over long video sequences and that it gives more accurate 2D human pose tracking than the particle filter and 2D pose estimation.
在本文中,我们提出了一种结合监督学习和粒子滤波的混合方法来跟踪单目视频序列中人体主体的二维姿态。我们的方法,我们称之为监督粒子滤波方法,包括两个步骤:训练步骤和跟踪步骤。在训练步骤中,我们使用监督学习方法来训练以轮廓描述符为输入并产生2D姿态作为输出的回归器。在跟踪步骤中,将回归量估计的输出姿态与粒子滤波相结合,跟踪每个视频帧中的二维姿态。与粒子过滤器不同,我们的方法不需要任何手动初始化。我们使用HumanEva视频数据集测试了我们的方法,并将其与标准粒子过滤器和单个帧的2D姿态估计进行了比较。我们的实验结果表明,我们的方法可以成功地跟踪长视频序列的姿态,并且比粒子滤波和二维姿态估计更准确地跟踪二维人体姿态。
{"title":"Supervised particle filter for tracking 2D human pose in monocular video","authors":"S. Sedai, D. Huynh, Bennamoun","doi":"10.1109/WACV.2011.5711527","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711527","url":null,"abstract":"In this paper, we propose a hybrid method that combines supervised learning and particle filtering to track the 2D pose of a human subject in monocular video sequences. Our approach, which we call a supervised particle filter method, consists of two steps: the training step and the tracking step. In the training step, we use a supervised learning method to train the regressors that take the silhouette descriptors as input and produce the 2D poses as output. In the tracking step, the output pose estimated from the regressors is combined with the particle filter to track the 2D pose in each video frame. Unlike the particle filter, our method does not require any manual initialization. We have tested our approach using the HumanEva video datasets and compared it with the standard particle filter and 2D pose estimation on individual frames. Our experimental results show that our approach can successfully track the pose over long video sequences and that it gives more accurate 2D human pose tracking than the particle filter and 2D pose estimation.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"144 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127298903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A performance study of an intelligent headlight control system 智能前照灯控制系统的性能研究
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711537
Ying Li, Sharath Pankanti
In this paper, we first present the architecture of an intelligent headlight control (IHC) system that we developed in our earlier work. This IHC system aims to automatically control a vehicle's beam state (high beam or low beam) during a night-time drive. A three-level decision framework built around a support vector machine (SVM) learning engine is then briefly discussed. Next, we switch our focus to the study of system performance by varying the SVM feature set, as well as by exploiting various SVM training options and adjustments through a set of experiments. We believe that what we learned from this performance study can provide readers useful guidelines on extracting effective SVM features within the IHC problem domain, as well as on training an effective SVM learning engine for more generalized applications.
在本文中,我们首先介绍了我们在早期工作中开发的智能前照灯控制(IHC)系统的架构。这个IHC系统的目的是在夜间驾驶时自动控制车辆的光束状态(远光灯或远光灯)。然后简要讨论了围绕支持向量机(SVM)学习引擎构建的三层决策框架。接下来,我们通过改变支持向量机特征集,以及利用各种支持向量机训练选项和通过一组实验进行调整,将重点转移到系统性能的研究上。我们相信,我们从这项性能研究中学到的东西可以为读者提供有用的指导,帮助他们在IHC问题领域中提取有效的SVM特征,以及为更广泛的应用训练有效的SVM学习引擎。
{"title":"A performance study of an intelligent headlight control system","authors":"Ying Li, Sharath Pankanti","doi":"10.1109/WACV.2011.5711537","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711537","url":null,"abstract":"In this paper, we first present the architecture of an intelligent headlight control (IHC) system that we developed in our earlier work. This IHC system aims to automatically control a vehicle's beam state (high beam or low beam) during a night-time drive. A three-level decision framework built around a support vector machine (SVM) learning engine is then briefly discussed. Next, we switch our focus to the study of system performance by varying the SVM feature set, as well as by exploiting various SVM training options and adjustments through a set of experiments. We believe that what we learned from this performance study can provide readers useful guidelines on extracting effective SVM features within the IHC problem domain, as well as on training an effective SVM learning engine for more generalized applications.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131540033","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Using visibility cameras to estimate atmospheric light extinction 使用能见度相机估算大气光消
Pub Date : 2011-01-05 DOI: 10.1109/WACV.2011.5711556
Nathan Graves, S. Newsam
We describe methods for estimating the coefficient of atmospheric light extinction using visibility cameras. We use a standard haze image formation model to estimate atmospheric transmission using local contrast features as well as a recently proposed dark channel prior. A log-linear model is then used to relate transmission and extinction. We train and evaluate our model using an extensive set of ground truth images acquired over a year long period from two visibility cameras in the Phoenix, Arizona region. We present informative results which are particularly accurate for a visibility index used in long-term haze studies.
我们描述了利用能见度相机估算大气光消系数的方法。我们使用标准的雾霾图像形成模型来估计大气传输,使用局部对比度特征以及最近提出的暗通道先验。然后使用对数线性模型来联系传输和消光。我们使用亚利桑那州凤凰城地区两台能见度相机在一年中获得的大量地面真实图像来训练和评估我们的模型。我们提供了信息丰富的结果,这些结果对于长期雾霾研究中使用的能见度指数特别准确。
{"title":"Using visibility cameras to estimate atmospheric light extinction","authors":"Nathan Graves, S. Newsam","doi":"10.1109/WACV.2011.5711556","DOIUrl":"https://doi.org/10.1109/WACV.2011.5711556","url":null,"abstract":"We describe methods for estimating the coefficient of atmospheric light extinction using visibility cameras. We use a standard haze image formation model to estimate atmospheric transmission using local contrast features as well as a recently proposed dark channel prior. A log-linear model is then used to relate transmission and extinction. We train and evaluate our model using an extensive set of ground truth images acquired over a year long period from two visibility cameras in the Phoenix, Arizona region. We present informative results which are particularly accurate for a visibility index used in long-term haze studies.","PeriodicalId":424724,"journal":{"name":"2011 IEEE Workshop on Applications of Computer Vision (WACV)","volume":"59 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-01-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131873945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
期刊
2011 IEEE Workshop on Applications of Computer Vision (WACV)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1