首页 > 最新文献

2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance最新文献

英文 中文
A Method Based on the Indirect Approach for Counting People in Crowded Scenes 一种基于间接方法的拥挤场景人口计数方法
Donatello Conte, P. Foggia, G. Percannella, M. Vento
This paper presents a method for counting people in a scene by establishing a mapping between some scene features and the number of people avoiding the complex foreground detection problem. The method is based on the use of SURF features and of an [left] [right]-SVR regressor to provide an estimate of this count. The algorithm takes specifically into account problems due to partial occlusions and to perspective.
本文提出了一种通过建立场景特征与人数之间的映射关系来计算场景中人数的方法,避免了复杂的前景检测问题。该方法基于SURF特征和[左][右]-SVR回归量的使用,以提供该计数的估计。该算法特别考虑了由于局部遮挡和透视造成的问题。
{"title":"A Method Based on the Indirect Approach for Counting People in Crowded Scenes","authors":"Donatello Conte, P. Foggia, G. Percannella, M. Vento","doi":"10.1109/AVSS.2010.86","DOIUrl":"https://doi.org/10.1109/AVSS.2010.86","url":null,"abstract":"This paper presents a method for counting people in a scene by establishing a mapping between some scene features and the number of people avoiding the complex foreground detection problem. The method is based on the use of SURF features and of an [left] [right]-SVR regressor to provide an estimate of this count. The algorithm takes specifically into account problems due to partial occlusions and to perspective.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121884096","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Soccer Player Activity Recognition by a Multivariate Features Integration 基于多元特征集成的足球运动员活动识别
T. D’orazio, Marco Leo, P. Mazzeo, P. Spagnolo
Human action recognition is an important research areain the field of computer vision having a great number ofreal-world applications. This paper presents a multi-viewaction recognition framework that extracts human silhouetteclues from different cameras, analyzes scene dynamicsand interprets human behaviors by the integration of multivariatedata in fuzzy rule-based system. Different featureshave been considered for the player action recognition someof them concerning the human silhouette analysis, and someothers related to the ball and player kinematics. Experimentswere carried out on a multi view image sequences ofa public soccer data set.
人体动作识别是计算机视觉领域的一个重要研究领域,在现实世界中有着广泛的应用。本文提出了一种基于模糊规则系统的多视角动作识别框架,该框架从不同的摄像机中提取人体轮廓,分析场景动态,并通过多变量数据的集成来解释人体行为。不同的特征已经被考虑用于球员的动作识别,其中一些涉及到人体轮廓分析,和一些有关的球和球员的运动学。对某公共足球数据集的多视点图像序列进行了实验。
{"title":"Soccer Player Activity Recognition by a Multivariate Features Integration","authors":"T. D’orazio, Marco Leo, P. Mazzeo, P. Spagnolo","doi":"10.1109/AVSS.2010.62","DOIUrl":"https://doi.org/10.1109/AVSS.2010.62","url":null,"abstract":"Human action recognition is an important research areain the field of computer vision having a great number ofreal-world applications. This paper presents a multi-viewaction recognition framework that extracts human silhouetteclues from different cameras, analyzes scene dynamicsand interprets human behaviors by the integration of multivariatedata in fuzzy rule-based system. Different featureshave been considered for the player action recognition someof them concerning the human silhouette analysis, and someothers related to the ball and player kinematics. Experimentswere carried out on a multi view image sequences ofa public soccer data set.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125008170","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A Spatiotemporal Motion-Vector Filter for Object Tracking on Compressed Video 一种用于压缩视频目标跟踪的时空运动矢量滤波器
Ronaldo C. Moura, E. M. Hemerly
In this paper, a novel filter for real-time object trackingfrom compressed domain is presented and evaluated. Thefilter significantly reduces the noisy motion vectors, that donot represent a real object movement, from Mpeg familycompressed videos. The filter analyses the spatial (neighborhood)and temporal coherence of block motion vectorsto determine if they are likely to represent true motion fromthe recorded scene. Qualitative and quantitative experimentsare performed displaying that the proposed spatiotemporalfilter (STF) outperforms the currently widelyused vector median filter. The results obtained with the spatiotemporalfilter make it suitable as a first step of any systemthat aims to detect and track objects from compressedvideo using its motion vectors.
提出了一种新的压缩域实时目标跟踪滤波器,并对其进行了评价。该滤波器显著减少了Mpeg家族压缩视频中不代表真实物体运动的噪声运动向量。该滤波器分析块运动矢量的空间(邻域)和时间相干性,以确定它们是否可能代表记录场景中的真实运动。定性和定量实验表明,所提出的时空滤波器(STF)优于目前广泛使用的矢量中值滤波器。使用时空滤波器获得的结果使其适合作为任何系统的第一步,旨在利用其运动向量从压缩视频中检测和跟踪对象。
{"title":"A Spatiotemporal Motion-Vector Filter for Object Tracking on Compressed Video","authors":"Ronaldo C. Moura, E. M. Hemerly","doi":"10.1109/AVSS.2010.82","DOIUrl":"https://doi.org/10.1109/AVSS.2010.82","url":null,"abstract":"In this paper, a novel filter for real-time object trackingfrom compressed domain is presented and evaluated. Thefilter significantly reduces the noisy motion vectors, that donot represent a real object movement, from Mpeg familycompressed videos. The filter analyses the spatial (neighborhood)and temporal coherence of block motion vectorsto determine if they are likely to represent true motion fromthe recorded scene. Qualitative and quantitative experimentsare performed displaying that the proposed spatiotemporalfilter (STF) outperforms the currently widelyused vector median filter. The results obtained with the spatiotemporalfilter make it suitable as a first step of any systemthat aims to detect and track objects from compressedvideo using its motion vectors.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121053955","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Real Time Human Action Recognition in a Long Video Sequence 长视频序列中的实时人体动作识别
Ping Guo, Z. Miao, Yuan Shen, Heng-Da Cheng
In recent years, most action recognition researches focuson isolated action analysis for short videos, but ignore theissue of continuous action recognition for a long videosequence in real time. This paper proposes a novelapproach for human action recognition in a video sequencewith whatever length, which, unlike previous works,requires no annotations and no pre-temporal-segmentations.Based on the bag of words representation and theprobabilistic Latent Semantic Analysis (pLSA) model, therecognition process goes frame by frame and the decisionupdates from time to time. Experimental results show thatthis approach is effective to recognize both isolated actionsand continuous actions no matter how long a videosequence is. This is very useful for real time applicationslike video surveillance. Besides, we also test our approachfor real time temporal video segmentation and real time keyframe extraction.
近年来,大多数动作识别研究都集中在对短视频的孤立动作分析上,而忽略了对长视频序列的实时连续动作识别问题。本文提出了一种在任意长度的视频序列中识别人类动作的新方法,与以往的工作不同,该方法不需要注释和预时间分割。基于词包表示和概率潜在语义分析(pLSA)模型,识别过程逐帧进行,决策不断更新。实验结果表明,无论视频序列有多长,该方法都能有效地识别孤立动作和连续动作。这对于视频监控等实时应用非常有用。此外,我们还测试了我们的方法用于实时视频分割和实时关键帧提取。
{"title":"Real Time Human Action Recognition in a Long Video Sequence","authors":"Ping Guo, Z. Miao, Yuan Shen, Heng-Da Cheng","doi":"10.1109/AVSS.2010.44","DOIUrl":"https://doi.org/10.1109/AVSS.2010.44","url":null,"abstract":"In recent years, most action recognition researches focuson isolated action analysis for short videos, but ignore theissue of continuous action recognition for a long videosequence in real time. This paper proposes a novelapproach for human action recognition in a video sequencewith whatever length, which, unlike previous works,requires no annotations and no pre-temporal-segmentations.Based on the bag of words representation and theprobabilistic Latent Semantic Analysis (pLSA) model, therecognition process goes frame by frame and the decisionupdates from time to time. Experimental results show thatthis approach is effective to recognize both isolated actionsand continuous actions no matter how long a videosequence is. This is very useful for real time applicationslike video surveillance. Besides, we also test our approachfor real time temporal video segmentation and real time keyframe extraction.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133875431","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
A Bayesian Framework for Online Interaction Classification 在线交互分类的贝叶斯框架
S. Maludrottu, M. Beoldo, M. Alvarez, C. Regazzoni
Real-time automatic human behavior recognition is oneof the most challenging tasks for intelligent surveillancesystems. Its importance lies in the possibility of robust detectionof suspicious behaviors in order to prevent possiblethreats. The widespread integration of tracking algorithmsinto modern surveillance systems makes it possible to acquiredescriptive motion patterns of different human activities.In this work, a statistical framework for human interactionrecognition based on Dynamic Bayesian Networks(DBNs) is presented: the environment is partitioned by atopological algorithm into a set of zones that are used to definethe state of the DBNs. Interactive and non-interactivebehaviors are described in terms of sequences of significantmotion events in the topological map of the environment.Finally, by means of an incremental classification measure,a scenario can be classified while it is currently evolving.In this way an autonomous surveillance system can detectand cope with potential threats in real-time.
实时自动识别人类行为是智能监控系统中最具挑战性的任务之一。它的重要性在于有可能对可疑行为进行强有力的检测,以防止可能的威胁。跟踪算法与现代监视系统的广泛集成使得获取不同人类活动的描述性运动模式成为可能。在这项工作中,提出了一个基于动态贝叶斯网络(dbn)的人类交互识别的统计框架:通过拓扑算法将环境划分为一组用于定义dbn状态的区域。交互和非交互行为是根据环境拓扑图中的重要运动事件序列来描述的。最后,通过增量分类度量,可以在场景正在发展时对其进行分类。通过这种方式,自主监视系统可以实时检测和应对潜在威胁。
{"title":"A Bayesian Framework for Online Interaction Classification","authors":"S. Maludrottu, M. Beoldo, M. Alvarez, C. Regazzoni","doi":"10.1109/AVSS.2010.56","DOIUrl":"https://doi.org/10.1109/AVSS.2010.56","url":null,"abstract":"Real-time automatic human behavior recognition is oneof the most challenging tasks for intelligent surveillancesystems. Its importance lies in the possibility of robust detectionof suspicious behaviors in order to prevent possiblethreats. The widespread integration of tracking algorithmsinto modern surveillance systems makes it possible to acquiredescriptive motion patterns of different human activities.In this work, a statistical framework for human interactionrecognition based on Dynamic Bayesian Networks(DBNs) is presented: the environment is partitioned by atopological algorithm into a set of zones that are used to definethe state of the DBNs. Interactive and non-interactivebehaviors are described in terms of sequences of significantmotion events in the topological map of the environment.Finally, by means of an incremental classification measure,a scenario can be classified while it is currently evolving.In this way an autonomous surveillance system can detectand cope with potential threats in real-time.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129751417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Person Re-identification Using Spatial Covariance Regions of Human Body Parts 基于人体部位空间协方差区域的人物再识别
Sławomir Bąk, E. Corvée, F. Brémond, M. Thonnat
In many surveillance systems there is a requirement todetermine whether a given person of interest has alreadybeen observed over a network of cameras. This is the personre-identification problem. The human appearance obtainedin one camera is usually different from the ones obtained inanother camera. In order to re-identify people the humansignature should handle difference in illumination, pose andcamera parameters. We propose a new appearance modelbased on spatial covariance regions extracted from humanbody parts. The new spatial pyramid scheme is applied tocapture the correlation between human body parts in orderto obtain a discriminative human signature. The humanbody parts are automatically detected using Histograms ofOriented Gradients (HOG). The method is evaluated usingbenchmark video sequences from i-LIDS Multiple-CameraTracking Scenario data set. The re-identification performanceis presented using the cumulative matching characteristic(CMC) curve. Finally, we show that the proposedapproach outperforms state of the art methods.
在许多监视系统中,需要确定是否已经通过摄像机网络观察到某个特定的感兴趣的人。这就是个人识别问题。在一个摄像机中获得的人的外观通常与在另一个摄像机中获得的不同。为了重新识别人,人的签名应该处理光照、姿势和相机参数的差异。提出了一种基于人体部位空间协方差区域提取的外观模型。利用新的空间金字塔图来捕捉人体部位之间的相关性,从而获得具有区别性的人体特征。使用直方图定向梯度(HOG)自动检测人体部位。使用i-LIDS多摄像机跟踪场景数据集的基准视频序列对该方法进行了评估。利用累积匹配特性(CMC)曲线描述了该方法的再识别性能。最后,我们证明了所提出的方法优于最先进的方法。
{"title":"Person Re-identification Using Spatial Covariance Regions of Human Body Parts","authors":"Sławomir Bąk, E. Corvée, F. Brémond, M. Thonnat","doi":"10.1109/AVSS.2010.34","DOIUrl":"https://doi.org/10.1109/AVSS.2010.34","url":null,"abstract":"In many surveillance systems there is a requirement todetermine whether a given person of interest has alreadybeen observed over a network of cameras. This is the personre-identification problem. The human appearance obtainedin one camera is usually different from the ones obtained inanother camera. In order to re-identify people the humansignature should handle difference in illumination, pose andcamera parameters. We propose a new appearance modelbased on spatial covariance regions extracted from humanbody parts. The new spatial pyramid scheme is applied tocapture the correlation between human body parts in orderto obtain a discriminative human signature. The humanbody parts are automatically detected using Histograms ofOriented Gradients (HOG). The method is evaluated usingbenchmark video sequences from i-LIDS Multiple-CameraTracking Scenario data set. The re-identification performanceis presented using the cumulative matching characteristic(CMC) curve. Finally, we show that the proposedapproach outperforms state of the art methods.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133190388","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 273
Human Localization in a Cluttered Space Using Multiple Cameras 使用多个摄像头在混乱空间中进行人类定位
Jiali Shen, Weiqi Yan, P. Miller, Huiyu Zhou
The use of single and dual-camera approaches tolocating a subject in a 3-D cluttered space is investigated.Specifically, we investigate the case where the lowerportion of the body may be occluded, e.g., by a chair on abus. Experiments were conducted involving elevensubjects moving along a pre-designated route within acluttered space. For each time instant the position of eachsubject was manually estimated and compared to thatproduced automatically. The dual camera approach wasfound to give significantly better performance than thesingle camera approach. It was found that inaccuratebounding of the lowest part of the subject, due toocclusion, led to localisation errors in range as large as10m for the latter. Using the side bounds of the detectedobject, which were found to be robust, accurate azimuthestimates can be obtained for a single camera. The dualcameraapproach exploits the greater degree of accuracyin azimuth to estimate the range through triangulation,giving average localisation errors of 40cm over the spaceof interest.
使用单摄像头和双摄像头的方法来定位一个对象在一个三维杂乱的空间进行了研究。具体来说,我们调查的情况下,身体的下半部分可能被闭塞,例如,在走动的椅子。实验涉及11名受试者沿着预先指定的路线在拥挤的空间中移动。在每个时间瞬间,每个受试者的位置都是人工估计的,并与自动产生的位置进行比较。双摄像头方法被发现比单摄像头方法提供了明显更好的性能。研究发现,由于结论,受试者最低部分的边界不准确,导致后者的定位误差范围高达10米。利用被检测物体的边界是鲁棒的,可以获得单个相机的精确方位估计。双摄像头的方法利用更高的方位角精度,通过三角测量来估计距离,在感兴趣的空间上给出40厘米的平均定位误差。
{"title":"Human Localization in a Cluttered Space Using Multiple Cameras","authors":"Jiali Shen, Weiqi Yan, P. Miller, Huiyu Zhou","doi":"10.1109/AVSS.2010.60","DOIUrl":"https://doi.org/10.1109/AVSS.2010.60","url":null,"abstract":"The use of single and dual-camera approaches tolocating a subject in a 3-D cluttered space is investigated.Specifically, we investigate the case where the lowerportion of the body may be occluded, e.g., by a chair on abus. Experiments were conducted involving elevensubjects moving along a pre-designated route within acluttered space. For each time instant the position of eachsubject was manually estimated and compared to thatproduced automatically. The dual camera approach wasfound to give significantly better performance than thesingle camera approach. It was found that inaccuratebounding of the lowest part of the subject, due toocclusion, led to localisation errors in range as large as10m for the latter. Using the side bounds of the detectedobject, which were found to be robust, accurate azimuthestimates can be obtained for a single camera. The dualcameraapproach exploits the greater degree of accuracyin azimuth to estimate the range through triangulation,giving average localisation errors of 40cm over the spaceof interest.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127739505","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Learning Dense Optical-Flow Trajectory Patterns for Video Object Extraction 学习用于视频对象提取的密集光流轨迹模式
Wang-Chou Lu, Y. Wang, Chu-Song Chen
We proposes an unsupervised method to address videoobject extraction (VOE) in uncontrolled videos, i.e. videoscaptured by low-resolution and freely moving cameras. Weadvocate the use of dense optical-flow trajectories (DOTs),which are obtained by propagating the optical flow informationat the pixel level. Therefore, no interest point extractionis required in our framework. To integrate colorand and shape information of moving objects, we groupthe DOTs at the super-pixel level to extract co-motion regions,and use the associated pyramid histogram of orientedgradients (PHOG) descriptors to extract objects of interestacross video frames. Our approach for VOE is easy to implement,and the use of DOTs for both motion segmentationand object tracking is more robust than existing trajectorybasedmethods. Experiments on several video sequencesexhibit the feasibility of our proposed VOE framework.
我们提出了一种无监督的方法来解决非受控视频中的视频对象提取(VOE)问题,即由低分辨率和自由移动的摄像机捕获的视频。我们提倡使用密集光流轨迹(DOTs),它是通过在像素级传播光流信息获得的。因此,在我们的框架中不需要提取兴趣点。为了整合运动物体的颜色和形状信息,我们在超像素级别对DOTs进行分组以提取共同运动区域,并使用相关的定向梯度金字塔直方图(PHOG)描述符在视频帧中提取感兴趣的物体。我们的VOE方法很容易实现,并且在运动分割和目标跟踪中使用DOTs比现有的基于轨迹的方法更健壮。在多个视频序列上的实验证明了我们提出的VOE框架的可行性。
{"title":"Learning Dense Optical-Flow Trajectory Patterns for Video Object Extraction","authors":"Wang-Chou Lu, Y. Wang, Chu-Song Chen","doi":"10.1109/AVSS.2010.79","DOIUrl":"https://doi.org/10.1109/AVSS.2010.79","url":null,"abstract":"We proposes an unsupervised method to address videoobject extraction (VOE) in uncontrolled videos, i.e. videoscaptured by low-resolution and freely moving cameras. Weadvocate the use of dense optical-flow trajectories (DOTs),which are obtained by propagating the optical flow informationat the pixel level. Therefore, no interest point extractionis required in our framework. To integrate colorand and shape information of moving objects, we groupthe DOTs at the super-pixel level to extract co-motion regions,and use the associated pyramid histogram of orientedgradients (PHOG) descriptors to extract objects of interestacross video frames. Our approach for VOE is easy to implement,and the use of DOTs for both motion segmentationand object tracking is more robust than existing trajectorybasedmethods. Experiments on several video sequencesexhibit the feasibility of our proposed VOE framework.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121104183","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Bringing Richer Information with Reliability to Automated Traffic Monitoring from the Fusion of Multiple Cameras, Inductive Loops and Road Maps 通过融合多个摄像头、感应回路和道路地图,为自动交通监控带来更丰富、更可靠的信息
Kostia Robert
This paper presents a novel, deterministic framework toextract the traffic state of an intersection with high reliabilityand in real-time. The multiple video cameras and inductiveloops at the intersection are fused on a common planewhich consists of a satellite map. The sensors are registeredfrom a CAD map of the intersection that is aligned on thesatellite map. The cameras are calibrated to provide themapping equations that project the detected vehicle positionsonto the coordinate system of the satellite map. Weuse a night time vehicle detection algorithm to process thecamera frames. The inductive loops confirm or reject thevehicle tracks measured by the cameras, and the fusion ofcamera and loop provides an additional feature : the vehiclelength. A Kalman filter linearly tracks the vehicles alongthe lanes. Over time, this filter reduces the noise presentin the measurements. The advantage of this approach isthat the detected vehicles and their parameters acquire avery high confidence, which brings almost 100% accuracyof the traffic state. An empirical evaluation is performedon a testbed intersection. We show the improvement of thisframework over single sensor frameworks.
本文提出了一种新颖的、确定性的、高可靠的、实时的交叉口交通状态提取框架。交叉路口的多个摄像机和感应摄像机融合在一个由卫星地图组成的公共平面上。这些传感器是根据在卫星地图上对齐的十字路口的CAD地图注册的。摄像机经过校准,以提供映射方程,将检测到的车辆位置投影到卫星地图的坐标系统中。我们使用夜间车辆检测算法来处理摄像机帧。感应回路确认或拒绝由摄像头测量的车辆轨迹,摄像头和回路的融合提供了一个额外的功能:车辆长度。卡尔曼滤波线性跟踪车道上的车辆。随着时间的推移,该滤波器减少了测量中的噪声。该方法的优点是被检测车辆及其参数具有很高的置信度,几乎可以达到100%的交通状态准确性。在一个试验台交叉点上进行了经验评价。我们展示了该框架相对于单一传感器框架的改进。
{"title":"Bringing Richer Information with Reliability to Automated Traffic Monitoring from the Fusion of Multiple Cameras, Inductive Loops and Road Maps","authors":"Kostia Robert","doi":"10.1109/AVSS.2010.67","DOIUrl":"https://doi.org/10.1109/AVSS.2010.67","url":null,"abstract":"This paper presents a novel, deterministic framework toextract the traffic state of an intersection with high reliabilityand in real-time. The multiple video cameras and inductiveloops at the intersection are fused on a common planewhich consists of a satellite map. The sensors are registeredfrom a CAD map of the intersection that is aligned on thesatellite map. The cameras are calibrated to provide themapping equations that project the detected vehicle positionsonto the coordinate system of the satellite map. Weuse a night time vehicle detection algorithm to process thecamera frames. The inductive loops confirm or reject thevehicle tracks measured by the cameras, and the fusion ofcamera and loop provides an additional feature : the vehiclelength. A Kalman filter linearly tracks the vehicles alongthe lanes. Over time, this filter reduces the noise presentin the measurements. The advantage of this approach isthat the detected vehicles and their parameters acquire avery high confidence, which brings almost 100% accuracyof the traffic state. An empirical evaluation is performedon a testbed intersection. We show the improvement of thisframework over single sensor frameworks.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124427559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Real-Time 3D Human Pose Estimation from Monocular View with Applications to Event Detection and Video Gaming 单目实时三维人体姿态估计及其在事件检测和视频游戏中的应用
Shian-Ru Ke, Liang-Jia Zhu, Jenq-Neng Hwang, Hung-I Pai, Kung-Ming Lan, C. Liao
We present an effective real-time approach forautomatically estimating 3D human body poses frommonocular video sequences. In this approach, human bodyis automatically detected from video sequence, then imagefeatures such as silhouette, edge and color are extractedand integrated to infer 3D human poses by iterativelyminimizing the cost function defined between 2D featuresderived from the projected 3D model and those extractedfrom video sequence. In addition, 2D locations of head,hands, and feet are tracked to facilitate 3D tracking. Whentracking failure happens, the approach can detect andrecover from failures quickly. Finally, the efficiency androbustness of the proposed approach is shown in two realapplications: human event detection and video gaming.
我们提出了一种有效的实时方法,用于从单目视频序列中自动估计3D人体姿势。在该方法中,从视频序列中自动检测人体,然后通过迭代最小化从投影3D模型中提取的2D特征与从视频序列中提取的2D特征之间定义的代价函数,提取和集成图像特征(如轮廓、边缘和颜色)来推断3D人体姿势。此外,头部,手和脚的2D位置被跟踪,以方便3D跟踪。当跟踪故障发生时,该方法可以快速检测并从故障中恢复。最后,在人类事件检测和视频游戏两个实际应用中证明了该方法的有效性和可靠性。
{"title":"Real-Time 3D Human Pose Estimation from Monocular View with Applications to Event Detection and Video Gaming","authors":"Shian-Ru Ke, Liang-Jia Zhu, Jenq-Neng Hwang, Hung-I Pai, Kung-Ming Lan, C. Liao","doi":"10.1109/AVSS.2010.80","DOIUrl":"https://doi.org/10.1109/AVSS.2010.80","url":null,"abstract":"We present an effective real-time approach forautomatically estimating 3D human body poses frommonocular video sequences. In this approach, human bodyis automatically detected from video sequence, then imagefeatures such as silhouette, edge and color are extractedand integrated to infer 3D human poses by iterativelyminimizing the cost function defined between 2D featuresderived from the projected 3D model and those extractedfrom video sequence. In addition, 2D locations of head,hands, and feet are tracked to facilitate 3D tracking. Whentracking failure happens, the approach can detect andrecover from failures quickly. Finally, the efficiency androbustness of the proposed approach is shown in two realapplications: human event detection and video gaming.","PeriodicalId":415758,"journal":{"name":"2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance","volume":"16 2","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-08-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121001785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
期刊
2010 7th IEEE International Conference on Advanced Video and Signal Based Surveillance
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1