首页 > 最新文献

2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance最新文献

英文 中文
Visual Surveillance for Aircraft Activity Monitoring 用于飞机活动监测的视觉监视
D. Thirde, M. Borg, V. Valentin, F. Fusier, J. Aguilera, J. Ferryman, F. Brémond, M. Thonnat, M. Kampel
This paper presents a visual surveillance system for the automatic scene interpretation of airport aprons. The system comprises two modules - scene tracking and scene understanding. The scene tracking module, comprising a bottom-up methodology, and the scene understanding module, comprising a video event representation and recognition scheme, have been demonstrated to be a valid approach for apron monitoring
介绍了一种用于机场停机坪场景自动判读的视觉监控系统。该系统包括场景跟踪和场景理解两个模块。场景跟踪模块(包括自下而上的方法)和场景理解模块(包括视频事件表示和识别方案)已被证明是一种有效的停机坪监控方法
{"title":"Visual Surveillance for Aircraft Activity Monitoring","authors":"D. Thirde, M. Borg, V. Valentin, F. Fusier, J. Aguilera, J. Ferryman, F. Brémond, M. Thonnat, M. Kampel","doi":"10.1109/VSPETS.2005.1570923","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570923","url":null,"abstract":"This paper presents a visual surveillance system for the automatic scene interpretation of airport aprons. The system comprises two modules - scene tracking and scene understanding. The scene tracking module, comprising a bottom-up methodology, and the scene understanding module, comprising a video event representation and recognition scheme, have been demonstrated to be a valid approach for apron monitoring","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115821315","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Towards intelligent camera networks: a virtual vision approach 迈向智能摄像头网络:一种虚拟视觉方法
F. Qureshi, Demetri Terzopoulos
The goals of this paper are two-fold: (i) to present our initial efforts towards the realization of a fully autonomous sensor network of dynamic video cameras capable of providing perceptive coverage of a large public space, and (ii) to further the cause of exploiting visually and behaviorally realistic virtual environments in the development and testing of machine vision systems. In particular, our proposed sensor network employs techniques that enable a collection of active (pan-tilt-zoom) cameras to collaborate in performing various visual surveillance tasks, such as keeping one or more pedestrians within view, with minimal reliance on a human operator. The network features local and global autonomy and lacks any central controller, which entails robustness and scalability. Its functionality is the result of local decision-making capabilities at each camera node and communication between the nodes. We demonstrate our surveillance system in a virtual train station environment populated by autonomous, lifelike virtual pedestrians. Our readily reconfigurable virtual cameras generate synthetic video feeds that emulate those generated by real surveillance cameras monitoring public spaces. This type of research would be difficult in the real world given the costs of deploying and experimenting with an appropriately complex camera network in a large public space the size of a train station.
本文的目标有两个方面:(i)展示我们为实现一个能够提供大型公共空间感知覆盖的完全自主的动态视频摄像机传感器网络所做的初步努力,以及(ii)在机器视觉系统的开发和测试中进一步开发视觉和行为逼真的虚拟环境。特别是,我们提出的传感器网络采用技术,使一组活动(泛倾斜变焦)摄像机能够协作执行各种视觉监视任务,例如在视线内保持一个或多个行人,而对人工操作员的依赖最小。该网络具有局部和全局自治的特点,没有任何中央控制器,这需要鲁棒性和可扩展性。它的功能是每个相机节点的本地决策能力和节点之间的通信的结果。我们在一个由自主的、逼真的虚拟行人组成的虚拟火车站环境中演示了我们的监控系统。我们易于重新配置的虚拟摄像机生成合成的视频馈送,模拟由监控公共空间的真实监控摄像机生成的视频馈送。考虑到在火车站大小的大型公共空间部署和试验一个适当复杂的摄像头网络的成本,这种类型的研究在现实世界中是困难的。
{"title":"Towards intelligent camera networks: a virtual vision approach","authors":"F. Qureshi, Demetri Terzopoulos","doi":"10.1109/VSPETS.2005.1570913","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570913","url":null,"abstract":"The goals of this paper are two-fold: (i) to present our initial efforts towards the realization of a fully autonomous sensor network of dynamic video cameras capable of providing perceptive coverage of a large public space, and (ii) to further the cause of exploiting visually and behaviorally realistic virtual environments in the development and testing of machine vision systems. In particular, our proposed sensor network employs techniques that enable a collection of active (pan-tilt-zoom) cameras to collaborate in performing various visual surveillance tasks, such as keeping one or more pedestrians within view, with minimal reliance on a human operator. The network features local and global autonomy and lacks any central controller, which entails robustness and scalability. Its functionality is the result of local decision-making capabilities at each camera node and communication between the nodes. We demonstrate our surveillance system in a virtual train station environment populated by autonomous, lifelike virtual pedestrians. Our readily reconfigurable virtual cameras generate synthetic video feeds that emulate those generated by real surveillance cameras monitoring public spaces. This type of research would be difficult in the real world given the costs of deploying and experimenting with an appropriately complex camera network in a large public space the size of a train station.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"116 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122604104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 32
Vehicle Class Recognition from Video-Based on 3D Curve Probes 基于三维曲线探针的视频车辆类别识别
Dongjin Han, Matthew J. Leotta, D. Cooper, J. Mundy
A new approach is presented to vehicle-class recognition from a video clip. Two new concepts introduced are: probes consisting of local 3D curve-groups which when projected into video frames are features for recognizing vehicle classes in video clips; and Bayesian recognition based on class probability densities for groups of 3D distances between pairs of 3D probes. The most stable image features for vehicle class recognition appear to be image curves associate with 3D ridges on the vehicle surface. These ridges are mostly those occurring at metal/glass interfaces, two-surface intersections such as back and side, and self occluding contours such as wheel wells or vehicle-body apparent contours, i.e., silhouettes. There are other detectable surface curves, but most do not provide useful discriminatory features, and many of these are clutter, i.e., due to reflections from the somewhat shiny vehicle surface. Models are built and used for the considerable variability that exists in the features used. A Bayesian recognizer is then used for vehicle class recognition from a sequence of frames. The ultimate goal is a recognizer to deal with essentially all classes of civilian vehicles seen from arbitrary directions, at a broad range of distances and under the broad range of lighting ranging from sunny to cloudy. Experiments are run with a small set of classes to prove feasibility. This work uses estimated knowledge of the motion and position of the vehicle. We briefly indicate one way of inferring that information which uses ID projectivity invariance.
提出了一种从视频片段中识别车辆类别的新方法。引入了两个新概念:由局部三维曲线组组成的探针,当投影到视频帧中时,它是识别视频片段中车辆类别的特征;以及基于类概率密度的三维探针对之间三维距离组的贝叶斯识别。对于车辆类别识别而言,最稳定的图像特征似乎是与车辆表面三维脊相关联的图像曲线。这些脊线主要发生在金属/玻璃界面,双面交叉处,如背面和侧面,以及自遮挡轮廓,如轮井或车身表观轮廓,即轮廓。还有其他可检测的表面曲线,但大多数没有提供有用的区分特征,其中许多是杂波,即由于反射从一些闪亮的车辆表面。模型是为所使用的特征中存在的相当大的可变性而建立和使用的。然后使用贝叶斯识别器从一系列帧中进行车辆类别识别。最终目标是让识别器能够处理从任意方向、大范围距离和大范围光照(从晴天到阴天)下看到的基本上所有类别的民用车辆。为了证明可行性,我们对一小部分班级进行了实验。这项工作使用了对车辆运动和位置的估计知识。我们简要地指出了一种利用ID投影不变性来推断该信息的方法。
{"title":"Vehicle Class Recognition from Video-Based on 3D Curve Probes","authors":"Dongjin Han, Matthew J. Leotta, D. Cooper, J. Mundy","doi":"10.1109/VSPETS.2005.1570927","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570927","url":null,"abstract":"A new approach is presented to vehicle-class recognition from a video clip. Two new concepts introduced are: probes consisting of local 3D curve-groups which when projected into video frames are features for recognizing vehicle classes in video clips; and Bayesian recognition based on class probability densities for groups of 3D distances between pairs of 3D probes. The most stable image features for vehicle class recognition appear to be image curves associate with 3D ridges on the vehicle surface. These ridges are mostly those occurring at metal/glass interfaces, two-surface intersections such as back and side, and self occluding contours such as wheel wells or vehicle-body apparent contours, i.e., silhouettes. There are other detectable surface curves, but most do not provide useful discriminatory features, and many of these are clutter, i.e., due to reflections from the somewhat shiny vehicle surface. Models are built and used for the considerable variability that exists in the features used. A Bayesian recognizer is then used for vehicle class recognition from a sequence of frames. The ultimate goal is a recognizer to deal with essentially all classes of civilian vehicles seen from arbitrary directions, at a broad range of distances and under the broad range of lighting ranging from sunny to cloudy. Experiments are run with a small set of classes to prove feasibility. This work uses estimated knowledge of the motion and position of the vehicle. We briefly indicate one way of inferring that information which uses ID projectivity invariance.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"72 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114465437","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Background initialization with a new robust statistical approach 一种新的鲁棒统计方法的后台初始化
Hanzi Wang, D. Suter
Initializing a background model requires robust statistical methods as the task should be robust against random occurrences of foreground objects, as well as against general image noise. The median has been employed for the problem of background initialization. However, the median has only a breakdown point of 50%. In this paper, we propose a new robust method which can tolerate more than 50% of noise and foreground pixels in the background initialization process. We compare our new method with five others and give quantitative evaluations on background initialization. Experiments show that the proposed method achieves very promising results in background initialization.
初始化背景模型需要稳健的统计方法,因为该任务应该对前景对象的随机出现以及一般图像噪声具有鲁棒性。中值已用于背景初始化问题。然而,中位数只有50%的分解点。本文提出了一种新的鲁棒方法,该方法可以在背景初始化过程中容忍超过50%的噪声和前景像素。我们将新方法与其他五种方法进行了比较,并对背景初始化进行了定量评价。实验表明,该方法在后台初始化方面取得了很好的效果。
{"title":"Background initialization with a new robust statistical approach","authors":"Hanzi Wang, D. Suter","doi":"10.1109/VSPETS.2005.1570910","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570910","url":null,"abstract":"Initializing a background model requires robust statistical methods as the task should be robust against random occurrences of foreground objects, as well as against general image noise. The median has been employed for the problem of background initialization. However, the median has only a breakdown point of 50%. In this paper, we propose a new robust method which can tolerate more than 50% of noise and foreground pixels in the background initialization process. We compare our new method with five others and give quantitative evaluations on background initialization. Experiments show that the proposed method achieves very promising results in background initialization.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126692031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Modeling background from compressed video 从压缩视频建模背景
Weiqiang Wang, Datong Chen, Wen Gao, Jie Yang
Background models have been widely used for video surveillance and other applications. Methods for constructing background models and associated application algorithms are mainly studied in the spatial domain (pixel level). Many video sources, however, are in a compressed format before processing. In this paper, we propose an approach to construct background models directly from compressed video. The proposed approach utilizes the information from DCT coefficients at block level to construct accurate background models at pixel level. We implemented three representative algorithms of background models in the compressed domain, and theoretically explored their properties and the relationship with their counterparts in the spatial domain. We also present some general technical improvements to make them more capable for a wide range of applications. The proposed method can achieve the same accuracy as the methods that construct background models from the spatial domain with much lower computational cost (50% on average) and more compact storages.
背景模型已广泛应用于视频监控和其他应用。本文主要研究了空间域(像素级)背景模型的构建方法及其应用算法。然而,许多视频源在处理之前都是压缩格式的。本文提出了一种直接从压缩视频中构建背景模型的方法。该方法利用块级DCT系数信息在像素级构建精确的背景模型。在压缩域实现了三种具有代表性的背景模型算法,并从理论上探讨了它们在空间域的性质及其相互关系。我们还介绍了一些一般的技术改进,使它们能够更广泛地应用。该方法可以达到与从空间域构建背景模型的方法相同的精度,但计算成本更低(平均为50%),存储空间更紧凑。
{"title":"Modeling background from compressed video","authors":"Weiqiang Wang, Datong Chen, Wen Gao, Jie Yang","doi":"10.1109/VSPETS.2005.1570911","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570911","url":null,"abstract":"Background models have been widely used for video surveillance and other applications. Methods for constructing background models and associated application algorithms are mainly studied in the spatial domain (pixel level). Many video sources, however, are in a compressed format before processing. In this paper, we propose an approach to construct background models directly from compressed video. The proposed approach utilizes the information from DCT coefficients at block level to construct accurate background models at pixel level. We implemented three representative algorithms of background models in the compressed domain, and theoretically explored their properties and the relationship with their counterparts in the spatial domain. We also present some general technical improvements to make them more capable for a wide range of applications. The proposed method can achieve the same accuracy as the methods that construct background models from the spatial domain with much lower computational cost (50% on average) and more compact storages.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"82 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132325153","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Object Tracking using Color Correlogram 使用颜色相关图进行对象跟踪
Qi Zhao, Hai Tao
Color histogram based representations have been widely used for blob tracking. In this paper, a new color histogram based approach for object representation is proposed. By using a simplified version of color correlogram as object feature, spatial information is incorporated into object representation, which allows variations of rotation to be detected throughout the tracking therefore rotational objects can be more accurately tracked. The gradient decent method mean shift algorithm is adopted as the central computational module and further extended to a 3D domain to find the mostprobable target position and orientation simultaneously. The capability of the tracker to tolerate appearance changes like orientation changes, small scale changes, partial occlusions and background scene changes is demonstrated using real image sequences.
基于颜色直方图的表示已广泛用于斑点跟踪。本文提出了一种新的基于颜色直方图的对象表示方法。通过使用简化版本的颜色相关图作为对象特征,将空间信息纳入对象表示,从而可以在整个跟踪过程中检测到旋转的变化,从而可以更准确地跟踪旋转的对象。采用梯度体面法均值移位算法作为中心计算模块,进一步扩展到三维域,同时寻找最可能的目标位置和方向。跟踪器容忍外观变化的能力,如方向变化、小规模变化、部分遮挡和背景场景变化,使用真实图像序列进行了演示。
{"title":"Object Tracking using Color Correlogram","authors":"Qi Zhao, Hai Tao","doi":"10.1109/VSPETS.2005.1570924","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570924","url":null,"abstract":"Color histogram based representations have been widely used for blob tracking. In this paper, a new color histogram based approach for object representation is proposed. By using a simplified version of color correlogram as object feature, spatial information is incorporated into object representation, which allows variations of rotation to be detected throughout the tracking therefore rotational objects can be more accurately tracked. The gradient decent method mean shift algorithm is adopted as the central computational module and further extended to a 3D domain to find the mostprobable target position and orientation simultaneously. The capability of the tracker to tolerate appearance changes like orientation changes, small scale changes, partial occlusions and background scene changes is demonstrated using real image sequences.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130126178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
Tracking objects in occluding environments using temporal spatio-velocity transform 基于时空速度变换的遮挡环境中目标跟踪
K. Sato, J. Aggarwal
This paper presents a methodology for tracking moving objects in an occluded environment when occlusion occurs. We analyze sequences in which physical obstacles such as fences and trees divide an object into several blobs, both spatially and temporally. Our system successfully tracks the divided blobs as one object and reconstructs the whole object. We use the temporal spatio-velocity (TSV) transform and a cylinder model of object trajectories. The TSV transform extracts pixels with stable velocities and removes noisy pixels with unstable velocities. The cylinder model connects several blobs into one object and associates blobs that are occluded for a long period of time. We present results in which moving persons and vehicles occluded by fences and trees are successfully tracked even when the occlusion lasts for as long as 100 frames.
本文提出了一种在遮挡环境中跟踪运动物体的方法。我们分析了物理障碍物(如篱笆和树木)在空间上和时间上将物体分成几个团的序列。我们的系统成功地将分割的斑点作为一个对象进行跟踪,并重建整个对象。我们使用时间-空间-速度(TSV)变换和物体轨迹的圆柱体模型。TSV变换提取速度稳定的像素,去除速度不稳定的噪声像素。圆柱体模型将几个斑点连接到一个对象中,并将长时间遮挡的斑点联系起来。我们展示的结果表明,即使遮挡持续长达100帧,被围栏和树木遮挡的移动人员和车辆也能成功跟踪。
{"title":"Tracking objects in occluding environments using temporal spatio-velocity transform","authors":"K. Sato, J. Aggarwal","doi":"10.1109/VSPETS.2005.1570902","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570902","url":null,"abstract":"This paper presents a methodology for tracking moving objects in an occluded environment when occlusion occurs. We analyze sequences in which physical obstacles such as fences and trees divide an object into several blobs, both spatially and temporally. Our system successfully tracks the divided blobs as one object and reconstructs the whole object. We use the temporal spatio-velocity (TSV) transform and a cylinder model of object trajectories. The TSV transform extracts pixels with stable velocities and removes noisy pixels with unstable velocities. The cylinder model connects several blobs into one object and associates blobs that are occluded for a long period of time. We present results in which moving persons and vehicles occluded by fences and trees are successfully tracked even when the occlusion lasts for as long as 100 frames.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125847415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Efficient occlusion handling for multiple agent tracking by reasoning with surveillance event primitives 基于监视事件原语推理的多智能体跟踪的高效遮挡处理
P. Guha, A. Mukerjee, K. Venkatesh
Tracking multiple agents in a monocular visual surveillance system is often challenged by the phenomenon of occlusions. Agents entering the field of view can undergo two different forms of occlusions, either caused by crowding or due to obstructions by background objects at finite distances from the camera. The agents are primarily detected as foreground blobs and are characterized by their motion history and weighted color histograms. These features are further used for localizing them in subsequent frames through motion prediction assisted mean shift tracking. A number of Boolean predicates are evaluated based on the fractional overlaps between the localized regions and foreground blobs. We construct predicates describing a comprehensive set of possible surveillance event primitives including entry/exit, partial or complete occlusions by background objects, crowding, splitting of agents and algorithm failures resulting from track loss. Instantiation of these event primitives followed by selective feature updates enables us to develop an effective scheme for tracking multiple agents in relatively unconstrained environments.
在单目视觉监控系统中,跟踪多个智能体经常受到遮挡现象的挑战。进入视场的主体可以经历两种不同形式的遮挡,一种是由于拥挤造成的,另一种是由于距离相机有限距离的背景物体的阻碍造成的。这些代理主要被检测为前景斑点,并通过它们的运动历史和加权颜色直方图来表征。这些特征通过运动预测辅助平均移位跟踪进一步用于在后续帧中定位它们。基于局部区域和前景blob之间的分数重叠来评估许多布尔谓词。我们构建了描述一组全面的可能的监视事件原语的谓词,包括进入/退出、背景物体的部分或完全遮挡、拥挤、代理分裂和由轨迹丢失导致的算法失败。这些事件原语的实例化以及选择性的特征更新使我们能够开发一种有效的方案,用于在相对不受约束的环境中跟踪多个代理。
{"title":"Efficient occlusion handling for multiple agent tracking by reasoning with surveillance event primitives","authors":"P. Guha, A. Mukerjee, K. Venkatesh","doi":"10.1109/VSPETS.2005.1570897","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570897","url":null,"abstract":"Tracking multiple agents in a monocular visual surveillance system is often challenged by the phenomenon of occlusions. Agents entering the field of view can undergo two different forms of occlusions, either caused by crowding or due to obstructions by background objects at finite distances from the camera. The agents are primarily detected as foreground blobs and are characterized by their motion history and weighted color histograms. These features are further used for localizing them in subsequent frames through motion prediction assisted mean shift tracking. A number of Boolean predicates are evaluated based on the fractional overlaps between the localized regions and foreground blobs. We construct predicates describing a comprehensive set of possible surveillance event primitives including entry/exit, partial or complete occlusions by background objects, crowding, splitting of agents and algorithm failures resulting from track loss. Instantiation of these event primitives followed by selective feature updates enables us to develop an effective scheme for tracking multiple agents in relatively unconstrained environments.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131498386","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Object tracking with dynamic feature graph 基于动态特征图的目标跟踪
Feng Tang, Hai Tao
Two major problems for model-based object tracking are: 1) how to represent an object so that it can effectively be discriminated with background and other objects; 2) how to dynamically update the model to accommodate the object appearance and structure changes. Traditional appearance based representations (like color histogram) fails when the object has rich texture. In this paper, we present a novel feature based object representation attributed relational graph (ARG) for reliable object tracking. The object is modeled with invariant features (SIFT) and their relationship is encoded in the form of an ARG that can effectively distinguish itself from background and other objects. We adopt a competitive and efficient dynamic model to adoptively update the object model by adding new stable features as well as deleting inactive features. A relaxation labeling method is used to match the model graph with the observation to gel the best object position. Experiments show that our method can get reliable track even under dramatic appearance changes, occlusions, etc.
基于模型的目标跟踪的两个主要问题是:1)如何表示一个目标,使其能够有效地与背景和其他目标区分;2)如何动态更新模型以适应对象外观和结构的变化。当对象具有丰富的纹理时,传统的基于外观的表示(如颜色直方图)就失效了。本文提出了一种新的基于特征的对象表示属性关系图(ARG),用于可靠的目标跟踪。用不变特征(SIFT)对目标进行建模,并将它们之间的关系以ARG的形式进行编码,从而有效地将目标与背景和其他目标区分开来。我们采用一种竞争高效的动态模型,通过增加新的稳定特征和删除不活跃特征来自适应地更新对象模型。采用松弛标记法将模型图与观测值进行匹配,得到最佳目标位置。实验表明,该方法可以在剧烈的外观变化、遮挡等情况下获得可靠的跟踪结果。
{"title":"Object tracking with dynamic feature graph","authors":"Feng Tang, Hai Tao","doi":"10.1109/VSPETS.2005.1570894","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570894","url":null,"abstract":"Two major problems for model-based object tracking are: 1) how to represent an object so that it can effectively be discriminated with background and other objects; 2) how to dynamically update the model to accommodate the object appearance and structure changes. Traditional appearance based representations (like color histogram) fails when the object has rich texture. In this paper, we present a novel feature based object representation attributed relational graph (ARG) for reliable object tracking. The object is modeled with invariant features (SIFT) and their relationship is encoded in the form of an ARG that can effectively distinguish itself from background and other objects. We adopt a competitive and efficient dynamic model to adoptively update the object model by adding new stable features as well as deleting inactive features. A relaxation labeling method is used to match the model graph with the observation to gel the best object position. Experiments show that our method can get reliable track even under dramatic appearance changes, occlusions, etc.","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115301007","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 65
Can the Surveillance System Run Pose Variant Face Recognition in Real Time? 监控系统能否实时进行姿态变异人脸识别?
Hung-Son Le, Haibo Li
This paper presents an approach for face recognition across pose variations when only one sample image per person is available. From a near frontal face sample image, virtual views at different off-frontal angles were generated and used for the system training task. The manual work and computation burden, thus, are put on the offline training process, that makes it possible to build a real-time face recognition surveillance system. Our work exploited the inherent advantages of "single" HMM scheme, which is based on an ID discrete hidden Markov model (ID-DHMM) and is designed to avoid the need of retraining the system whenever it is provided new image(s). Experiment results on the CMU PIE face database demonstrate that the proposed scheme improves significantly the recognition performance
本文提出了一种在每个人只有一张样本图像时,跨姿态变化的人脸识别方法。从近正面人脸样本图像中,生成不同正面角度的虚拟视图,并用于系统训练任务。从而将人工劳动和计算负担转移到离线训练过程中,使构建实时人脸识别监控系统成为可能。我们的工作利用了“单一”HMM方案的固有优势,该方案基于ID离散隐马尔可夫模型(ID- dhmm),旨在避免在提供新图像时对系统进行再训练的需要。在CMU PIE人脸数据库上的实验结果表明,该方法显著提高了人脸识别性能
{"title":"Can the Surveillance System Run Pose Variant Face Recognition in Real Time?","authors":"Hung-Son Le, Haibo Li","doi":"10.1109/VSPETS.2005.1570917","DOIUrl":"https://doi.org/10.1109/VSPETS.2005.1570917","url":null,"abstract":"This paper presents an approach for face recognition across pose variations when only one sample image per person is available. From a near frontal face sample image, virtual views at different off-frontal angles were generated and used for the system training task. The manual work and computation burden, thus, are put on the offline training process, that makes it possible to build a real-time face recognition surveillance system. Our work exploited the inherent advantages of \"single\" HMM scheme, which is based on an ID discrete hidden Markov model (ID-DHMM) and is designed to avoid the need of retraining the system whenever it is provided new image(s). Experiment results on the CMU PIE face database demonstrate that the proposed scheme improves significantly the recognition performance","PeriodicalId":435841,"journal":{"name":"2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2005-10-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129268252","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2005 IEEE International Workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1