首页 > 最新文献

2007 IEEE Conference on Advanced Video and Signal Based Surveillance最新文献

英文 中文
Combination of self-organization map and kernel mutual subspace method for video surveillance 自组织映射与核互子空间相结合的视频监控方法
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425297
Bailing Zhang, Junbum Park, Hanseok Ko
This paper addresses the video surveillance issue of automatically identifying moving vehicles and people from continuous observation of image sequences. With a single far-field surveillance camera, moving objects are first segmented by simple background subtraction. To reduce the redundancy and select the representative prototypes from input video streams, the self-organizing feature map (SOM) is applied for both training and testing sequences. The recognition scheme is designed based on the recently proposed kernel mutual subspace (KMS) model. As an alternative to some probability-based models, KMS does not make assumptions about the data sampling processing and offers an efficient and robust classifier. Experiments demonstrated a highly accurate recognition result, showing the model's applicability in real-world surveillance system.
本文研究了从连续观察图像序列中自动识别移动车辆和人员的视频监控问题。对于单个远场监控摄像机,首先通过简单的背景减法分割运动物体。为了减少冗余并从输入视频流中选择具有代表性的原型,将自组织特征映射(SOM)应用于训练序列和测试序列。基于最近提出的核互子空间(KMS)模型设计了识别方案。作为一些基于概率的模型的替代方案,KMS不对数据采样处理进行假设,并提供了高效和鲁棒的分类器。实验结果表明,该模型具有较高的识别精度,在实际监控系统中具有一定的适用性。
{"title":"Combination of self-organization map and kernel mutual subspace method for video surveillance","authors":"Bailing Zhang, Junbum Park, Hanseok Ko","doi":"10.1109/AVSS.2007.4425297","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425297","url":null,"abstract":"This paper addresses the video surveillance issue of automatically identifying moving vehicles and people from continuous observation of image sequences. With a single far-field surveillance camera, moving objects are first segmented by simple background subtraction. To reduce the redundancy and select the representative prototypes from input video streams, the self-organizing feature map (SOM) is applied for both training and testing sequences. The recognition scheme is designed based on the recently proposed kernel mutual subspace (KMS) model. As an alternative to some probability-based models, KMS does not make assumptions about the data sampling processing and offers an efficient and robust classifier. Experiments demonstrated a highly accurate recognition result, showing the model's applicability in real-world surveillance system.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134340106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Tracking of two acoustic sources in reverberant environments using a particle swarm optimizer 用粒子群优化器跟踪混响环境中的两个声源
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425373
F. Antonacci, Davide Riva, A. Sarti, M. Tagliasacchi, S. Tubaro
In this paper we consider the problem of tracking multiple acoustic sources in reverberant environments. The solution that we propose is based on the combination of two techniques. A blind source separation (BSS) method known as TRINICON [5] is applied to the signals acquired by the microphone arrays. The TRINICON de-mixing filters are used to obtain the Time Differences of Arrival (TDOAs), which are related to the source location through a nonlinear function. A particle filter is then applied in order to localize the sources. Particles move according to a swarm-like dynamics, which significatively reduces the number of particles involved with respect to traditional particle filter. We discuss results for the case of two sources and four microphone pairs. In addition, we propose a method, based on detecting source inactivity, which overcomes the ambiguities that intrinsically arise when only two microphone pairs are used. Experimental results demonstrate that the average localization error on a variety of pseudo-random trajectories is around 40 cm when the T60 reverberation time is 0.6s.
本文研究了混响环境中多声源的跟踪问题。我们提出的解决方案是基于两种技术的结合。对麦克风阵列采集的信号采用盲源分离(blind source separation, BSS)方法TRINICON[5]。利用TRINICON解混滤波器获得与源位置有关的到达时间差(TDOAs), TDOAs是一个非线性函数。然后应用粒子滤波器来定位源。粒子的运动以一种类似于群体的动态方式进行,与传统的粒子滤波相比,这大大减少了粒子的数量。我们讨论了两个声源和四个传声器对情况下的结果。此外,我们提出了一种基于检测源不活动的方法,该方法克服了仅使用两个麦克风对时固有的模糊性。实验结果表明,当T60混响时间为0.6s时,各种伪随机轨迹的平均定位误差在40 cm左右。
{"title":"Tracking of two acoustic sources in reverberant environments using a particle swarm optimizer","authors":"F. Antonacci, Davide Riva, A. Sarti, M. Tagliasacchi, S. Tubaro","doi":"10.1109/AVSS.2007.4425373","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425373","url":null,"abstract":"In this paper we consider the problem of tracking multiple acoustic sources in reverberant environments. The solution that we propose is based on the combination of two techniques. A blind source separation (BSS) method known as TRINICON [5] is applied to the signals acquired by the microphone arrays. The TRINICON de-mixing filters are used to obtain the Time Differences of Arrival (TDOAs), which are related to the source location through a nonlinear function. A particle filter is then applied in order to localize the sources. Particles move according to a swarm-like dynamics, which significatively reduces the number of particles involved with respect to traditional particle filter. We discuss results for the case of two sources and four microphone pairs. In addition, we propose a method, based on detecting source inactivity, which overcomes the ambiguities that intrinsically arise when only two microphone pairs are used. Experimental results demonstrate that the average localization error on a variety of pseudo-random trajectories is around 40 cm when the T60 reverberation time is 0.6s.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117173865","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
What are customers looking at? 顾客在看什么?
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425345
Xiaoming Liu, N. Krahnstoever, Ting Yu, P. Tu
Computer vision approaches for retail applications can provide value far beyond the common domain of loss prevention. Gaining insight into the movement and behaviors of shoppers is of high interest for marketing, merchandizing, store operations and data mining. Of particular interest is the process of purchase decision making. What catches a customers attention? What products go unnoticed? What does a customer look at before making a final decision? Towards this goal we presents a system that detects and tracks both the location and gaze of shoppers in retail environments. While networks of standard overhead store cameras are used for tracking the location of customers, small in-shelf cameras are used for estimating customer gaze. The presented system operates robustly in real-time and can be deployed in a variety of retail applications.
零售应用的计算机视觉方法提供的价值远远超出了常见的防损领域。深入了解购物者的活动和行为对于市场营销、商品销售、商店运营和数据挖掘都是非常重要的。特别有趣的是购买决策的过程。什么能吸引顾客的注意力?哪些产品不被注意?顾客在做最后决定之前会看什么?为了实现这一目标,我们提出了一个系统,可以检测和跟踪零售环境中购物者的位置和目光。标准的商店顶部摄像头网络用于跟踪顾客的位置,而小型的货架内摄像头用于估计顾客的目光。所提出的系统运行健壮,实时,可部署在各种零售应用。
{"title":"What are customers looking at?","authors":"Xiaoming Liu, N. Krahnstoever, Ting Yu, P. Tu","doi":"10.1109/AVSS.2007.4425345","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425345","url":null,"abstract":"Computer vision approaches for retail applications can provide value far beyond the common domain of loss prevention. Gaining insight into the movement and behaviors of shoppers is of high interest for marketing, merchandizing, store operations and data mining. Of particular interest is the process of purchase decision making. What catches a customers attention? What products go unnoticed? What does a customer look at before making a final decision? Towards this goal we presents a system that detects and tracks both the location and gaze of shoppers in retail environments. While networks of standard overhead store cameras are used for tracking the location of customers, small in-shelf cameras are used for estimating customer gaze. The presented system operates robustly in real-time and can be deployed in a variety of retail applications.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121631708","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 52
Classification of gait types based on the duty-factor 基于责任因子的步态类型分类
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425330
P. Fihl, T. Moeslund
This paper deals with classification of human gait types based on the notion that different gait types are in fact different types of locomotion, i.e., running is not simply walking done faster. We present the duty-factor, which is a descriptor based on this notion. The duty-factor is independent on the speed of the human, the cameras setup etc. and hence a robust descriptor for gait classification. The duty-factor is basically a matter of measuring the ground support of the feet with respect to the stride. We estimate this by comparing the incoming silhouettes to a database of silhouettes with known ground support. Silhouettes are extracted using the codebook method and represented using shape contexts. The matching with database silhouettes is done using the Hungarian method. While manually estimated duty-factors show a clear classification the presented system contains misclassifications due to silhouette noise and ambiguities in the database silhouettes.
本文基于不同的步态类型实际上是不同类型的运动的概念来处理人类步态类型的分类,即跑步不是简单地走得更快。我们提出了责任因子,它是基于这个概念的一个描述符。责任因子独立于人的速度,相机设置等,因此是步态分类的鲁棒描述符。责任系数基本上是测量双脚相对于步幅的地面支撑力的问题。我们通过将传入的轮廓与已知地面支持的轮廓数据库进行比较来估计这一点。使用代码本方法提取轮廓,并使用形状上下文表示轮廓。与数据库轮廓的匹配使用匈牙利方法完成。虽然人工估计的责任因子显示出清晰的分类,但由于轮廓噪声和数据库轮廓的模糊性,所提出的系统存在错误分类。
{"title":"Classification of gait types based on the duty-factor","authors":"P. Fihl, T. Moeslund","doi":"10.1109/AVSS.2007.4425330","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425330","url":null,"abstract":"This paper deals with classification of human gait types based on the notion that different gait types are in fact different types of locomotion, i.e., running is not simply walking done faster. We present the duty-factor, which is a descriptor based on this notion. The duty-factor is independent on the speed of the human, the cameras setup etc. and hence a robust descriptor for gait classification. The duty-factor is basically a matter of measuring the ground support of the feet with respect to the stride. We estimate this by comparing the incoming silhouettes to a database of silhouettes with known ground support. Silhouettes are extracted using the codebook method and represented using shape contexts. The matching with database silhouettes is done using the Hungarian method. While manually estimated duty-factors show a clear classification the presented system contains misclassifications due to silhouette noise and ambiguities in the database silhouettes.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123873226","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A system for face detection and tracking in unconstrained environments 无约束环境下的人脸检测和跟踪系统
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425361
Augusto Destrero, F. Odone, A. Verri
We describe a trainable system for face detection and tracking. The structure of the system is based on multiple cues that discard non face areas as soon as possible: we combine motion, skin, and face detection. The latter is the core of our system and consists of a hierarchy of small SVM classifiers built on the output of an automatic feature selection procedure. Our feature selection is entirely data-driven and allows us to obtain powerful descriptions from a relatively small set of data. Finally, a Kalman tracking on the face region optimizes detection results over time. We present an experimental analysis of the face detection module and results obtained with the whole system on the specific task of counting people entering the scene.
我们描述了一个可训练的人脸检测和跟踪系统。该系统的结构基于多个线索,这些线索会尽快丢弃非面部区域:我们将运动、皮肤和面部检测结合起来。后者是我们系统的核心,由基于自动特征选择过程输出的小SVM分类器组成。我们的特征选择完全是数据驱动的,允许我们从相对较小的数据集中获得强大的描述。最后,人脸区域的卡尔曼跟踪随着时间的推移优化检测结果。我们对人脸检测模块进行了实验分析,并结合整个系统对进入场景的人数进行了统计。
{"title":"A system for face detection and tracking in unconstrained environments","authors":"Augusto Destrero, F. Odone, A. Verri","doi":"10.1109/AVSS.2007.4425361","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425361","url":null,"abstract":"We describe a trainable system for face detection and tracking. The structure of the system is based on multiple cues that discard non face areas as soon as possible: we combine motion, skin, and face detection. The latter is the core of our system and consists of a hierarchy of small SVM classifiers built on the output of an automatic feature selection procedure. Our feature selection is entirely data-driven and allows us to obtain powerful descriptions from a relatively small set of data. Finally, a Kalman tracking on the face region optimizes detection results over time. We present an experimental analysis of the face detection module and results obtained with the whole system on the specific task of counting people entering the scene.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"35 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123299359","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Single camera calibration for trajectory-based behavior analysis 基于轨迹行为分析的单摄像机标定
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425301
N. Anjum, A. Cavallaro
Perspective deformations on the image plane make the analysis of object behaviors difficult in surveillance video. In this paper, we improve the results of trajectory-based scene analysis by using single camera calibration for perspective rectification. First, the ground-plane view is estimated from perspective images captured from a single camera. Next, unsupervised fuzzy clustering is applied on the transformed trajectories to group similar behaviors and to isolate outliers. We evaluate the proposed approach on real outdoor surveillance scenarios with standard datasets and show that perspective rectification improves the accuracy of the trajectory clustering results.
在监控视频中,图像平面上的透视变形给物体行为分析带来困难。在本文中,我们改进了基于轨迹的场景分析结果,使用单摄像机校准进行视角校正。首先,从单个摄像机拍摄的透视图像估计地平面视图。其次,将无监督模糊聚类应用于变换后的轨迹,对相似行为进行分组并分离异常值。我们在标准数据集的真实室外监控场景中评估了所提出的方法,并表明视角校正提高了轨迹聚类结果的准确性。
{"title":"Single camera calibration for trajectory-based behavior analysis","authors":"N. Anjum, A. Cavallaro","doi":"10.1109/AVSS.2007.4425301","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425301","url":null,"abstract":"Perspective deformations on the image plane make the analysis of object behaviors difficult in surveillance video. In this paper, we improve the results of trajectory-based scene analysis by using single camera calibration for perspective rectification. First, the ground-plane view is estimated from perspective images captured from a single camera. Next, unsupervised fuzzy clustering is applied on the transformed trajectories to group similar behaviors and to isolate outliers. We evaluate the proposed approach on real outdoor surveillance scenarios with standard datasets and show that perspective rectification improves the accuracy of the trajectory clustering results.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128840919","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
Detection of temporarily static regions by processing video at different frame rates 通过以不同帧率处理视频来检测临时静态区域
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425316
F. Porikli
This paper presents an abandoned item and illegally parked vehicle detection method for single static camera video surveillance applications. By processing the input video at different frame rates, two backgrounds are constructed; one for short-term and another for long-term. Each of these backgrounds is defined as a mixture of Gaussian models, which are adapted using online Bayesian update. Two binary foreground maps are estimated by comparing the current frame with the backgrounds, and motion statistics are aggregated in a likelihood image by applying a set of heuristics to the foreground maps. Likelihood image is then used to differentiate between the pixels that belong to moving objects, temporarily static regions and scene background. Depending on the application, the temporary static regions indicate abandoned items, illegally parked vehicles, objects removed from the scene, etc. The presented pixel-wise method does not require object tracking, thus its performance is not upper-bounded to error prone detection and correspondence tasks that usually fail for crowded scenes. It accurately segments objects even if they are fully occluded. It can also be effectively implemented on a parallel processing architecture.
提出了一种适用于单静态摄像机视频监控应用的废弃物品和非法停放车辆检测方法。通过对输入视频进行不同帧率的处理,构造两个背景;一个是短期的,另一个是长期的。每个背景都被定义为高斯模型的混合,这些模型使用在线贝叶斯更新进行调整。通过比较当前帧和背景来估计两个二元前景图,并通过对前景图应用一组启发式算法将运动统计信息聚合在似然图像中。然后使用似然图像来区分属于运动物体、临时静态区域和场景背景的像素。根据应用程序的不同,临时静态区域表示废弃物品、非法停放的车辆、从现场移走的物体等。所提出的逐像素方法不需要对象跟踪,因此它的性能不受容易出错的检测和通信任务的上限,而这些任务通常在拥挤的场景中失败。即使物体被完全遮挡,它也能准确地分割物体。它也可以在并行处理架构上有效地实现。
{"title":"Detection of temporarily static regions by processing video at different frame rates","authors":"F. Porikli","doi":"10.1109/AVSS.2007.4425316","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425316","url":null,"abstract":"This paper presents an abandoned item and illegally parked vehicle detection method for single static camera video surveillance applications. By processing the input video at different frame rates, two backgrounds are constructed; one for short-term and another for long-term. Each of these backgrounds is defined as a mixture of Gaussian models, which are adapted using online Bayesian update. Two binary foreground maps are estimated by comparing the current frame with the backgrounds, and motion statistics are aggregated in a likelihood image by applying a set of heuristics to the foreground maps. Likelihood image is then used to differentiate between the pixels that belong to moving objects, temporarily static regions and scene background. Depending on the application, the temporary static regions indicate abandoned items, illegally parked vehicles, objects removed from the scene, etc. The presented pixel-wise method does not require object tracking, thus its performance is not upper-bounded to error prone detection and correspondence tasks that usually fail for crowded scenes. It accurately segments objects even if they are fully occluded. It can also be effectively implemented on a parallel processing architecture.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129385668","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
Adaptive summarisation of surveillance video sequences 监控视频序列的自适应摘要
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425369
Jian Li, S. G. Nikolov, C. Benton, N. Scott-Samuel
We describe our studies on summarising surveillance videos using optical flow information. The proposed method incorporates motion analysis into a video skimming scheme in which the playback speed is determined by the detectability of interesting motion behaviours according to prior information. A psycho-visual experiment was conducted to compare human performance and viewing strategy for summarised videos using standard video skimming techniques and a proposed motion-based adaptive summarisation technique.
我们描述了利用光流信息对监控视频进行总结的研究。该方法将运动分析结合到视频浏览方案中,其中播放速度由根据先验信息的有趣运动行为的可检测性决定。通过一项心理视觉实验,比较了使用标准视频浏览技术和提出的基于动作的自适应摘要技术的人类对摘要视频的表现和观看策略。
{"title":"Adaptive summarisation of surveillance video sequences","authors":"Jian Li, S. G. Nikolov, C. Benton, N. Scott-Samuel","doi":"10.1109/AVSS.2007.4425369","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425369","url":null,"abstract":"We describe our studies on summarising surveillance videos using optical flow information. The proposed method incorporates motion analysis into a video skimming scheme in which the playback speed is determined by the detectability of interesting motion behaviours according to prior information. A psycho-visual experiment was conducted to compare human performance and viewing strategy for summarised videos using standard video skimming techniques and a proposed motion-based adaptive summarisation technique.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129542810","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Detecting shopper groups in video sequences 在视频序列中检测购物者组
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425347
A. Leykin, M. Tuceryan
We present a generalized extensible framework for automated recognition of swarming activities in video sequences. The trajectory of each individual is produced by the visual tracking sub-system and is further analyzed to detect certain types of high-level grouping behavior. We utilize recent findings in swarming behavior analysis to formulate a problem in terms of the specific distance function that we subsequently apply as part of the two-stage agglomerative clustering method to create a set of swarming events followed by a set of swarming activities. In this paper we present results for one particular type of swarming: shopper grouping. As part of this work the events detected in a relatively short time interval are further integrated into activities, the manifestation of prolonged high-level swarming behavior. The results demonstrate the ability of our method to detect such activities in congested surveillance videos. In particular in three hours of indoor retail store video, our method has correctly identified over85% of valid '"shopper-groups'" with a very low level of false positives, validated against human coded ground truth.
我们提出了一个通用的可扩展框架,用于自动识别视频序列中的群集活动。每个个体的轨迹由视觉跟踪子系统产生,并进一步分析以检测某些类型的高级分组行为。我们利用最近在蜂群行为分析方面的发现,根据特定距离函数来制定一个问题,我们随后将其作为两阶段凝聚聚类方法的一部分,以创建一组蜂群事件,随后是一组蜂群活动。在本文中,我们给出了一种特殊类型的蜂群:购物者分组的结果。作为这项工作的一部分,在相对较短的时间间隔内检测到的事件被进一步整合到活动中,这是长期高水平群体行为的表现。结果证明了我们的方法在拥挤的监控视频中检测此类活动的能力。特别是在三个小时的室内零售商店视频中,我们的方法正确识别了超过85%的有效“购物者群体”,假阳性水平非常低,与人类编码的基本事实相对照。
{"title":"Detecting shopper groups in video sequences","authors":"A. Leykin, M. Tuceryan","doi":"10.1109/AVSS.2007.4425347","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425347","url":null,"abstract":"We present a generalized extensible framework for automated recognition of swarming activities in video sequences. The trajectory of each individual is produced by the visual tracking sub-system and is further analyzed to detect certain types of high-level grouping behavior. We utilize recent findings in swarming behavior analysis to formulate a problem in terms of the specific distance function that we subsequently apply as part of the two-stage agglomerative clustering method to create a set of swarming events followed by a set of swarming activities. In this paper we present results for one particular type of swarming: shopper grouping. As part of this work the events detected in a relatively short time interval are further integrated into activities, the manifestation of prolonged high-level swarming behavior. The results demonstrate the ability of our method to detect such activities in congested surveillance videos. In particular in three hours of indoor retail store video, our method has correctly identified over85% of valid '\"shopper-groups'\" with a very low level of false positives, validated against human coded ground truth.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126879031","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
A framework for track matching across disjoint cameras using robust shape and appearance features 基于鲁棒形状和外观特征的跨不相交摄像机的轨迹匹配框架
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425308
Christopher S. Madden, M. Piccardi
This paper presents a framework based on robust shape and appearance features for matching the various tracks generated by a single individual moving within a surveillance system. Each track is first automatically analysed in order to detect and remove the frames affected by large segmentation errors and drastic changes in illumination. The object's features computed over the remaining frames prove more robust and capable of supporting correct matching of tracks even in the case of significantly disjointed camera views. The shape and appearance features used include a height estimate as well as illumination-tolerant colour representation of the individual's global colours and the colours of the upper and lower portions of clothing. The results of a test from a real surveillance system show that the combination of these four features can provide a probability of matching as high as 91 percent with 5 percent probability of false alarms under views which have significantly differing illumination levels and suffer from significant segmentation errors in as many as 1 in 4 frames.
本文提出了一种基于鲁棒形状和外观特征的框架,用于匹配监视系统中单个移动个体产生的各种轨迹。每个轨道首先被自动分析,以检测和去除受大分割错误和光照剧烈变化影响的帧。在剩余的帧上计算的物体特征证明更健壮,并且能够支持正确的轨迹匹配,即使在明显脱节的相机视图的情况下。所使用的形状和外观特征包括身高估计,以及个人整体颜色和衣服上下部分颜色的耐光照颜色表示。来自真实监控系统的测试结果表明,这四个特征的组合可以提供高达91%的匹配概率和5%的假警报概率,在具有显着不同的照明水平和遭受多达1 / 4帧的显著分割错误的视图下。
{"title":"A framework for track matching across disjoint cameras using robust shape and appearance features","authors":"Christopher S. Madden, M. Piccardi","doi":"10.1109/AVSS.2007.4425308","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425308","url":null,"abstract":"This paper presents a framework based on robust shape and appearance features for matching the various tracks generated by a single individual moving within a surveillance system. Each track is first automatically analysed in order to detect and remove the frames affected by large segmentation errors and drastic changes in illumination. The object's features computed over the remaining frames prove more robust and capable of supporting correct matching of tracks even in the case of significantly disjointed camera views. The shape and appearance features used include a height estimate as well as illumination-tolerant colour representation of the individual's global colours and the colours of the upper and lower portions of clothing. The results of a test from a real surveillance system show that the combination of these four features can provide a probability of matching as high as 91 percent with 5 percent probability of false alarms under views which have significantly differing illumination levels and suffer from significant segmentation errors in as many as 1 in 4 frames.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127391501","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
期刊
2007 IEEE Conference on Advanced Video and Signal Based Surveillance
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1