首页 > 最新文献

2007 IEEE Conference on Advanced Video and Signal Based Surveillance最新文献

英文 中文
Passive sensor based dynamic object association with particle filtering 基于粒子滤波的被动传感器动态目标关联
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425311
S. Cho, Jinseok Lee, Sangjin Hong
This paper develops and evaluates the threshold based algorithm proposed in [S.H. Cho, J. Lee, and S. Hong, "Passive Sensor Based Dynamic Object Association Method in Wireless Sensor Network," Proceedings of MWSCAS07 and NEWCAS07, Aug. 2007. ] for dynamic data association in wireless sensor networks. The sensor node incorporates RFID reader and acoustic sensor where the signals are fused for tracking and associating multiple objects. The RFID tag is used for object identification and acoustic sensor is used for estimating object movement. For the better data association, we apply the particle filtering for the prediction of an object. The algorithm with the particle filtering has an effect on increasing the association case where even objects overlap. The simulation result is compared to that using only the original algorithm. The association performance under single node coverage and multiple node coverage is evaluated as a function of sampling time.
本文对基于阈值的算法进行了改进和评价赵、李、洪,“基于无源传感器的无线传感器网络动态目标关联方法”,MWSCAS07和NEWCAS07, 2007年8月。]用于无线传感器网络中的动态数据关联。传感器节点包含RFID读取器和声学传感器,其中融合信号用于跟踪和关联多个对象。RFID标签用于识别物体,声学传感器用于估计物体的运动。为了获得更好的数据关联,我们采用粒子滤波对目标进行预测。采用粒子滤波的算法在偶数物体重叠的情况下,增加了关联情况。仿真结果与仅使用原始算法的仿真结果进行了比较。将单节点覆盖和多节点覆盖下的关联性能作为采样时间的函数进行评估。
{"title":"Passive sensor based dynamic object association with particle filtering","authors":"S. Cho, Jinseok Lee, Sangjin Hong","doi":"10.1109/AVSS.2007.4425311","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425311","url":null,"abstract":"This paper develops and evaluates the threshold based algorithm proposed in [S.H. Cho, J. Lee, and S. Hong, \"Passive Sensor Based Dynamic Object Association Method in Wireless Sensor Network,\" Proceedings of MWSCAS07 and NEWCAS07, Aug. 2007. ] for dynamic data association in wireless sensor networks. The sensor node incorporates RFID reader and acoustic sensor where the signals are fused for tracking and associating multiple objects. The RFID tag is used for object identification and acoustic sensor is used for estimating object movement. For the better data association, we apply the particle filtering for the prediction of an object. The algorithm with the particle filtering has an effect on increasing the association case where even objects overlap. The simulation result is compared to that using only the original algorithm. The association performance under single node coverage and multiple node coverage is evaluated as a function of sampling time.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"126 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131219512","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Image enhancement in multi-resolution multi-sensor fusion 多分辨率多传感器融合中的图像增强
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425325
J. Jang, Yong Sun Kim, J. Ra
In multi-sensor image fusion, multi-resolution approaches became popular because they can preserve detailed information well. Among them, the gradient-based multi-resolution (GBMR) algorithm is known to effectively reduce ringing artifacts near edges compared with the discrete wavelet transform (DWT)-based algorithm. However, since the GBMR algorithm does not consider the diagonal direction, the ringing artifacts reduction is not satisfactory at diagonal edges. In this paper, we generalize the GBMR algorithm by adopting the wavelet structure. Thereby, the proposed algorithm improves the fusion process in high-frequency sub-bands so as to preserve details of input images. Meanwhile, the algorithm fuses the low-frequency sub-band by considering the overall contrast in the output image. To evaluate the proposed algorithm, we compare it with the DWT-based and GBMR algorithms. Experimental results clearly demonstrate that the proposed algorithm effectively reduces ringing artifacts for edges of all directions and greatly enhances the overall contrast while minimizing visual information loss.
在多传感器图像融合中,多分辨率方法因其能很好地保留细节信息而受到青睐。其中,基于梯度的多分辨率(GBMR)算法与基于离散小波变换(DWT)的算法相比,可以有效地减少边缘附近的环形伪影。然而,由于GBMR算法没有考虑对角方向,因此对角边缘处的环形伪影抑制效果不理想。本文采用小波结构对GBMR算法进行了推广。因此,该算法改进了高频子带的融合过程,以保持输入图像的细节。同时,考虑输出图像的整体对比度,对低频子带进行融合。为了评估该算法,我们将其与基于dwt和GBMR的算法进行了比较。实验结果清楚地表明,该算法有效地减少了各个方向边缘的环形伪影,大大提高了整体对比度,同时最大限度地减少了视觉信息损失。
{"title":"Image enhancement in multi-resolution multi-sensor fusion","authors":"J. Jang, Yong Sun Kim, J. Ra","doi":"10.1109/AVSS.2007.4425325","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425325","url":null,"abstract":"In multi-sensor image fusion, multi-resolution approaches became popular because they can preserve detailed information well. Among them, the gradient-based multi-resolution (GBMR) algorithm is known to effectively reduce ringing artifacts near edges compared with the discrete wavelet transform (DWT)-based algorithm. However, since the GBMR algorithm does not consider the diagonal direction, the ringing artifacts reduction is not satisfactory at diagonal edges. In this paper, we generalize the GBMR algorithm by adopting the wavelet structure. Thereby, the proposed algorithm improves the fusion process in high-frequency sub-bands so as to preserve details of input images. Meanwhile, the algorithm fuses the low-frequency sub-band by considering the overall contrast in the output image. To evaluate the proposed algorithm, we compare it with the DWT-based and GBMR algorithms. Experimental results clearly demonstrate that the proposed algorithm effectively reduces ringing artifacts for edges of all directions and greatly enhances the overall contrast while minimizing visual information loss.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"192 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132794631","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A 2D+3D face identification system for surveillance applications 用于监控应用的2D+3D人脸识别系统
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425309
F. Tsalakanidou, S. Malassiotis, M. Strintzis
A novel surveillance system integrating 2D and 3D facial data is presented in this paper, based on a low-cost sensor capable of real-time acquisition of 3D images and associated color images of a scene. Depth data is used for robust face detection, localization and 3D pose estimation, as well as for compensating pose and illumination variations of facial images prior to classification . The proposed system was tested under an open-set identification scenario for surveillance of humans passing through a relatively constrained area. Experimental results demonstrate the accuracy and robustness of the system under a variety of conditions usually encountered in surveillance applications.
本文提出了一种基于低成本传感器的二维和三维人脸数据集成监控系统,该系统能够实时采集场景的三维图像和相关彩色图像。深度数据用于鲁棒人脸检测、定位和3D姿态估计,以及在分类之前补偿面部图像的姿态和光照变化。所提出的系统在一个开放集识别场景下进行了测试,用于监视通过相对受限区域的人类。实验结果证明了该系统在各种监控应用条件下的准确性和鲁棒性。
{"title":"A 2D+3D face identification system for surveillance applications","authors":"F. Tsalakanidou, S. Malassiotis, M. Strintzis","doi":"10.1109/AVSS.2007.4425309","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425309","url":null,"abstract":"A novel surveillance system integrating 2D and 3D facial data is presented in this paper, based on a low-cost sensor capable of real-time acquisition of 3D images and associated color images of a scene. Depth data is used for robust face detection, localization and 3D pose estimation, as well as for compensating pose and illumination variations of facial images prior to classification . The proposed system was tested under an open-set identification scenario for surveillance of humans passing through a relatively constrained area. Experimental results demonstrate the accuracy and robustness of the system under a variety of conditions usually encountered in surveillance applications.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132932225","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Verbal aggression detection in complex social environments 复杂社会环境下的言语攻击检测
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425279
P. V. Hengel, T. Andringa
The paper presents a knowledge-based system designed to detect evidence of aggression by means of audio analysis. The detection is based on the way sounds are analyzed and how they attract attention in the human auditory system. The performance achieved is comparable to human performance in complex social environments. The SIgard system has been deployed in a number of different real-life situations and was tested extensively in the inner city of Groningen. Experienced police observers have annotated ~1400 recordings with various degrees of shouting, which were used for optimization. All essential events and a small number of nonessential aggressive events were detected. The system produces only a few false alarms (non-shouts) per microphone per year and misses no incidents. This makes it the first successful detection system for a non-trivial target in an unconstrained environment.
本文提出了一种基于知识的基于音频分析的攻击证据检测系统。这种检测是基于声音被分析的方式,以及它们如何在人类听觉系统中吸引注意力。所取得的成绩与人类在复杂社会环境中的表现相当。SIgard系统已经部署在许多不同的现实环境中,并在格罗宁根市中心进行了广泛的测试。有经验的警察观察员注释了大约1400段不同程度的喊叫录音,用于优化。检测到所有必要事件和少量非必要的侵略性事件。该系统每年每个麦克风只产生几次假警报(非喊叫声),并且没有遗漏任何事件。这使它成为第一个在无约束环境中成功检测非平凡目标的系统。
{"title":"Verbal aggression detection in complex social environments","authors":"P. V. Hengel, T. Andringa","doi":"10.1109/AVSS.2007.4425279","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425279","url":null,"abstract":"The paper presents a knowledge-based system designed to detect evidence of aggression by means of audio analysis. The detection is based on the way sounds are analyzed and how they attract attention in the human auditory system. The performance achieved is comparable to human performance in complex social environments. The SIgard system has been deployed in a number of different real-life situations and was tested extensively in the inner city of Groningen. Experienced police observers have annotated ~1400 recordings with various degrees of shouting, which were used for optimization. All essential events and a small number of nonessential aggressive events were detected. The system produces only a few false alarms (non-shouts) per microphone per year and misses no incidents. This makes it the first successful detection system for a non-trivial target in an unconstrained environment.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122291414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Human body gesture recognition using adapted auxiliary particle filtering 基于自适应辅助粒子滤波的人体手势识别
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425351
A. Oikonomopoulos, M. Pantic
In this paper we propose a tracking scheme specifically tailored for tracking human body parts in cluttered scenes. We model the background and the human skin using Gaussian mixture models and we combine these estimates to localize the features to be tracked. We further use these estimates to determine the pixels which belong to the background and those which belong to the subject's skin and we incorporate this information in the observation model of the used tracking scheme. For handling self-occlusion (i.e., when one body part occludes another), we incorporate the information about the direction of the observed motion into the propagation model of the used tracking scheme. We demonstrate that the proposed method outperforms the conventional condensation and auxiliary particle filtering when the hands and the head are the tracked body features. For the purposes of human body gesture recognition, we use a variant of the longest common subsequence algorithm (LCSS) in order to acquire a distance measure between the acquired trajectories and we use this measure in order to define new kernels for a relevance vector machine (RVM) classification scheme. We present results on real image sequences from a small database depicting people performing 15 aerobic exercises.
在本文中,我们提出了一种专门针对在混乱场景中跟踪人体部位的跟踪方案。我们使用高斯混合模型对背景和人体皮肤进行建模,并结合这些估计来定位要跟踪的特征。我们进一步使用这些估计来确定属于背景的像素和属于受试者皮肤的像素,并将这些信息合并到所用跟踪方案的观察模型中。对于处理自遮挡(即当一个身体部位遮挡另一个身体部位时),我们将观察到的运动方向信息合并到所用跟踪方案的传播模型中。实验结果表明,当手部和头部是被跟踪的身体特征时,该方法优于传统的粒子滤波和辅助粒子滤波。为了人体手势识别的目的,我们使用最长公共子序列算法(LCSS)的一种变体来获取所获取的轨迹之间的距离度量,并使用该度量来定义相关向量机(RVM)分类方案的新核。我们展示了来自一个小型数据库的真实图像序列的结果,描绘了人们进行15种有氧运动。
{"title":"Human body gesture recognition using adapted auxiliary particle filtering","authors":"A. Oikonomopoulos, M. Pantic","doi":"10.1109/AVSS.2007.4425351","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425351","url":null,"abstract":"In this paper we propose a tracking scheme specifically tailored for tracking human body parts in cluttered scenes. We model the background and the human skin using Gaussian mixture models and we combine these estimates to localize the features to be tracked. We further use these estimates to determine the pixels which belong to the background and those which belong to the subject's skin and we incorporate this information in the observation model of the used tracking scheme. For handling self-occlusion (i.e., when one body part occludes another), we incorporate the information about the direction of the observed motion into the propagation model of the used tracking scheme. We demonstrate that the proposed method outperforms the conventional condensation and auxiliary particle filtering when the hands and the head are the tracked body features. For the purposes of human body gesture recognition, we use a variant of the longest common subsequence algorithm (LCSS) in order to acquire a distance measure between the acquired trajectories and we use this measure in order to define new kernels for a relevance vector machine (RVM) classification scheme. We present results on real image sequences from a small database depicting people performing 15 aerobic exercises.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130180077","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Efficient side information encoding for text hardcopy documents 文本硬拷贝文档的有效侧信息编码
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425370
P. Borges, E. Izquierdo, J. Mayer
This paper proposes a new coding method that increases significantly the signal-to-watermark ratio in document watermarking algorithms. A possible approach to text document watermarking is to consider text characters as a data structure consisting of several modifiable features such as size, shape, position, luminance, among others. In existing algorithms, these features can be modified sequentially according to bit values to be embedded. In contrast, the solution proposed here uses a positional information coding approach to embed information. Using this approach, the information is related to the position of modified characters, and not to the bit embedded on each character. This coding is based on combinatorial analysis and it can embed more bits in comparison to the usual methods, given a distortion constraint. An analysis showing the superior performance of positional coding for this type of application is presented. Experiments validate the analysis and the applicability of the method.
本文提出了一种新的编码方法,可以显著提高文档水印算法中的信水印比。文本文档水印的一种可能方法是将文本字符视为由若干可修改的特征(如大小、形状、位置、亮度等)组成的数据结构。在现有的算法中,这些特征可以根据要嵌入的位值顺序修改。相比之下,本文提出的解决方案使用位置信息编码方法来嵌入信息。使用这种方法,信息与修改字符的位置有关,而与每个字符上嵌入的位无关。这种编码基于组合分析,与通常的方法相比,在给定失真约束的情况下,它可以嵌入更多的比特。分析显示了位置编码在这类应用中的优越性能。实验验证了分析结果和方法的适用性。
{"title":"Efficient side information encoding for text hardcopy documents","authors":"P. Borges, E. Izquierdo, J. Mayer","doi":"10.1109/AVSS.2007.4425370","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425370","url":null,"abstract":"This paper proposes a new coding method that increases significantly the signal-to-watermark ratio in document watermarking algorithms. A possible approach to text document watermarking is to consider text characters as a data structure consisting of several modifiable features such as size, shape, position, luminance, among others. In existing algorithms, these features can be modified sequentially according to bit values to be embedded. In contrast, the solution proposed here uses a positional information coding approach to embed information. Using this approach, the information is related to the position of modified characters, and not to the bit embedded on each character. This coding is based on combinatorial analysis and it can embed more bits in comparison to the usual methods, given a distortion constraint. An analysis showing the superior performance of positional coding for this type of application is presented. Experiments validate the analysis and the applicability of the method.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129036472","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A fast algorithm for adaptive background model construction using parzen density estimation 基于parzen密度估计的自适应背景模型快速构建算法
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425366
T. Tanaka, Atsushi Shimada, Daisaku Arita, R. Taniguchi
Non-parametric representation of pixel intensity distribution is quite effective to construct proper background model and to detect foreground objects accurately. However, from the viewpoint of practical application, the computation cost of the distribution estimation should be reduced. In this paper, we present fast estimation of the probability density function (PDF) of pixel value using Parzen density estimation and foreground object detection based on the estimated PDF. Here, the PDF is computed by partially updating the PDF estimated at the previous frame, and it greatly reduces the computation cost of the PDF estimation. Thus, the background model adapts quickly to changes in the scene and, therefore, foreground objects can be robustly detected. Several experiments show the effectiveness of our approach.
像素强度分布的非参数化表示对于构建合适的背景模型和准确检测前景目标是非常有效的。然而,从实际应用的角度来看,应该降低分布估计的计算成本。本文提出了一种基于Parzen密度估计的像素值概率密度函数(PDF)的快速估计方法,并基于估计的PDF进行前景目标检测。本文通过对前一帧估计的PDF进行部分更新来计算PDF,大大降低了PDF估计的计算成本。因此,背景模型可以快速适应场景的变化,从而可以鲁棒地检测前景物体。几个实验证明了我们方法的有效性。
{"title":"A fast algorithm for adaptive background model construction using parzen density estimation","authors":"T. Tanaka, Atsushi Shimada, Daisaku Arita, R. Taniguchi","doi":"10.1109/AVSS.2007.4425366","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425366","url":null,"abstract":"Non-parametric representation of pixel intensity distribution is quite effective to construct proper background model and to detect foreground objects accurately. However, from the viewpoint of practical application, the computation cost of the distribution estimation should be reduced. In this paper, we present fast estimation of the probability density function (PDF) of pixel value using Parzen density estimation and foreground object detection based on the estimated PDF. Here, the PDF is computed by partially updating the PDF estimated at the previous frame, and it greatly reduces the computation cost of the PDF estimation. Thus, the background model adapts quickly to changes in the scene and, therefore, foreground objects can be robustly detected. Several experiments show the effectiveness of our approach.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129514091","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Improved one-class SVM classifier for sounds classification 改进的单类SVM分类器用于声音分类
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425296
A. Rabaoui, M. Davy, S. Rossignol, Z. Lachiri, N. Ellouze
This paper proposes to apply optimized one-class support vector machines (1-SVMs) as a discriminative framework in order to address a specific audio classification problem. First, since SVM-based classifier with gaussian RBF kernel is sensitive to the kernel width, the width will be scaled in a distribution-dependent way permitting to avoid under-fitting and over-fitting problems. Moreover, an advanced dissimilarity measure will be introduced. We illustrate the performance of these methods on an audio database containing environmental sounds that may be of great importance for surveillance and security applications. The experiments conducted on a multi-class problem show that by choosing adequately the SVM parameters, we can efficiently address a sounds classification problem characterized by complex real-world datasets.
为了解决特定的音频分类问题,本文提出将优化的一类支持向量机(1- svm)作为判别框架。首先,由于基于svm的高斯RBF核分类器对核宽度敏感,因此宽度将以分布相关的方式缩放,以避免欠拟合和过拟合问题。此外,还将引入一种先进的不相似度度量方法。我们举例说明了这些方法在包含环境声音的音频数据库上的性能,这些声音可能对监视和安全应用非常重要。对一个多类问题的实验表明,通过选择适当的支持向量机参数,可以有效地解决现实世界复杂数据集的声音分类问题。
{"title":"Improved one-class SVM classifier for sounds classification","authors":"A. Rabaoui, M. Davy, S. Rossignol, Z. Lachiri, N. Ellouze","doi":"10.1109/AVSS.2007.4425296","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425296","url":null,"abstract":"This paper proposes to apply optimized one-class support vector machines (1-SVMs) as a discriminative framework in order to address a specific audio classification problem. First, since SVM-based classifier with gaussian RBF kernel is sensitive to the kernel width, the width will be scaled in a distribution-dependent way permitting to avoid under-fitting and over-fitting problems. Moreover, an advanced dissimilarity measure will be introduced. We illustrate the performance of these methods on an audio database containing environmental sounds that may be of great importance for surveillance and security applications. The experiments conducted on a multi-class problem show that by choosing adequately the SVM parameters, we can efficiently address a sounds classification problem characterized by complex real-world datasets.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129042596","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 46
Detection of abnormal behaviors using a mixture of Von Mises distributions 使用Von Mises分布的混合检测异常行为
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425300
S. Calderara, R. Cucchiara, A. Prati
This paper proposes the use of a mixture of Von Mises distributions to detect abnormal behaviors of moving people. The mixture is created from an unsupervised training set by exploiting k-medoids clustering algorithm based on Bhattacharyya distance between distributions. The extracted medoids are used as modes in the multi-modal mixture whose weights are the priors of the specific medoid. Given the mixture model a new trajectory is verified on the model by considering each direction composing it as independent. Experiments over a real scenario composed of multiple, partially-overlapped cameras are reported.
本文提出使用混合的Von Mises分布来检测移动人群的异常行为。利用基于分布间Bhattacharyya距离的k- medioids聚类算法,从无监督训练集生成混合物。将提取的介质用作多模态混合物中的模态,其权重为特定介质的先验。给定混合模型,通过考虑组成混合模型的各个方向是独立的,从而在模型上验证新的轨迹。本文报道了由多个部分重叠的摄像机组成的真实场景的实验。
{"title":"Detection of abnormal behaviors using a mixture of Von Mises distributions","authors":"S. Calderara, R. Cucchiara, A. Prati","doi":"10.1109/AVSS.2007.4425300","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425300","url":null,"abstract":"This paper proposes the use of a mixture of Von Mises distributions to detect abnormal behaviors of moving people. The mixture is created from an unsupervised training set by exploiting k-medoids clustering algorithm based on Bhattacharyya distance between distributions. The extracted medoids are used as modes in the multi-modal mixture whose weights are the priors of the specific medoid. Given the mixture model a new trajectory is verified on the model by considering each direction composing it as independent. Experiments over a real scenario composed of multiple, partially-overlapped cameras are reported.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"18 11","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120856372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 55
A profile of MPEG-7 for visual surveillance 用于视觉监控的MPEG-7概述
Pub Date : 2007-09-05 DOI: 10.1109/AVSS.2007.4425358
J. Annesley, A. Colombo, J. Orwell, S. Velastín
This paper builds on previous work to propose a meta-data standard for video surveillance. The motivation is to promote interoperability. The starting point is the set of requirements under consideration for a Multimedia Application Format. These requirements cover a description of the surveillance system and of the activity in the scene. In addition to this set, appropriate descriptions for the relation between camera and scene are also considered. To improve interoperability between systems and between components of a system, two types of restrictions are proposed. The first proposal is a restricted subset of the MPEG-7 elements that are applicable to the surveillance domain. The second proposal is to use the MPEG-7 tools to include domain-specific taxonomies to restrict the names of elements used in the semantic descriptions. Both proposals are incorporated into examples which demonstrate the use of the standard.
本文在前人工作的基础上提出了视频监控的元数据标准。其动机是促进互操作性。起点是考虑多媒体应用程序格式的一组需求。这些要求包括对监视系统和现场活动的描述。在此基础上,还考虑了对摄像机与场景关系的适当描述。为了提高系统之间和系统组件之间的互操作性,提出了两种类型的限制。第一个建议是适用于监视领域的MPEG-7元素的受限子集。第二个建议是使用MPEG-7工具来包含特定于领域的分类法,以限制语义描述中使用的元素名称。这两种建议都纳入了示范使用标准的例子。
{"title":"A profile of MPEG-7 for visual surveillance","authors":"J. Annesley, A. Colombo, J. Orwell, S. Velastín","doi":"10.1109/AVSS.2007.4425358","DOIUrl":"https://doi.org/10.1109/AVSS.2007.4425358","url":null,"abstract":"This paper builds on previous work to propose a meta-data standard for video surveillance. The motivation is to promote interoperability. The starting point is the set of requirements under consideration for a Multimedia Application Format. These requirements cover a description of the surveillance system and of the activity in the scene. In addition to this set, appropriate descriptions for the relation between camera and scene are also considered. To improve interoperability between systems and between components of a system, two types of restrictions are proposed. The first proposal is a restricted subset of the MPEG-7 elements that are applicable to the surveillance domain. The second proposal is to use the MPEG-7 tools to include domain-specific taxonomies to restrict the names of elements used in the semantic descriptions. Both proposals are incorporated into examples which demonstrate the use of the standard.","PeriodicalId":371050,"journal":{"name":"2007 IEEE Conference on Advanced Video and Signal Based Surveillance","volume":"171 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2007-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121305510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
2007 IEEE Conference on Advanced Video and Signal Based Surveillance
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1