首页 > 最新文献

2008 IEEE 10th Workshop on Multimedia Signal Processing最新文献

英文 中文
Optimal camera selection in vision networks for shape approximation 基于形状逼近的视觉网络最佳摄像机选择
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665047
M. Morbée, L. Tessens, Huang Lee, W. Philips, H. Aghajan
Within a camera network, the contribution of a camera to the observation of a scene depends on its viewpoint and on the scene configuration. This is a dynamic property, as the scene content is subject to change over time. An automatic selection of a subset of cameras that significantly contributes to the desired observation of a scene can be of great value for the reduction of the amount of transmitted or stored image data. In this work, we propose low data rate schemes to select from a vision network a subset of cameras that provides a good frontal observation of the persons in the scene and allows for the best approximation of their 3D shape. We also investigate to what degree low data rates trade off quality of reconstructed 3D shapes.
在摄像机网络中,摄像机对观察场景的贡献取决于其视点和场景配置。这是一个动态属性,因为场景内容会随着时间而变化。自动选择显著有助于对场景进行所需观察的相机子集,对于减少传输或存储的图像数据量具有很大价值。在这项工作中,我们提出了低数据速率方案,从视觉网络中选择一个摄像头子集,该子集提供了对场景中人物的良好正面观察,并允许最佳近似其3D形状。我们还研究了低数据速率在多大程度上影响了重建三维形状的质量。
{"title":"Optimal camera selection in vision networks for shape approximation","authors":"M. Morbée, L. Tessens, Huang Lee, W. Philips, H. Aghajan","doi":"10.1109/MMSP.2008.4665047","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665047","url":null,"abstract":"Within a camera network, the contribution of a camera to the observation of a scene depends on its viewpoint and on the scene configuration. This is a dynamic property, as the scene content is subject to change over time. An automatic selection of a subset of cameras that significantly contributes to the desired observation of a scene can be of great value for the reduction of the amount of transmitted or stored image data. In this work, we propose low data rate schemes to select from a vision network a subset of cameras that provides a good frontal observation of the persons in the scene and allows for the best approximation of their 3D shape. We also investigate to what degree low data rates trade off quality of reconstructed 3D shapes.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"64 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134258022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Pedestrian detection based on multi-modal cooperation 基于多模态合作的行人检测
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665065
Yan-ning Zhang, Xiao-min Tong, Xiu-wei Zhang, Jiang-bin Zheng, Jun Zhou, Si-wei You
Pedestrian detection plays an important role in automated surveillance system. However, it is challenging to detect pedestrian robustly and accurately in a cluttered environment. In this paper, we propose a new cooperative pedestrian detection method using both colour and thermal image sequences, which is compared with the method using only colour image sequence and that using multi-modal fusion. Experiment results show that our cooperative detection mechanism could get more accurate pedestrian areas, a lower false alarm rate and a higher detection precision. Therefore, it has broad application prospects in the field of industry and military.
行人检测在自动监控系统中起着重要的作用。然而,在混乱的环境中,如何鲁棒准确地检测行人是一个挑战。本文提出了一种基于彩色和热图像序列的行人协同检测方法,并与仅基于彩色图像序列和基于多模态融合的行人协同检测方法进行了比较。实验结果表明,我们的协同检测机制可以获得更准确的行人区域,更低的虚警率和更高的检测精度。因此,在工业和军事领域具有广阔的应用前景。
{"title":"Pedestrian detection based on multi-modal cooperation","authors":"Yan-ning Zhang, Xiao-min Tong, Xiu-wei Zhang, Jiang-bin Zheng, Jun Zhou, Si-wei You","doi":"10.1109/MMSP.2008.4665065","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665065","url":null,"abstract":"Pedestrian detection plays an important role in automated surveillance system. However, it is challenging to detect pedestrian robustly and accurately in a cluttered environment. In this paper, we propose a new cooperative pedestrian detection method using both colour and thermal image sequences, which is compared with the method using only colour image sequence and that using multi-modal fusion. Experiment results show that our cooperative detection mechanism could get more accurate pedestrian areas, a lower false alarm rate and a higher detection precision. Therefore, it has broad application prospects in the field of industry and military.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124842571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
A new low complex reference free video quality predictor 一种新的低复杂度无参考视频质量预测器
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665177
A. Rossholm, B. Lövström
In many applications and environments for mobile communication there is a need for reference free perceptual quality measurements. In this paper a method for prediction of a number of quality metrics is proposed, where the input to the prediction is readily available parameters at the receiver side of a communications channel. Since the parameters are extracted from the coded video bit stream the model can be used in user scenarios where it is normally difficult to estimate the quality due to the reference not being available, as in streaming video and mobile TV applications. The predictor turns out to give good results for both the PSNR and the PEVQ metrics.
在移动通信的许多应用和环境中,需要无参考的感知质量测量。本文提出了一种预测若干质量指标的方法,其中预测的输入是通信信道接收端随时可用的参数。由于参数是从编码的视频比特流中提取的,因此该模型可以用于通常由于无法获得参考而难以估计质量的用户场景,例如流媒体视频和移动电视应用。预测器对于PSNR和PEVQ指标都给出了很好的结果。
{"title":"A new low complex reference free video quality predictor","authors":"A. Rossholm, B. Lövström","doi":"10.1109/MMSP.2008.4665177","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665177","url":null,"abstract":"In many applications and environments for mobile communication there is a need for reference free perceptual quality measurements. In this paper a method for prediction of a number of quality metrics is proposed, where the input to the prediction is readily available parameters at the receiver side of a communications channel. Since the parameters are extracted from the coded video bit stream the model can be used in user scenarios where it is normally difficult to estimate the quality due to the reference not being available, as in streaming video and mobile TV applications. The predictor turns out to give good results for both the PSNR and the PEVQ metrics.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125078572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Image annotation with parametric mixture model based multi-class multi-labeling 基于参数混合模型的多类多标记图像标注
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665153
Zhiyong Wang, W. Siu, D. Feng
Image annotation, which labels an image with a set of semantic terms so as to bridge the semantic gap between low level features and high level semantics in visual information retrieval, is generally posed as a classification problem. Recently, multi-label classification has been investigated for image annotation since an image presents rich contents and can be associated with multiple concepts (i.e. labels). In this paper, a parametric mixture model based multi-class multi-labeling approach is proposed to tackle image annotation. Instead of building classifiers to learn individual labels exclusively, we model images with parametric mixture models so that the mixture characteristics of labels can be simultaneously exploited in both training and annotation processes. Our proposed method has been benchmarked with several state-of-the-art methods and achieved promising results.
图像标注通常被认为是一个分类问题,用一组语义术语对图像进行标注,以弥补视觉信息检索中低级特征和高级语义之间的语义差距。由于图像内容丰富,可以与多个概念(即标签)相关联,近年来,多标签分类成为图像标注的研究热点。本文提出了一种基于参数混合模型的多类多标记方法来解决图像标注问题。我们使用参数混合模型对图像进行建模,从而可以在训练和标注过程中同时利用标签的混合特征,而不是构建分类器来单独学习单个标签。我们提出的方法已经与几个最先进的方法进行了基准测试,并取得了可喜的结果。
{"title":"Image annotation with parametric mixture model based multi-class multi-labeling","authors":"Zhiyong Wang, W. Siu, D. Feng","doi":"10.1109/MMSP.2008.4665153","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665153","url":null,"abstract":"Image annotation, which labels an image with a set of semantic terms so as to bridge the semantic gap between low level features and high level semantics in visual information retrieval, is generally posed as a classification problem. Recently, multi-label classification has been investigated for image annotation since an image presents rich contents and can be associated with multiple concepts (i.e. labels). In this paper, a parametric mixture model based multi-class multi-labeling approach is proposed to tackle image annotation. Instead of building classifiers to learn individual labels exclusively, we model images with parametric mixture models so that the mixture characteristics of labels can be simultaneously exploited in both training and annotation processes. Our proposed method has been benchmarked with several state-of-the-art methods and achieved promising results.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125142511","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Hybrid frame-recursive block-based distortion estimation model for wireless video transmission 基于帧递归分块的无线视频传输混合失真估计模型
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665179
Werayut Saesue, Jian Zhang, C. Chou
In wireless environments, video quality can be severely degraded due to channel errors. Improving error robustness towards the impact of packet loss in error-prone network is considered as a critical concern in wireless video networking research. Data partitioning (DP) is an efficient error-resilient tool in video codec that is capable of reducing the effect of transmission errors by reorganizing the coded video bitstream into different partitions with different levels of importance. Significant video performance improvement can be achieved if DP is jointly optimized with unequal error protection (UEP). This paper proposes a fast and accurate frame-recursive block-based distortion estimation model for the DP tool in H.264.AVC. The accuracy of our model comes from appropriately approximating the error-concealment cross-correlation term (which is neglected in earlier work in order to reduce computation burden) as a function of the first moment of decoded pixels.Without increasing computation complexity, our proposed distortion model can be applied to both fixed and variable block size intra-prediction and motion compensation. Extensive simulation results are presented to show the accuracy of our estimation algorithm.
在无线环境中,由于信道错误,视频质量可能会严重下降。提高对易出错网络中丢包影响的错误鲁棒性是无线视频网络研究中的一个关键问题。数据分区(Data partitioning, DP)是视频编解码器中一种有效的容错工具,它通过将编码后的视频比特流重新组织成不同重要程度的分区来减少传输错误的影响。将非等错保护(UEP)与数据传输数据(DP)联合优化,可以显著提高视频性能。针对H.264.AVC中DP工具,提出了一种快速准确的基于帧递归分块的失真估计模型。我们模型的准确性来自于将错误隐藏相互关联项(在早期的工作中为了减少计算负担而忽略)作为解码像素的第一矩的函数适当地逼近。在不增加计算复杂度的前提下,我们提出的失真模型可以同时用于固定和可变块大小的内部预测和运动补偿。大量的仿真结果表明了我们估计算法的准确性。
{"title":"Hybrid frame-recursive block-based distortion estimation model for wireless video transmission","authors":"Werayut Saesue, Jian Zhang, C. Chou","doi":"10.1109/MMSP.2008.4665179","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665179","url":null,"abstract":"In wireless environments, video quality can be severely degraded due to channel errors. Improving error robustness towards the impact of packet loss in error-prone network is considered as a critical concern in wireless video networking research. Data partitioning (DP) is an efficient error-resilient tool in video codec that is capable of reducing the effect of transmission errors by reorganizing the coded video bitstream into different partitions with different levels of importance. Significant video performance improvement can be achieved if DP is jointly optimized with unequal error protection (UEP). This paper proposes a fast and accurate frame-recursive block-based distortion estimation model for the DP tool in H.264.AVC. The accuracy of our model comes from appropriately approximating the error-concealment cross-correlation term (which is neglected in earlier work in order to reduce computation burden) as a function of the first moment of decoded pixels.Without increasing computation complexity, our proposed distortion model can be applied to both fixed and variable block size intra-prediction and motion compensation. Extensive simulation results are presented to show the accuracy of our estimation algorithm.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127725081","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Sparse human movement representation and recognition 稀疏人体运动表征与识别
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665068
Nikolaos Gkalelis, A. Tefas, I. Pitas
In this paper a novel method for human movement representation and recognition is proposed. A movement type is regarded as a unique combination of basic movement patterns, the so-called dynemes. The fuzzy c-mean (FCM) algorithm is used to identify the dynemes in the input space and allow the expression of a posture in terms of these dynemes. In the so-called dyneme space, the sparse posture representations of a movement are combined to represent the movement as a single point in that space, and linear discriminant analysis (LDA) is further employed to increase movement type discrimination and compactness of representation. This method allows for simple Mahalanobis or cosine distance comparison of movements, taking implicitly into account time shifts and internal speed variations, and, thus, aiding the design of a real-time movement recognition algorithm.
本文提出了一种新的人体运动表征与识别方法。一种运动类型被认为是基本运动模式的独特组合,即所谓的动力。使用模糊c均值(FCM)算法来识别输入空间中的动力,并允许根据这些动力来表达姿态。在所谓的动态空间中,将运动的稀疏姿态表示组合为该空间中的单个点,并进一步使用线性判别分析(LDA)来增加运动类型的区分和表示的紧凑性。该方法允许对运动进行简单的马氏或余弦距离比较,隐含地考虑到时间偏移和内部速度变化,从而帮助设计实时运动识别算法。
{"title":"Sparse human movement representation and recognition","authors":"Nikolaos Gkalelis, A. Tefas, I. Pitas","doi":"10.1109/MMSP.2008.4665068","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665068","url":null,"abstract":"In this paper a novel method for human movement representation and recognition is proposed. A movement type is regarded as a unique combination of basic movement patterns, the so-called dynemes. The fuzzy c-mean (FCM) algorithm is used to identify the dynemes in the input space and allow the expression of a posture in terms of these dynemes. In the so-called dyneme space, the sparse posture representations of a movement are combined to represent the movement as a single point in that space, and linear discriminant analysis (LDA) is further employed to increase movement type discrimination and compactness of representation. This method allows for simple Mahalanobis or cosine distance comparison of movements, taking implicitly into account time shifts and internal speed variations, and, thus, aiding the design of a real-time movement recognition algorithm.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129586129","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
On the detection and localization of facial occlusions and its use within different scenarios 面部咬合的检测与定位及其在不同场景中的应用
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665146
Lutz Goldmann, A. Rama, T. Sikora, F. Tarrés
Face analysis is a very active research field, due to its large variety of applications and the different challenges (illumination, pose, expressions or occlusions) the methods need to cope with. Facial occlusions are one of the biggest challenges since they are difficult to model and have a large influence on the performance of subsequent analysis modules. This paper describes a face detection/classification module that allows to detect and localize faces and present occlusions and discusses the use of this additional information within different application scenarios. The approach is evaluated on two databases with realistic occlusions and performs very well for the different detection/classification tasks. It achieves a f-measure of over 97% for face detection and around 86% for component detection. Regarding the occlusion detection, the proposed approach reaches a recognition rate above 91% for both faces and components.
人脸分析是一个非常活跃的研究领域,因为它有各种各样的应用和不同的挑战(照明、姿势、表情或遮挡)需要处理的方法。面部遮挡是最大的挑战之一,因为它们很难建模,并且对后续分析模块的性能有很大影响。本文描述了一个人脸检测/分类模块,该模块允许检测和定位人脸和呈现遮挡,并讨论了在不同应用场景中使用这些附加信息。该方法在两个具有真实遮挡的数据库上进行了评估,并在不同的检测/分类任务中表现良好。人脸检测的f值超过97%,成分检测的f值约为86%。在遮挡检测方面,该方法对人脸和部件的识别率均达到91%以上。
{"title":"On the detection and localization of facial occlusions and its use within different scenarios","authors":"Lutz Goldmann, A. Rama, T. Sikora, F. Tarrés","doi":"10.1109/MMSP.2008.4665146","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665146","url":null,"abstract":"Face analysis is a very active research field, due to its large variety of applications and the different challenges (illumination, pose, expressions or occlusions) the methods need to cope with. Facial occlusions are one of the biggest challenges since they are difficult to model and have a large influence on the performance of subsequent analysis modules. This paper describes a face detection/classification module that allows to detect and localize faces and present occlusions and discusses the use of this additional information within different application scenarios. The approach is evaluated on two databases with realistic occlusions and performs very well for the different detection/classification tasks. It achieves a f-measure of over 97% for face detection and around 86% for component detection. Regarding the occlusion detection, the proposed approach reaches a recognition rate above 91% for both faces and components.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121188234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Seamless MDVC in P2P: A transform-domain approach P2P中的无缝MDVC:一种转换域方法
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665091
Shuyuan Zhu, B. Zeng
Multiple-description coding (MDC) provides an effective way to mitigate the effects of packet errors/loses by making use of multiple channels. Perhaps, the most attractive application of MDC is in the peer-to-peer (P2P) scenario to support simultaneous video streaming to a large population of clients. To this end, a number of multiple-description video coding (MDVC) schemes (both non-scalable and scalable) were proposed in the past few years. However, almost all non-scalable schemes would suffer from the prediction mismatch between the references used at the encoder and decoder sides; whereas all scalable schemes (involving a base-layer and some enhancement layers) would suffer from the inter-dependency within the enhancement-layer information. In this paper, we propose a transform-domain MDVC method that can solve these problems and at the same time offer some other interesting features.
多描述编码(multi -description coding, MDC)通过使用多个信道提供了一种有效的方法来减轻数据包错误/丢失的影响。也许,MDC最吸引人的应用是在点对点(P2P)场景中,支持向大量客户端同时传输视频流。为此,在过去几年中提出了许多多描述视频编码(MDVC)方案(包括不可扩展和可扩展)。然而,几乎所有不可伸缩的方案都会受到编码器和解码器端使用的引用之间的预测不匹配的影响;然而,所有可扩展方案(涉及一个基础层和一些增强层)都会受到增强层信息内部相互依赖的影响。在本文中,我们提出了一种变换域MDVC方法,可以解决这些问题,同时提供了一些其他有趣的特性。
{"title":"Seamless MDVC in P2P: A transform-domain approach","authors":"Shuyuan Zhu, B. Zeng","doi":"10.1109/MMSP.2008.4665091","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665091","url":null,"abstract":"Multiple-description coding (MDC) provides an effective way to mitigate the effects of packet errors/loses by making use of multiple channels. Perhaps, the most attractive application of MDC is in the peer-to-peer (P2P) scenario to support simultaneous video streaming to a large population of clients. To this end, a number of multiple-description video coding (MDVC) schemes (both non-scalable and scalable) were proposed in the past few years. However, almost all non-scalable schemes would suffer from the prediction mismatch between the references used at the encoder and decoder sides; whereas all scalable schemes (involving a base-layer and some enhancement layers) would suffer from the inter-dependency within the enhancement-layer information. In this paper, we propose a transform-domain MDVC method that can solve these problems and at the same time offer some other interesting features.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"88 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115619627","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A video analysis framework for surveillance system 用于监控系统的视频分析框架
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665195
N. Suvonvorn
An on-line video processing for surveillance system is a very challenging problem. The computational complexity of video analysis algorithms and the massive amount of data to be analyzed must be considered under real-time constraints. Moreover it needs to satisfy different criteria of application domain, such as, scalability, re-configurability, and quality of service. In this paper we propose a flexible/efficient video analysis framework for surveillance system which is a component-based architecture. The video acquisition, re-configurable video analysis, and video storage are some of the basic components. The component execution and inter-components synchronization are designed for supporting the multi-cores and multi-processors architecture with multi-threading implementation on .NET Framework. Experimental results on real-time motion tracking are presented with discussion.
监控系统的在线视频处理是一个非常具有挑战性的问题。视频分析算法的计算复杂度和需要分析的海量数据必须在实时性约束下加以考虑。此外,它还需要满足应用领域的不同标准,如可扩展性、可重构性和服务质量。本文提出了一种灵活高效的基于组件的监控系统视频分析框架。视频采集、可重构视频分析和视频存储是系统的基本组成部分。组件执行和组件间同步是为了支持。net框架上多线程实现的多核多处理器架构而设计的。给出了实时运动跟踪的实验结果,并进行了讨论。
{"title":"A video analysis framework for surveillance system","authors":"N. Suvonvorn","doi":"10.1109/MMSP.2008.4665195","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665195","url":null,"abstract":"An on-line video processing for surveillance system is a very challenging problem. The computational complexity of video analysis algorithms and the massive amount of data to be analyzed must be considered under real-time constraints. Moreover it needs to satisfy different criteria of application domain, such as, scalability, re-configurability, and quality of service. In this paper we propose a flexible/efficient video analysis framework for surveillance system which is a component-based architecture. The video acquisition, re-configurable video analysis, and video storage are some of the basic components. The component execution and inter-components synchronization are designed for supporting the multi-cores and multi-processors architecture with multi-threading implementation on .NET Framework. Experimental results on real-time motion tracking are presented with discussion.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123165632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Camera motion-constraint video codec selection 摄像机运动约束视频编解码器选择
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665049
A. Krutz, S. Knorr, M. Kunter, T. Sikora
In recent years advanced video codecs have been developed, such as standardized in MPEG-4. The latest video codec H.264/AVC provides compression performance superior to previous standards, but is based on the same basic motion-compensated-DCT architecture. However, for certain types of video, it has been shown that it is possible to outperform the H.264/AVC using an object-based video codec. Towards a general-purpose object-based video coding system we present an automated approach to separate a video sequences into sub-sequences regarding its camera motion type. Then, the sub-sequences are coded either with an object-based codec or the common H.264/AVC. Applying different video codecs for different kinds of camera motion, we achieve a higher overall coding gain for the video sequence. In first experimental evaluations, we demonstrate the excellence performance of this approach on two test sequences.
近年来,先进的视频编解码器得到了发展,如标准化的MPEG-4。最新的视频编解码器H.264/AVC提供了优于以前标准的压缩性能,但基于相同的基本运动补偿dct架构。然而,对于某些类型的视频,已经证明使用基于对象的视频编解码器可以胜过H.264/AVC。针对一种通用的基于对象的视频编码系统,我们提出了一种自动将视频序列根据其摄像机运动类型分离成子序列的方法。然后,用基于对象的编解码器或常用的H.264/AVC对子序列进行编码。针对不同类型的摄像机运动采用不同的视频编解码器,实现了视频序列较高的整体编码增益。在第一次实验评估中,我们证明了该方法在两个测试序列上的卓越性能。
{"title":"Camera motion-constraint video codec selection","authors":"A. Krutz, S. Knorr, M. Kunter, T. Sikora","doi":"10.1109/MMSP.2008.4665049","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665049","url":null,"abstract":"In recent years advanced video codecs have been developed, such as standardized in MPEG-4. The latest video codec H.264/AVC provides compression performance superior to previous standards, but is based on the same basic motion-compensated-DCT architecture. However, for certain types of video, it has been shown that it is possible to outperform the H.264/AVC using an object-based video codec. Towards a general-purpose object-based video coding system we present an automated approach to separate a video sequences into sub-sequences regarding its camera motion type. Then, the sub-sequences are coded either with an object-based codec or the common H.264/AVC. Applying different video codecs for different kinds of camera motion, we achieve a higher overall coding gain for the video sequence. In first experimental evaluations, we demonstrate the excellence performance of this approach on two test sequences.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131445379","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2008 IEEE 10th Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1