首页 > 最新文献

2012 IEEE International Conference on Multimedia and Expo Workshops最新文献

英文 中文
A Novel Edge Detection Framework by Component Tree Construction 通过构件树构建新颖的边缘检测框架
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.99
Zhijun Dai, Yihong Wu, Youji Feng
This paper proposes a new edge detection framework with component tree construction. This open framework is efficient for edge property computation and convenient for subsequent image processing. We detect edges according to the properties which are customized by framework rules. Experiments on using the framework for a new efficient implementation of Canny edge detector are reported. The results demonstrate that the tree construction is efficient and the framework is flexible.
提出了一种新的基于构件树结构的边缘检测框架。这种开放的框架可以有效地计算边缘属性,方便后续的图像处理。我们根据框架规则自定义的属性来检测边缘。本文报道了利用该框架高效实现Canny边缘检测器的实验。结果表明,树形结构是高效的,框架是灵活的。
{"title":"A Novel Edge Detection Framework by Component Tree Construction","authors":"Zhijun Dai, Yihong Wu, Youji Feng","doi":"10.1109/ICMEW.2012.99","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.99","url":null,"abstract":"This paper proposes a new edge detection framework with component tree construction. This open framework is efficient for edge property computation and convenient for subsequent image processing. We detect edges according to the properties which are customized by framework rules. Experiments on using the framework for a new efficient implementation of Canny edge detector are reported. The results demonstrate that the tree construction is efficient and the framework is flexible.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129855348","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
3D Pose Estimation of Front Vehicle Towards a Better Driver Assistance System 面向更好的驾驶员辅助系统的前车三维姿态估计
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.97
Yu Peng, Jesse S. Jin, S. Luo, Min Xu, Yue Cui
Driver assistance system enhances traffic safety and efficiency. Accurate 3D pose of front vehicle can help driver to make right decisions on road. We propose a novel real-time system to estimate 3D pose of the front vehicle. This system consists of two parallel threads: vehicle rear tracking and mapping. Vehicle rear is firstly identified in the video captured by an on-board camera, after license plate localization and foreground extraction. 3D pose estimation technique is then employed with respect to extracted vehicle rear. Most 3D pose estimation techniques need prior models or a stereo initialization with user cooperation. It is extremely difficult to obtain prior models due to various appearances of vehicle rears. Moreover, it is unsafe to ask for driver's cooperation when vehicle is running. In our system, two initial key frames for stereo algorithm are automatically extracted by vehicle rear detection and tracking. Map points are defined as a collection of point features extracted from vehicle rear with their 3D information. These map points are inferences that relating 2D features detected in following vehicle rears with 3D world. Relative 3D Pose between current vehicle rear and on-board camera is then estimated through mapping that matches map points with current point features. We demonstrate the abilities of our system by augmented reality, which needs accurate and real-time 3D pose estimation.
驾驶员辅助系统提高交通安全和效率。前方车辆准确的三维姿态可以帮助驾驶员在道路上做出正确的决策。提出了一种新型的前方车辆三维姿态实时估计系统。该系统由两个并行线程组成:车辆后方跟踪和测绘。首先在车载摄像头拍摄的视频中识别车辆后方,然后进行车牌定位和前景提取。然后对提取的车辆尾部采用三维姿态估计技术。大多数三维姿态估计技术需要预先建立模型或在用户配合下进行立体初始化。由于车辆后部的各种外观,获得先前的模型是极其困难的。此外,在车辆行驶时要求驾驶员的配合是不安全的。在我们的系统中,通过车辆后方检测和跟踪自动提取立体算法的两个初始关键帧。地图点定义为从车辆尾部提取的点特征及其三维信息的集合。这些地图点是将后续车辆尾部检测到的2D特征与3D世界联系起来的推论。然后,通过将地图点与当前点特征匹配的映射,估计当前车辆后置和车载摄像头之间的相对3D姿态。我们通过增强现实展示了我们系统的能力,这需要精确和实时的3D姿态估计。
{"title":"3D Pose Estimation of Front Vehicle Towards a Better Driver Assistance System","authors":"Yu Peng, Jesse S. Jin, S. Luo, Min Xu, Yue Cui","doi":"10.1109/ICMEW.2012.97","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.97","url":null,"abstract":"Driver assistance system enhances traffic safety and efficiency. Accurate 3D pose of front vehicle can help driver to make right decisions on road. We propose a novel real-time system to estimate 3D pose of the front vehicle. This system consists of two parallel threads: vehicle rear tracking and mapping. Vehicle rear is firstly identified in the video captured by an on-board camera, after license plate localization and foreground extraction. 3D pose estimation technique is then employed with respect to extracted vehicle rear. Most 3D pose estimation techniques need prior models or a stereo initialization with user cooperation. It is extremely difficult to obtain prior models due to various appearances of vehicle rears. Moreover, it is unsafe to ask for driver's cooperation when vehicle is running. In our system, two initial key frames for stereo algorithm are automatically extracted by vehicle rear detection and tracking. Map points are defined as a collection of point features extracted from vehicle rear with their 3D information. These map points are inferences that relating 2D features detected in following vehicle rears with 3D world. Relative 3D Pose between current vehicle rear and on-board camera is then estimated through mapping that matches map points with current point features. We demonstrate the abilities of our system by augmented reality, which needs accurate and real-time 3D pose estimation.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129968414","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Lossless Compression of Stereo Disparity Maps for 3D 无损压缩立体视差地图为3D
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.113
M. Zamarin, Søren Forchhammer
Efficient compression of disparity data is important for accurate view synthesis purposes in multi-view communication systems based on the "texture plus depth" format, including the stereo case. In this paper a novel technique for loss less compression of stereo disparity images is presented. The coding algorithm is based on bit-plane coding, disparity prediction via disparity warping and context-based arithmetic coding exploiting predicted disparity data. Experimental results show that the proposed compression scheme achieves average compression factors of about 48:1 for high resolution disparity maps for stereo pairs and outperforms different standard solutions for loss less still image compression. Moreover, it provides a progressive representation of disparity data as well as a parallelizable structure.
在基于“纹理+深度”格式的多视点通信系统中,视差数据的有效压缩对于精确的视点合成至关重要。提出了一种对立体视差图像进行无损压缩的新方法。该编码算法基于位平面编码,通过视差扭曲进行视差预测,利用预测的视差数据进行基于上下文的算术编码。实验结果表明,该压缩方案对高分辨率立体视差图的平均压缩系数约为48:1,在低损失静态图像压缩方面优于其他标准压缩方案。此外,它还提供了视差数据的渐进式表示以及可并行化的结构。
{"title":"Lossless Compression of Stereo Disparity Maps for 3D","authors":"M. Zamarin, Søren Forchhammer","doi":"10.1109/ICMEW.2012.113","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.113","url":null,"abstract":"Efficient compression of disparity data is important for accurate view synthesis purposes in multi-view communication systems based on the \"texture plus depth\" format, including the stereo case. In this paper a novel technique for loss less compression of stereo disparity images is presented. The coding algorithm is based on bit-plane coding, disparity prediction via disparity warping and context-based arithmetic coding exploiting predicted disparity data. Experimental results show that the proposed compression scheme achieves average compression factors of about 48:1 for high resolution disparity maps for stereo pairs and outperforms different standard solutions for loss less still image compression. Moreover, it provides a progressive representation of disparity data as well as a parallelizable structure.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121935318","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Video Summarization with Global and Local Features 具有全局和局部特征的视频摘要
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.105
Genliang Guan, Zhiyong Wang, Kaimin Yu, Shaohui Mei, Mingyi He, D. Feng
Video summarization has been crucial for effective and efficient access of video content due to the ever increasing amount of video data. Most of the existing key frame based summarization approaches represent individual frames with global features, which neglects the local details of visual content. Considering that a video generally depicts a story with a number of scenes in different temporal order and shooting angles, we formulate scene summarization as identifying a set of frames which best covers the key point pool constructed from the scene. Therefore, our approach is a two-step process, identifying scenes and selecting representative content for each scene. Global features are utilized to identify scenes through clustering due to the visual similarity among video frames of the same scene, and local features to summarize each scene. We develop a key point based key frame selection method to identify representative content of a scene, which allows users to flexibly tune summarization length. Our preliminary results indicate that the proposed approach is very promising and potentially robust to clustering based scene identification.
随着视频数据量的不断增加,视频摘要对视频内容的有效访问至关重要。现有的基于关键帧的摘要方法大多是用全局特征表示单个帧,忽略了视觉内容的局部细节。考虑到一个视频通常用不同时间顺序和拍摄角度的多个场景来描述一个故事,我们将场景摘要定义为识别一组最能覆盖由场景构建的关键点池的帧。因此,我们的方法是一个两步的过程,识别场景和为每个场景选择代表性内容。利用全局特征通过聚类来识别同一场景的视频帧之间的视觉相似性,利用局部特征来总结每个场景。我们开发了一种基于关键点的关键帧选择方法来识别场景的代表性内容,使用户可以灵活地调整摘要长度。我们的初步结果表明,该方法在基于聚类的场景识别中非常有前途,具有潜在的鲁棒性。
{"title":"Video Summarization with Global and Local Features","authors":"Genliang Guan, Zhiyong Wang, Kaimin Yu, Shaohui Mei, Mingyi He, D. Feng","doi":"10.1109/ICMEW.2012.105","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.105","url":null,"abstract":"Video summarization has been crucial for effective and efficient access of video content due to the ever increasing amount of video data. Most of the existing key frame based summarization approaches represent individual frames with global features, which neglects the local details of visual content. Considering that a video generally depicts a story with a number of scenes in different temporal order and shooting angles, we formulate scene summarization as identifying a set of frames which best covers the key point pool constructed from the scene. Therefore, our approach is a two-step process, identifying scenes and selecting representative content for each scene. Global features are utilized to identify scenes through clustering due to the visual similarity among video frames of the same scene, and local features to summarize each scene. We develop a key point based key frame selection method to identify representative content of a scene, which allows users to flexibly tune summarization length. Our preliminary results indicate that the proposed approach is very promising and potentially robust to clustering based scene identification.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116383070","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
An Improved Pruning Method Based on the Number of States Possessed by Hypotheses 一种基于假设状态数的改进剪枝方法
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.106
Junyao Shao, Gang Liu, Zhiyuan Guo, Baoxiang Li, Yueming Lu
This paper presents an improved pruning method taking into account of the number of states possessed by hypotheses in some certain frames. With conventional pruning strategy, the hypotheses with a low score or a bad ranking will be discarded. However, it neglects a fact that the hypotheses several states ahead of or behind the right hypothesis in the prefix tree, which should be discarded, have similar scores and rankings with the right hypothesis. If a state is part of a partial path hypothesis, we say it is possessed by the hypothesis. So in a speech frame, we can deduce that the hypotheses which possess the most states and the hypotheses which possess the least states have little chance to be the right hypothesis. The proposed method analysis the range of the number of the states possessed by the hypotheses, and discards the hypotheses that possess too many or too few states. According to the experiments, This method could effectively improve the performance of the ASR.
本文提出了一种改进的剪枝方法,该方法考虑了某些框架中假设所拥有的状态数。在传统的修剪策略下,得分低或排名差的假设将被丢弃。然而,它忽略了一个事实,即在前缀树中正确假设前面或后面几个状态的假设应该被丢弃,它们与正确假设具有相似的分数和排名。如果一个状态是部分路径假设的一部分,我们说它被这个假设所拥有。所以在一个演讲框架中,我们可以推断出拥有最多状态的假设和拥有最少状态的假设都不太可能是正确的假设。该方法分析了假设所拥有的状态数的范围,并抛弃了拥有过多或过少状态的假设。实验结果表明,该方法可以有效地提高ASR的性能。
{"title":"An Improved Pruning Method Based on the Number of States Possessed by Hypotheses","authors":"Junyao Shao, Gang Liu, Zhiyuan Guo, Baoxiang Li, Yueming Lu","doi":"10.1109/ICMEW.2012.106","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.106","url":null,"abstract":"This paper presents an improved pruning method taking into account of the number of states possessed by hypotheses in some certain frames. With conventional pruning strategy, the hypotheses with a low score or a bad ranking will be discarded. However, it neglects a fact that the hypotheses several states ahead of or behind the right hypothesis in the prefix tree, which should be discarded, have similar scores and rankings with the right hypothesis. If a state is part of a partial path hypothesis, we say it is possessed by the hypothesis. So in a speech frame, we can deduce that the hypotheses which possess the most states and the hypotheses which possess the least states have little chance to be the right hypothesis. The proposed method analysis the range of the number of the states possessed by the hypotheses, and discards the hypotheses that possess too many or too few states. According to the experiments, This method could effectively improve the performance of the ASR.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116469018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
L-infinite Coding of 3D Representations of Human Affect 人类情感三维表征的l -无限编码
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.14
Ruxandra-Marina Florea, Leon Denis, J. Lievens, P. Schelkens, A. Munteanu
Off-line scanning, coding, transmission and remote animation of the human affect represents a possible processing pipeline for providing 3D immersion in virtual worlds. In this paper we target applications that make use of compact and scalable 3D representations of human affect and require close control over the local error introduced by lossy coding of the mesh geometry. To satisfy this requirement, we propose a novel L-infinite wavelet-based semi-regular mesh coding system. The system lies in contrast with classical mesh coding approaches which make use of the L-2 distortion metric. Specifically, in contrast to an L-2 driven implementation, the proposed system provides a bound on the local error on each vertex resulting from scalar embedded quantization of the wavelet coefficients. The experiments show that the proposed system provides scalability in L-infinite sense and that it outperforms the state-of-the art in L-infinite mesh coding.
人类情感的离线扫描、编码、传输和远程动画代表了在虚拟世界中提供3D沉浸感的可能处理管道。在本文中,我们的目标应用是利用紧凑和可扩展的人类情感的3D表示,并需要对网格几何的有损编码引入的局部误差进行严密控制。为了满足这一要求,我们提出了一种新颖的基于l无限小波的半正则网格编码系统。该系统与传统的利用L-2失真度量的网格编码方法形成了对比。具体来说,与L-2驱动的实现相比,所提出的系统提供了每个顶点上由小波系数的标量嵌入量化产生的局部误差的界限。实验表明,该系统具有l -无限意义上的可扩展性,并且在l -无限网格编码方面优于目前的技术水平。
{"title":"L-infinite Coding of 3D Representations of Human Affect","authors":"Ruxandra-Marina Florea, Leon Denis, J. Lievens, P. Schelkens, A. Munteanu","doi":"10.1109/ICMEW.2012.14","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.14","url":null,"abstract":"Off-line scanning, coding, transmission and remote animation of the human affect represents a possible processing pipeline for providing 3D immersion in virtual worlds. In this paper we target applications that make use of compact and scalable 3D representations of human affect and require close control over the local error introduced by lossy coding of the mesh geometry. To satisfy this requirement, we propose a novel L-infinite wavelet-based semi-regular mesh coding system. The system lies in contrast with classical mesh coding approaches which make use of the L-2 distortion metric. Specifically, in contrast to an L-2 driven implementation, the proposed system provides a bound on the local error on each vertex resulting from scalar embedded quantization of the wavelet coefficients. The experiments show that the proposed system provides scalability in L-infinite sense and that it outperforms the state-of-the art in L-infinite mesh coding.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127188528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Hybrid Coded Block Patterns Based Fast Mode Decision in H.264/AVC H.264/AVC中基于混合编码块模式的快速模式判定
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.10
Zhiru Shi, W. Fernando, A. Kondoz
The video coding standard H.264/AVC although possesses the advantage of higher coding efficiency compared to previous ones, it would result in a high computational complexity due to various block sizes motion estimation for multi-modes decision. In this paper, a hybrid inter-mode decision algorithm is presented, combining coded block pattern (CBP) with motion activity and Rate-Distortion (RD) cost. In this algorithm, CBP and CBP4×4, which indicate none-zero coefficient block, is used to determine candidate modes at macroblock level and sub-macroblock level. Further, the normalized motion activity is used to identify the homogeneity of the block and target the candidate modes set more accurate. An early termination is also made between macroblock level and sub-macroblock level by RD cost comparing. The experimental results show that the proposed algorithm achieves approximately 60% of computational complexity saving in terms of encoding time, while with negligible quality degradation, compared to the conventional method in H.264/AVC.
视频编码标准H.264/AVC虽然比以往的编码标准具有更高的编码效率,但由于多模式决策的块大小运动估计不同,导致计算复杂度较高。提出了一种将编码块模式(CBP)与运动活度和率失真(RD)代价相结合的混合模式间决策算法。该算法利用CBP和CBP4×4表示非零系数块来确定宏块级和子宏块级的候选模式。进一步,使用归一化的运动活动来识别块的均匀性,并更准确地针对候选模式设置。通过RD成本比较,在宏块级和子宏块级之间提前终止。实验结果表明,与传统的H.264/AVC编码方法相比,该算法在编码时间方面节省了约60%的计算复杂度,而质量下降可以忽略不计。
{"title":"A Hybrid Coded Block Patterns Based Fast Mode Decision in H.264/AVC","authors":"Zhiru Shi, W. Fernando, A. Kondoz","doi":"10.1109/ICMEW.2012.10","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.10","url":null,"abstract":"The video coding standard H.264/AVC although possesses the advantage of higher coding efficiency compared to previous ones, it would result in a high computational complexity due to various block sizes motion estimation for multi-modes decision. In this paper, a hybrid inter-mode decision algorithm is presented, combining coded block pattern (CBP) with motion activity and Rate-Distortion (RD) cost. In this algorithm, CBP and CBP4×4, which indicate none-zero coefficient block, is used to determine candidate modes at macroblock level and sub-macroblock level. Further, the normalized motion activity is used to identify the homogeneity of the block and target the candidate modes set more accurate. An early termination is also made between macroblock level and sub-macroblock level by RD cost comparing. The experimental results show that the proposed algorithm achieves approximately 60% of computational complexity saving in terms of encoding time, while with negligible quality degradation, compared to the conventional method in H.264/AVC.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127423267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Distributed Area of Interest Management for Large-Scale Immersive Video Conferencing 大规模沉浸式视频会议的分布式兴趣区管理
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.31
Pedram Pourashraf, F. Safaei, D. Franklin
Although video conferencing and its related applications have grown into a significant research area, the limited scalability of conference size is still a major problem. In this paper, a range of strategies for real-time area of interest (AOI) management in a 3D immersive video conference (IVC) are evaluated with the objective of minimising the required video transmission capacity and hence maximising the number of concurrent users. The paper shows that with judicious application of these techniques, the download capacity requirements of clients can be reduced by as much as 90% in a crowded virtual space.
虽然视频会议及其相关应用已成为一个重要的研究领域,但会议规模的可扩展性有限仍然是一个主要问题。本文对三维沉浸式视频会议(IVC)中实时感兴趣区域(AOI)管理的一系列策略进行了评估,其目标是最小化所需的视频传输容量,从而最大化并发用户数量。本文表明,通过明智地应用这些技术,在拥挤的虚拟空间中,客户机的下载容量需求可以减少多达90%。
{"title":"Distributed Area of Interest Management for Large-Scale Immersive Video Conferencing","authors":"Pedram Pourashraf, F. Safaei, D. Franklin","doi":"10.1109/ICMEW.2012.31","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.31","url":null,"abstract":"Although video conferencing and its related applications have grown into a significant research area, the limited scalability of conference size is still a major problem. In this paper, a range of strategies for real-time area of interest (AOI) management in a 3D immersive video conference (IVC) are evaluated with the objective of minimising the required video transmission capacity and hence maximising the number of concurrent users. The paper shows that with judicious application of these techniques, the download capacity requirements of clients can be reduced by as much as 90% in a crowded virtual space.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114889971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Vehicle Type Classification Using PCA with Self-Clustering 基于PCA的自聚类车辆类型分类
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.73
Yu Peng, Jesse S. Jin, S. Luo, Min Xu, Yue Cui
Different conditions, such as occlusions, changes of lighting, shadows and rotations, make vehicle type classification still a challenging task, especially for real-time applications. Most existing methods rely on presumptions on certain conditions, such as lighting conditions and special camera settings. However, these presumptions usually do not work for applications in real world. In this paper, we propose a robust vehicle type classification method based on adaptive multi-class Principal Components Analysis (PCA). We treat car images captured at daytime and night-time separately. Vehicle front is extracted by examining vehicle front width and the location of license plate. Then, after generating eigenvectors to represent extracted vehicle fronts, we propose a PCA method with self-clustering to classify vehicle type. The comparison experiments with the state of art methods and real-time evaluations demonstrate the promising performance of our proposed method. Moreover, as we do not find any public database including sufficient desired images, we built up online our own database including 4924 high-resolution images of vehicle front view for further research on this topic.
不同的条件,如遮挡、光照、阴影和旋转的变化,使得车辆类型分类仍然是一项具有挑战性的任务,特别是在实时应用中。大多数现有的方法都依赖于对某些条件的假设,比如照明条件和特殊的相机设置。然而,这些假设通常不适用于现实世界中的应用程序。本文提出了一种基于自适应多类主成分分析(PCA)的鲁棒车型分类方法。我们分别处理白天和夜间拍摄的汽车图像。通过检测车辆前方宽度和车牌位置提取车辆前方信息。然后,在生成特征向量来表示提取的车辆车头后,我们提出了一种自聚类的PCA方法来对车辆进行分类。通过与现有方法和实时评估方法的对比实验,证明了该方法具有良好的性能。此外,由于我们没有找到任何包含足够所需图像的公共数据库,因此我们在网上建立了自己的数据库,其中包含4924张汽车前视图的高分辨率图像,以进一步研究该主题。
{"title":"Vehicle Type Classification Using PCA with Self-Clustering","authors":"Yu Peng, Jesse S. Jin, S. Luo, Min Xu, Yue Cui","doi":"10.1109/ICMEW.2012.73","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.73","url":null,"abstract":"Different conditions, such as occlusions, changes of lighting, shadows and rotations, make vehicle type classification still a challenging task, especially for real-time applications. Most existing methods rely on presumptions on certain conditions, such as lighting conditions and special camera settings. However, these presumptions usually do not work for applications in real world. In this paper, we propose a robust vehicle type classification method based on adaptive multi-class Principal Components Analysis (PCA). We treat car images captured at daytime and night-time separately. Vehicle front is extracted by examining vehicle front width and the location of license plate. Then, after generating eigenvectors to represent extracted vehicle fronts, we propose a PCA method with self-clustering to classify vehicle type. The comparison experiments with the state of art methods and real-time evaluations demonstrate the promising performance of our proposed method. Moreover, as we do not find any public database including sufficient desired images, we built up online our own database including 4924 high-resolution images of vehicle front view for further research on this topic.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133536192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
A Multi-User Interaction System Based on Kinect and Wii Remote 基于Kinect和Wii遥控器的多用户交互系统
Pub Date : 2012-07-09 DOI: 10.1109/ICMEW.2012.123
Yihua Lou, Wenjun Wu, Hui Zhang, Haikuo Zhang, Yongquan Chen
We will demonstrate a multi-user interaction system that uses Kinect and Wii Remote for manipulating windows in both desktop and wall-sized environment. This system combines the gesture information collected by Kinect and other sensor information such as acceleration from Wii Remote, therefore providing a more accurate control and a more nature experience for users.
我们将演示一个多用户交互系统,使用Kinect和Wii遥控器在桌面和墙壁大小的环境中操作窗口。该系统将Kinect收集的手势信息与Wii Remote的加速度等传感器信息相结合,为用户提供更精确的控制和更自然的体验。
{"title":"A Multi-User Interaction System Based on Kinect and Wii Remote","authors":"Yihua Lou, Wenjun Wu, Hui Zhang, Haikuo Zhang, Yongquan Chen","doi":"10.1109/ICMEW.2012.123","DOIUrl":"https://doi.org/10.1109/ICMEW.2012.123","url":null,"abstract":"We will demonstrate a multi-user interaction system that uses Kinect and Wii Remote for manipulating windows in both desktop and wall-sized environment. This system combines the gesture information collected by Kinect and other sensor information such as acceleration from Wii Remote, therefore providing a more accurate control and a more nature experience for users.","PeriodicalId":385797,"journal":{"name":"2012 IEEE International Conference on Multimedia and Expo Workshops","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2012-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128499643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2012 IEEE International Conference on Multimedia and Expo Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1