首页 > 最新文献

2013 Visual Communications and Image Processing (VCIP)最新文献

英文 中文
Measuring orderliness based on social force model in collective motions 集体运动中基于社会力量模型的有序度度量
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706431
Yu Bai, Yi Xu, Xiaokang Yang, Qing Yan
Collective motions, one of the coordinated behaviors in crowd system, widely exist in nature. Orderliness characterizes how well an individual will move smoothly and consistently with his neighbors in collective motions. It is still an open problem in computer vision. In this paper, we propose an orderliness descriptor based on correlation of interactive social force between individuals. In order to include the force correlation between two individuals in a distance, we propose a Social Force Correlation Propagation algorithm to calculate orderliness of every individual effectively and efficiently. We validate the effectiveness of the proposed orderliness descriptor on synthetic simulation. Experimental results on challenging videos of real scene crowds demonstrate that orderliness descriptor can perceive motion with low smoothness and locate disorder.
群体运动是群体系统中的一种协调行为,在自然界中广泛存在。井然有序的特点是一个人在集体运动中与他的邻居平稳一致地移动的程度。这在计算机视觉中仍然是一个开放的问题。本文提出了一种基于个体间互动社会力相关性的有序描述符。为了包含距离上两个个体之间的力相关性,我们提出了一种社会力相关性传播算法来有效地计算每个个体的有序度。在综合仿真中验证了所提出的有序描述符的有效性。在真实场景人群挑战性视频上的实验结果表明,有序描述符可以感知低平滑度的运动并定位无序。
{"title":"Measuring orderliness based on social force model in collective motions","authors":"Yu Bai, Yi Xu, Xiaokang Yang, Qing Yan","doi":"10.1109/VCIP.2013.6706431","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706431","url":null,"abstract":"Collective motions, one of the coordinated behaviors in crowd system, widely exist in nature. Orderliness characterizes how well an individual will move smoothly and consistently with his neighbors in collective motions. It is still an open problem in computer vision. In this paper, we propose an orderliness descriptor based on correlation of interactive social force between individuals. In order to include the force correlation between two individuals in a distance, we propose a Social Force Correlation Propagation algorithm to calculate orderliness of every individual effectively and efficiently. We validate the effectiveness of the proposed orderliness descriptor on synthetic simulation. Experimental results on challenging videos of real scene crowds demonstrate that orderliness descriptor can perceive motion with low smoothness and locate disorder.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"137 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131719547","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Efficient key picture and single loop decoding scheme for SHVC 高效的SHVC密钥图像和单循环解码方案
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706453
K. Rapaka, Jianle Chen, M. Karczewicz
Scalable video coding has been a popular research topic for many years. As one of its key objectives, it aims to support different receiving devices connected through a network structure using a single bitstream. Scalable video coding extension of HEVC, also called as SHVC, is being developed by Joint Collaborative Team on Video Coding (JCT-VC) of ISO/IEC MPEG and ITU-T VCEG. Compared to previous standardized scalable video coding technologies, SHVC employs multi-loop decoding design with no low-level changes within any given layer compared to HEVC. With such a simplified extension it aims at solving some of the problems of previous scalable extensions that haven't been successful, and at the same time, aims at supporting all design features that are of vital importance for the success of SHVC. Supporting lightweight and finely tunable bandwidth adaptation is one such vital design feature important for the success of SHVC. This paper proposes novel high level syntax mechanism for SHVC quality scalability to support: (a) using the decoded pictures from higher quality layer as reference for lower layer pictures and key pictures concept to reduce drift; (b) single loop decoding design with encoder only constraints without introducing any normative low-level changes to the normal multi-loop decoding process. Experimental results based on SHVC reference software (SHM 2.0) show that the proposed key picture method achieves an average of 2.9% luma BD-rate reduction in multi-loop framework and an average of 4.4% luma BD-rate loss to attain the capability of single loop decoding.
多年来,可扩展视频编码一直是一个热门的研究课题。作为其主要目标之一,它旨在支持使用单个比特流通过网络结构连接的不同接收设备。HEVC的可扩展视频编码扩展,也称为SHVC,正在由ISO/IEC MPEG和ITU-T VCEG的视频编码联合协作小组(JCT-VC)开发。与以往标准化的可扩展视频编码技术相比,SHVC采用多环路解码设计,与HEVC相比,在任何给定层内都没有底层变化。有了这样一个简化的扩展,它的目标是解决以前的可伸缩扩展没有成功的一些问题,同时,它的目标是支持对SHVC的成功至关重要的所有设计特性。支持轻量级和精细可调的带宽适应是SHVC成功的重要设计特性之一。本文提出了一种新的SHVC质量可扩展性高级语法机制,以支持:(a)使用高质量层的解码图像作为低质量层图像和关键图像概念的参考,以减少漂移;(b)单循环解码设计,仅具有编码器约束,而不对正常的多循环解码过程引入任何规范的低级更改。基于SHVC参考软件(SHM 2.0)的实验结果表明,所提出的关键图像方法在多环框架下平均降低了2.9%的亮度帧率,平均降低了4.4%的亮度帧率,达到了单环解码的能力。
{"title":"Efficient key picture and single loop decoding scheme for SHVC","authors":"K. Rapaka, Jianle Chen, M. Karczewicz","doi":"10.1109/VCIP.2013.6706453","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706453","url":null,"abstract":"Scalable video coding has been a popular research topic for many years. As one of its key objectives, it aims to support different receiving devices connected through a network structure using a single bitstream. Scalable video coding extension of HEVC, also called as SHVC, is being developed by Joint Collaborative Team on Video Coding (JCT-VC) of ISO/IEC MPEG and ITU-T VCEG. Compared to previous standardized scalable video coding technologies, SHVC employs multi-loop decoding design with no low-level changes within any given layer compared to HEVC. With such a simplified extension it aims at solving some of the problems of previous scalable extensions that haven't been successful, and at the same time, aims at supporting all design features that are of vital importance for the success of SHVC. Supporting lightweight and finely tunable bandwidth adaptation is one such vital design feature important for the success of SHVC. This paper proposes novel high level syntax mechanism for SHVC quality scalability to support: (a) using the decoded pictures from higher quality layer as reference for lower layer pictures and key pictures concept to reduce drift; (b) single loop decoding design with encoder only constraints without introducing any normative low-level changes to the normal multi-loop decoding process. Experimental results based on SHVC reference software (SHM 2.0) show that the proposed key picture method achieves an average of 2.9% luma BD-rate reduction in multi-loop framework and an average of 4.4% luma BD-rate loss to attain the capability of single loop decoding.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"84 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131791237","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On secondary transforms for scalable video coding 关于可扩展视频编码的二次变换
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706392
A. Saxena, Felix C. A. Fernandes
In this paper, we present a secondary transform scheme for inter-layer prediction residue in scalable video coding (SVC). Efficient prediction of the co-located blocks from the base layer (BL) can significantly improve the enhancement layer (EL) coding in SVC, especially when the temporal information from previous EL frames is less correlated than the co-located BL information. However, Guo et al. showed that because of the peculiar frequency characteristics of EL residuals, the conventional DCT Type-2 transform is suboptimal and is often outperformed by either the DCT Type-3, or DST Type-3 when these transforms are applied to the EL residuals. However, their proposed technique requires upto 8 additional transform cores, two of which are of size 32×32. Here, in this work, we propose a secondary transform scheme, where the proposed transform is applied only to the lower 8x8 frequency coefficients after DCT, for block sizes 8×8 to 32×32. Our proposed transform scheme requires at most only 2 additional cores. We also propose a low-complexity 8x8 Rotational Transform as a special case of secondary transforms in this paper. Simulation results show that the proposed transform scheme provides significant BD-Rate improvement over the conventional DCT-based coding scheme for video sequences in the ongoing scalable extensions of HEVC standardization.
本文提出了一种可扩展视频编码(SVC)中层间预测残差的二次变换方案。对基帧(BL)中共定位块的有效预测可以显著改善SVC中增强层(EL)编码,特别是当来自前EL帧的时间信息与共定位BL信息的相关性较低时。然而,Guo等人表明,由于EL残差特有的频率特性,当将这些变换应用于EL残差时,传统的DCT Type-2变换是次优的,并且通常被DCT Type-3或DST Type-3优于。然而,他们提出的技术需要多达8个额外的变换核心,其中两个大小为32×32。在这里,在这项工作中,我们提出了一种二次变换方案,其中所提议的变换仅应用于DCT后较低的8x8频率系数,用于块大小8×8到32×32。我们提出的转换方案最多只需要2个额外的核心。本文还提出了一种低复杂度的8x8旋转变换作为二次变换的特例。仿真结果表明,在HEVC标准化的持续扩展中,所提出的变换方案比传统的基于dct的视频序列编码方案具有显著的BD-Rate改进。
{"title":"On secondary transforms for scalable video coding","authors":"A. Saxena, Felix C. A. Fernandes","doi":"10.1109/VCIP.2013.6706392","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706392","url":null,"abstract":"In this paper, we present a secondary transform scheme for inter-layer prediction residue in scalable video coding (SVC). Efficient prediction of the co-located blocks from the base layer (BL) can significantly improve the enhancement layer (EL) coding in SVC, especially when the temporal information from previous EL frames is less correlated than the co-located BL information. However, Guo et al. showed that because of the peculiar frequency characteristics of EL residuals, the conventional DCT Type-2 transform is suboptimal and is often outperformed by either the DCT Type-3, or DST Type-3 when these transforms are applied to the EL residuals. However, their proposed technique requires upto 8 additional transform cores, two of which are of size 32×32. Here, in this work, we propose a secondary transform scheme, where the proposed transform is applied only to the lower 8x8 frequency coefficients after DCT, for block sizes 8×8 to 32×32. Our proposed transform scheme requires at most only 2 additional cores. We also propose a low-complexity 8x8 Rotational Transform as a special case of secondary transforms in this paper. Simulation results show that the proposed transform scheme provides significant BD-Rate improvement over the conventional DCT-based coding scheme for video sequences in the ongoing scalable extensions of HEVC standardization.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129406104","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Scene-aware perceptual video coding 场景感知感知视频编码
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706427
Fei Liang, Xiulian Peng, Jizheng Xu
The mean-square-error (MSE) distortion criterion used in the state-of-the-art video coding standards, e.g. H.264/AVC and the High Efficiency Video Coding (HEVC) under standardization, is widely criticized for poor measurement of perceived visual quality. Existing research on perceptual video coding mainly employs low-level features of images/video, which cannot take into account the big picture people see. This paper proposes a scene-aware perceptual video coding scheme (SAPC), which accommodates human visual perception of the scene by reconstructing the scene from video and perform scene-based bits allocation. To be specific, more bits are allocated to the foreground object and its boundaries considering that people tend to pay more attention to the foreground and object boundaries are prone to blur at low bitrates for object occlusion. The structure from motion (SFM) technology is employed for scene reconstruction. Experiments taking HEVC as the benchmark show that our algorithm can give better visual quality than the original HEVC encoder at the same bitrate.
目前最先进的视频编码标准,如H.264/AVC和标准化的高效视频编码(HEVC)中使用的均方误差(MSE)失真标准,因其无法衡量感知视觉质量而受到广泛批评。现有的感知视频编码研究主要利用图像/视频的底层特征,无法考虑到人们看到的大局观。本文提出了一种场景感知感知视频编码方案(SAPC),该方案通过从视频中重构场景并进行基于场景的比特分配来适应人类对场景的视觉感知。具体来说,考虑到人们往往更关注前景,并且在低比特率下物体遮挡容易使物体边界模糊,因此分配给前景物体及其边界的比特数更多。采用运动结构(SFM)技术进行场景重建。以HEVC为基准的实验表明,在相同比特率下,我们的算法比原来的HEVC编码器具有更好的视觉质量。
{"title":"Scene-aware perceptual video coding","authors":"Fei Liang, Xiulian Peng, Jizheng Xu","doi":"10.1109/VCIP.2013.6706427","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706427","url":null,"abstract":"The mean-square-error (MSE) distortion criterion used in the state-of-the-art video coding standards, e.g. H.264/AVC and the High Efficiency Video Coding (HEVC) under standardization, is widely criticized for poor measurement of perceived visual quality. Existing research on perceptual video coding mainly employs low-level features of images/video, which cannot take into account the big picture people see. This paper proposes a scene-aware perceptual video coding scheme (SAPC), which accommodates human visual perception of the scene by reconstructing the scene from video and perform scene-based bits allocation. To be specific, more bits are allocated to the foreground object and its boundaries considering that people tend to pay more attention to the foreground and object boundaries are prone to blur at low bitrates for object occlusion. The structure from motion (SFM) technology is employed for scene reconstruction. Experiments taking HEVC as the benchmark show that our algorithm can give better visual quality than the original HEVC encoder at the same bitrate.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130955513","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Neighboring block based disparity vector derivation for 3D-AVC 基于邻块的3D-AVC视差矢量推导
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706401
Li Zhang, Jewon Kang, Xin Zhao, Ying Chen, R. Joshi
3D-AVC, being developed under Joint Collaborative Team on 3D Video Coding (JCT-3V), significantly outperforms the Multiview Video Coding plus Depth (MVC+D) which has no new macroblock level coding tools compared to Multiview video coding extension of H.264/AVC (MVC). However, for multiview compatible configuration, i.e., when texture views are decoded without accessing depth information, the performance of the current 3D-AVC is only marginally better than MVC+D. The problem is caused by the lack of disparity vectors which can be obtained only from the coded depth views in 3D-AVC. In this paper, a disparity vector derivation method is proposed by using the motion information of neighboring blocks and applied along with existing coding tools in 3D-AVC. The proposed method improves 3D-AVC in the multiview compatible mode substantially, resulting in about 20% bitrate reduction for texture coding. When enabling the so-called view synthesis prediction to further refine the disparity vectors, the performance of the proposed method is 31% better than MVC+D and even better than 3D-AVC under the best performing 3D-AVC configuration.
3D-AVC由3D视频编码联合协作团队(JCT-3V)开发,与H.264/AVC (MVC)的多视图视频编码扩展相比,明显优于没有新的宏块级编码工具的多视图视频编码加深度(MVC+D)。然而,对于多视图兼容配置,即当纹理视图在不访问深度信息的情况下进行解码时,当前3D-AVC的性能仅略好于MVC+D。这个问题是由于缺乏视差矢量造成的,而视差矢量只能从3D-AVC编码的深度视图中获得。本文提出了一种利用相邻块的运动信息推导视差矢量的方法,并结合现有的编码工具应用于3D-AVC中。该方法在多视图兼容模式下对3D-AVC进行了大幅度改进,使纹理编码的比特率降低了约20%。当启用所谓的视图综合预测来进一步细化视差向量时,在性能最好的3D-AVC配置下,该方法的性能比MVC+D提高31%,甚至优于3D-AVC。
{"title":"Neighboring block based disparity vector derivation for 3D-AVC","authors":"Li Zhang, Jewon Kang, Xin Zhao, Ying Chen, R. Joshi","doi":"10.1109/VCIP.2013.6706401","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706401","url":null,"abstract":"3D-AVC, being developed under Joint Collaborative Team on 3D Video Coding (JCT-3V), significantly outperforms the Multiview Video Coding plus Depth (MVC+D) which has no new macroblock level coding tools compared to Multiview video coding extension of H.264/AVC (MVC). However, for multiview compatible configuration, i.e., when texture views are decoded without accessing depth information, the performance of the current 3D-AVC is only marginally better than MVC+D. The problem is caused by the lack of disparity vectors which can be obtained only from the coded depth views in 3D-AVC. In this paper, a disparity vector derivation method is proposed by using the motion information of neighboring blocks and applied along with existing coding tools in 3D-AVC. The proposed method improves 3D-AVC in the multiview compatible mode substantially, resulting in about 20% bitrate reduction for texture coding. When enabling the so-called view synthesis prediction to further refine the disparity vectors, the performance of the proposed method is 31% better than MVC+D and even better than 3D-AVC under the best performing 3D-AVC configuration.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130984406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Transform coefficient coding design for AVS2 video coding standard AVS2视频编码标准的变换系数编码设计
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706447
Jing Wang, Xiaofeng Wang, Tianying Ji, Dake He
AVS2 is a next-generation audio and video coding standard currently under development by the Audio Video Coding Standard Workgroup of China. In this paper, a coefficient-group based transform coefficient coding design for AVS2 video coding standard is presented, which includes two main coding tools, namely, two-level coefficient coding and intra-mode based context design. The two-level coefficient coding scheme allows accurate coefficient position information to be used in the context model design and improves the coding efficiency. It also helps increase the entropy coding throughput and facilitate parallel implementation. The intra-mode based context design further improves coding performance by utilizing the intra-prediction mode information in the context model. The two coding tools combined provide consistent rate-distortion performance gains under standard test conditions. Both tools were adopted into the AVS2 working draft. Furthermore, an improved rate-distortion optimized quantization algorithm is designed based on the proposed scheme, which significantly reduces the encoder complexity.
AVS2是中国音视频编码标准工作组正在开发的下一代音视频编码标准。本文提出了一种基于系数组的AVS2视频编码标准变换系数编码设计方法,该方法包括两种主要的编码工具,即两级系数编码和基于模内的上下文设计。两级系数编码方案可以在上下文模型设计中使用准确的系数位置信息,提高编码效率。它还有助于提高熵编码吞吐量和促进并行实现。基于模式内的上下文设计通过利用上下文模型中的模式内预测信息进一步提高了编码性能。这两种编码工具在标准测试条件下提供一致的率失真性能增益。这两个工具都被纳入了AVS2工作草案。在此基础上设计了一种改进的率失真优化量化算法,显著降低了编码器的复杂度。
{"title":"Transform coefficient coding design for AVS2 video coding standard","authors":"Jing Wang, Xiaofeng Wang, Tianying Ji, Dake He","doi":"10.1109/VCIP.2013.6706447","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706447","url":null,"abstract":"AVS2 is a next-generation audio and video coding standard currently under development by the Audio Video Coding Standard Workgroup of China. In this paper, a coefficient-group based transform coefficient coding design for AVS2 video coding standard is presented, which includes two main coding tools, namely, two-level coefficient coding and intra-mode based context design. The two-level coefficient coding scheme allows accurate coefficient position information to be used in the context model design and improves the coding efficiency. It also helps increase the entropy coding throughput and facilitate parallel implementation. The intra-mode based context design further improves coding performance by utilizing the intra-prediction mode information in the context model. The two coding tools combined provide consistent rate-distortion performance gains under standard test conditions. Both tools were adopted into the AVS2 working draft. Furthermore, an improved rate-distortion optimized quantization algorithm is designed based on the proposed scheme, which significantly reduces the encoder complexity.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"82 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133543161","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Effective stereo matching using reliable points based graph cut 有效的立体匹配使用可靠的点为基础的图切
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706415
Haoqian Wang, M. Wu, Yongbing Zhang, Lei Zhang
In this paper, we propose an effective stereo matching algorithm using reliable points and region-based graph cut. Firstly, the initial disparity maps are calculated via local windowbased method. Secondly, the unreliable points are detected according to the DSI(Disparity Space Image) and the estimated disparity values of each unreliable point are obtained by considering its surrounding points. Then, the scheme of reliable points is introduced in region-based graph cut framework to optimize the initial result. Finally, remaining errors in the disparity results are effectively handled in a multi-step refinement process. Experiment results show that the proposed algorithm achieves a significant reduction in computation cost and guarantee high matching quality.
本文提出了一种基于可靠点和区域图割的立体匹配算法。首先,采用基于局部窗口的方法计算初始视差图;其次,根据DSI(视差空间图像)检测不可靠点,并结合每个不可靠点周围的点得到估计的视差值;然后,在基于区域的图割框架中引入可靠点方案,对初始结果进行优化。最后,通过多步细化过程有效处理视差结果中的剩余误差。实验结果表明,该算法显著降低了计算量,保证了较高的匹配质量。
{"title":"Effective stereo matching using reliable points based graph cut","authors":"Haoqian Wang, M. Wu, Yongbing Zhang, Lei Zhang","doi":"10.1109/VCIP.2013.6706415","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706415","url":null,"abstract":"In this paper, we propose an effective stereo matching algorithm using reliable points and region-based graph cut. Firstly, the initial disparity maps are calculated via local windowbased method. Secondly, the unreliable points are detected according to the DSI(Disparity Space Image) and the estimated disparity values of each unreliable point are obtained by considering its surrounding points. Then, the scheme of reliable points is introduced in region-based graph cut framework to optimize the initial result. Finally, remaining errors in the disparity results are effectively handled in a multi-step refinement process. Experiment results show that the proposed algorithm achieves a significant reduction in computation cost and guarantee high matching quality.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132592622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
A coding unit classification based AVC-to-HEVC transcoding with background modeling for surveillance videos 基于avc - hevc转码的监控视频背景建模编码单元分类
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706393
Peiyin Xing, Yonghong Tian, Xianguo Zhang, Yaowei Wang, Tiejun Huang
To save the storage and transmission cost, it is applicable now to develop fast and efficient methods to transcode the perennial surveillance videos to HEVC ones, since HEVC has doubled the compression ratio. Considering the long-time static background characteristic of surveillance videos, this paper presents a coding unit (CU) classification based AVC-to-HEVC transcoding method with background modeling. In our method, the background frame modeled from originally decoded frames is firstly transcoded into HEVC stream as long-term reference to enhance the prediction efficiency. Afterwards, a CU classification algorithm which employs decoded motion vectors and the modeled background frame as input is proposed to divide the decoded data into background, foreground and hybrid CUs. Following this, different transcoding strategies of CU partition termination, prediction unit candidate selection and motion estimation simplification are adopted for different CU categories to reduce the complexity. Experimental results show our method can achieve 45% bit saving and 50% complexity reduction against traditional AVC-to-HEVC transcoding.
为了节省存储和传输成本,开发快速高效的方法将常年监控视频转码为HEVC, HEVC使压缩比提高了一倍。针对监控视频长时间静态背景的特点,提出了一种基于背景建模的基于编码单元(CU)分类的avc - hevc转码方法。在我们的方法中,首先将原始解码帧建模的背景帧转编码为HEVC流作为长期参考,以提高预测效率。然后,提出了一种以解码后的运动矢量和建模后的背景帧为输入的CU分类算法,将解码后的数据分为背景CU、前景CU和混合CU。然后,针对不同的CU类别,采用不同的CU分区终止转码策略、预测单元候选选择策略和运动估计简化策略来降低复杂性。实验结果表明,与传统的avc - hevc转码相比,该方法可以节省45%的比特,降低50%的复杂度。
{"title":"A coding unit classification based AVC-to-HEVC transcoding with background modeling for surveillance videos","authors":"Peiyin Xing, Yonghong Tian, Xianguo Zhang, Yaowei Wang, Tiejun Huang","doi":"10.1109/VCIP.2013.6706393","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706393","url":null,"abstract":"To save the storage and transmission cost, it is applicable now to develop fast and efficient methods to transcode the perennial surveillance videos to HEVC ones, since HEVC has doubled the compression ratio. Considering the long-time static background characteristic of surveillance videos, this paper presents a coding unit (CU) classification based AVC-to-HEVC transcoding method with background modeling. In our method, the background frame modeled from originally decoded frames is firstly transcoded into HEVC stream as long-term reference to enhance the prediction efficiency. Afterwards, a CU classification algorithm which employs decoded motion vectors and the modeled background frame as input is proposed to divide the decoded data into background, foreground and hybrid CUs. Following this, different transcoding strategies of CU partition termination, prediction unit candidate selection and motion estimation simplification are adopted for different CU categories to reduce the complexity. Experimental results show our method can achieve 45% bit saving and 50% complexity reduction against traditional AVC-to-HEVC transcoding.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115364796","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Progressive motion vector resolution for HEVC HEVC的渐进运动矢量分辨率
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706363
Juncheng Ma, Jicheng An, Kai Zhang, Siwei Ma, S. Lei
This paper proposes a progressive motion vector resolution (PMVR) method for High Efficiency Video Coding (HEVC). In the proposed scheme, high motion vector (MV) resolutions, e.g. 1/4 or 1/8 pixel resolution, are employed for MVs near to the motion vector predictor (MVP) and low MV resolutions are employed for MVs far from the MVP. The range of each MV resolution is indicated by a threshold parameter. And a new motion vector difference (MVD) derivation method is designed to encode MVD efficiently. Experimental results show that PMVR with 1/8 pixel motion search can achieve a BD-rate gain up to 16% with almost the same coding time with HM8.0, and for PMVR without 1/8 pixel motion search, up to 6.1% BD-rate gain can be achieved with 9% encoding time saving on average.
提出了一种用于高效视频编码(HEVC)的渐进式运动矢量分辨率(PMVR)方法。在提出的方案中,对靠近运动矢量预测器(MVP)的MV采用高运动矢量(MV)分辨率,例如1/4或1/8像素分辨率,而对远离MVP的MV采用低MV分辨率。每个MV分辨率的范围由一个阈值参数表示。设计了一种新的运动矢量差分(MVD)推导方法,对MVD进行高效编码。实验结果表明,采用1/8像素运动搜索的PMVR在与HM8.0编码时间几乎相同的情况下,BD-rate增益可达16%;不采用1/8像素运动搜索的PMVR在编码时间平均节省9%的情况下,BD-rate增益可达6.1%。
{"title":"Progressive motion vector resolution for HEVC","authors":"Juncheng Ma, Jicheng An, Kai Zhang, Siwei Ma, S. Lei","doi":"10.1109/VCIP.2013.6706363","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706363","url":null,"abstract":"This paper proposes a progressive motion vector resolution (PMVR) method for High Efficiency Video Coding (HEVC). In the proposed scheme, high motion vector (MV) resolutions, e.g. 1/4 or 1/8 pixel resolution, are employed for MVs near to the motion vector predictor (MVP) and low MV resolutions are employed for MVs far from the MVP. The range of each MV resolution is indicated by a threshold parameter. And a new motion vector difference (MVD) derivation method is designed to encode MVD efficiently. Experimental results show that PMVR with 1/8 pixel motion search can achieve a BD-rate gain up to 16% with almost the same coding time with HM8.0, and for PMVR without 1/8 pixel motion search, up to 6.1% BD-rate gain can be achieved with 9% encoding time saving on average.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"200 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124255510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Wavelet based smoke detection method with RGB Contrast-image and shape constrain 基于RGB对比图像和形状约束的小波烟雾检测方法
Pub Date : 2013-11-01 DOI: 10.1109/VCIP.2013.6706406
Jia Chen, Yaowei Wang, Yonghong Tian, Tiejun Huang
Smoke detection in video surveillance is very important for early fire detection. A general viewpoint assumes that smoke is a low frequency signal which may smoothen the background. However, some pure-color objects also have this characteristic, and smoke also produces high frequency signal because the rich edge information of its contour. In order to solve these problems, an improved smoke detection method with RGB Contrast-image and shape constrain is proposed. In this method, wavelet transformation is implemented based on the RGB Contrast-image to distinguish smoke from other low frequency signals, and the existence of smoke is determined by analyzing the combination of the shape and the energy change of the region. Experimental results show our method outperforms the conventional methods remarkably.
视频监控中的烟雾探测对于火灾的早期探测非常重要。一般的观点认为烟雾是一种低频信号,可以使背景变得平滑。然而,一些纯色物体也具有这一特性,而烟雾也因为其轮廓丰富的边缘信息而产生高频信号。为了解决这些问题,提出了一种改进的基于RGB对比图像和形状约束的烟雾检测方法。该方法基于RGB对比图像进行小波变换,将烟雾与其他低频信号区分开来,结合区域的形状变化和能量变化进行分析,判断烟雾是否存在。实验结果表明,该方法明显优于传统方法。
{"title":"Wavelet based smoke detection method with RGB Contrast-image and shape constrain","authors":"Jia Chen, Yaowei Wang, Yonghong Tian, Tiejun Huang","doi":"10.1109/VCIP.2013.6706406","DOIUrl":"https://doi.org/10.1109/VCIP.2013.6706406","url":null,"abstract":"Smoke detection in video surveillance is very important for early fire detection. A general viewpoint assumes that smoke is a low frequency signal which may smoothen the background. However, some pure-color objects also have this characteristic, and smoke also produces high frequency signal because the rich edge information of its contour. In order to solve these problems, an improved smoke detection method with RGB Contrast-image and shape constrain is proposed. In this method, wavelet transformation is implemented based on the RGB Contrast-image to distinguish smoke from other low frequency signals, and the existence of smoke is determined by analyzing the combination of the shape and the energy change of the region. Experimental results show our method outperforms the conventional methods remarkably.","PeriodicalId":407080,"journal":{"name":"2013 Visual Communications and Image Processing (VCIP)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114570181","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
期刊
2013 Visual Communications and Image Processing (VCIP)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1