首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
A novel upsampling scheme for depth map compression in 3DTV system 一种新的3DTV系统深度图压缩上采样方案
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702456
Yanjie Li, Lifeng Sun
In 3D video transmission, the depth map is normally compressed by resolution reduction to save bandwidth. The lost information in resolution reduction is recovered by an appropriate upsampling algorithm in decoding step. Most previous work considers the depth upsampling problem as common 2D map upsampling problem and do not take the intrinsic property of depth map into consideration. In this paper, we propose a novel two-step depth map upsampling scheme to address the problem on 3D videos. The first step utilizes the full resolution 2D color map to direct the reconstruction of a more accurate full resolution depth map. And the second step further flats the reconstructed depth map to ensure its local uniformity. Test results show that the proposed novel upsam-pling scheme achieves up to 2dB coding gains for the rendering of free-viewpoint video, and improves its perceptual quality significantly.
在3D视频传输中,深度图通常通过降低分辨率来压缩以节省带宽。在解码步骤中,通过适当的上采样算法恢复分辨率降低过程中丢失的信息。以往的研究大都将深度上采样问题视为普通的二维地图上采样问题,没有考虑深度图的固有特性。在本文中,我们提出了一种新的两步深度图上采样方案来解决三维视频的问题。第一步利用全分辨率2D彩色图来指导更精确的全分辨率深度图的重建。第二步将重建的深度图进一步平化,以保证局部均匀性。测试结果表明,所提出的上采样方案可实现高达2dB的编码增益,并显著提高了自由视点视频的感知质量。
{"title":"A novel upsampling scheme for depth map compression in 3DTV system","authors":"Yanjie Li, Lifeng Sun","doi":"10.1109/PCS.2010.5702456","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702456","url":null,"abstract":"In 3D video transmission, the depth map is normally compressed by resolution reduction to save bandwidth. The lost information in resolution reduction is recovered by an appropriate upsampling algorithm in decoding step. Most previous work considers the depth upsampling problem as common 2D map upsampling problem and do not take the intrinsic property of depth map into consideration. In this paper, we propose a novel two-step depth map upsampling scheme to address the problem on 3D videos. The first step utilizes the full resolution 2D color map to direct the reconstruction of a more accurate full resolution depth map. And the second step further flats the reconstructed depth map to ensure its local uniformity. Test results show that the proposed novel upsam-pling scheme achieves up to 2dB coding gains for the rendering of free-viewpoint video, and improves its perceptual quality significantly.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126607198","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Recent advances in video coding using static background models 使用静态背景模型的视频编码的最新进展
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702536
A. Krutz, A. Glantz, T. Sikora
Sprite coding, as standardized in MPEG-4 Visual, can result in superior performance compared to common hybrid video codecs both objectively and subjectively. However, state-of-the-art video coding standard H.264/AVC clearly outperforms MPEG-4 Visual sprite coding in broad bit rate ranges. Based on the sprite coding idea, this paper proposes a video coding technique that merges the advantages of H.264/AVC and sprite coding. For that, sophisticated algorithms for global motion estimation, sprite generation and object segmentation — all needed for thorough sprite coding — are incorporated into an H.264/AVC coding environment. The proposed approach outperforms H.264/AVC especially in lower bit rate ranges. Savings up to 21% can be achieved.
在MPEG-4 Visual中标准化的精灵编码在客观上和主观上都比普通的混合视频编解码器产生更好的性能。然而,最先进的视频编码标准H.264/AVC在宽比特率范围内明显优于MPEG-4视觉精灵编码。基于精灵编码的思想,提出了一种融合H.264/AVC和精灵编码优点的视频编码技术。为此,用于全局运动估计、精灵生成和对象分割的复杂算法(所有这些都需要彻底的精灵编码)被整合到H.264/AVC编码环境中。该方法在较低比特率范围内优于H.264/AVC。可节省高达21%的费用。
{"title":"Recent advances in video coding using static background models","authors":"A. Krutz, A. Glantz, T. Sikora","doi":"10.1109/PCS.2010.5702536","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702536","url":null,"abstract":"Sprite coding, as standardized in MPEG-4 Visual, can result in superior performance compared to common hybrid video codecs both objectively and subjectively. However, state-of-the-art video coding standard H.264/AVC clearly outperforms MPEG-4 Visual sprite coding in broad bit rate ranges. Based on the sprite coding idea, this paper proposes a video coding technique that merges the advantages of H.264/AVC and sprite coding. For that, sophisticated algorithms for global motion estimation, sprite generation and object segmentation — all needed for thorough sprite coding — are incorporated into an H.264/AVC coding environment. The proposed approach outperforms H.264/AVC especially in lower bit rate ranges. Savings up to 21% can be achieved.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126175139","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Low complexity video coding and the emerging HEVC standard 低复杂度视频编码和新兴的HEVC标准
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702540
K. Ugur, K. Andersson, A. Fuldseth, G. Bjøntegaard, L. P. Endresen, J. Lainema, A. Hallapuro, J. Ridge, D. Rusanovskyy, Cixun Zhang, A. Norkin, C. Priddle, T. Rusert, Jonatan Samuelsson, Rickard Sjöberg, Zhuangfei Wu
This paper describes a low complexity video codec with high coding efficiency. It was proposed to the High Efficiency Video Coding (HEVC) standardization effort of MPEG and VCEG, and has been partially adopted into the initial HEVC Test Model under Consideration design. The proposal utilizes a quad-tree structure with a support of large macroblocks of size 64×64 and 32×32, in addition to macroblocks of size 16×16. The entropy coding is done using a low complexity variable length coding based scheme with improved context adaptation over the H.264/AVC design. In addition, the proposal includes improved interpolation and deblocking filters, giving better coding efficiency while having low complexity. Finally, an improved intra coding method is presented. The subjective quality of the proposal is evaluated extensively and the results show that the proposed method achieves similar visual quality as H.264/AVC High Profile anchors with around 50% and 35% bit rate reduction for low delay and random-access experiments respectively at high definition sequences. This is achieved with less complexity than H.264/AVC Baseline Profile, making the proposal especially suitable for resource constrained environments.
本文介绍了一种编码效率高、复杂度低的视频编解码器。针对MPEG和VCEG的高效视频编码(High Efficiency Video Coding, HEVC)标准化工作提出了该方法,并在初步HEVC测试模型中部分采用。该提案利用四叉树结构,支持大小为64×64和32×32的大型宏块,以及大小为16×16的宏块。在H.264/AVC设计基础上,熵编码采用了一种低复杂度的变长编码方案,改进了上下文适应性。此外,该方案还包括改进的插值和去块滤波器,在降低复杂度的同时提高了编码效率。最后,提出了一种改进的帧内编码方法。对该方法的主观质量进行了广泛的评估,结果表明,该方法在高清晰度序列下的低延迟和随机访问实验中分别获得了与H.264/AVC High Profile锚点相似的视觉质量,比特率分别降低了50%和35%左右。这比H.264/AVC基线配置文件的复杂性更低,使该提议特别适合资源受限的环境。
{"title":"Low complexity video coding and the emerging HEVC standard","authors":"K. Ugur, K. Andersson, A. Fuldseth, G. Bjøntegaard, L. P. Endresen, J. Lainema, A. Hallapuro, J. Ridge, D. Rusanovskyy, Cixun Zhang, A. Norkin, C. Priddle, T. Rusert, Jonatan Samuelsson, Rickard Sjöberg, Zhuangfei Wu","doi":"10.1109/PCS.2010.5702540","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702540","url":null,"abstract":"This paper describes a low complexity video codec with high coding efficiency. It was proposed to the High Efficiency Video Coding (HEVC) standardization effort of MPEG and VCEG, and has been partially adopted into the initial HEVC Test Model under Consideration design. The proposal utilizes a quad-tree structure with a support of large macroblocks of size 64×64 and 32×32, in addition to macroblocks of size 16×16. The entropy coding is done using a low complexity variable length coding based scheme with improved context adaptation over the H.264/AVC design. In addition, the proposal includes improved interpolation and deblocking filters, giving better coding efficiency while having low complexity. Finally, an improved intra coding method is presented. The subjective quality of the proposal is evaluated extensively and the results show that the proposed method achieves similar visual quality as H.264/AVC High Profile anchors with around 50% and 35% bit rate reduction for low delay and random-access experiments respectively at high definition sequences. This is achieved with less complexity than H.264/AVC Baseline Profile, making the proposal especially suitable for resource constrained environments.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126473079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Automatic moving object extraction using x-means clustering 基于x均值聚类的自动运动目标提取
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702477
K. Imamura, Naoki Kubo, H. Hashimoto
The present paper proposes an automatic extraction technique of moving objects using x-means clustering. The proposed technique is an extended k-means clustering and can determine the optimal number of clusters based on the Bayesian Information Criterion(BIC). In the proposed method, the feature points are extracted from a current frame, and x-means clustering classifies the feature points based on their estimated affine motion parameters. A label is assigned to the segmented region, which is obtained by morphological watershed, by voting for the feature point cluster in each region. The labeling result represents the moving object extraction. Experimental results reveal that the proposed method provides extraction results with the suitable object number.
提出了一种基于x均值聚类的运动目标自动提取技术。该方法是一种扩展的k-means聚类方法,可以根据贝叶斯信息准则(BIC)确定最优聚类数量。在该方法中,从当前帧中提取特征点,并根据估计的仿射运动参数对特征点进行x均值聚类分类。通过形态学分水岭对每个区域的特征点聚类进行投票,为分割后的区域分配标签。标记结果表示运动目标的提取。实验结果表明,该方法能够提供具有合适目标数的提取结果。
{"title":"Automatic moving object extraction using x-means clustering","authors":"K. Imamura, Naoki Kubo, H. Hashimoto","doi":"10.1109/PCS.2010.5702477","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702477","url":null,"abstract":"The present paper proposes an automatic extraction technique of moving objects using x-means clustering. The proposed technique is an extended k-means clustering and can determine the optimal number of clusters based on the Bayesian Information Criterion(BIC). In the proposed method, the feature points are extracted from a current frame, and x-means clustering classifies the feature points based on their estimated affine motion parameters. A label is assigned to the segmented region, which is obtained by morphological watershed, by voting for the feature point cluster in each region. The labeling result represents the moving object extraction. Experimental results reveal that the proposed method provides extraction results with the suitable object number.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125197747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
The dependence of visual noise perception on background color and luminance 视觉噪声感知对背景颜色和亮度的依赖性
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702573
M. Shohara, K. Kotani
This paper describes the dependency of noise perception on background color and luminance of noise quantitatively. We conduct subjective and quantitative experiments for three noise models, using a modified grayscale method. The subjective experiment results show the perceived color noise depends on the background color, but the perceived luminance noise does not. The most sensitive background colors for color noises are yellow and purple. The perceived noises against background gray level show the similar trend between noise models. When the background gray level is L∗∼25, we perceive the noise best. In addition, the perceived chromatic noise level is about 8 times smaller than the calculated color noise using CIELAB Euclidean distance.
本文定量地描述了噪声感知对背景颜色和噪声亮度的依赖关系。采用改进的灰度法对三种噪声模型进行了主观和定量实验。主观实验结果表明,感知到的颜色噪声依赖于背景颜色,而感知到的亮度噪声不依赖于背景颜色。对彩色噪声最敏感的背景色是黄色和紫色。不同噪声模型对背景灰度值的感知趋势相似。当背景灰度为L * ~ 25时,我们对噪声的感知效果最好。此外,使用CIELAB欧氏距离计算得到的颜色噪声比感知到的颜色噪声小约8倍。
{"title":"The dependence of visual noise perception on background color and luminance","authors":"M. Shohara, K. Kotani","doi":"10.1109/PCS.2010.5702573","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702573","url":null,"abstract":"This paper describes the dependency of noise perception on background color and luminance of noise quantitatively. We conduct subjective and quantitative experiments for three noise models, using a modified grayscale method. The subjective experiment results show the perceived color noise depends on the background color, but the perceived luminance noise does not. The most sensitive background colors for color noises are yellow and purple. The perceived noises against background gray level show the similar trend between noise models. When the background gray level is L∗∼25, we perceive the noise best. In addition, the perceived chromatic noise level is about 8 times smaller than the calculated color noise using CIELAB Euclidean distance.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116128333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Separable Wiener filter based adaptive in-loop filter for video coding 基于可分离维纳滤波器的视频编码自适应环内滤波器
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702581
Mischa Siekmann, S. Bosse, H. Schwarz, T. Wiegand
Recent investigations have shown that a non-separable Wiener filter, that is applied inside the motion-compensation loop, can improve the coding efficiency of hybrid video coding designs. In this paper, we study the application of separable Wiener filters. Our design includes the possibility to adaptively choose between the application of the vertical, horizontal, or combined filter. The simulation results verify that a separable in-loop Wiener filter is capable of providing virtually the same increase in coding efficiency as a non-separable Wiener filter, but at a significantly reduced decoder complexity.
近年来的研究表明,在运动补偿环路中加入不可分离维纳滤波器可以提高混合视频编码设计的编码效率。本文研究了可分离维纳滤波器的应用。我们的设计包括自适应地选择应用垂直、水平或组合滤波器的可能性。仿真结果验证了可分离环内维纳滤波器能够提供与不可分离维纳滤波器几乎相同的编码效率提高,但显著降低了解码器的复杂性。
{"title":"Separable Wiener filter based adaptive in-loop filter for video coding","authors":"Mischa Siekmann, S. Bosse, H. Schwarz, T. Wiegand","doi":"10.1109/PCS.2010.5702581","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702581","url":null,"abstract":"Recent investigations have shown that a non-separable Wiener filter, that is applied inside the motion-compensation loop, can improve the coding efficiency of hybrid video coding designs. In this paper, we study the application of separable Wiener filters. Our design includes the possibility to adaptively choose between the application of the vertical, horizontal, or combined filter. The simulation results verify that a separable in-loop Wiener filter is capable of providing virtually the same increase in coding efficiency as a non-separable Wiener filter, but at a significantly reduced decoder complexity.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"105 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128061175","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Power-aware complexity-scalable multiview video coding for mobile devices 用于移动设备的功率感知复杂可扩展的多视图视频编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702506
M. Shafique, B. Zatt, S. Bampi, J. Henkel
We propose a novel power-aware scheme for complexity-scalable multiview video coding on mobile devices. Our scheme exploits the asymmetric view quality which is based on the binocular suppression theory. Our scheme employs different quality-complexity classes (QCCs) and adapts at run time depending upon the current battery state. It thereby enables a run-time tradeoff between complexity and video quality. The experimental results show that our scheme is superior to state-of-the-art and it provides an up to 87% complexity reduction while keeping the PSNR close to the exhaustive mode decision. We have demonstrated the power-aware adaptivity between different QCCs using a laptop with battery charging and discharging scenarios.
针对移动设备上复杂度可扩展的多视点视频编码,提出了一种新的功率感知方案。我们的方案利用了基于双目抑制理论的非对称视点质量。我们的方案采用不同的质量复杂度类(qcc),并根据当前电池状态在运行时进行调整。因此,它可以在复杂性和视频质量之间进行运行时权衡。实验结果表明,我们的方案优于最先进的方案,它在保持PSNR接近穷举模式决策的同时,提供了高达87%的复杂性降低。我们已经使用笔记本电脑演示了不同qc之间的功率感知适应性,并使用电池充电和放电场景。
{"title":"Power-aware complexity-scalable multiview video coding for mobile devices","authors":"M. Shafique, B. Zatt, S. Bampi, J. Henkel","doi":"10.1109/PCS.2010.5702506","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702506","url":null,"abstract":"We propose a novel power-aware scheme for complexity-scalable multiview video coding on mobile devices. Our scheme exploits the asymmetric view quality which is based on the binocular suppression theory. Our scheme employs different quality-complexity classes (QCCs) and adapts at run time depending upon the current battery state. It thereby enables a run-time tradeoff between complexity and video quality. The experimental results show that our scheme is superior to state-of-the-art and it provides an up to 87% complexity reduction while keeping the PSNR close to the exhaustive mode decision. We have demonstrated the power-aware adaptivity between different QCCs using a laptop with battery charging and discharging scenarios.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"2016 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127360240","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Coding efficient improvement by adaptive search center definition 自适应搜索中心定义提高编码效率
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702571
Kyohei Oba, Takahiro Bandou, Tian Song, T. Shimamoto
In this paper, an efficient search center definition algorithm is proposed for H.264/AVC. H.264/AVC achieved high coding efficiency by introducing some new coding tools including a new definition of the search center. However, the definition of the search center is not efficient in the case of significant motions. This work proposes some new search center candidates using spatial and temporal correlations of motion vectors to improve the coding efficiency. Simulation results show that the proposed search centers can achieve very much bit saving but induced high computation complexity. Additional complexity reduction algorithm is also introduced to improve the trade off between bit saving and implementation performance. This work realized a maximum bit saving of 19%.
本文针对H.264/AVC,提出了一种高效的搜索中心定义算法。H.264/AVC通过引入新的编码工具,包括对搜索中心的新定义,实现了较高的编码效率。然而,在重大运动的情况下,搜索中心的定义并不有效。本文利用运动矢量的时空相关性,提出了一些新的候选搜索中心,以提高编码效率。仿真结果表明,所提出的搜索中心可以节省大量的比特,但会导致较高的计算复杂度。此外,还引入了额外的复杂性降低算法,以改善比特节省和实现性能之间的权衡。这项工作最大节省了19%的钻头。
{"title":"Coding efficient improvement by adaptive search center definition","authors":"Kyohei Oba, Takahiro Bandou, Tian Song, T. Shimamoto","doi":"10.1109/PCS.2010.5702571","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702571","url":null,"abstract":"In this paper, an efficient search center definition algorithm is proposed for H.264/AVC. H.264/AVC achieved high coding efficiency by introducing some new coding tools including a new definition of the search center. However, the definition of the search center is not efficient in the case of significant motions. This work proposes some new search center candidates using spatial and temporal correlations of motion vectors to improve the coding efficiency. Simulation results show that the proposed search centers can achieve very much bit saving but induced high computation complexity. Additional complexity reduction algorithm is also introduced to improve the trade off between bit saving and implementation performance. This work realized a maximum bit saving of 19%.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132796554","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive direct vector derivation for video coding 视频编码的自适应直接矢量推导
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702458
Yusuke Itani, Shun-ichi Sekiguchi, Y. Yamada
This paper proposes a new method for improving direct prediction scheme that has been employed in conventional video coding standards such as AVC/H.264. We extend direct prediction concept to achieve better adaptation to local statistics of video source with the assumption of the use of larger motion blocks than conventional macroblock size. Firstly, our direct prediction method introduces block adaptive spatio-temporal estimation of direct motion vector, in order to compensate loss of accuracy of motion vector estimation for large motion blocks. This estimation method is performed without explicit signaling by employing decode-side collaborative decision. Then, adaptive selection of two reference pictures is performed to improve direct prediction efficiency where reliability of estimated direct motion vector is poor. Experimental results show the proposed method provides up to 3.3% bitrate saving and 1.5% in average in low-bitrate coding.
本文提出了一种改进AVC/H.264等传统视频编码标准中直接预测方案的新方法。我们扩展了直接预测概念,假设使用比传统宏块更大的运动块,以更好地适应视频源的局部统计。首先,我们的直接预测方法引入了直接运动矢量的分块自适应时空估计,以弥补大运动块时运动矢量估计精度的损失。该估计方法采用解码侧协同决策,无需显式信令。然后,在直接运动矢量估计可靠性较差的情况下,对两幅参考图片进行自适应选择,提高直接预测效率;实验结果表明,该方法在低比特率编码中可节省3.3%的比特率,平均节省1.5%的比特率。
{"title":"Adaptive direct vector derivation for video coding","authors":"Yusuke Itani, Shun-ichi Sekiguchi, Y. Yamada","doi":"10.1109/PCS.2010.5702458","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702458","url":null,"abstract":"This paper proposes a new method for improving direct prediction scheme that has been employed in conventional video coding standards such as AVC/H.264. We extend direct prediction concept to achieve better adaptation to local statistics of video source with the assumption of the use of larger motion blocks than conventional macroblock size. Firstly, our direct prediction method introduces block adaptive spatio-temporal estimation of direct motion vector, in order to compensate loss of accuracy of motion vector estimation for large motion blocks. This estimation method is performed without explicit signaling by employing decode-side collaborative decision. Then, adaptive selection of two reference pictures is performed to improve direct prediction efficiency where reliability of estimated direct motion vector is poor. Experimental results show the proposed method provides up to 3.3% bitrate saving and 1.5% in average in low-bitrate coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"3 23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129684953","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Improved FMO based H.264 frame layer rate control for low bit rate video transmission 改进的基于FMO的H.264帧层速率控制,用于低比特率视频传输
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702529
R. Cajote, S. Aramvith
The use of Flexible Macroblock Ordering (FMO) in H.264/AVC as an error-resilient tool incurs extra overhead bits that reduces coding efficiency at low bit rate. To improve coding efficiency, we present an improved frame-layer H.264/AVC rate control that takes into consideration the effects of using FMO for video transmission. In this paper, we propose a new header bits model, an enhanced frame complexity measure and a quantization parameter (QP) adjustment scheme. Simulation results show that the proposed method performed better than the existing frame layer rate control with FMO enabled using different number of slice groups.
在H.264/AVC中使用柔性宏块排序(FMO)作为纠错工具会导致额外的开销比特,从而降低低比特率下的编码效率。为了提高编码效率,我们提出了一种改进的帧层H.264/AVC速率控制方法,该方法考虑了FMO对视频传输的影响。在本文中,我们提出了一种新的报头位模型,一种增强的帧复杂度度量和量化参数(QP)调整方案。仿真结果表明,该方法比使用不同切片组数使能FMO的现有帧层速率控制方法效果更好。
{"title":"Improved FMO based H.264 frame layer rate control for low bit rate video transmission","authors":"R. Cajote, S. Aramvith","doi":"10.1109/PCS.2010.5702529","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702529","url":null,"abstract":"The use of Flexible Macroblock Ordering (FMO) in H.264/AVC as an error-resilient tool incurs extra overhead bits that reduces coding efficiency at low bit rate. To improve coding efficiency, we present an improved frame-layer H.264/AVC rate control that takes into consideration the effects of using FMO for video transmission. In this paper, we propose a new header bits model, an enhanced frame complexity measure and a quantization parameter (QP) adjustment scheme. Simulation results show that the proposed method performed better than the existing frame layer rate control with FMO enabled using different number of slice groups.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125113556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1