首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Joint Optimization of Rate, Distortion, and Maximum Absolute Error for Compression of Medical Volumes Using HEVC Intra 基于HEVC Intra的医疗卷压缩速率、失真和最大绝对误差联合优化
Pub Date : 2018-06-24 DOI: 10.1109/PCS.2018.8456282
Karina Jaskolka, André Kaup
Many visual quality metrics are used to measure the quality of lossy compressed images and videos, and are integrated in the rate-distortion optimization of hybrid video codecs. However, most of the metrics focus on the average objective quality in a picture. In certain applications, like medical image processing, the maximum absolute error should be more weighted. In this paper, the rate-distortion optimization of HEVC is extended by integrating this error metric. Thus, rate, average error, and maximum absolute error are jointly optimized. Furthermore, a weighting factor α is included into the calculation of the optimization for balancing the ratio between average and maximum absolute error. For HEVC intra with $alpha =0.25$ an average maximum absolute error reduction of −25.63 can be achieved, while the bitrate increases slightly by 0.59%. Furthermore, the visual quality of the medical volumes improves and the data fidelity increases, i.e. less block artifacts appear and less structure disappear.
许多视觉质量指标被用来衡量有损压缩图像和视频的质量,并被集成到混合视频编解码器的率失真优化中。然而,大多数度量标准关注的是图片的平均客观质量。在某些应用中,如医学图像处理,最大绝对误差应该更加权。本文通过对该误差度量进行积分,扩展了HEVC的率失真优化。从而对错误率、平均误差和最大绝对误差进行了联合优化。此外,在优化计算中加入加权因子α来平衡平均和最大绝对误差的比值。对于$alpha =0.25$的HEVC intra,可以实现平均最大绝对误差减少- 25.63,而比特率略微增加0.59%。此外,医疗体的视觉质量得到改善,数据保真度得到提高,即更少的块伪影出现,更少的结构消失。
{"title":"Joint Optimization of Rate, Distortion, and Maximum Absolute Error for Compression of Medical Volumes Using HEVC Intra","authors":"Karina Jaskolka, André Kaup","doi":"10.1109/PCS.2018.8456282","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456282","url":null,"abstract":"Many visual quality metrics are used to measure the quality of lossy compressed images and videos, and are integrated in the rate-distortion optimization of hybrid video codecs. However, most of the metrics focus on the average objective quality in a picture. In certain applications, like medical image processing, the maximum absolute error should be more weighted. In this paper, the rate-distortion optimization of HEVC is extended by integrating this error metric. Thus, rate, average error, and maximum absolute error are jointly optimized. Furthermore, a weighting factor α is included into the calculation of the optimization for balancing the ratio between average and maximum absolute error. For HEVC intra with $alpha =0.25$ an average maximum absolute error reduction of −25.63 can be achieved, while the bitrate increases slightly by 0.59%. Furthermore, the visual quality of the medical volumes improves and the data fidelity increases, i.e. less block artifacts appear and less structure disappear.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130713454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Future Video Coding Technologies: A Performance Evaluation of AV1, JEM, VP9, and HM 未来视频编码技术:AV1、JEM、VP9和HM的性能评估
Pub Date : 2018-06-24 DOI: 10.1109/PCS.2018.8456289
Tung Nguyen, D. Marpe
This paper presents and discusses an objective performance evaluation of two lines of development for the next generation of video coding schemes: AV1 of the Alliance for Open Media (AOM) and JEM of the Joint Video Exploration Team (JVET). Their already established predecessors, as given by VP9 for AV1 and High Efficiency Video Coding (HEVC) for JEM, serve as references for this evaluation process. A large testset of 28 video sequences with different content, various resolution, and different frame rate has been used as the common data basis. For each sequence and each test candidate, four different rate points associated with varying qualities of reconstruction have been generated. When evaluated in this way relative to their corresponding predecessors, AV1 and JEM show a quite significant coding-efficiency improvement in terms of averaged Bjøntegaard-Delta (BD) bit rate of about 23% and about 32%, respectively. Moreover, this performance study also reveals that there exist different trade-offs between compression efficiency and computational complexity in terms of encoder run time. While the JEM encoder requires a factor of about 8.5 in run time relative to HM, the corresponding run time factor for the AV1 encoder is about 55.8 relative to VP9. When comparing the next-generation coding schemes among each other, this study reveals that AV1 requires an average BD bit-rate overhead of about 49% at the same objective quality while consuming 2.47 times the run time of the JEM encoder.
本文提出并讨论了对下一代视频编码方案的两条发展路线的客观性能评估:开放媒体联盟(AOM)的AV1和联合视频探索小组(JVET)的JEM。它们已经建立的前辈,如AV1的VP9和JEM的高效视频编码(HEVC),可以作为评估过程的参考。使用28个不同内容、不同分辨率、不同帧率的视频序列的大型测试集作为通用数据基础。对于每个序列和每个候选测试,生成了与不同重建质量相关的四个不同速率点。当以这种方式进行评估时,AV1和JEM在平均Bjøntegaard-Delta (BD)比特率方面分别显示出相当显著的编码效率提高,分别约为23%和32%。此外,这项性能研究还表明,在编码器运行时间方面,压缩效率和计算复杂性之间存在不同的权衡。相对于HM, JEM编码器需要8.5倍的运行时间,而AV1编码器对应的运行时间因子相对于VP9大约是55.8倍。通过对下一代编码方案的比较,本研究发现,在相同的物象质量下,AV1的平均BD比特率开销约为49%,而其运行时间是JEM编码器的2.47倍。
{"title":"Future Video Coding Technologies: A Performance Evaluation of AV1, JEM, VP9, and HM","authors":"Tung Nguyen, D. Marpe","doi":"10.1109/PCS.2018.8456289","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456289","url":null,"abstract":"This paper presents and discusses an objective performance evaluation of two lines of development for the next generation of video coding schemes: AV1 of the Alliance for Open Media (AOM) and JEM of the Joint Video Exploration Team (JVET). Their already established predecessors, as given by VP9 for AV1 and High Efficiency Video Coding (HEVC) for JEM, serve as references for this evaluation process. A large testset of 28 video sequences with different content, various resolution, and different frame rate has been used as the common data basis. For each sequence and each test candidate, four different rate points associated with varying qualities of reconstruction have been generated. When evaluated in this way relative to their corresponding predecessors, AV1 and JEM show a quite significant coding-efficiency improvement in terms of averaged Bjøntegaard-Delta (BD) bit rate of about 23% and about 32%, respectively. Moreover, this performance study also reveals that there exist different trade-offs between compression efficiency and computational complexity in terms of encoder run time. While the JEM encoder requires a factor of about 8.5 in run time relative to HM, the corresponding run time factor for the AV1 encoder is about 55.8 relative to VP9. When comparing the next-generation coding schemes among each other, this study reveals that AV1 requires an average BD bit-rate overhead of about 49% at the same objective quality while consuming 2.47 times the run time of the JEM encoder.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"220 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115527640","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
Region-based models for motion compensation in video compression 视频压缩中基于区域的运动补偿模型
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456258
Jean Bégaint, Franck Galpin, P. Guillotel, C. Guillemot
Video codecs are primarily designed assuming that rigid, block-based, two-dimensional displacements are suitable models to describe the motion taking place in a scene. However, translational models are not sufficient to handle real world motion types such as camera zoom, shake, pan, shearing or changes in aspect ratio. We present here a region-based interprediction scheme to compensate such motion. The proposed mode is able to estimate multiple homography models in order to predict complex scene motion. We also introduce an affine photometric correction to each geometric model. Experiments on targeted sequences with complex motion demonstrate the efficiency of the proposed approach compared to the state-of-the-art HEVC video codec.
视频编解码器的主要设计假设刚性,基于块的二维位移是描述场景中发生的运动的合适模型。然而,平移模型不足以处理真实世界的运动类型,如相机缩放、晃动、平移、剪切或宽高比变化。我们提出了一种基于区域的互预测方案来补偿这种运动。该模型能够估计多个单应性模型,从而预测复杂的场景运动。我们还对每个几何模型引入了仿射光度校正。在复杂运动目标序列上的实验表明,与目前最先进的HEVC视频编解码器相比,该方法具有更高的效率。
{"title":"Region-based models for motion compensation in video compression","authors":"Jean Bégaint, Franck Galpin, P. Guillotel, C. Guillemot","doi":"10.1109/PCS.2018.8456258","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456258","url":null,"abstract":"Video codecs are primarily designed assuming that rigid, block-based, two-dimensional displacements are suitable models to describe the motion taking place in a scene. However, translational models are not sufficient to handle real world motion types such as camera zoom, shake, pan, shearing or changes in aspect ratio. We present here a region-based interprediction scheme to compensate such motion. The proposed mode is able to estimate multiple homography models in order to predict complex scene motion. We also introduce an affine photometric correction to each geometric model. Experiments on targeted sequences with complex motion demonstrate the efficiency of the proposed approach compared to the state-of-the-art HEVC video codec.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"57 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115087782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Motion-Distribution based Dynamic Texture Synthesis for Video Coding 基于运动分布的视频编码动态纹理合成
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456271
Olena Chubach, Patrick Garus, M. Wien, J. Ohm
In this paper, a new approach for an improved video coding scheme is presented, which combines hybrid video coding and texture synthesis based on motion distribution statistics. Considering that the utilized texture synthesis approach provides high-quality visual results, while it is developed only for synthe- sizing the identified dynamic textures within a certain area, a new framework is presented, which allows to identify of areas for synthesis and combine conventional coding with synthesis. Also, a new representation and compression of synthesis parameters is presented, which is required due to the updated coding structure. When combining the proposed approach with conventional en- coder (HEVC reference software, HM 16.6), significantly reduced bit rates of the compressed video sequences with the texture replaced can be obtained. Moreover, because the synthesized textures have similar perceptual characteristics to those of the original textures, the video sequences with the texture replaced are also visually similar to the original sequences. Video results are provided online to allow assessing the visual quality of the tested content.
本文提出了一种改进的视频编码方案,将混合视频编码与基于运动分布统计的纹理合成相结合。考虑到现有的纹理合成方法仅用于对识别出的某一区域内的动态纹理进行合成,而提供了高质量的视觉效果,提出了一种新的框架,可以识别出需要合成的区域,并将传统编码与合成相结合。同时,由于编码结构的更新,提出了一种新的合成参数表示和压缩方法。将该方法与传统编码器(HEVC参考软件HM 16.6)相结合,可以显著降低替换纹理后的压缩视频序列的比特率。此外,由于合成的纹理具有与原始纹理相似的感知特性,因此替换纹理后的视频序列在视觉上也与原始序列相似。在线提供视频结果,以便评估测试内容的视觉质量。
{"title":"Motion-Distribution based Dynamic Texture Synthesis for Video Coding","authors":"Olena Chubach, Patrick Garus, M. Wien, J. Ohm","doi":"10.1109/PCS.2018.8456271","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456271","url":null,"abstract":"In this paper, a new approach for an improved video coding scheme is presented, which combines hybrid video coding and texture synthesis based on motion distribution statistics. Considering that the utilized texture synthesis approach provides high-quality visual results, while it is developed only for synthe- sizing the identified dynamic textures within a certain area, a new framework is presented, which allows to identify of areas for synthesis and combine conventional coding with synthesis. Also, a new representation and compression of synthesis parameters is presented, which is required due to the updated coding structure. When combining the proposed approach with conventional en- coder (HEVC reference software, HM 16.6), significantly reduced bit rates of the compressed video sequences with the texture replaced can be obtained. Moreover, because the synthesized textures have similar perceptual characteristics to those of the original textures, the video sequences with the texture replaced are also visually similar to the original sequences. Video results are provided online to allow assessing the visual quality of the tested content.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121858228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Geometry-based Partitioning for Predictive Video Coding with Transform Adaptation 基于变换自适应的预测视频编码几何分割
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456238
Max Bläser, J. Schneider, Johannes Sauer, M. Wien
Rectangular block partitioning as it is used in state of the art video codecs such as HEVC can produce visually displeasing artifacts at low bitrates. This effect is particularly noticeable at moving object boundaries. This contribution presents a comprehensive geometry-based block partitioning framework in a post-HEVC codec for motion compensated prediction, intra-prediction and transform coding as a solution. The method is evaluated on the set of sequences defined by the Joint Call for Proposals on Video Compression with Capabilities beyond HEVC [1]. Our contribution aims at visually improving the quality of object boundaries and provides an objective BD-rate gain of 0.82% on average compared to the reference Joint Video Exploration Team (JVET) test model (JEM 7.0).
矩形块分割,因为它被用于最先进的视频编解码器,如HEVC,可以在低比特率下产生视觉上不愉快的工件。这种效果在移动物体边界时尤其明显。这一贡献提出了一个全面的基于几何的块划分框架,用于后hevc编解码器中的运动补偿预测,内部预测和转换编码作为解决方案。该方法在具有超越HEVC功能的视频压缩联合提案(Joint Call for Proposals on Video Compression)[1]定义的序列集上进行评估。我们的贡献旨在从视觉上提高目标边界的质量,与参考的联合视频勘探小组(JVET)测试模型(JEM 7.0)相比,提供了平均0.82%的客观bd率增益。
{"title":"Geometry-based Partitioning for Predictive Video Coding with Transform Adaptation","authors":"Max Bläser, J. Schneider, Johannes Sauer, M. Wien","doi":"10.1109/PCS.2018.8456238","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456238","url":null,"abstract":"Rectangular block partitioning as it is used in state of the art video codecs such as HEVC can produce visually displeasing artifacts at low bitrates. This effect is particularly noticeable at moving object boundaries. This contribution presents a comprehensive geometry-based block partitioning framework in a post-HEVC codec for motion compensated prediction, intra-prediction and transform coding as a solution. The method is evaluated on the set of sequences defined by the Joint Call for Proposals on Video Compression with Capabilities beyond HEVC [1]. Our contribution aims at visually improving the quality of object boundaries and provides an objective BD-rate gain of 0.82% on average compared to the reference Joint Video Exploration Team (JVET) test model (JEM 7.0).","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121341528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Complexity-Constrained Video Encoding and Delivery using Configuration Transfer Matrix 基于配置传输矩阵的复杂性约束视频编码与传输
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456269
Saverio G. Blasi, André Seixas Dias, M. Mrak, Shenglan Huang, E. Izquierdo
Many applications require video content to be encoded and uploaded under specific complexity constraints. While many speed-ups are available in practical video encoder implementations, it is difficult to predict the impact of such techniques on the actual content being encoded and therefore select the best configuration to meet the given constraints. A method is proposed in this paper to automatically select the encoder configuration in order to meet complexity constraints in terms of encoding and uploading time, using a pre-trained encoder configuration transfer matrix. The algorithm ensures that the content is processed within the specified targets, as presented in the experimental evaluation, where it is shown that the encoder can accurately meet specific constraints under a variety of conditions.
许多应用程序需要在特定的复杂性约束下对视频内容进行编码和上传。虽然在实际的视频编码器实现中可以使用许多加速,但很难预测这些技术对正在编码的实际内容的影响,因此选择最佳配置来满足给定的约束。本文提出了一种利用预训练的编码器配置传输矩阵自动选择编码器配置以满足编码和上传时间方面的复杂性约束的方法。该算法确保在指定的目标范围内处理内容,如实验评估所示,在各种条件下,编码器可以准确地满足特定的约束。
{"title":"Complexity-Constrained Video Encoding and Delivery using Configuration Transfer Matrix","authors":"Saverio G. Blasi, André Seixas Dias, M. Mrak, Shenglan Huang, E. Izquierdo","doi":"10.1109/PCS.2018.8456269","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456269","url":null,"abstract":"Many applications require video content to be encoded and uploaded under specific complexity constraints. While many speed-ups are available in practical video encoder implementations, it is difficult to predict the impact of such techniques on the actual content being encoded and therefore select the best configuration to meet the given constraints. A method is proposed in this paper to automatically select the encoder configuration in order to meet complexity constraints in terms of encoding and uploading time, using a pre-trained encoder configuration transfer matrix. The algorithm ensures that the content is processed within the specified targets, as presented in the experimental evaluation, where it is shown that the encoder can accurately meet specific constraints under a variety of conditions.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"163 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124469815","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Joint Energy Optimization of Video Encoding and Transmission 视频编码与传输的联合能量优化
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456245
Ziyu Ye, Rana Hegazy, Wei Zhou, P. Cosman, L. Milstein
Disposable wireless video sensors have many potential applications but are subject to stringent energy constraints. We studied the minimization of end-to-end distortion under an total energy constraint, by means of optimizing FEC code rate, number of source bits, and energy allocation between video encoding and wireless transmission. A two-step approach is employed. First, the FEC rate is optimized by exhaustive search. Then a binary-search-based algorithm is proposed to optimize the energy allocation and number of source bits. Experiments show that the algorithm achieves a PSNR gain up to 1dB over some reasonable baselines. A simpler suboptimal algorithm is also tested and exhibits similar performance.
一次性无线视频传感器有许多潜在的应用,但受到严格的能源限制。通过优化FEC码率、源比特数以及视频编码和无线传输之间的能量分配,研究了总能量约束下端到端失真的最小化问题。采用了两步方法。首先,通过穷举搜索优化FEC率。然后提出了一种基于二叉搜索的算法来优化能量分配和源比特数。实验表明,该算法在合理的基准上实现了高达1dB的PSNR增益。本文还测试了一种更简单的次优算法,并显示出类似的性能。
{"title":"Joint Energy Optimization of Video Encoding and Transmission","authors":"Ziyu Ye, Rana Hegazy, Wei Zhou, P. Cosman, L. Milstein","doi":"10.1109/PCS.2018.8456245","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456245","url":null,"abstract":"Disposable wireless video sensors have many potential applications but are subject to stringent energy constraints. We studied the minimization of end-to-end distortion under an total energy constraint, by means of optimizing FEC code rate, number of source bits, and energy allocation between video encoding and wireless transmission. A two-step approach is employed. First, the FEC rate is optimized by exhaustive search. Then a binary-search-based algorithm is proposed to optimize the energy allocation and number of source bits. Experiments show that the algorithm achieves a PSNR gain up to 1dB over some reasonable baselines. A simpler suboptimal algorithm is also tested and exhibits similar performance.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"84 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132631291","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Low-Complexity Intra Prediction Refinements for Video Coding 视频编码的低复杂度内部预测改进
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456305
Xin Zhao, V. Seregin, A. Said, Kai Zhang, Hilmi E. Egilmez, M. Karczewicz
In existing video coding standards such as H.264/AVC and HEVC, the intra prediction is typically derived using fixed, symmetric prediction filters along the prediction direction, e.g., in planar mode, top-right and bottom-left samples are predicted using symmetric prediction filters. However, in case ofasymmetric availability of neighboring reference samples, the performance of intra prediction filters designed in HEVC may not be optimal. To further refine the intra prediction and achieve higher accuracy of prediction samples, this paper proposes low-complexity refinements over HEVC intra prediction, which are applied on frequently used planar, DC, horizontal and vertical modes. The proposed method only requires simple addition and bit-shift operations on top of HEVC’s intra prediction implementation. Experimental results show that, an average of 0.7% coding gain is achieved for intra coding with no increase in run-time complexity.
在现有的视频编码标准(如H.264/AVC和HEVC)中,通常使用沿预测方向的固定对称预测滤波器推导出帧内预测,例如,在平面模式下,使用对称预测滤波器预测右上和左下的样本。然而,在邻近参考样本可用性不对称的情况下,HEVC中设计的内预测滤波器的性能可能不是最优的。为了进一步细化帧内预测,提高预测样本的精度,本文提出了HEVC帧内预测的低复杂度细化方法,分别应用于常用的平面、直流、水平和垂直模式。该方法只需要在HEVC的帧内预测实现基础上进行简单的加法和位移操作。实验结果表明,在不增加运行时复杂度的情况下,帧内编码的平均编码增益为0.7%。
{"title":"Low-Complexity Intra Prediction Refinements for Video Coding","authors":"Xin Zhao, V. Seregin, A. Said, Kai Zhang, Hilmi E. Egilmez, M. Karczewicz","doi":"10.1109/PCS.2018.8456305","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456305","url":null,"abstract":"In existing video coding standards such as H.264/AVC and HEVC, the intra prediction is typically derived using fixed, symmetric prediction filters along the prediction direction, e.g., in planar mode, top-right and bottom-left samples are predicted using symmetric prediction filters. However, in case ofasymmetric availability of neighboring reference samples, the performance of intra prediction filters designed in HEVC may not be optimal. To further refine the intra prediction and achieve higher accuracy of prediction samples, this paper proposes low-complexity refinements over HEVC intra prediction, which are applied on frequently used planar, DC, horizontal and vertical modes. The proposed method only requires simple addition and bit-shift operations on top of HEVC’s intra prediction implementation. Experimental results show that, an average of 0.7% coding gain is achieved for intra coding with no increase in run-time complexity.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124324014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Decoding Energy Modeling For The Next Generation Video Codec Based On Jem 基于Jem的下一代视频编解码器解码能量建模
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456244
Christian Herglotz, Matthias Kränzler, André Kaup
This paper shows that the processing energy of the decoder software for the next generation video codec can be accurately estimated using a feature based model. Therefore, a model from the literature is taken and extended to account for a high amount of the newly introduced coding modes. It is shown that using a selected set of 60 features, for a large set of more than 800 coded bit streams, a mean estimation error below 5% can be reached. Using the trained parameters of the model, the energy consumption of the decoder can be analyzed in detail such that, e.g., the coding modes consuming most processing energy can be identified. The model can be used inside the encoder for decoding- energy-rate-distortion optimization to generate decoding energy saving bit streams.
本文表明,采用基于特征的模型可以准确估计下一代视频编解码器的解码器软件的处理能量。因此,采用文献中的模型并进行扩展,以解释大量新引入的编码模式。结果表明,选用60个特征,对于800多个编码比特流的大数据集,平均估计误差可达5%以下。利用模型的训练参数,可以详细分析解码器的能量消耗,例如,可以识别出消耗最多处理能量的编码模式。该模型可用于编码器内部的解码能量率失真优化,以产生解码节能比特流。
{"title":"Decoding Energy Modeling For The Next Generation Video Codec Based On Jem","authors":"Christian Herglotz, Matthias Kränzler, André Kaup","doi":"10.1109/PCS.2018.8456244","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456244","url":null,"abstract":"This paper shows that the processing energy of the decoder software for the next generation video codec can be accurately estimated using a feature based model. Therefore, a model from the literature is taken and extended to account for a high amount of the newly introduced coding modes. It is shown that using a selected set of 60 features, for a large set of more than 800 coded bit streams, a mean estimation error below 5% can be reached. Using the trained parameters of the model, the energy consumption of the decoder can be analyzed in detail such that, e.g., the coding modes consuming most processing energy can be identified. The model can be used inside the encoder for decoding- energy-rate-distortion optimization to generate decoding energy saving bit streams.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114728902","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Scene-based KLT for Intra Coding in HEVC HEVC中基于场景的内部编码KLT
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456240
Yiqun Liu, J. Ostermann
Transform coding and quantization are part of the cornerstones in the current High Efficiency Video Coding (HEVC) standard. They are applied on the residuals from inter-frame or intra predictions. With specified transform matrices, HEVC enhances the coding efficiency vastly compared to Advanced Video Coding (AVC). However, there is still room for improvement. It is observed that the coding of transform coefficients occupies the majority of the bit rate in the stream, since transform matrices in HEVC can not offer the best energy compaction for prediction errors, especially for diagonal features. We introduce scene-based Karhunen-Loeve transform (KLT) in place of the conventional transform for the intra-predicted data for 8 × 8 and 16 × 16 Transform Units (TU). The transform matrices are adaptively designed and later applied according to the prediction modes, quantization steps as well as sizes. The simulation shows great prospect of reducing the bit rate further with KLT, as we gain 3.23%, 7.18% and 6.25% in terms of BD-Rate against HM-16.15 on average for class B, class C and BVI textures respectively, with All-Intra configuration.
变换编码和量化是当前高效视频编码(HEVC)标准的基础。它们被应用于帧间或帧内预测的残差上。HEVC与高级视频编码(AVC)相比,通过指定变换矩阵,大大提高了编码效率。然而,仍有改进的余地。观察到变换系数的编码占据了流中大部分比特率,因为HEVC中的变换矩阵不能为预测误差提供最佳的能量压缩,特别是对角特征。对于8 × 8和16 × 16变换单元(TU),我们引入了基于场景的Karhunen-Loeve变换(KLT)来代替传统的预测内数据变换。变换矩阵根据预测模式、量化步骤和大小自适应设计和应用。仿真结果表明,使用KLT进一步降低比特率的前景很大,在All-Intra配置下,我们在B类、C类和BVI纹理上的平均比特率分别比HM-16.15提高了3.23%、7.18%和6.25%。
{"title":"Scene-based KLT for Intra Coding in HEVC","authors":"Yiqun Liu, J. Ostermann","doi":"10.1109/PCS.2018.8456240","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456240","url":null,"abstract":"Transform coding and quantization are part of the cornerstones in the current High Efficiency Video Coding (HEVC) standard. They are applied on the residuals from inter-frame or intra predictions. With specified transform matrices, HEVC enhances the coding efficiency vastly compared to Advanced Video Coding (AVC). However, there is still room for improvement. It is observed that the coding of transform coefficients occupies the majority of the bit rate in the stream, since transform matrices in HEVC can not offer the best energy compaction for prediction errors, especially for diagonal features. We introduce scene-based Karhunen-Loeve transform (KLT) in place of the conventional transform for the intra-predicted data for 8 × 8 and 16 × 16 Transform Units (TU). The transform matrices are adaptively designed and later applied according to the prediction modes, quantization steps as well as sizes. The simulation shows great prospect of reducing the bit rate further with KLT, as we gain 3.23%, 7.18% and 6.25% in terms of BD-Rate against HM-16.15 on average for class B, class C and BVI textures respectively, with All-Intra configuration.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121816043","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1