首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Compression of Dynamic Medical CT Data Using Motion Compensated Wavelet Lifting with Denoised Update 运动补偿小波提升与去噪更新的医学CT动态数据压缩
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456262
Daniela Lanz, Jürgen Seiler, Karina Jaskolka, André Kaup
For the lossless compression of dynamic $3-mathrm {D}+mathrm {t}$ volumes as produced by medical devices like Computed Tomography, various coding schemes can be applied. This paper shows that 3-D subband coding outperforms lossless HEVC coding and additionally provides a scalable representation, which is often required in telemedicine applications. However, the resulting lowpass subband, which shall be used as a downscaled representative of the whole original sequence, contains a lot of ghosting artifacts. This can be alleviated by incorporating motion compensation methods into the subband coder. This results in a high quality lowpass subband but also leads to a lower compression ratio. In order to cope with this, we introduce a new approach for improving the compression efficiency of compensated 3-D wavelet lifting by performing denoising in the update step. We are able to reduce the file size of the lowpass subband by up to 1.64%, while the lowpass subband is still applicable for being used as a downscaled representative of the whole original sequence.
对于由计算机断层扫描等医疗设备产生的动态$3- mathm {D}+ mathm {t}$卷的无损压缩,可以采用各种编码方案。本文表明,三维子带编码优于无损HEVC编码,并且还提供了远程医疗应用中经常需要的可扩展表示。然而,得到的低通子带,作为整个原始序列的缩小代表,包含了大量的鬼影伪影。这可以通过在子带编码器中加入运动补偿方法来缓解。这导致高质量的低通子带,但也导致较低的压缩比。为了解决这个问题,我们提出了一种新的方法,通过在更新步骤中进行去噪来提高补偿三维小波提升的压缩效率。我们能够将低通子带的文件大小减少1.64%,同时低通子带仍然可以作为整个原始序列的缩小代表。
{"title":"Compression of Dynamic Medical CT Data Using Motion Compensated Wavelet Lifting with Denoised Update","authors":"Daniela Lanz, Jürgen Seiler, Karina Jaskolka, André Kaup","doi":"10.1109/PCS.2018.8456262","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456262","url":null,"abstract":"For the lossless compression of dynamic $3-mathrm {D}+mathrm {t}$ volumes as produced by medical devices like Computed Tomography, various coding schemes can be applied. This paper shows that 3-D subband coding outperforms lossless HEVC coding and additionally provides a scalable representation, which is often required in telemedicine applications. However, the resulting lowpass subband, which shall be used as a downscaled representative of the whole original sequence, contains a lot of ghosting artifacts. This can be alleviated by incorporating motion compensation methods into the subband coder. This results in a high quality lowpass subband but also leads to a lower compression ratio. In order to cope with this, we introduce a new approach for improving the compression efficiency of compensated 3-D wavelet lifting by performing denoising in the update step. We are able to reduce the file size of the lowpass subband by up to 1.64%, while the lowpass subband is still applicable for being used as a downscaled representative of the whole original sequence.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"104 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122579673","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Backward Compatible Layered Video Coding for 360° Video Broadcast 360°视频广播的向后兼容分层视频编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456301
T. Biatek, J. Travers, Pierre-Loup Cabarat, W. Hamidouche
Recently, coding of 360° video contents has been investigated in the context of over-the-top streaming services. To be delivered using terrestrial broadcast, it is required to provide backward compatibility of such content to legacy receivers. In this paper, a novel layered coding scheme is proposed to address the delivery of 360° video content over terrestrial broadcast networks. One or several views are extracted from the 360° video and coded as base layers using standard HEVC encoding. Inter-layer reference pictures are built based on projected base-layers and are used in the enhancement layer to encode the 360° video. Experimental results show that the proposed approach provides substantial coding gains of 14.99% compared to simulcast coding and enables limited coding overhead of 5.15% compared to 360° single-layer coding.
最近,360°视频内容的编码已经在超顶级流媒体服务的背景下进行了研究。要使用地面广播传送,就需要向传统接收器提供这些内容的向后兼容性。本文提出了一种新颖的分层编码方案,以解决地面广播网络上360°视频内容的传输问题。从360°视频中提取一个或多个视图,并使用标准HEVC编码作为基础层进行编码。基于投影基层构建层间参考图像,用于增强层对360°视频进行编码。实验结果表明,与360°单层编码相比,该方法的编码增益可达14.99%,编码开销仅为5.15%。
{"title":"Backward Compatible Layered Video Coding for 360° Video Broadcast","authors":"T. Biatek, J. Travers, Pierre-Loup Cabarat, W. Hamidouche","doi":"10.1109/PCS.2018.8456301","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456301","url":null,"abstract":"Recently, coding of 360° video contents has been investigated in the context of over-the-top streaming services. To be delivered using terrestrial broadcast, it is required to provide backward compatibility of such content to legacy receivers. In this paper, a novel layered coding scheme is proposed to address the delivery of 360° video content over terrestrial broadcast networks. One or several views are extracted from the 360° video and coded as base layers using standard HEVC encoding. Inter-layer reference pictures are built based on projected base-layers and are used in the enhancement layer to encode the 360° video. Experimental results show that the proposed approach provides substantial coding gains of 14.99% compared to simulcast coding and enables limited coding overhead of 5.15% compared to 360° single-layer coding.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130475971","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
An MSE Approach For Training And Coding Steered Mixtures Of Experts 一种训练和编码操纵混合专家的MSE方法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456250
M. Tok, Rolf Jongebloed, Lieven Lange, Erik Bochinski, T. Sikora
Previous research has shown the interesting properties and potential of Steered Mixtures-of-Experts (SMoE) for image representation, approximation, and compression based on EM optimization. In this paper we introduce an MSE optimization method based on Gradient Descent for training SMoEs. This allows improved optimization towards PSNR and SSIM and de-coupling of experts and gates. In consequence we can now generate very high quality SMoE models with significantly reduced model complexity compared to previous work and much improved edge representations. Uased on this strategy a block-based image coder was developed using Mixture-of-Experts that uses very simple experts with very few model parameters. Experimental evaluations shows that a significant compression gain can be achieved compared to JPEG for low bit rates.
先前的研究已经显示了基于EM优化的转向混合专家(SMoE)在图像表示、逼近和压缩方面的有趣特性和潜力。本文介绍了一种基于梯度下降的最小均方误差优化方法。这允许改进对PSNR和SSIM的优化以及专家和门的解耦。因此,我们现在可以生成非常高质量的SMoE模型,与以前的工作相比,模型复杂性显著降低,边缘表示也得到了很大改善。基于这一策略,使用混合专家开发了基于块的图像编码器,使用非常简单的专家和很少的模型参数。实验评估表明,与低比特率的JPEG相比,可以获得显着的压缩增益。
{"title":"An MSE Approach For Training And Coding Steered Mixtures Of Experts","authors":"M. Tok, Rolf Jongebloed, Lieven Lange, Erik Bochinski, T. Sikora","doi":"10.1109/PCS.2018.8456250","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456250","url":null,"abstract":"Previous research has shown the interesting properties and potential of Steered Mixtures-of-Experts (SMoE) for image representation, approximation, and compression based on EM optimization. In this paper we introduce an MSE optimization method based on Gradient Descent for training SMoEs. This allows improved optimization towards PSNR and SSIM and de-coupling of experts and gates. In consequence we can now generate very high quality SMoE models with significantly reduced model complexity compared to previous work and much improved edge representations. Uased on this strategy a block-based image coder was developed using Mixture-of-Experts that uses very simple experts with very few model parameters. Experimental evaluations shows that a significant compression gain can be achieved compared to JPEG for low bit rates.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127721468","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Extending HEVC with a Texture Synthesis Framework using Detail-aware Image Decomposition 使用细节感知图像分解的纹理合成框架扩展HEVC
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456248
Bastian Wandt, Thorsten Laude, B. Rosenhahn, J. Ostermann
In recent years, there has been a tremendous improvement in video coding algorithms. This improvement resulted in 2013 in the standardization of the first version of High Efficiency Video Coding (HEVC) which now forms the state-of-theart with superior coding efficiency. Nevertheless, the development of video coding algorithms did not stop as HEVC still has its limitations. Especially for complex textures HEVC reveals one of its limitations. As these textures are hard to predict, very high bit rates are required to achieve a high quality. Texture synthesis was proposed as solution for this limitation in previous works. However, previous texture synthesis frameworks only prevailed if the decomposition into synthesizable and non-synthesizable regions was either known or very easy. In this paper, we address this scenario with a texture synthesis framework based on detail-aware image decomposition techniques. Our techniques are based on a multiple-steps coarse-to-fine approach in which an initial decomposition is refined with awareness for small details. The efficiency of our approach is evaluated objectively and subjectively: BD-rate gains of up to 28.81% over HEVC and up to 12.75% over the closest related work were achieved. Our subjective tests indicate an improved visual quality in addition to the bit rate savings.
近年来,视频编码算法有了很大的改进。这一改进导致了2013年第一版高效视频编码(HEVC)的标准化,现在形成了具有卓越编码效率的最新状态。然而,视频编码算法的发展并没有停止,HEVC仍然有它的局限性。特别是对于复杂的纹理,HEVC揭示了它的局限性之一。由于这些纹理很难预测,因此需要非常高的比特率来实现高质量。纹理合成在以前的工作中被提出来解决这一限制。然而,以前的纹理合成框架只有在已知或非常容易分解为可合成和不可合成区域的情况下才会流行。在本文中,我们使用基于细节感知图像分解技术的纹理合成框架来解决这种情况。我们的技术是基于一个多步骤的从粗到精的方法,在这个方法中,一个初始的分解被细化到小细节。我们的方法的效率得到了客观和主观的评价:与HEVC相比,bd率提高了28.81%,与最接近的相关工作相比,bd率提高了12.75%。我们的主观测试表明,除了比特率节省之外,视觉质量也得到了改善。
{"title":"Extending HEVC with a Texture Synthesis Framework using Detail-aware Image Decomposition","authors":"Bastian Wandt, Thorsten Laude, B. Rosenhahn, J. Ostermann","doi":"10.1109/PCS.2018.8456248","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456248","url":null,"abstract":"In recent years, there has been a tremendous improvement in video coding algorithms. This improvement resulted in 2013 in the standardization of the first version of High Efficiency Video Coding (HEVC) which now forms the state-of-theart with superior coding efficiency. Nevertheless, the development of video coding algorithms did not stop as HEVC still has its limitations. Especially for complex textures HEVC reveals one of its limitations. As these textures are hard to predict, very high bit rates are required to achieve a high quality. Texture synthesis was proposed as solution for this limitation in previous works. However, previous texture synthesis frameworks only prevailed if the decomposition into synthesizable and non-synthesizable regions was either known or very easy. In this paper, we address this scenario with a texture synthesis framework based on detail-aware image decomposition techniques. Our techniques are based on a multiple-steps coarse-to-fine approach in which an initial decomposition is refined with awareness for small details. The efficiency of our approach is evaluated objectively and subjectively: BD-rate gains of up to 28.81% over HEVC and up to 12.75% over the closest related work were achieved. Our subjective tests indicate an improved visual quality in addition to the bit rate savings.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127878247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Geometry-Corrected Deblocking Filter for 360° Video Coding using Cube Representation 使用立方体表示的360°视频编码的几何校正去块滤波器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456263
Johannes Sauer, M. Wien, J. Schneider, Max Bläser
In 360° video, a complete scene is captured, as it can be seen from a single point in any direction. Since the captured 360 images are spherical, they cannot be converted to planar images without introducing geometric distortions. The nature of these distortion depends on the used projection format.This paper introduces an approach to reduce artifacts occurring when encoding 360° video which has been projected to the faces of a cube. In order to achieve this, the operation of the deblocking filter is modified such that the correct pixels with respect to the 3D geometry are used for filtering of edges.The method is evaluated on the set of sequences defined by the Joint Call for Proposals on Video Compression with Capability beyond HEVC. While the method has almost no impact on the objective coding performance, the visual quality is still clearly enhanced. Edges of the cube, previously visible as coding artifacts, are mostly removed with the proposed method.
在360°视频中,一个完整的场景被捕获,因为它可以从任何方向的一个点看到。由于捕获的360度图像是球形的,因此在不引入几何畸变的情况下无法将其转换为平面图像。这些失真的性质取决于所使用的投影格式。本文介绍了一种减少投影到立方体表面的360°视频编码时产生伪影的方法。为了实现这一点,对去块滤波器的操作进行修改,以便使用相对于3D几何形状的正确像素来过滤边缘。在HEVC以外的视频压缩联合提案中定义的序列集上对该方法进行了评估。虽然该方法对目标编码性能几乎没有影响,但视觉质量仍然明显增强。以前作为编码工件可见的多维数据集边缘,使用所提出的方法大部分被删除。
{"title":"Geometry-Corrected Deblocking Filter for 360° Video Coding using Cube Representation","authors":"Johannes Sauer, M. Wien, J. Schneider, Max Bläser","doi":"10.1109/PCS.2018.8456263","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456263","url":null,"abstract":"In 360° video, a complete scene is captured, as it can be seen from a single point in any direction. Since the captured 360 images are spherical, they cannot be converted to planar images without introducing geometric distortions. The nature of these distortion depends on the used projection format.This paper introduces an approach to reduce artifacts occurring when encoding 360° video which has been projected to the faces of a cube. In order to achieve this, the operation of the deblocking filter is modified such that the correct pixels with respect to the 3D geometry are used for filtering of edges.The method is evaluated on the set of sequences defined by the Joint Call for Proposals on Video Compression with Capability beyond HEVC. While the method has almost no impact on the objective coding performance, the visual quality is still clearly enhanced. Edges of the cube, previously visible as coding artifacts, are mostly removed with the proposed method.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115246734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Detecting Source Video Artifacts with Supervised Sparse Filters 用监督稀疏滤波器检测源视频伪影
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456303
T. Goodall, A. Bovik
A variety of powerful picture quality predictors are available that rely on neuro-statistical models of distortion perception. We extend these principles to video source inspection, by coupling spatial divisive normalization with a filterbank tuned for artifact detection, implemented in an augmented sparse functional form. We call this method the Video Impairment Detection by SParse Error CapTure (VIDSPECT). We configure VIDSPECT to create state-of-the-art detectors of two kinds of commonly encountered source video artifacts: upscaling and combing. The system detects upscaling, identifies upscaling type, and predicts the native video resolution. It also detects combing artifacts arising from interlacing. Our approach is simple, highly generalizable, and yields better accuracy than competing methods. A software release of VIDSPECT is available online: http://live.ece.utexas.edu/research/quality/VIDSPECT release.zip for public use and evaluation.
各种强大的图像质量预测是可用的,依赖于扭曲感知的神经统计模型。我们将这些原则扩展到视频源检测,通过将空间分裂归一化与调整为伪影检测的滤波器组耦合,以增强的稀疏函数形式实现。我们称这种方法为稀疏错误捕获视频损伤检测(VIDSPECT)。我们配置VIDSPECT来创建两种常见的源视频伪影的最先进的检测器:升级和梳理。系统检测升级,识别升级类型,预测原生视频分辨率。它还可以检测由交错产生的梳理伪影。我们的方法简单,可高度概括,并且比竞争对手的方法产生更好的准确性。VIDSPECT的软件版本可在网上获得:http://live.ece.utexas.edu/research/quality/VIDSPECT release.zip供公众使用和评估。
{"title":"Detecting Source Video Artifacts with Supervised Sparse Filters","authors":"T. Goodall, A. Bovik","doi":"10.1109/PCS.2018.8456303","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456303","url":null,"abstract":"A variety of powerful picture quality predictors are available that rely on neuro-statistical models of distortion perception. We extend these principles to video source inspection, by coupling spatial divisive normalization with a filterbank tuned for artifact detection, implemented in an augmented sparse functional form. We call this method the Video Impairment Detection by SParse Error CapTure (VIDSPECT). We configure VIDSPECT to create state-of-the-art detectors of two kinds of commonly encountered source video artifacts: upscaling and combing. The system detects upscaling, identifies upscaling type, and predicts the native video resolution. It also detects combing artifacts arising from interlacing. Our approach is simple, highly generalizable, and yields better accuracy than competing methods. A software release of VIDSPECT is available online: http://live.ece.utexas.edu/research/quality/VIDSPECT release.zip for public use and evaluation.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114178203","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Adaptive Paired-Comparison Method for Subjective Video Quality Assessment on Mobile Devices 移动设备主观视频质量评价的自适应配对比较方法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456273
Katherine Storrs, S. V. Leuven, S. Kojder, Lucas Theis, Ferenc Huszár
To effectively evaluate subjective visual quality in weakly-controlled environments, we propose an Adaptive Paired Comparison method based on particle filtering. As our approach requires each sample to be rated only once, the test time compared to regular paired comparison can be reduced. The method works with non-experts and improves reliability compared to MOS and DS-MOS methods.
为了有效地评价弱控制环境下的主观视觉质量,提出了一种基于粒子滤波的自适应配对比较方法。由于我们的方法只需要对每个样本进行一次评级,因此与常规配对比较相比,可以减少测试时间。该方法适用于非专家,与MOS和DS-MOS方法相比,提高了可靠性。
{"title":"Adaptive Paired-Comparison Method for Subjective Video Quality Assessment on Mobile Devices","authors":"Katherine Storrs, S. V. Leuven, S. Kojder, Lucas Theis, Ferenc Huszár","doi":"10.1109/PCS.2018.8456273","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456273","url":null,"abstract":"To effectively evaluate subjective visual quality in weakly-controlled environments, we propose an Adaptive Paired Comparison method based on particle filtering. As our approach requires each sample to be rated only once, the test time compared to regular paired comparison can be reduced. The method works with non-experts and improves reliability compared to MOS and DS-MOS methods.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"120 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115193637","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Enhanced Homogeneous Motion Discovery Oriented Prediction for Key Intermediate Frames 面向关键中间帧的增强均匀运动发现预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456251
Ashek Ahmmed, A. Naman, D. Taubman
Conventional video compression systems use motion model to approximate the geometry of moving object boundaries. Motion model can be relieved from describing discontinuities in the underlying motion field, by employing motion hint that exploits the spatial structure of reference frames to infer appropriate boundaries for the future ones. However, estimation of highly accurate motion hint is computationally demanding, in particular for high resolution video sequences. Leveraging on the advantages of homogeneous motion discovery oriented prediction, in this paper, we propose to tune the intra-domain motion uniformity for B-frames as per the frame’s reference utility. Experimental results show an improved bit rate savings compared to the approach where no such selective tuning is enforced.
传统的视频压缩系统使用运动模型来近似运动物体边界的几何形状。运动模型可以从描述底层运动场的不连续中解脱出来,通过使用运动暗示来利用参考帧的空间结构来推断未来的适当边界。然而,高度精确的运动提示估计是计算要求很高的,特别是对于高分辨率的视频序列。利用均匀运动发现导向预测的优势,在本文中,我们建议根据帧的参考效用调整b帧的域内运动均匀性。实验结果表明,与没有强制执行这种选择性调优的方法相比,该方法可以提高比特率。
{"title":"Enhanced Homogeneous Motion Discovery Oriented Prediction for Key Intermediate Frames","authors":"Ashek Ahmmed, A. Naman, D. Taubman","doi":"10.1109/PCS.2018.8456251","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456251","url":null,"abstract":"Conventional video compression systems use motion model to approximate the geometry of moving object boundaries. Motion model can be relieved from describing discontinuities in the underlying motion field, by employing motion hint that exploits the spatial structure of reference frames to infer appropriate boundaries for the future ones. However, estimation of highly accurate motion hint is computationally demanding, in particular for high resolution video sequences. Leveraging on the advantages of homogeneous motion discovery oriented prediction, in this paper, we propose to tune the intra-domain motion uniformity for B-frames as per the frame’s reference utility. Experimental results show an improved bit rate savings compared to the approach where no such selective tuning is enforced.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116271217","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Efficient Rate-distortion Approximation and Transform Type Selection using Laplacian Operators 基于拉普拉斯算子的有效率失真近似和变换类型选择
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456313
Keng-Shih Lu, Antonio Ortega, D. Mukherjee, Yue Chen
Rate-distortion (RD) optimization is an important tool in many video compression standards and can be used for transform selection. However, this is typically very computationally demanding because a full RD search involves the computation of transform co-efficients for each candidate transform. In this paper, we propose an approach that uses sparse Laplacian operators to estimate the RD cost by computing a weighted squared sum of transform coefficients, without having to compute the actual transform coefficients. We demonstrate experimentally how our method can be applied for transform selection. Implemented in the AV1 encoder, our approach yields a significant speed-up in encoding time with a small increase in bitrate.
率失真(RD)优化是许多视频压缩标准中的一个重要工具,可用于变换选择。然而,这通常是非常需要计算的,因为一个完整的RD搜索涉及到计算每个候选变换的变换系数。在本文中,我们提出了一种使用稀疏拉普拉斯算子通过计算变换系数的加权平方和来估计RD成本的方法,而无需计算实际的变换系数。我们通过实验证明了我们的方法如何应用于变换选择。在AV1编码器中实现,我们的方法在编码时间上产生了显着的加速,比特率略有增加。
{"title":"Efficient Rate-distortion Approximation and Transform Type Selection using Laplacian Operators","authors":"Keng-Shih Lu, Antonio Ortega, D. Mukherjee, Yue Chen","doi":"10.1109/PCS.2018.8456313","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456313","url":null,"abstract":"Rate-distortion (RD) optimization is an important tool in many video compression standards and can be used for transform selection. However, this is typically very computationally demanding because a full RD search involves the computation of transform co-efficients for each candidate transform. In this paper, we propose an approach that uses sparse Laplacian operators to estimate the RD cost by computing a weighted squared sum of transform coefficients, without having to compute the actual transform coefficients. We demonstrate experimentally how our method can be applied for transform selection. Implemented in the AV1 encoder, our approach yields a significant speed-up in encoding time with a small increase in bitrate.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126180720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Effective Inter Transform Method Based on QTBT Structure for Future Video Coding 面向未来视频编码的基于QTBT结构的有效间变换方法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456266
Liqiang Wang, Benben Niu, Yun He
Transform, a crucial module for hybrid video coding framework, has been selecting Discrete Cosine Transform (DCT) for several decades. Recently, Singular Value Decomposition (SVD) and Enhanced Multiple Transform (EMT) are proposed to improve transform efficiency. However, the perspectives of SVD and EMT are different. SVD enhances transform efficiency by utilizing the similarity of prediction block and inter residual block. EMT adopts some new sinusoidal transform cores to accommodate the larger prediction errors closer to the boundary of prediction unit. In this paper, the proposed method mainly has two key contributions. First, SVD and EMT are combined skillfully. Second, non-square SVD is newly introduced to the original algorithm. By extensive experiments, averages 1.07%, 1.06% and 0.65% BD-rate saving for Y, U and V are achieved compared to JEM5.0.1 with some coding tools off, up to 5.87%, 4.28% and 4.47%.
变换是混合视频编码框架的关键模块,几十年来一直选择离散余弦变换(DCT)。近年来,为了提高变换效率,提出了奇异值分解(SVD)和增强多重变换(EMT)。然而,SVD和EMT的视角不同。奇异值分解利用预测块和残差块之间的相似性来提高变换效率。EMT采用了一些新的正弦变换核,以适应更靠近预测单元边界的较大预测误差。在本文中,所提出的方法主要有两个关键贡献。首先,将SVD和EMT巧妙地结合起来。其次,在原有算法中引入了非平方奇异值分解。通过大量的实验,与关闭部分编码工具的JEM5.0.1相比,Y、U和V的平均钻速分别节省了1.07%、1.06%和0.65%,最高可达5.87%、4.28%和4.47%。
{"title":"Effective Inter Transform Method Based on QTBT Structure for Future Video Coding","authors":"Liqiang Wang, Benben Niu, Yun He","doi":"10.1109/PCS.2018.8456266","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456266","url":null,"abstract":"Transform, a crucial module for hybrid video coding framework, has been selecting Discrete Cosine Transform (DCT) for several decades. Recently, Singular Value Decomposition (SVD) and Enhanced Multiple Transform (EMT) are proposed to improve transform efficiency. However, the perspectives of SVD and EMT are different. SVD enhances transform efficiency by utilizing the similarity of prediction block and inter residual block. EMT adopts some new sinusoidal transform cores to accommodate the larger prediction errors closer to the boundary of prediction unit. In this paper, the proposed method mainly has two key contributions. First, SVD and EMT are combined skillfully. Second, non-square SVD is newly introduced to the original algorithm. By extensive experiments, averages 1.07%, 1.06% and 0.65% BD-rate saving for Y, U and V are achieved compared to JEM5.0.1 with some coding tools off, up to 5.87%, 4.28% and 4.47%.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122156836","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1