首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Quality Assessment of Thumbnail and Billboard Images on Mobile Devices 移动设备上缩略图和广告牌图像的质量评估
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456285
Zeina Sinno, Anush K. Moorthy, J. D. Cock, Zhi Li, A. Bovik
Objective image quality assessment (IQA) research entails developing algorithms that predict human judgments of picture quality. Validating performance entails evaluating algorithms under conditions similar to where they are deployed. Hence, creating image quality databases representative of target use cases is an important endeavor. Here we present a database that relates to quality assessment of billboard images commonly displayed on mobile devices. Billboard images are a subset of thumbnail images, that extend across a display screen, representing things like album covers, banners, or frames or artwork. We conducted a subjective study of the quality of billboard images distorted by processes like compression, scaling and chroma-subsampling, and compared high-performance quality prediction models on the images and subjective data.
客观图像质量评估(IQA)研究需要开发预测人类对图像质量判断的算法。验证性能需要在与部署算法相似的条件下评估算法。因此,创建代表目标用例的图像质量数据库是一项重要的工作。在这里,我们提出了一个数据库,涉及到通常显示在移动设备上的广告牌图像的质量评估。广告牌图像是缩略图图像的一个子集,它在显示屏幕上延伸,代表诸如专辑封面、横幅、框架或艺术品之类的东西。我们对经过压缩、缩放和色度子采样等处理后的广告牌图像进行了主观质量研究,并比较了基于图像和主观数据的高性能质量预测模型。
{"title":"Quality Assessment of Thumbnail and Billboard Images on Mobile Devices","authors":"Zeina Sinno, Anush K. Moorthy, J. D. Cock, Zhi Li, A. Bovik","doi":"10.1109/PCS.2018.8456285","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456285","url":null,"abstract":"Objective image quality assessment (IQA) research entails developing algorithms that predict human judgments of picture quality. Validating performance entails evaluating algorithms under conditions similar to where they are deployed. Hence, creating image quality databases representative of target use cases is an important endeavor. Here we present a database that relates to quality assessment of billboard images commonly displayed on mobile devices. Billboard images are a subset of thumbnail images, that extend across a display screen, representing things like album covers, banners, or frames or artwork. We conducted a subjective study of the quality of billboard images distorted by processes like compression, scaling and chroma-subsampling, and compared high-performance quality prediction models on the images and subjective data.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"10 8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127088882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Perceptually-Aligned Frame Rate Selection Using Spatio-Temporal Features 基于时空特征的感知对齐帧率选择
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456274
Angeliki V. Katsenou, Di Ma, D. Bull
During recent years, the standardisation committees on video compression and broadcast formats have worked on extending practical video frame rates up to 120 frames per second. Generally, increased video frame rates have been shown to improve immersion, but at the cost of higher bit rates. Taking into consideration that the benefits of high frame rates are content dependent, a decision mechanism that recommends the appropriate frame rate for the specific content would provide benefits prior to compression and transmission. Furthermore, this decision mechanism must take account of the perceived video quality. The proposed method extracts and selects suitable spatio-temporal features and uses a supervised machine learning technique to build a model that is able to predict, with high accuracy, the lowest frame rate for which the perceived video quality is indistinguishable from that of video at the acquisition frame rate. The results show that it is a promising tool for prior to compression and delivery processing of videos, such as content-aware frame rate adaptation.
近年来,视频压缩和广播格式标准化委员会一直致力于将实际视频帧率提高到每秒120帧。一般来说,提高视频帧率已经被证明可以提高沉浸感,但代价是更高的比特率。考虑到高帧率的好处依赖于内容,为特定内容推荐适当帧率的决策机制将在压缩和传输之前提供好处。此外,这种决策机制必须考虑到感知到的视频质量。该方法提取并选择合适的时空特征,并使用监督机器学习技术构建一个模型,该模型能够高精度地预测在获取帧率下感知到的视频质量与视频质量无法区分的最低帧率。结果表明,它是一种很有前途的工具,用于视频的预先压缩和传输处理,如内容感知帧率自适应。
{"title":"Perceptually-Aligned Frame Rate Selection Using Spatio-Temporal Features","authors":"Angeliki V. Katsenou, Di Ma, D. Bull","doi":"10.1109/PCS.2018.8456274","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456274","url":null,"abstract":"During recent years, the standardisation committees on video compression and broadcast formats have worked on extending practical video frame rates up to 120 frames per second. Generally, increased video frame rates have been shown to improve immersion, but at the cost of higher bit rates. Taking into consideration that the benefits of high frame rates are content dependent, a decision mechanism that recommends the appropriate frame rate for the specific content would provide benefits prior to compression and transmission. Furthermore, this decision mechanism must take account of the perceived video quality. The proposed method extracts and selects suitable spatio-temporal features and uses a supervised machine learning technique to build a model that is able to predict, with high accuracy, the lowest frame rate for which the perceived video quality is indistinguishable from that of video at the acquisition frame rate. The results show that it is a promising tool for prior to compression and delivery processing of videos, such as content-aware frame rate adaptation.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114304125","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Temporal Adaptive Quantization using Accurate Estimations of Inter and Skip Probabilities 使用精确估计间隔和跳过概率的时间自适应量化
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456275
Maxime Bichon, J. L. Tanou, M. Ropert, W. Hamidouche, L. Morin, Lu Zhang
Hybrid video coding systems use spatial and temporal predictions in order to remove redundancies within the video source signal. These predictions create coding-scheme-related dependencies, often neglected for sake of simplicity. The R-D Spatio-Temporal Adaptive Quantization (RDSTQ) solution uses such dependencies to achieve better coding efficiency. It models the temporal distortion propagation by estimating the probability of a Coding Unit (CU) to be Inter coded. Uased on this probability, each CU is given a weight depending on its relative importance compared to other CUs. However, the initial approach roughly estimates the Inter probability and does not take into account the Skip mode characteristics in the propagation. It induces important Target uitrate Deviation (TBD) compared to the reference target rate. This paper provides undeniable improvements of the original RDSTQ model in using a more accurate estimation of the Inter probability. Then a new analytical solution for local quantizers is obtained by introducing the Skip probability of a CU into the temporal distortion propagation model. The proposed solution brings −2.05% BD-BR gain in average over the RDSTQ at low rate, which corresponds to −13.54% BD-BR gain in average against no local quantization. Moreover, the TBD is reduced from 38% to 14%.
混合视频编码系统使用空间和时间预测来消除视频源信号中的冗余。这些预测创建了与编码方案相关的依赖项,为了简单起见,这些依赖项经常被忽略。R-D时空自适应量化(RDSTQ)解决方案利用这种依赖关系来实现更好的编码效率。它通过估计编码单元(CU)被互编码的概率来建立时间失真传播模型。根据这个概率,每个CU根据其相对于其他CU的重要性被赋予权重。然而,最初的方法粗略地估计了Inter概率,而没有考虑传播中的跳过模式特征。与参考目标速率相比,它会产生重要的目标速率偏差(TBD)。本文在使用更准确的Inter概率估计方面对原始RDSTQ模型进行了不可否认的改进。然后,在时域失真传播模型中引入CU的跳过概率,得到了局部量化的解析解。该方案在低速率下比RDSTQ平均获得−2.05%的BD-BR增益,在不进行局部量化的情况下平均获得−13.54%的BD-BR增益。此外,TBD从38%降至14%。
{"title":"Temporal Adaptive Quantization using Accurate Estimations of Inter and Skip Probabilities","authors":"Maxime Bichon, J. L. Tanou, M. Ropert, W. Hamidouche, L. Morin, Lu Zhang","doi":"10.1109/PCS.2018.8456275","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456275","url":null,"abstract":"Hybrid video coding systems use spatial and temporal predictions in order to remove redundancies within the video source signal. These predictions create coding-scheme-related dependencies, often neglected for sake of simplicity. The R-D Spatio-Temporal Adaptive Quantization (RDSTQ) solution uses such dependencies to achieve better coding efficiency. It models the temporal distortion propagation by estimating the probability of a Coding Unit (CU) to be Inter coded. Uased on this probability, each CU is given a weight depending on its relative importance compared to other CUs. However, the initial approach roughly estimates the Inter probability and does not take into account the Skip mode characteristics in the propagation. It induces important Target uitrate Deviation (TBD) compared to the reference target rate. This paper provides undeniable improvements of the original RDSTQ model in using a more accurate estimation of the Inter probability. Then a new analytical solution for local quantizers is obtained by introducing the Skip probability of a CU into the temporal distortion propagation model. The proposed solution brings −2.05% BD-BR gain in average over the RDSTQ at low rate, which corresponds to −13.54% BD-BR gain in average against no local quantization. Moreover, the TBD is reduced from 38% to 14%.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122234098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Single Layer Progressive Coding for High Dynamic Range Videos 高动态范围视频的单层渐进编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456314
H. Kadu, Qing Song, Guan-Ming Su
There are different kinds of high dynamic range (HDR) displays in the market today. These displays have different HDR specifications, like, pealddark brightness levels, electro-optical transfer functions (EOTF), color spaces etc. For the best visual experience on a given HDR screen, colorists have to grade videos for that specific display’s luminance range. Uut simultaneous transmission of multiple video bitstreams graded at different luminance ranges, is inefficient in terms of network utility and server storage. To overcome this problem, we propose transmitting our progressive metadata with a base layer video bitstream. This embedding allows different overlapping portions of metadata to scale the base video to progressively wider luminance ranges. Our progressive metadata format provides a significant design improvement over the existing architectures, preserves colorist intent at all the supported brightness ranges and still keeps the bandwidth or storage overhead minimal.
目前市场上有不同种类的高动态范围(HDR)显示器。这些显示器具有不同的HDR规格,例如,pealddark亮度级别,光电传递函数(etf),色彩空间等。为了在给定的HDR屏幕上获得最佳的视觉体验,色彩师必须根据特定显示器的亮度范围对视频进行分级。但是,同时传输在不同亮度范围内分级的多个视频比特流,在网络利用率和服务器存储方面效率低下。为了克服这个问题,我们提出用基础层视频比特流传输渐进式元数据。这种嵌入允许元数据的不同重叠部分将基本视频扩展到逐渐扩大的亮度范围。我们的渐进式元数据格式在现有架构上提供了重大的设计改进,在所有支持的亮度范围内保留色彩器意图,并且仍然保持最小的带宽或存储开销。
{"title":"Single Layer Progressive Coding for High Dynamic Range Videos","authors":"H. Kadu, Qing Song, Guan-Ming Su","doi":"10.1109/PCS.2018.8456314","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456314","url":null,"abstract":"There are different kinds of high dynamic range (HDR) displays in the market today. These displays have different HDR specifications, like, pealddark brightness levels, electro-optical transfer functions (EOTF), color spaces etc. For the best visual experience on a given HDR screen, colorists have to grade videos for that specific display’s luminance range. Uut simultaneous transmission of multiple video bitstreams graded at different luminance ranges, is inefficient in terms of network utility and server storage. To overcome this problem, we propose transmitting our progressive metadata with a base layer video bitstream. This embedding allows different overlapping portions of metadata to scale the base video to progressively wider luminance ranges. Our progressive metadata format provides a significant design improvement over the existing architectures, preserves colorist intent at all the supported brightness ranges and still keeps the bandwidth or storage overhead minimal.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123176520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Region-Wise Super-Resolution Algorithm Based On the Viewpoint Distribution 基于视点分布的区域超分辨算法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456295
Kazunori Uruma, Shunsuke Takasu, Keiko Masuda, S. Hangai
Recently, super-resolution techniques have been energetically studied for the purpose of reusing the low resolution image contents. Although a lot of approaches to achieve the appropriate super-resolution have been proposed such as non-linear filtering, total variation regularization, deep learning etc., the characteristic of the viewpoint distribution of the observer has not been effectively utilized. Because applying super-resolution to unimportant regions in an image may hinder the observer’s attention to seeing the display, it leads to a low subjective evaluation. This paper proposes the region-wise super-resolution algorithm based on the view-point distribution of observer. However, we cannot obtain the viewpoint distribution map for an image without the pre-experiment using the device such as eye mark recorder, therefore, the saliency map is utilized in this paper. Numerical examples show that the proposed algorithm using saliency map achieves a higher subjective evaluation than the previous study based on the non-linear filtering based super-resolution. Furthermore, in numerical examples, the proposed algorithm using the saliency map is shown to give the similar results of the algorithm using the viewpoint distribution map obtained by the pre-experiment using eye mark recorder.
近年来,为了实现低分辨率图像内容的重复利用,超分辨率技术得到了大力的研究。尽管人们提出了非线性滤波、全变分正则化、深度学习等方法来达到适当的超分辨率,但并没有有效地利用观测器视点分布的特性。由于将超分辨率应用于图像中不重要的区域可能会阻碍观察者的注意力,从而导致主观评价较低。提出了一种基于观测器视点分布的区域超分辨算法。然而,如果没有使用眼痕记录仪等设备进行预实验,我们无法获得图像的视点分布图,因此本文采用显著性图。数值算例表明,与基于非线性滤波的超分辨率算法相比,基于显著性映射的算法获得了更高的主观评价。在数值算例中,利用显著性图的算法与利用眼痕记录仪预实验所得的视点分布图算法的结果相似。
{"title":"Region-Wise Super-Resolution Algorithm Based On the Viewpoint Distribution","authors":"Kazunori Uruma, Shunsuke Takasu, Keiko Masuda, S. Hangai","doi":"10.1109/PCS.2018.8456295","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456295","url":null,"abstract":"Recently, super-resolution techniques have been energetically studied for the purpose of reusing the low resolution image contents. Although a lot of approaches to achieve the appropriate super-resolution have been proposed such as non-linear filtering, total variation regularization, deep learning etc., the characteristic of the viewpoint distribution of the observer has not been effectively utilized. Because applying super-resolution to unimportant regions in an image may hinder the observer’s attention to seeing the display, it leads to a low subjective evaluation. This paper proposes the region-wise super-resolution algorithm based on the view-point distribution of observer. However, we cannot obtain the viewpoint distribution map for an image without the pre-experiment using the device such as eye mark recorder, therefore, the saliency map is utilized in this paper. Numerical examples show that the proposed algorithm using saliency map achieves a higher subjective evaluation than the previous study based on the non-linear filtering based super-resolution. Furthermore, in numerical examples, the proposed algorithm using the saliency map is shown to give the similar results of the algorithm using the viewpoint distribution map obtained by the pre-experiment using eye mark recorder.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123436034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PCS 2018 Copyright Page PCS 2018版权页面
Pub Date : 2018-06-01 DOI: 10.1109/pcs.2018.8456294
{"title":"PCS 2018 Copyright Page","authors":"","doi":"10.1109/pcs.2018.8456294","DOIUrl":"https://doi.org/10.1109/pcs.2018.8456294","url":null,"abstract":"","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116193735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Multiple Feature-based Classifications Adaptive Loop Filter 基于多特征的分类自适应环路滤波器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456264
Johannes Erfurt, Wang-Q Lim, H. Schwarz, D. Marpe, T. Wiegand
In video coding, adaptive loop filter (ALF) has attracted attention due to its increasing coding performances. Recently ALF has been further developed for its extension, which introduces geometry transformation-based adaptive loop filter (GALF) outperforming the existing ALF techniques. The main idea of ALF is to apply a classification to obtain multiple classes, which gives a partition of a set of all pixel locations. After that, a Wiener filter is applied for each class. Therefore, the performance of ALF essentially relies on how its classification behaves. In this paper, we introduce a novel classification method, Multiple feature-based Classifications ALF (MCALF) extending a classification in GALF and show that it increases coding efficiency while only marginally raising encoding complexity. The key idea is to apply more than one classifier at the encoder to group all reconstructed samples and then to select a classifier with the best RD-performance to carry out the classification process. Simulation results show that around 2% bit rate reduction can be achieved on top of GALF for some selected test sequences.
在视频编码中,自适应环路滤波器(ALF)因其不断提高的编码性能而备受关注。近年来,基于几何变换的自适应环路滤波器(GALF)在原有滤波技术的基础上得到了进一步的发展。ALF的主要思想是应用一个分类来获得多个类,它给出了一组所有像素位置的分区。之后,对每个类应用维纳过滤器。因此,ALF的性能本质上依赖于它的分类行为。本文提出了一种新的分类方法——基于多特征的分类ALF (Multiple feature-based Classifications ALF, MCALF),该方法对基于多特征的分类ALF进行了扩展,在提高编码效率的同时,只略微提高了编码复杂度。关键思想是在编码器上应用多个分类器对所有重构样本进行分组,然后选择一个rd性能最好的分类器进行分类。仿真结果表明,对于一些选定的测试序列,在GALF的基础上可以实现约2%的比特率降低。
{"title":"Multiple Feature-based Classifications Adaptive Loop Filter","authors":"Johannes Erfurt, Wang-Q Lim, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS.2018.8456264","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456264","url":null,"abstract":"In video coding, adaptive loop filter (ALF) has attracted attention due to its increasing coding performances. Recently ALF has been further developed for its extension, which introduces geometry transformation-based adaptive loop filter (GALF) outperforming the existing ALF techniques. The main idea of ALF is to apply a classification to obtain multiple classes, which gives a partition of a set of all pixel locations. After that, a Wiener filter is applied for each class. Therefore, the performance of ALF essentially relies on how its classification behaves. In this paper, we introduce a novel classification method, Multiple feature-based Classifications ALF (MCALF) extending a classification in GALF and show that it increases coding efficiency while only marginally raising encoding complexity. The key idea is to apply more than one classifier at the encoder to group all reconstructed samples and then to select a classifier with the best RD-performance to carry out the classification process. Simulation results show that around 2% bit rate reduction can be achieved on top of GALF for some selected test sequences.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121032200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Rotational Motion Compensated Prediction in HEVC Based Omnidirectional Video Coding 基于HEVC的全向视频编码旋转运动补偿预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456296
B. Vishwanath, K. Rose, Yuwen He, Yan Ye
Spherical video is becoming prevalent in virtual and augmented reality applications. With the increased field of view, spherical video needs enormous amounts of data, obviously demanding efficient compression. Existing approaches simply project the spherical content onto a plane to facilitate the use of standard video coders. Earlier work at UCSB was motivated by the realization that existing approaches are suboptimal due to warping introduced by the projection, yielding complex non-linear motion that is not captured by the simple translational motion model employed in standard coders. Moreover, motion vectors in the projected domain do not offer a physically meaningful model. The proposed remedy was to capture the motion directly on the sphere with a rotational motion model, in terms of sphere rotations along geodesics. The rotational motion model preserves the shape and size of objects on the sphere. This paper implements and tests the main ideas from the previous work [1] in the context of a full-fledged, unconstrained coder including, in particular, bi-prediction, multiple reference frames and motion vector refinement. Experimental results provide evidence for considerable gains over HEVC.
球形视频在虚拟和增强现实应用中越来越流行。随着视野的扩大,球形视频需要大量的数据,显然需要高效的压缩。现有的方法只是将球形内容投影到平面上,以方便标准视频编码器的使用。UCSB的早期工作是由于认识到现有的方法是次优的,因为投影引入了扭曲,产生了复杂的非线性运动,而标准编码器中使用的简单平移运动模型无法捕获这些运动。此外,投影域中的运动向量不能提供物理上有意义的模型。提出的补救办法是用球体沿测地线旋转的旋转运动模型直接捕捉球体上的运动。旋转运动模型保留了球体上物体的形状和大小。本文在一个成熟的、无约束的编码器的背景下实现并测试了先前工作[1]的主要思想,特别是包括双预测、多参考帧和运动矢量细化。实验结果证明了HEVC的显著增益。
{"title":"Rotational Motion Compensated Prediction in HEVC Based Omnidirectional Video Coding","authors":"B. Vishwanath, K. Rose, Yuwen He, Yan Ye","doi":"10.1109/PCS.2018.8456296","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456296","url":null,"abstract":"Spherical video is becoming prevalent in virtual and augmented reality applications. With the increased field of view, spherical video needs enormous amounts of data, obviously demanding efficient compression. Existing approaches simply project the spherical content onto a plane to facilitate the use of standard video coders. Earlier work at UCSB was motivated by the realization that existing approaches are suboptimal due to warping introduced by the projection, yielding complex non-linear motion that is not captured by the simple translational motion model employed in standard coders. Moreover, motion vectors in the projected domain do not offer a physically meaningful model. The proposed remedy was to capture the motion directly on the sphere with a rotational motion model, in terms of sphere rotations along geodesics. The rotational motion model preserves the shape and size of objects on the sphere. This paper implements and tests the main ideas from the previous work [1] in the context of a full-fledged, unconstrained coder including, in particular, bi-prediction, multiple reference frames and motion vector refinement. Experimental results provide evidence for considerable gains over HEVC.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121571741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Wavefront Parallel Processing for AV1 Encoder AV1编码器的波前并行处理
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456283
Yikai Zhao, Jiangtao Wen
The emerging AV1 coding standard brings even higher computational complexity than current coding standards, but does not support traditional Wavefront Parallel Processing (WPP) approach due to the lacking of syntax support. In this paper we introduced a novel framework to implement WPP for AV1 encoder that is compatible with current decoder without additional bitstream syntax support, where mode selection is processed in wavefront parallel before entropy encoding and entropy contexts for rate-distortion optimization are predicted. Based on this framework, context prediction algorithms that use same data dependency model as previous works in H.264 and HEVC are implemented. Furthermore, we proposed an optimal context prediction algorithm specifically for AV1. Experimental results showed that our framework with proposed optimal algorithm yields good parallelism and scalability (over 10x speed-up with 16 threads for 4k sequences) with little coding performance loss (less than 0.2% bitrate increasing).
新兴的AV1编码标准带来了比现有编码标准更高的计算复杂度,但由于缺乏语法支持而不支持传统的波前并行处理(WPP)方法。在本文中,我们引入了一种新的框架来实现AV1编码器的WPP,该框架与当前的解码器兼容,无需额外的比特流语法支持,其中模式选择在波前并行处理,然后预测熵编码和熵上下文用于率失真优化。基于该框架,实现了与H.264和HEVC中使用相同数据依赖模型的上下文预测算法。此外,我们提出了一种针对AV1的最佳上下文预测算法。实验结果表明,我们的框架具有良好的并行性和可扩展性(4k序列16线程加速超过10倍),编码性能损失很小(比特率增加不到0.2%)。
{"title":"Wavefront Parallel Processing for AV1 Encoder","authors":"Yikai Zhao, Jiangtao Wen","doi":"10.1109/PCS.2018.8456283","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456283","url":null,"abstract":"The emerging AV1 coding standard brings even higher computational complexity than current coding standards, but does not support traditional Wavefront Parallel Processing (WPP) approach due to the lacking of syntax support. In this paper we introduced a novel framework to implement WPP for AV1 encoder that is compatible with current decoder without additional bitstream syntax support, where mode selection is processed in wavefront parallel before entropy encoding and entropy contexts for rate-distortion optimization are predicted. Based on this framework, context prediction algorithms that use same data dependency model as previous works in H.264 and HEVC are implemented. Furthermore, we proposed an optimal context prediction algorithm specifically for AV1. Experimental results showed that our framework with proposed optimal algorithm yields good parallelism and scalability (over 10x speed-up with 16 threads for 4k sequences) with little coding performance loss (less than 0.2% bitrate increasing).","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"40 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133163323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Compression Performance Comparison of x264, x265, libvpx and aomenc for On-Demand Adaptive Streaming Applications x264, x265, libvpx和aomenc在按需自适应流媒体应用中的压缩性能比较
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456302
Liwei Guo, J. D. Cock, A. Aaron
Video compression standard H.264/AVC was released in 2003, and has been dominating the industry for the past decade. Over the last few years, a number of next-generation standards/formats like VP9 (2012), H.265/HEVC (2013) and AV1 (2018) were introduced, all claiming significant improvement over H.264/AVC. In this paper, we present our evaluation of the performance of these compression standards. Our evaluation is conducted using open-source encoder implementations of these standards, x264 (for H.264/AVC), x265 (for H.265/HEVC), libvpx (for VP9) and aomenc (for AV1). The process is designed to evaluate the attainable compression efficiency for on-demand adaptive streaming applications. Results with two different quality metrics, PSNR and VMAF, are reported. Our results reveal that x265, libvpx and aomenc all achieve substantial compression efficiency improvement over x264.
视频压缩标准H.264/AVC于2003年发布,在过去的十年里一直主导着这个行业。在过去的几年中,推出了许多下一代标准/格式,如VP9 (2012), H.265/HEVC(2013)和AV1(2018),都声称比H.264/AVC有了重大改进。在本文中,我们提出了我们的性能评估这些压缩标准。我们的评估使用这些标准的开源编码器实现进行,x264(用于H.264/AVC), x265(用于H.265/HEVC), libvpx(用于VP9)和aomenc(用于AV1)。该过程旨在评估按需自适应流媒体应用程序可实现的压缩效率。报告了两种不同质量指标PSNR和VMAF的结果。我们的结果表明,x265、libvpx和aomenc都比x264实现了显著的压缩效率提高。
{"title":"Compression Performance Comparison of x264, x265, libvpx and aomenc for On-Demand Adaptive Streaming Applications","authors":"Liwei Guo, J. D. Cock, A. Aaron","doi":"10.1109/PCS.2018.8456302","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456302","url":null,"abstract":"Video compression standard H.264/AVC was released in 2003, and has been dominating the industry for the past decade. Over the last few years, a number of next-generation standards/formats like VP9 (2012), H.265/HEVC (2013) and AV1 (2018) were introduced, all claiming significant improvement over H.264/AVC. In this paper, we present our evaluation of the performance of these compression standards. Our evaluation is conducted using open-source encoder implementations of these standards, x264 (for H.264/AVC), x265 (for H.265/HEVC), libvpx (for VP9) and aomenc (for AV1). The process is designed to evaluate the attainable compression efficiency for on-demand adaptive streaming applications. Results with two different quality metrics, PSNR and VMAF, are reported. Our results reveal that x265, libvpx and aomenc all achieve substantial compression efficiency improvement over x264.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127613113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1