首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Benchmarking of Objective Quality Metrics for Colorless Point Clouds 无色点云客观质量指标的基准测试
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456252
E. Alexiou, T. Ebrahimi
Recent advances in depth sensing and display technologies, along with the significant growth of interest for augmented and virtual reality applications, lay the foundation for the rapid evolution of applications that provide immersive experiences. In such applications, advanced content representations are required in order to increase the engagement of the user with the displayed imageries. Point clouds have emerged as a promising solution to this aim, due to their efficiency in capturing, storing, delivering and rendering of 3D immersive contents. As in any type of imaging, the evaluation of point clouds in terms of visual quality is essential. In this paper, benchmarking results of the state-of-the-art objective metrics in geometry-only point clouds are reported and analyzed under two different types of geometry degradations, namely Gaussian noise and octree- based compression. Human ratings obtained from two subjective experiments are used as the ground truth. Our results show that most objective quality metrics perform well in the presence of noise, whereas one particular method has high predictive power and outperforms the others after octree-based encoding.
深度传感和显示技术的最新进展,以及对增强现实和虚拟现实应用的兴趣的显着增长,为提供沉浸式体验的应用的快速发展奠定了基础。在这样的应用程序中,为了增加用户对所显示图像的参与,需要高级内容表示。由于点云在捕获、存储、传输和渲染3D沉浸式内容方面的效率,它已经成为实现这一目标的一个有希望的解决方案。与任何类型的成像一样,从视觉质量方面对点云进行评估是必不可少的。本文报道并分析了纯几何点云在高斯噪声和基于八叉树的压缩两种不同几何退化类型下最先进的客观指标的基准测试结果。从两个主观实验中获得的人类评分被用作基础事实。我们的研究结果表明,大多数客观质量指标在存在噪声的情况下表现良好,而一种特定的方法具有很高的预测能力,并且在基于八叉树的编码后优于其他方法。
{"title":"Benchmarking of Objective Quality Metrics for Colorless Point Clouds","authors":"E. Alexiou, T. Ebrahimi","doi":"10.1109/PCS.2018.8456252","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456252","url":null,"abstract":"Recent advances in depth sensing and display technologies, along with the significant growth of interest for augmented and virtual reality applications, lay the foundation for the rapid evolution of applications that provide immersive experiences. In such applications, advanced content representations are required in order to increase the engagement of the user with the displayed imageries. Point clouds have emerged as a promising solution to this aim, due to their efficiency in capturing, storing, delivering and rendering of 3D immersive contents. As in any type of imaging, the evaluation of point clouds in terms of visual quality is essential. In this paper, benchmarking results of the state-of-the-art objective metrics in geometry-only point clouds are reported and analyzed under two different types of geometry degradations, namely Gaussian noise and octree- based compression. Human ratings obtained from two subjective experiments are used as the ground truth. Our results show that most objective quality metrics perform well in the presence of noise, whereas one particular method has high predictive power and outperforms the others after octree-based encoding.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"41 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120918293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Analysis and Prediction of JND-Based Video Quality Model 基于jnd的视频质量模型分析与预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456243
Haiqiang Wang, Xinfeng Zhang, Chao Yang, C.-C. Jay Kuo
The just-noticeable-difference (JND) visual perception property has received much attention in characterizing human subjective viewing experience of compressed video. In this work, we quantity the JND-based video quality assessment model using the satisfied user ratio (SUR) curve, and show that the SUR model can be greatly simplified since the JND points of multiple subjects for the same content in the VideoSet can be well modeled by the normal distribution. Then, we design an SUR prediction method with video quality degradation features and masking features and use them to predict the first, second and the third JND points and their corresponding SUR curves. Finally, we verify the performance of the proposed SUR prediction method with different configurations on the VideoSet. The experimental results demonstrate that the proposed SUR prediction method achieves good performance in various resolutions with the mean absolute error (MAE) of the SUR smaller than 0.05 on average.
just- visible -difference (JND)视觉感知特性在描述压缩视频的主观观看体验方面受到了广泛关注。在这项工作中,我们使用满意用户比例(SUR)曲线对基于JND的视频质量评估模型进行了量化,并表明由于VideoSet中相同内容的多个主题的JND点可以通过正态分布很好地建模,因此SUR模型可以大大简化。然后,我们设计了一种包含视频质量退化特征和掩蔽特征的SUR预测方法,并利用它们预测第一、第二和第三个JND点及其对应的SUR曲线。最后,在VideoSet上验证了不同配置下所提出的SUR预测方法的性能。实验结果表明,所提出的SUR预测方法在不同分辨率下均取得了较好的预测效果,平均绝对误差(MAE)均小于0.05。
{"title":"Analysis and Prediction of JND-Based Video Quality Model","authors":"Haiqiang Wang, Xinfeng Zhang, Chao Yang, C.-C. Jay Kuo","doi":"10.1109/PCS.2018.8456243","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456243","url":null,"abstract":"The just-noticeable-difference (JND) visual perception property has received much attention in characterizing human subjective viewing experience of compressed video. In this work, we quantity the JND-based video quality assessment model using the satisfied user ratio (SUR) curve, and show that the SUR model can be greatly simplified since the JND points of multiple subjects for the same content in the VideoSet can be well modeled by the normal distribution. Then, we design an SUR prediction method with video quality degradation features and masking features and use them to predict the first, second and the third JND points and their corresponding SUR curves. Finally, we verify the performance of the proposed SUR prediction method with different configurations on the VideoSet. The experimental results demonstrate that the proposed SUR prediction method achieves good performance in various resolutions with the mean absolute error (MAE) of the SUR smaller than 0.05 on average.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127684293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Rotational Motion Compensated Prediction in HEVC Based Omnidirectional Video Coding 基于HEVC的全向视频编码旋转运动补偿预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456296
B. Vishwanath, K. Rose, Yuwen He, Yan Ye
Spherical video is becoming prevalent in virtual and augmented reality applications. With the increased field of view, spherical video needs enormous amounts of data, obviously demanding efficient compression. Existing approaches simply project the spherical content onto a plane to facilitate the use of standard video coders. Earlier work at UCSB was motivated by the realization that existing approaches are suboptimal due to warping introduced by the projection, yielding complex non-linear motion that is not captured by the simple translational motion model employed in standard coders. Moreover, motion vectors in the projected domain do not offer a physically meaningful model. The proposed remedy was to capture the motion directly on the sphere with a rotational motion model, in terms of sphere rotations along geodesics. The rotational motion model preserves the shape and size of objects on the sphere. This paper implements and tests the main ideas from the previous work [1] in the context of a full-fledged, unconstrained coder including, in particular, bi-prediction, multiple reference frames and motion vector refinement. Experimental results provide evidence for considerable gains over HEVC.
球形视频在虚拟和增强现实应用中越来越流行。随着视野的扩大,球形视频需要大量的数据,显然需要高效的压缩。现有的方法只是将球形内容投影到平面上,以方便标准视频编码器的使用。UCSB的早期工作是由于认识到现有的方法是次优的,因为投影引入了扭曲,产生了复杂的非线性运动,而标准编码器中使用的简单平移运动模型无法捕获这些运动。此外,投影域中的运动向量不能提供物理上有意义的模型。提出的补救办法是用球体沿测地线旋转的旋转运动模型直接捕捉球体上的运动。旋转运动模型保留了球体上物体的形状和大小。本文在一个成熟的、无约束的编码器的背景下实现并测试了先前工作[1]的主要思想,特别是包括双预测、多参考帧和运动矢量细化。实验结果证明了HEVC的显著增益。
{"title":"Rotational Motion Compensated Prediction in HEVC Based Omnidirectional Video Coding","authors":"B. Vishwanath, K. Rose, Yuwen He, Yan Ye","doi":"10.1109/PCS.2018.8456296","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456296","url":null,"abstract":"Spherical video is becoming prevalent in virtual and augmented reality applications. With the increased field of view, spherical video needs enormous amounts of data, obviously demanding efficient compression. Existing approaches simply project the spherical content onto a plane to facilitate the use of standard video coders. Earlier work at UCSB was motivated by the realization that existing approaches are suboptimal due to warping introduced by the projection, yielding complex non-linear motion that is not captured by the simple translational motion model employed in standard coders. Moreover, motion vectors in the projected domain do not offer a physically meaningful model. The proposed remedy was to capture the motion directly on the sphere with a rotational motion model, in terms of sphere rotations along geodesics. The rotational motion model preserves the shape and size of objects on the sphere. This paper implements and tests the main ideas from the previous work [1] in the context of a full-fledged, unconstrained coder including, in particular, bi-prediction, multiple reference frames and motion vector refinement. Experimental results provide evidence for considerable gains over HEVC.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121571741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Multiple Feature-based Classifications Adaptive Loop Filter 基于多特征的分类自适应环路滤波器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456264
Johannes Erfurt, Wang-Q Lim, H. Schwarz, D. Marpe, T. Wiegand
In video coding, adaptive loop filter (ALF) has attracted attention due to its increasing coding performances. Recently ALF has been further developed for its extension, which introduces geometry transformation-based adaptive loop filter (GALF) outperforming the existing ALF techniques. The main idea of ALF is to apply a classification to obtain multiple classes, which gives a partition of a set of all pixel locations. After that, a Wiener filter is applied for each class. Therefore, the performance of ALF essentially relies on how its classification behaves. In this paper, we introduce a novel classification method, Multiple feature-based Classifications ALF (MCALF) extending a classification in GALF and show that it increases coding efficiency while only marginally raising encoding complexity. The key idea is to apply more than one classifier at the encoder to group all reconstructed samples and then to select a classifier with the best RD-performance to carry out the classification process. Simulation results show that around 2% bit rate reduction can be achieved on top of GALF for some selected test sequences.
在视频编码中,自适应环路滤波器(ALF)因其不断提高的编码性能而备受关注。近年来,基于几何变换的自适应环路滤波器(GALF)在原有滤波技术的基础上得到了进一步的发展。ALF的主要思想是应用一个分类来获得多个类,它给出了一组所有像素位置的分区。之后,对每个类应用维纳过滤器。因此,ALF的性能本质上依赖于它的分类行为。本文提出了一种新的分类方法——基于多特征的分类ALF (Multiple feature-based Classifications ALF, MCALF),该方法对基于多特征的分类ALF进行了扩展,在提高编码效率的同时,只略微提高了编码复杂度。关键思想是在编码器上应用多个分类器对所有重构样本进行分组,然后选择一个rd性能最好的分类器进行分类。仿真结果表明,对于一些选定的测试序列,在GALF的基础上可以实现约2%的比特率降低。
{"title":"Multiple Feature-based Classifications Adaptive Loop Filter","authors":"Johannes Erfurt, Wang-Q Lim, H. Schwarz, D. Marpe, T. Wiegand","doi":"10.1109/PCS.2018.8456264","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456264","url":null,"abstract":"In video coding, adaptive loop filter (ALF) has attracted attention due to its increasing coding performances. Recently ALF has been further developed for its extension, which introduces geometry transformation-based adaptive loop filter (GALF) outperforming the existing ALF techniques. The main idea of ALF is to apply a classification to obtain multiple classes, which gives a partition of a set of all pixel locations. After that, a Wiener filter is applied for each class. Therefore, the performance of ALF essentially relies on how its classification behaves. In this paper, we introduce a novel classification method, Multiple feature-based Classifications ALF (MCALF) extending a classification in GALF and show that it increases coding efficiency while only marginally raising encoding complexity. The key idea is to apply more than one classifier at the encoder to group all reconstructed samples and then to select a classifier with the best RD-performance to carry out the classification process. Simulation results show that around 2% bit rate reduction can be achieved on top of GALF for some selected test sequences.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121032200","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
PCS 2018 Copyright Page PCS 2018版权页面
Pub Date : 2018-06-01 DOI: 10.1109/pcs.2018.8456294
{"title":"PCS 2018 Copyright Page","authors":"","doi":"10.1109/pcs.2018.8456294","DOIUrl":"https://doi.org/10.1109/pcs.2018.8456294","url":null,"abstract":"","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116193735","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Temporal Adaptive Quantization using Accurate Estimations of Inter and Skip Probabilities 使用精确估计间隔和跳过概率的时间自适应量化
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456275
Maxime Bichon, J. L. Tanou, M. Ropert, W. Hamidouche, L. Morin, Lu Zhang
Hybrid video coding systems use spatial and temporal predictions in order to remove redundancies within the video source signal. These predictions create coding-scheme-related dependencies, often neglected for sake of simplicity. The R-D Spatio-Temporal Adaptive Quantization (RDSTQ) solution uses such dependencies to achieve better coding efficiency. It models the temporal distortion propagation by estimating the probability of a Coding Unit (CU) to be Inter coded. Uased on this probability, each CU is given a weight depending on its relative importance compared to other CUs. However, the initial approach roughly estimates the Inter probability and does not take into account the Skip mode characteristics in the propagation. It induces important Target uitrate Deviation (TBD) compared to the reference target rate. This paper provides undeniable improvements of the original RDSTQ model in using a more accurate estimation of the Inter probability. Then a new analytical solution for local quantizers is obtained by introducing the Skip probability of a CU into the temporal distortion propagation model. The proposed solution brings −2.05% BD-BR gain in average over the RDSTQ at low rate, which corresponds to −13.54% BD-BR gain in average against no local quantization. Moreover, the TBD is reduced from 38% to 14%.
混合视频编码系统使用空间和时间预测来消除视频源信号中的冗余。这些预测创建了与编码方案相关的依赖项,为了简单起见,这些依赖项经常被忽略。R-D时空自适应量化(RDSTQ)解决方案利用这种依赖关系来实现更好的编码效率。它通过估计编码单元(CU)被互编码的概率来建立时间失真传播模型。根据这个概率,每个CU根据其相对于其他CU的重要性被赋予权重。然而,最初的方法粗略地估计了Inter概率,而没有考虑传播中的跳过模式特征。与参考目标速率相比,它会产生重要的目标速率偏差(TBD)。本文在使用更准确的Inter概率估计方面对原始RDSTQ模型进行了不可否认的改进。然后,在时域失真传播模型中引入CU的跳过概率,得到了局部量化的解析解。该方案在低速率下比RDSTQ平均获得−2.05%的BD-BR增益,在不进行局部量化的情况下平均获得−13.54%的BD-BR增益。此外,TBD从38%降至14%。
{"title":"Temporal Adaptive Quantization using Accurate Estimations of Inter and Skip Probabilities","authors":"Maxime Bichon, J. L. Tanou, M. Ropert, W. Hamidouche, L. Morin, Lu Zhang","doi":"10.1109/PCS.2018.8456275","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456275","url":null,"abstract":"Hybrid video coding systems use spatial and temporal predictions in order to remove redundancies within the video source signal. These predictions create coding-scheme-related dependencies, often neglected for sake of simplicity. The R-D Spatio-Temporal Adaptive Quantization (RDSTQ) solution uses such dependencies to achieve better coding efficiency. It models the temporal distortion propagation by estimating the probability of a Coding Unit (CU) to be Inter coded. Uased on this probability, each CU is given a weight depending on its relative importance compared to other CUs. However, the initial approach roughly estimates the Inter probability and does not take into account the Skip mode characteristics in the propagation. It induces important Target uitrate Deviation (TBD) compared to the reference target rate. This paper provides undeniable improvements of the original RDSTQ model in using a more accurate estimation of the Inter probability. Then a new analytical solution for local quantizers is obtained by introducing the Skip probability of a CU into the temporal distortion propagation model. The proposed solution brings −2.05% BD-BR gain in average over the RDSTQ at low rate, which corresponds to −13.54% BD-BR gain in average against no local quantization. Moreover, the TBD is reduced from 38% to 14%.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122234098","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Single Layer Progressive Coding for High Dynamic Range Videos 高动态范围视频的单层渐进编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456314
H. Kadu, Qing Song, Guan-Ming Su
There are different kinds of high dynamic range (HDR) displays in the market today. These displays have different HDR specifications, like, pealddark brightness levels, electro-optical transfer functions (EOTF), color spaces etc. For the best visual experience on a given HDR screen, colorists have to grade videos for that specific display’s luminance range. Uut simultaneous transmission of multiple video bitstreams graded at different luminance ranges, is inefficient in terms of network utility and server storage. To overcome this problem, we propose transmitting our progressive metadata with a base layer video bitstream. This embedding allows different overlapping portions of metadata to scale the base video to progressively wider luminance ranges. Our progressive metadata format provides a significant design improvement over the existing architectures, preserves colorist intent at all the supported brightness ranges and still keeps the bandwidth or storage overhead minimal.
目前市场上有不同种类的高动态范围(HDR)显示器。这些显示器具有不同的HDR规格,例如,pealddark亮度级别,光电传递函数(etf),色彩空间等。为了在给定的HDR屏幕上获得最佳的视觉体验,色彩师必须根据特定显示器的亮度范围对视频进行分级。但是,同时传输在不同亮度范围内分级的多个视频比特流,在网络利用率和服务器存储方面效率低下。为了克服这个问题,我们提出用基础层视频比特流传输渐进式元数据。这种嵌入允许元数据的不同重叠部分将基本视频扩展到逐渐扩大的亮度范围。我们的渐进式元数据格式在现有架构上提供了重大的设计改进,在所有支持的亮度范围内保留色彩器意图,并且仍然保持最小的带宽或存储开销。
{"title":"Single Layer Progressive Coding for High Dynamic Range Videos","authors":"H. Kadu, Qing Song, Guan-Ming Su","doi":"10.1109/PCS.2018.8456314","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456314","url":null,"abstract":"There are different kinds of high dynamic range (HDR) displays in the market today. These displays have different HDR specifications, like, pealddark brightness levels, electro-optical transfer functions (EOTF), color spaces etc. For the best visual experience on a given HDR screen, colorists have to grade videos for that specific display’s luminance range. Uut simultaneous transmission of multiple video bitstreams graded at different luminance ranges, is inefficient in terms of network utility and server storage. To overcome this problem, we propose transmitting our progressive metadata with a base layer video bitstream. This embedding allows different overlapping portions of metadata to scale the base video to progressively wider luminance ranges. Our progressive metadata format provides a significant design improvement over the existing architectures, preserves colorist intent at all the supported brightness ranges and still keeps the bandwidth or storage overhead minimal.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123176520","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Region-Wise Super-Resolution Algorithm Based On the Viewpoint Distribution 基于视点分布的区域超分辨算法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456295
Kazunori Uruma, Shunsuke Takasu, Keiko Masuda, S. Hangai
Recently, super-resolution techniques have been energetically studied for the purpose of reusing the low resolution image contents. Although a lot of approaches to achieve the appropriate super-resolution have been proposed such as non-linear filtering, total variation regularization, deep learning etc., the characteristic of the viewpoint distribution of the observer has not been effectively utilized. Because applying super-resolution to unimportant regions in an image may hinder the observer’s attention to seeing the display, it leads to a low subjective evaluation. This paper proposes the region-wise super-resolution algorithm based on the view-point distribution of observer. However, we cannot obtain the viewpoint distribution map for an image without the pre-experiment using the device such as eye mark recorder, therefore, the saliency map is utilized in this paper. Numerical examples show that the proposed algorithm using saliency map achieves a higher subjective evaluation than the previous study based on the non-linear filtering based super-resolution. Furthermore, in numerical examples, the proposed algorithm using the saliency map is shown to give the similar results of the algorithm using the viewpoint distribution map obtained by the pre-experiment using eye mark recorder.
近年来,为了实现低分辨率图像内容的重复利用,超分辨率技术得到了大力的研究。尽管人们提出了非线性滤波、全变分正则化、深度学习等方法来达到适当的超分辨率,但并没有有效地利用观测器视点分布的特性。由于将超分辨率应用于图像中不重要的区域可能会阻碍观察者的注意力,从而导致主观评价较低。提出了一种基于观测器视点分布的区域超分辨算法。然而,如果没有使用眼痕记录仪等设备进行预实验,我们无法获得图像的视点分布图,因此本文采用显著性图。数值算例表明,与基于非线性滤波的超分辨率算法相比,基于显著性映射的算法获得了更高的主观评价。在数值算例中,利用显著性图的算法与利用眼痕记录仪预实验所得的视点分布图算法的结果相似。
{"title":"Region-Wise Super-Resolution Algorithm Based On the Viewpoint Distribution","authors":"Kazunori Uruma, Shunsuke Takasu, Keiko Masuda, S. Hangai","doi":"10.1109/PCS.2018.8456295","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456295","url":null,"abstract":"Recently, super-resolution techniques have been energetically studied for the purpose of reusing the low resolution image contents. Although a lot of approaches to achieve the appropriate super-resolution have been proposed such as non-linear filtering, total variation regularization, deep learning etc., the characteristic of the viewpoint distribution of the observer has not been effectively utilized. Because applying super-resolution to unimportant regions in an image may hinder the observer’s attention to seeing the display, it leads to a low subjective evaluation. This paper proposes the region-wise super-resolution algorithm based on the view-point distribution of observer. However, we cannot obtain the viewpoint distribution map for an image without the pre-experiment using the device such as eye mark recorder, therefore, the saliency map is utilized in this paper. Numerical examples show that the proposed algorithm using saliency map achieves a higher subjective evaluation than the previous study based on the non-linear filtering based super-resolution. Furthermore, in numerical examples, the proposed algorithm using the saliency map is shown to give the similar results of the algorithm using the viewpoint distribution map obtained by the pre-experiment using eye mark recorder.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123436034","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Wavefront Parallel Processing for AV1 Encoder AV1编码器的波前并行处理
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456283
Yikai Zhao, Jiangtao Wen
The emerging AV1 coding standard brings even higher computational complexity than current coding standards, but does not support traditional Wavefront Parallel Processing (WPP) approach due to the lacking of syntax support. In this paper we introduced a novel framework to implement WPP for AV1 encoder that is compatible with current decoder without additional bitstream syntax support, where mode selection is processed in wavefront parallel before entropy encoding and entropy contexts for rate-distortion optimization are predicted. Based on this framework, context prediction algorithms that use same data dependency model as previous works in H.264 and HEVC are implemented. Furthermore, we proposed an optimal context prediction algorithm specifically for AV1. Experimental results showed that our framework with proposed optimal algorithm yields good parallelism and scalability (over 10x speed-up with 16 threads for 4k sequences) with little coding performance loss (less than 0.2% bitrate increasing).
新兴的AV1编码标准带来了比现有编码标准更高的计算复杂度,但由于缺乏语法支持而不支持传统的波前并行处理(WPP)方法。在本文中,我们引入了一种新的框架来实现AV1编码器的WPP,该框架与当前的解码器兼容,无需额外的比特流语法支持,其中模式选择在波前并行处理,然后预测熵编码和熵上下文用于率失真优化。基于该框架,实现了与H.264和HEVC中使用相同数据依赖模型的上下文预测算法。此外,我们提出了一种针对AV1的最佳上下文预测算法。实验结果表明,我们的框架具有良好的并行性和可扩展性(4k序列16线程加速超过10倍),编码性能损失很小(比特率增加不到0.2%)。
{"title":"Wavefront Parallel Processing for AV1 Encoder","authors":"Yikai Zhao, Jiangtao Wen","doi":"10.1109/PCS.2018.8456283","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456283","url":null,"abstract":"The emerging AV1 coding standard brings even higher computational complexity than current coding standards, but does not support traditional Wavefront Parallel Processing (WPP) approach due to the lacking of syntax support. In this paper we introduced a novel framework to implement WPP for AV1 encoder that is compatible with current decoder without additional bitstream syntax support, where mode selection is processed in wavefront parallel before entropy encoding and entropy contexts for rate-distortion optimization are predicted. Based on this framework, context prediction algorithms that use same data dependency model as previous works in H.264 and HEVC are implemented. Furthermore, we proposed an optimal context prediction algorithm specifically for AV1. Experimental results showed that our framework with proposed optimal algorithm yields good parallelism and scalability (over 10x speed-up with 16 threads for 4k sequences) with little coding performance loss (less than 0.2% bitrate increasing).","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"40 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133163323","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Compression Performance Comparison of x264, x265, libvpx and aomenc for On-Demand Adaptive Streaming Applications x264, x265, libvpx和aomenc在按需自适应流媒体应用中的压缩性能比较
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456302
Liwei Guo, J. D. Cock, A. Aaron
Video compression standard H.264/AVC was released in 2003, and has been dominating the industry for the past decade. Over the last few years, a number of next-generation standards/formats like VP9 (2012), H.265/HEVC (2013) and AV1 (2018) were introduced, all claiming significant improvement over H.264/AVC. In this paper, we present our evaluation of the performance of these compression standards. Our evaluation is conducted using open-source encoder implementations of these standards, x264 (for H.264/AVC), x265 (for H.265/HEVC), libvpx (for VP9) and aomenc (for AV1). The process is designed to evaluate the attainable compression efficiency for on-demand adaptive streaming applications. Results with two different quality metrics, PSNR and VMAF, are reported. Our results reveal that x265, libvpx and aomenc all achieve substantial compression efficiency improvement over x264.
视频压缩标准H.264/AVC于2003年发布,在过去的十年里一直主导着这个行业。在过去的几年中,推出了许多下一代标准/格式,如VP9 (2012), H.265/HEVC(2013)和AV1(2018),都声称比H.264/AVC有了重大改进。在本文中,我们提出了我们的性能评估这些压缩标准。我们的评估使用这些标准的开源编码器实现进行,x264(用于H.264/AVC), x265(用于H.265/HEVC), libvpx(用于VP9)和aomenc(用于AV1)。该过程旨在评估按需自适应流媒体应用程序可实现的压缩效率。报告了两种不同质量指标PSNR和VMAF的结果。我们的结果表明,x265、libvpx和aomenc都比x264实现了显著的压缩效率提高。
{"title":"Compression Performance Comparison of x264, x265, libvpx and aomenc for On-Demand Adaptive Streaming Applications","authors":"Liwei Guo, J. D. Cock, A. Aaron","doi":"10.1109/PCS.2018.8456302","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456302","url":null,"abstract":"Video compression standard H.264/AVC was released in 2003, and has been dominating the industry for the past decade. Over the last few years, a number of next-generation standards/formats like VP9 (2012), H.265/HEVC (2013) and AV1 (2018) were introduced, all claiming significant improvement over H.264/AVC. In this paper, we present our evaluation of the performance of these compression standards. Our evaluation is conducted using open-source encoder implementations of these standards, x264 (for H.264/AVC), x265 (for H.265/HEVC), libvpx (for VP9) and aomenc (for AV1). The process is designed to evaluate the attainable compression efficiency for on-demand adaptive streaming applications. Results with two different quality metrics, PSNR and VMAF, are reported. Our results reveal that x265, libvpx and aomenc all achieve substantial compression efficiency improvement over x264.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"34 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127613113","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1