首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
2D Video Coding of Volumetric Video Data 体积视频数据的二维视频编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456265
S. Schwarz, M. Hannuksela, Vida Fakour Sevom, Nahid Sheikhi-Pour
Due to the increased popularity of augmented and virtual reality experiences, the interest in representing the real world in an immersive fashion has never been higher. Distributing such representations enables users all over the world to freely navigate in never seen before media experiences. Unfortunately, such representations require a large amount of data, not feasible for transmission on today’s networks. Thus, efficient compression technologies are in high demand. This paper proposes an approach to compress 3D video data utilizing 2D video coding technology. The proposed solution was developed to address the needs of ‘tele-immersive’ applications, such as virtual (VR), augmented (AR) or mixed (MR) reality with Six Degrees of Freedom (6DoF) capabilities. Volumetric video data is projected on 2D image planes and compressed using standard 2D video coding solutions. A key benefit of this approach is its compatibility with readily available 2D video coding infrastructure. Furthermore, objective and subjective evaluation shows significant improvement in coding efficiency over reference technology.
由于增强现实和虚拟现实体验的日益普及,人们对以身临其境的方式呈现现实世界的兴趣从未如此高涨。分发这样的表现使世界各地的用户能够在从未见过的媒体体验中自由导航。不幸的是,这样的表示需要大量的数据,在今天的网络上传输是不可行的。因此,对高效压缩技术的需求很大。本文提出了一种利用二维视频编码技术对三维视频数据进行压缩的方法。提出的解决方案旨在满足“远程沉浸式”应用的需求,例如具有六自由度(6DoF)功能的虚拟(VR)、增强(AR)或混合(MR)现实。体积视频数据投影在二维图像平面上,并使用标准的二维视频编码解决方案进行压缩。这种方法的一个关键优点是它与现成的2D视频编码基础设施兼容。此外,客观和主观评价表明编码效率比参考技术有显著提高。
{"title":"2D Video Coding of Volumetric Video Data","authors":"S. Schwarz, M. Hannuksela, Vida Fakour Sevom, Nahid Sheikhi-Pour","doi":"10.1109/PCS.2018.8456265","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456265","url":null,"abstract":"Due to the increased popularity of augmented and virtual reality experiences, the interest in representing the real world in an immersive fashion has never been higher. Distributing such representations enables users all over the world to freely navigate in never seen before media experiences. Unfortunately, such representations require a large amount of data, not feasible for transmission on today’s networks. Thus, efficient compression technologies are in high demand. This paper proposes an approach to compress 3D video data utilizing 2D video coding technology. The proposed solution was developed to address the needs of ‘tele-immersive’ applications, such as virtual (VR), augmented (AR) or mixed (MR) reality with Six Degrees of Freedom (6DoF) capabilities. Volumetric video data is projected on 2D image planes and compressed using standard 2D video coding solutions. A key benefit of this approach is its compatibility with readily available 2D video coding infrastructure. Furthermore, objective and subjective evaluation shows significant improvement in coding efficiency over reference technology.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126520870","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
An Overview of Core Coding Tools in the AV1 Video Codec AV1视频编解码器的核心编码工具概述
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456249
Yue Chen, D. Mukherjee, Jingning Han, Adrian Grange, Yaowu Xu, Zoe Liu, Sarah Parker, Cheng Chen, Hui Su, Urvang Joshi, Ching-Han Chiang, Yunqing Wang, Paul Wilkins, Jim Bankoski, Luc N. Trudeau, N. Egge, J. Valin, T. Davies, Steinar Midtskogen, A. Norkin, Peter De Rivaz
AV1 is an emerging open-source and royalty-free video compression format, which is jointly developed and finalized in early 2018 by the Alliance for Open Media (AOMedia) industry consortium. The main goal of AV1 development is to achieve substantial compression gain over state-of-the-art codecs while maintaining practical decoding complexity and hardware feasibility. This paper provides a brief technical overview of key coding techniques in AV1 along with preliminary compression performance comparison against VP9 and HEVC.
AV1是一种新兴的开源和免版税视频压缩格式,由开放媒体联盟(AOMedia)行业联盟于2018年初共同开发并最终确定。AV1开发的主要目标是在保持实际解码复杂性和硬件可行性的同时,实现比最先进的编解码器更大的压缩增益。本文简要介绍了AV1中关键编码技术的技术概况,并与VP9和HEVC进行了初步的压缩性能比较。
{"title":"An Overview of Core Coding Tools in the AV1 Video Codec","authors":"Yue Chen, D. Mukherjee, Jingning Han, Adrian Grange, Yaowu Xu, Zoe Liu, Sarah Parker, Cheng Chen, Hui Su, Urvang Joshi, Ching-Han Chiang, Yunqing Wang, Paul Wilkins, Jim Bankoski, Luc N. Trudeau, N. Egge, J. Valin, T. Davies, Steinar Midtskogen, A. Norkin, Peter De Rivaz","doi":"10.1109/PCS.2018.8456249","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456249","url":null,"abstract":"AV1 is an emerging open-source and royalty-free video compression format, which is jointly developed and finalized in early 2018 by the Alliance for Open Media (AOMedia) industry consortium. The main goal of AV1 development is to achieve substantial compression gain over state-of-the-art codecs while maintaining practical decoding complexity and hardware feasibility. This paper provides a brief technical overview of key coding techniques in AV1 along with preliminary compression performance comparison against VP9 and HEVC.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125791572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 217
Deep Learning Based HEVC In-Loop Filtering for Decoder Quality Enhancement 基于深度学习的HEVC环内滤波解码器质量增强
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456278
Shiba Kuanar, C. Conly, K. Rao
High Efficiency Video Coding (HEVC), which is the latest video coding standard currently, achieves up to 50% bit rate reduction compared to previous H.264/AVC standard. While performing the block based video coding, these lossy compression techniques produce various artifacts like blurring, distortion, ringing, and contouring effects on output frames, especially at low bit rates. To reduce those compression artifacts HEVC adopted two post processing filtering technique namely de-blocking filter (DBF) and sample adaptive offset (SAO) on the decoder side. While DBF applies to samples located at block boundaries, SAO nonlinear operation applies adaptively to samples satisfying the gradient based conditions through a lookup table. Again SAO filter corrects the quantization errors by sending edge offset values to decoders. This operation consumes extra signaling bit and becomes an overhead to network. In this paper, we proposed a Convolutional Neural Network (CNN) based architecture for SAO in-loop filtering operation without modifying anything on encoding process. Our experimental results show that our proposed model outperformed previous state-of-the-art models in terms of BD-PSNR (0.408 dB) and BD-BR (3.44%), measured on a widely available standard video sequences.
高效视频编码(HEVC)是目前最新的视频编码标准,与之前的H.264/AVC标准相比,可实现高达50%的比特率降低。在执行基于块的视频编码时,这些有损压缩技术会在输出帧上产生各种伪影,如模糊、失真、振铃和轮廓效果,特别是在低比特率下。为了减少这些压缩伪影,HEVC在解码器侧采用了去块滤波(DBF)和采样自适应偏移(SAO)两种后处理滤波技术。DBF应用于位于块边界的样本,SAO非线性操作通过查找表自适应地应用于满足基于梯度条件的样本。同样,SAO滤波器通过向解码器发送边缘偏移值来纠正量化错误。此操作消耗额外的信令位,成为网络开销。本文提出了一种基于卷积神经网络(Convolutional Neural Network, CNN)的SAO环内滤波结构,在不改变编码过程的情况下进行SAO环内滤波。实验结果表明,在广泛使用的标准视频序列上,我们提出的模型在BD-PSNR (0.408 dB)和BD-BR(3.44%)方面优于先前的最先进模型。
{"title":"Deep Learning Based HEVC In-Loop Filtering for Decoder Quality Enhancement","authors":"Shiba Kuanar, C. Conly, K. Rao","doi":"10.1109/PCS.2018.8456278","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456278","url":null,"abstract":"High Efficiency Video Coding (HEVC), which is the latest video coding standard currently, achieves up to 50% bit rate reduction compared to previous H.264/AVC standard. While performing the block based video coding, these lossy compression techniques produce various artifacts like blurring, distortion, ringing, and contouring effects on output frames, especially at low bit rates. To reduce those compression artifacts HEVC adopted two post processing filtering technique namely de-blocking filter (DBF) and sample adaptive offset (SAO) on the decoder side. While DBF applies to samples located at block boundaries, SAO nonlinear operation applies adaptively to samples satisfying the gradient based conditions through a lookup table. Again SAO filter corrects the quantization errors by sending edge offset values to decoders. This operation consumes extra signaling bit and becomes an overhead to network. In this paper, we proposed a Convolutional Neural Network (CNN) based architecture for SAO in-loop filtering operation without modifying anything on encoding process. Our experimental results show that our proposed model outperformed previous state-of-the-art models in terms of BD-PSNR (0.408 dB) and BD-BR (3.44%), measured on a widely available standard video sequences.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124867899","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
A Hybrid Weighted Compound Motion Compensated Prediction for Video Compression 视频压缩的混合加权复合运动补偿预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456241
Cheng Chen, Jingning Han, Yaowu Xu
Compound motion compensated prediction that combines reconstructed reference blocks to exploit the temporal correlation is a major component in the hierarchical coding scheme. A uniform combination that applies equal weights to reference blocks regardless of distances towards the current frame is widely employed in mainstream codecs. Linear distance weighted combination, while reflecting the temporal correlation, is likely to ignore the quantization noise factor and hence degrade the prediction quality. This work builds on the premise that the compound prediction mode effectively embeds two functionalities - exploiting temporal correlation in the video signal and canceling the quantization noise from reference blocks. A modified distance weighting scheme is introduced to optimize the trade-off between these two factors. It quantizes the weights to limit the minimum contribution from both reference blocks for noise cancellation. We further introduces a hybrid scheme allowing the codec to switch between the proposed distance weighted compound mode and the averaging mode to provide more flexibility for the trade-off between temporal correlation and noise cancellation. The scheme is implemented in the AV1 codec as part of the syntax definition. It is experimentally demonstrated to provide on average 1.5% compression gains across a wide range of test sets.
复合运动补偿预测是分层编码方案的重要组成部分,它结合重构的参考块来利用时间相关性。在主流编解码器中广泛采用一种统一的组合,即对参考块应用相同的权重,而不管与当前帧的距离。线性距离加权组合虽然反映了时间相关性,但容易忽略量化噪声因素,从而降低预测质量。这项工作建立在复合预测模式有效嵌入两个功能的前提下-利用视频信号的时间相关性和消除参考块的量化噪声。引入了一种改进的距离加权方案来优化这两个因素之间的权衡。它量化权重以限制两个参考块对噪声消除的最小贡献。我们进一步介绍了一种混合方案,允许编解码器在提出的距离加权复合模式和平均模式之间切换,为时间相关和噪声消除之间的权衡提供更大的灵活性。该方案在AV1编解码器中作为语法定义的一部分实现。实验证明,在广泛的测试集范围内,它可以提供平均1.5%的压缩增益。
{"title":"A Hybrid Weighted Compound Motion Compensated Prediction for Video Compression","authors":"Cheng Chen, Jingning Han, Yaowu Xu","doi":"10.1109/PCS.2018.8456241","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456241","url":null,"abstract":"Compound motion compensated prediction that combines reconstructed reference blocks to exploit the temporal correlation is a major component in the hierarchical coding scheme. A uniform combination that applies equal weights to reference blocks regardless of distances towards the current frame is widely employed in mainstream codecs. Linear distance weighted combination, while reflecting the temporal correlation, is likely to ignore the quantization noise factor and hence degrade the prediction quality. This work builds on the premise that the compound prediction mode effectively embeds two functionalities - exploiting temporal correlation in the video signal and canceling the quantization noise from reference blocks. A modified distance weighting scheme is introduced to optimize the trade-off between these two factors. It quantizes the weights to limit the minimum contribution from both reference blocks for noise cancellation. We further introduces a hybrid scheme allowing the codec to switch between the proposed distance weighted compound mode and the averaging mode to provide more flexibility for the trade-off between temporal correlation and noise cancellation. The scheme is implemented in the AV1 codec as part of the syntax definition. It is experimentally demonstrated to provide on average 1.5% compression gains across a wide range of test sets.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128490448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
PCS 2018 Commentary pc 2018评论
Pub Date : 2018-06-01 DOI: 10.1109/pcs.2018.8456270
{"title":"PCS 2018 Commentary","authors":"","doi":"10.1109/pcs.2018.8456270","DOIUrl":"https://doi.org/10.1109/pcs.2018.8456270","url":null,"abstract":"","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"278 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127383308","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Machine learning as applied intrinsically to individual dimensions of HDR Display Quality 机器学习本质上应用于HDR显示质量的各个维度
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456284
A. Choudhury, S. Daly
This study builds on previous work exploring machine learning and perceptual transforms in predicting overall display quality as a function of image quality dimensions that correspond to physical display parameters. Previously, we found that the use of perceptually transformed parameters or machine learning exceeded the performance of predictors using just physical parameters and linear regression. Further, the combination of perceptually transformed parameters with machine learning allowed for robustness to parameters outside of the data set, both for cases of interpolation and extrapolation. Here we apply machine learning at a more intrinsic level. We first evaluate how well the machine learning can develop predictors of the individual dimensions of the overall quality, and then how well those individual predictors can be consolidated across themselves to predict the overall display quality. Having predictions of individual dimensions of quality that are closely related to specific hardware design choices enables more nimble cost trade-off design options.
本研究建立在先前探索机器学习和感知转换的工作基础上,以预测整体显示质量作为与物理显示参数对应的图像质量维度的函数。之前,我们发现使用感知转换参数或机器学习超过了仅使用物理参数和线性回归的预测器的性能。此外,感知转换参数与机器学习的结合允许对数据集之外的参数具有鲁棒性,无论是内插还是外推。在这里,我们将机器学习应用于更内在的层面。我们首先评估机器学习对整体质量的各个维度的预测能力,然后评估这些个体预测能力在预测整体显示质量方面的整合能力。预测与特定硬件设计选择密切相关的单个质量维度,可以实现更灵活的成本权衡设计选择。
{"title":"Machine learning as applied intrinsically to individual dimensions of HDR Display Quality","authors":"A. Choudhury, S. Daly","doi":"10.1109/PCS.2018.8456284","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456284","url":null,"abstract":"This study builds on previous work exploring machine learning and perceptual transforms in predicting overall display quality as a function of image quality dimensions that correspond to physical display parameters. Previously, we found that the use of perceptually transformed parameters or machine learning exceeded the performance of predictors using just physical parameters and linear regression. Further, the combination of perceptually transformed parameters with machine learning allowed for robustness to parameters outside of the data set, both for cases of interpolation and extrapolation. Here we apply machine learning at a more intrinsic level. We first evaluate how well the machine learning can develop predictors of the individual dimensions of the overall quality, and then how well those individual predictors can be consolidated across themselves to predict the overall display quality. Having predictions of individual dimensions of quality that are closely related to specific hardware design choices enables more nimble cost trade-off design options.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129406526","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Wavelet Decomposition Pre-processing for Spatial Scalability Video Compression Scheme 空间可扩展性视频压缩方案的小波分解预处理
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456307
Glenn Herrou, W. Hamidouche, L. Morin
Scalable video coding enables to compress the video at different formats within a single layered bitstream. SHVC, the scalable extension of the High Efficiency Video Coding (HEVC) standard, enables x2 spatial scalability, among other additional features. The closed-loop architecture of the SHVC codec is based on the use of multiple instances of the HEVC codec to encode the video layers, which considerably increases the encoding complexity. With the arrival of new immersive video formats, like 4K, 8K, High Frame Rate (HFR) and 360° videos, the quantity of data to compress is exploding, making the use of high-complexity coding algorithms unsuitable. In this paper, we propose a lowcomplexity scalable coding scheme based on the use of a single HEVC codec instance and a wavelet-based decomposition as pre-processing. The pre-encoding image decomposition relies on well-known simple Discrete Wavelet Transform (DWT) kernels, such as Haar or Le Gall 5/3. Compared to SHVC, the proposed architecture achieves a similar rate distortion performance with a coding complexity reduction of 50%.
可扩展的视频编码可以在单个层比特流中压缩不同格式的视频。SHVC是高效视频编码(HEVC)标准的可扩展扩展,支持x2空间可扩展性,以及其他附加功能。SHVC编解码器的闭环架构是基于使用多个HEVC编解码器实例对视频层进行编码,这大大增加了编码复杂度。随着新的沉浸式视频格式的出现,如4K、8K、高帧率(HFR)和360°视频,需要压缩的数据量呈爆炸式增长,这使得使用高复杂性的编码算法变得不合适。在本文中,我们提出了一种基于单个HEVC编解码器实例和基于小波分解作为预处理的低复杂度可扩展编码方案。预编码图像分解依赖于众所周知的简单离散小波变换(DWT)核,如Haar或Le Gall 5/3。与SHVC相比,该结构在编码复杂度降低50%的情况下实现了相似的率失真性能。
{"title":"Wavelet Decomposition Pre-processing for Spatial Scalability Video Compression Scheme","authors":"Glenn Herrou, W. Hamidouche, L. Morin","doi":"10.1109/PCS.2018.8456307","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456307","url":null,"abstract":"Scalable video coding enables to compress the video at different formats within a single layered bitstream. SHVC, the scalable extension of the High Efficiency Video Coding (HEVC) standard, enables x2 spatial scalability, among other additional features. The closed-loop architecture of the SHVC codec is based on the use of multiple instances of the HEVC codec to encode the video layers, which considerably increases the encoding complexity. With the arrival of new immersive video formats, like 4K, 8K, High Frame Rate (HFR) and 360° videos, the quantity of data to compress is exploding, making the use of high-complexity coding algorithms unsuitable. In this paper, we propose a lowcomplexity scalable coding scheme based on the use of a single HEVC codec instance and a wavelet-based decomposition as pre-processing. The pre-encoding image decomposition relies on well-known simple Discrete Wavelet Transform (DWT) kernels, such as Haar or Le Gall 5/3. Compared to SHVC, the proposed architecture achieves a similar rate distortion performance with a coding complexity reduction of 50%.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114063397","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Benchmarking of Objective Quality Metrics for Colorless Point Clouds 无色点云客观质量指标的基准测试
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456252
E. Alexiou, T. Ebrahimi
Recent advances in depth sensing and display technologies, along with the significant growth of interest for augmented and virtual reality applications, lay the foundation for the rapid evolution of applications that provide immersive experiences. In such applications, advanced content representations are required in order to increase the engagement of the user with the displayed imageries. Point clouds have emerged as a promising solution to this aim, due to their efficiency in capturing, storing, delivering and rendering of 3D immersive contents. As in any type of imaging, the evaluation of point clouds in terms of visual quality is essential. In this paper, benchmarking results of the state-of-the-art objective metrics in geometry-only point clouds are reported and analyzed under two different types of geometry degradations, namely Gaussian noise and octree- based compression. Human ratings obtained from two subjective experiments are used as the ground truth. Our results show that most objective quality metrics perform well in the presence of noise, whereas one particular method has high predictive power and outperforms the others after octree-based encoding.
深度传感和显示技术的最新进展,以及对增强现实和虚拟现实应用的兴趣的显着增长,为提供沉浸式体验的应用的快速发展奠定了基础。在这样的应用程序中,为了增加用户对所显示图像的参与,需要高级内容表示。由于点云在捕获、存储、传输和渲染3D沉浸式内容方面的效率,它已经成为实现这一目标的一个有希望的解决方案。与任何类型的成像一样,从视觉质量方面对点云进行评估是必不可少的。本文报道并分析了纯几何点云在高斯噪声和基于八叉树的压缩两种不同几何退化类型下最先进的客观指标的基准测试结果。从两个主观实验中获得的人类评分被用作基础事实。我们的研究结果表明,大多数客观质量指标在存在噪声的情况下表现良好,而一种特定的方法具有很高的预测能力,并且在基于八叉树的编码后优于其他方法。
{"title":"Benchmarking of Objective Quality Metrics for Colorless Point Clouds","authors":"E. Alexiou, T. Ebrahimi","doi":"10.1109/PCS.2018.8456252","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456252","url":null,"abstract":"Recent advances in depth sensing and display technologies, along with the significant growth of interest for augmented and virtual reality applications, lay the foundation for the rapid evolution of applications that provide immersive experiences. In such applications, advanced content representations are required in order to increase the engagement of the user with the displayed imageries. Point clouds have emerged as a promising solution to this aim, due to their efficiency in capturing, storing, delivering and rendering of 3D immersive contents. As in any type of imaging, the evaluation of point clouds in terms of visual quality is essential. In this paper, benchmarking results of the state-of-the-art objective metrics in geometry-only point clouds are reported and analyzed under two different types of geometry degradations, namely Gaussian noise and octree- based compression. Human ratings obtained from two subjective experiments are used as the ground truth. Our results show that most objective quality metrics perform well in the presence of noise, whereas one particular method has high predictive power and outperforms the others after octree-based encoding.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"41 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120918293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Analysis and Prediction of JND-Based Video Quality Model 基于jnd的视频质量模型分析与预测
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456243
Haiqiang Wang, Xinfeng Zhang, Chao Yang, C.-C. Jay Kuo
The just-noticeable-difference (JND) visual perception property has received much attention in characterizing human subjective viewing experience of compressed video. In this work, we quantity the JND-based video quality assessment model using the satisfied user ratio (SUR) curve, and show that the SUR model can be greatly simplified since the JND points of multiple subjects for the same content in the VideoSet can be well modeled by the normal distribution. Then, we design an SUR prediction method with video quality degradation features and masking features and use them to predict the first, second and the third JND points and their corresponding SUR curves. Finally, we verify the performance of the proposed SUR prediction method with different configurations on the VideoSet. The experimental results demonstrate that the proposed SUR prediction method achieves good performance in various resolutions with the mean absolute error (MAE) of the SUR smaller than 0.05 on average.
just- visible -difference (JND)视觉感知特性在描述压缩视频的主观观看体验方面受到了广泛关注。在这项工作中,我们使用满意用户比例(SUR)曲线对基于JND的视频质量评估模型进行了量化,并表明由于VideoSet中相同内容的多个主题的JND点可以通过正态分布很好地建模,因此SUR模型可以大大简化。然后,我们设计了一种包含视频质量退化特征和掩蔽特征的SUR预测方法,并利用它们预测第一、第二和第三个JND点及其对应的SUR曲线。最后,在VideoSet上验证了不同配置下所提出的SUR预测方法的性能。实验结果表明,所提出的SUR预测方法在不同分辨率下均取得了较好的预测效果,平均绝对误差(MAE)均小于0.05。
{"title":"Analysis and Prediction of JND-Based Video Quality Model","authors":"Haiqiang Wang, Xinfeng Zhang, Chao Yang, C.-C. Jay Kuo","doi":"10.1109/PCS.2018.8456243","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456243","url":null,"abstract":"The just-noticeable-difference (JND) visual perception property has received much attention in characterizing human subjective viewing experience of compressed video. In this work, we quantity the JND-based video quality assessment model using the satisfied user ratio (SUR) curve, and show that the SUR model can be greatly simplified since the JND points of multiple subjects for the same content in the VideoSet can be well modeled by the normal distribution. Then, we design an SUR prediction method with video quality degradation features and masking features and use them to predict the first, second and the third JND points and their corresponding SUR curves. Finally, we verify the performance of the proposed SUR prediction method with different configurations on the VideoSet. The experimental results demonstrate that the proposed SUR prediction method achieves good performance in various resolutions with the mean absolute error (MAE) of the SUR smaller than 0.05 on average.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"50 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127684293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Perceptual Quality Driven Adaptive Video Coding Using JND Estimation 基于JND估计的感知质量驱动自适应视频编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456297
Masaru Takeuchi, Shintaro Saika, Yusuke Sakamoto, Tatsuya Nagashima, Zhengxue Cheng, Kenji Kanai, J. Katto, Kaijin Wei, Ju Zengwei, Xu Wei
We introduce a perceptual video quality driven video encoding solution for optimized adaptive streaming. By using multiple bitrate/resolution encoding like MPEG-DASH, video streaming services can deliver the best video stream to a client, under the conditions of the client's available bandwidth and viewing device capability. However, conventional fixed encoding recipes (i.e., resolution-bitrate pairs) suffer from many problems, such as improper resolution selection and stream redundancy. To avoid these problems, we propose a novel video coding method, which generates multiple representations with constant JustNoticeable Difference (JND) interval. For this purpose, we developed a JND scale estimator using Support Vector Regression (SVR), and designed a pre-encoder which outputs an encoding recipe with constant JND interval in an adaptive manner to input video.
我们提出了一种感知视频质量驱动的视频编码方案,用于优化自适应流媒体。通过使用像MPEG-DASH这样的多比特率/分辨率编码,视频流服务可以在客户端可用带宽和观看设备能力的条件下向客户端提供最佳视频流。然而,传统的固定编码方法(即分辨率-比特率对)存在许多问题,例如不正确的分辨率选择和流冗余。为了避免这些问题,我们提出了一种新的视频编码方法,该方法以恒定的justvisible Difference (JND)间隔生成多个表示。为此,我们利用支持向量回归(SVR)开发了JND尺度估计器,并设计了一个预编码器,该预编码器以自适应方式输出恒定JND间隔的编码配方来输入视频。
{"title":"Perceptual Quality Driven Adaptive Video Coding Using JND Estimation","authors":"Masaru Takeuchi, Shintaro Saika, Yusuke Sakamoto, Tatsuya Nagashima, Zhengxue Cheng, Kenji Kanai, J. Katto, Kaijin Wei, Ju Zengwei, Xu Wei","doi":"10.1109/PCS.2018.8456297","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456297","url":null,"abstract":"We introduce a perceptual video quality driven video encoding solution for optimized adaptive streaming. By using multiple bitrate/resolution encoding like MPEG-DASH, video streaming services can deliver the best video stream to a client, under the conditions of the client's available bandwidth and viewing device capability. However, conventional fixed encoding recipes (i.e., resolution-bitrate pairs) suffer from many problems, such as improper resolution selection and stream redundancy. To avoid these problems, we propose a novel video coding method, which generates multiple representations with constant JustNoticeable Difference (JND) interval. For this purpose, we developed a JND scale estimator using Support Vector Regression (SVR), and designed a pre-encoder which outputs an encoding recipe with constant JND interval in an adaptive manner to input video.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125872696","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1