首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Physical High Dynamic Range Imaging with Conventional Sensors 物理高动态范围成像与传统传感器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456268
H. Meuel, H. Ackermann, B. Rosenhahn, J. Ostermann
This paper aims at simplified high dynamic range (HDR) image generation with non-modified, conventional camera sensors. One typical HDR approach is exposure bracketing, e.g. with varying shutter speeds. It requires to capture the same scene multiple times at different exposure times. These pictures are then merged into a single HDR picture which typically is converted back to an 8-bit image by using tone-mapping. Existing works on HDR imaging focus on image merging and tone mapping whereas we aim at simplified image acquisition. The proposed algorithm can be used in consumer-level cameras without hardware modifications at sensor level. Based on intermediate samplings of each sensor element during the total (pre-defined) exposure time, we extrapolate the luminance of sensor elements which are saturated after the total exposure time. Compared to existing HDR approaches which typically require three different images with carefully determined exposure times, we only take one image at the longest exposure time. The shortened total time between start and end of image acquisition can reduce ghosting artifacts. The experimental evaluation demonstrates the effectiveness of the algorithm.
本文的目的是简化高动态范围(HDR)图像生成与非修改,传统的相机传感器。一种典型的HDR方法是曝光包围,例如使用不同的快门速度。它需要在不同的曝光时间多次捕捉相同的场景。这些图片然后合并成一个单一的HDR图片,通常是转换回一个8位图像使用色调映射。现有的HDR成像工作主要集中在图像合并和色调映射,而我们的目标是简化图像获取。该算法可用于消费级相机,无需在传感器级进行硬件修改。基于每个传感器元件在总(预定)曝光时间内的中间采样,我们推断出在总曝光时间后饱和的传感器元件的亮度。现有的HDR方法通常需要三张不同的图像和仔细确定的曝光时间,相比之下,我们只需要在最长的曝光时间内拍摄一张图像。缩短图像采集开始和结束的总时间可以减少重影。实验验证了该算法的有效性。
{"title":"Physical High Dynamic Range Imaging with Conventional Sensors","authors":"H. Meuel, H. Ackermann, B. Rosenhahn, J. Ostermann","doi":"10.1109/PCS.2018.8456268","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456268","url":null,"abstract":"This paper aims at simplified high dynamic range (HDR) image generation with non-modified, conventional camera sensors. One typical HDR approach is exposure bracketing, e.g. with varying shutter speeds. It requires to capture the same scene multiple times at different exposure times. These pictures are then merged into a single HDR picture which typically is converted back to an 8-bit image by using tone-mapping. Existing works on HDR imaging focus on image merging and tone mapping whereas we aim at simplified image acquisition. The proposed algorithm can be used in consumer-level cameras without hardware modifications at sensor level. Based on intermediate samplings of each sensor element during the total (pre-defined) exposure time, we extrapolate the luminance of sensor elements which are saturated after the total exposure time. Compared to existing HDR approaches which typically require three different images with carefully determined exposure times, we only take one image at the longest exposure time. The shortened total time between start and end of image acquisition can reduce ghosting artifacts. The experimental evaluation demonstrates the effectiveness of the algorithm.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131689016","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
PCS 2018 List Committees 个人电脑2018名单委员会
Pub Date : 2018-06-01 DOI: 10.1109/pcs.2018.8456312
{"title":"PCS 2018 List Committees","authors":"","doi":"10.1109/pcs.2018.8456312","DOIUrl":"https://doi.org/10.1109/pcs.2018.8456312","url":null,"abstract":"","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131100022","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Method to Improve Perceptual Quality of Intra- Refresh-Enabled Low-Latency Video Coding 一种提高内刷新低延迟视频编码感知质量的方法
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456257
Wei Gao, I. Amer, Yang Liu, Gabor Sines
A typical video encoder includes into the generated bit stream Instantaneous Decoder Refresh (IDR) units. This allows random access playback at the receiver side as well as graceful recovery from potential channel errors. Such forced IDR units typically come in repetitive patterns, which may negatively impact the perceived subjective quality if not handled properly. The reason is that the restricted encoding process of an IDR unit results in a different (regardless higher or lower) quality of reconstructed signal compared to the surrounding non-IDR ones. This causes eye-capturing irritating periodical artifacts when it occurs in patterns. This phenomenon gets to be even more pronounced when the intra refresh feature is enabled, since it forces IDR and nonIDR units to co-exist within the same picture, making the quality difference more noticeable. This paper proposes a method to hide such undesired patterns that naturally accompany the intra refresh feature. Two ideas are presented; the first one imposes restrictions that prevent unwanted fluctuations in the quantization levels between different regions of the picture, while the second hides the repetitive pattern by randomly forcing IDR blocks within specific regions of the refreshed picture. Results show that the proposed method results in improvements in subjective quality.
典型的视频编码器在生成的比特流中包含瞬时解码器刷新(IDR)单元。这允许在接收端随机访问重放以及从潜在的信道错误中优雅地恢复。这种强制IDR单元通常以重复的模式出现,如果处理不当,可能会对感知到的主观质量产生负面影响。原因是IDR单元的编码过程受到限制,导致重构信号的质量与周围的非IDR信号不同(无论高或低)。当它出现在图案中时,这会引起眼球捕捉刺激的周期性伪影。当启用帧内刷新功能时,这种现象变得更加明显,因为它迫使IDR和非IDR单元在同一图像中共存,从而使质量差异更加明显。本文提出了一种方法来隐藏这些不希望的模式,这些模式自然伴随着内部刷新功能。提出了两个观点;第一种方法施加限制,防止图像不同区域之间量化水平的不必要波动,而第二种方法通过在刷新图像的特定区域内随机强制IDR块来隐藏重复模式。结果表明,该方法提高了主观素质。
{"title":"A Method to Improve Perceptual Quality of Intra- Refresh-Enabled Low-Latency Video Coding","authors":"Wei Gao, I. Amer, Yang Liu, Gabor Sines","doi":"10.1109/PCS.2018.8456257","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456257","url":null,"abstract":"A typical video encoder includes into the generated bit stream Instantaneous Decoder Refresh (IDR) units. This allows random access playback at the receiver side as well as graceful recovery from potential channel errors. Such forced IDR units typically come in repetitive patterns, which may negatively impact the perceived subjective quality if not handled properly. The reason is that the restricted encoding process of an IDR unit results in a different (regardless higher or lower) quality of reconstructed signal compared to the surrounding non-IDR ones. This causes eye-capturing irritating periodical artifacts when it occurs in patterns. This phenomenon gets to be even more pronounced when the intra refresh feature is enabled, since it forces IDR and nonIDR units to co-exist within the same picture, making the quality difference more noticeable. This paper proposes a method to hide such undesired patterns that naturally accompany the intra refresh feature. Two ideas are presented; the first one imposes restrictions that prevent unwanted fluctuations in the quantization levels between different regions of the picture, while the second hides the repetitive pattern by randomly forcing IDR blocks within specific regions of the refreshed picture. Results show that the proposed method results in improvements in subjective quality.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131329185","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Comparison of JEM and AV1 with HEVC: Coding Tools, Coding Efficiency and Complexity JEM和AV1与HEVC的比较:编码工具、编码效率和复杂度
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456291
Thorsten Laude, Y. G. Adhisantoso, Jan Voges, Marco Munderloh, J. Ostermann
The current state-of-the-art for standardized video codecs is High Efficiency Video Coding (HEVC) which was developed jointly by ISO/IEC and ITU-T. Recently, the development of two contenders for the next generation of standardized video codecs began: ISO/IEC and ITU-T advance the development of the Joint Exploration Model (JEM), a possible successor of HEVC, while the Alliance for Open Media pushes forward the video codec AV1. It is asserted by both groups that their codecs achieve superior coding efficiency over the state-of-the-art. In this paper, we discuss the distinguishing features of JEM and AV1 and evaluate their coding efficiency and computational complexity under well-defined and balanced test conditions. Our main findings are that JEM considerably outperforms HM and AV1 in terms of coding efficiency while AV1 cannot transform increased complexity into competitiveness in terms of coding efficiency with neither of the competitors except for the all-intra configuration.
目前最先进的标准化视频编解码器是由ISO/IEC和ITU-T联合开发的高效视频编码(HEVC)。最近,两个下一代标准化视频编解码器竞争者的开发开始了:ISO/IEC和ITU-T推进联合探索模型(JEM)的开发,这可能是HEVC的继任者,而开放媒体联盟则推动视频编解码器AV1的开发。这两组都声称他们的编解码器比最先进的编解码器具有更高的编码效率。本文讨论了JEM和AV1的区别特征,并在定义良好且平衡的测试条件下评估了它们的编码效率和计算复杂度。我们的主要发现是JEM在编码效率方面明显优于HM和AV1,而AV1在编码效率方面不能将增加的复杂性转化为竞争对手,除了全intra配置。
{"title":"A Comparison of JEM and AV1 with HEVC: Coding Tools, Coding Efficiency and Complexity","authors":"Thorsten Laude, Y. G. Adhisantoso, Jan Voges, Marco Munderloh, J. Ostermann","doi":"10.1109/PCS.2018.8456291","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456291","url":null,"abstract":"The current state-of-the-art for standardized video codecs is High Efficiency Video Coding (HEVC) which was developed jointly by ISO/IEC and ITU-T. Recently, the development of two contenders for the next generation of standardized video codecs began: ISO/IEC and ITU-T advance the development of the Joint Exploration Model (JEM), a possible successor of HEVC, while the Alliance for Open Media pushes forward the video codec AV1. It is asserted by both groups that their codecs achieve superior coding efficiency over the state-of-the-art. In this paper, we discuss the distinguishing features of JEM and AV1 and evaluate their coding efficiency and computational complexity under well-defined and balanced test conditions. Our main findings are that JEM considerably outperforms HM and AV1 in terms of coding efficiency while AV1 cannot transform increased complexity into competitiveness in terms of coding efficiency with neither of the competitors except for the all-intra configuration.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114414497","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Decoding JPEG XS on a GPU 在GPU上解码JPEG XS
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456310
Volker Bruns, T. Richter, Bilal Ahmed, J. Keinert, S. Fößel
JPEG XS is an upcoming lightweight image compression standard that is especially developed to meet the requirements of compressed video-over-IP use cases. It is designed with not only CPU, FPGA or ASIC platforms in mind, but explicitly also targets GPUs. Though not yet finished, the codec is now sufficiently mature to present a first NVIDIA CUDA-based GPU decoder architecture and preliminary performance results. On a 2014 mid-range GPU with 640 cores a 12 bit UHD 4:2:2 (4:4:4) can be decoded with 54 (42) fps. The algorithm scales very well: on a 2017 high-end GPU with 2560 cores the throughput increases to 190 (150) fps. In contrast, an optimized GPU-accelerated JPEG 2000 decoder takes 2x as long for high compression ratios that yield a PSNR of 40 dB and 3x as long for lower compression ratios with a PSNR of over 50 dB.
JPEG XS是一种即将推出的轻量级图像压缩标准,它是专门为满足ip视频压缩用例的要求而开发的。它的设计不仅考虑了CPU, FPGA或ASIC平台,而且明确地针对gpu。虽然还没有完成,编解码器现在已经足够成熟,可以展示第一个基于NVIDIA cuda的GPU解码器架构和初步性能结果。在2014年具有640核的中档GPU上,12位UHD 4:2:2(4:4:4)可以以54 (42)fps解码。该算法可扩展性非常好:在2017年具有2560核的高端GPU上,吞吐量增加到190 (150)fps。相比之下,经过优化的gpu加速的JPEG 2000解码器对于产生40 dB的PSNR的高压缩比需要2倍的时间,对于产生超过50 dB的PSNR的低压缩比需要3倍的时间。
{"title":"Decoding JPEG XS on a GPU","authors":"Volker Bruns, T. Richter, Bilal Ahmed, J. Keinert, S. Fößel","doi":"10.1109/PCS.2018.8456310","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456310","url":null,"abstract":"JPEG XS is an upcoming lightweight image compression standard that is especially developed to meet the requirements of compressed video-over-IP use cases. It is designed with not only CPU, FPGA or ASIC platforms in mind, but explicitly also targets GPUs. Though not yet finished, the codec is now sufficiently mature to present a first NVIDIA CUDA-based GPU decoder architecture and preliminary performance results. On a 2014 mid-range GPU with 640 cores a 12 bit UHD 4:2:2 (4:4:4) can be decoded with 54 (42) fps. The algorithm scales very well: on a 2017 high-end GPU with 2560 cores the throughput increases to 190 (150) fps. In contrast, an optimized GPU-accelerated JPEG 2000 decoder takes 2x as long for high compression ratios that yield a PSNR of 40 dB and 3x as long for lower compression ratios with a PSNR of over 50 dB.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126031557","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Optimized Non-local In-Loop Filter for Video Coding 优化的视频编码非局部环内滤波器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456299
Xuewei Meng, Chuanmin Jia, Shanshe Wang, Xiaozhen Zheng, Siwei Ma
In order to compensate the shortcomings of existing in-loop filters only based on local correlation in video coding standards, many non-local based loop filters with high coding performance and computational complexity are proposed. In this paper, we propose a fast block matching algorithm, adaptive two-step block matching algorithm, based on our previous work, structure-driven adaptive non-local filter (SANF) which is computationally intensive because of the high complexity of block matching and singular value decomposition (SVD). Our proposed algorithm based on image spatial statistical characteristics utilizes fixed template to select adaptive number of similar blocks according to image content, which can reduce up to 75.2% search candidates compared to exhaustive search in SANF and the adaptive determination strategy can remove blocks with less relation to reference block in similar block group which have little help for compression performance, and the remove of them can reduce the computational complexity of SVD. Our proposed optimization algorithm can save encoding and decoding time significantly with negligible performance loss, which achieves 70.7%, 84.4%, 80.82% and 81.95% decoding time saving with only 0.13%, 0.05%, 0.13% and 0.15% increases of BD-rate for AI, RA, LDB and LDP configurations, respectively compared to original SANF in JEM-7.0.
为了弥补现有视频编码标准中仅基于局部相关的环内滤波器的不足,提出了许多具有高编码性能和计算复杂度的非局部环内滤波器。本文提出了一种快速的块匹配算法——自适应两步块匹配算法,该算法是基于结构驱动的自适应非局部滤波器(SANF),该算法由于块匹配和奇异值分解(SVD)的高复杂度而导致计算量大。该算法基于图像空间统计特征,利用固定模板根据图像内容选择自适应数量的相似块,与SANF的穷举搜索相比,可减少75.2%的候选搜索量,并且自适应确定策略可以去除相似块组中与参考块关系较小且对压缩性能帮助不大的块,并且可以降低奇异值分解的计算复杂度。本文提出的优化算法可以显著节省编解码时间,性能损失可以忽略,在AI、RA、LDB和LDP配置下,译码时间分别节省70.7%、84.4%、80.82%和81.95%,译码率仅比JEM-7.0中原始SANF分别提高0.13%、0.05%、0.13%和0.15%。
{"title":"Optimized Non-local In-Loop Filter for Video Coding","authors":"Xuewei Meng, Chuanmin Jia, Shanshe Wang, Xiaozhen Zheng, Siwei Ma","doi":"10.1109/PCS.2018.8456299","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456299","url":null,"abstract":"In order to compensate the shortcomings of existing in-loop filters only based on local correlation in video coding standards, many non-local based loop filters with high coding performance and computational complexity are proposed. In this paper, we propose a fast block matching algorithm, adaptive two-step block matching algorithm, based on our previous work, structure-driven adaptive non-local filter (SANF) which is computationally intensive because of the high complexity of block matching and singular value decomposition (SVD). Our proposed algorithm based on image spatial statistical characteristics utilizes fixed template to select adaptive number of similar blocks according to image content, which can reduce up to 75.2% search candidates compared to exhaustive search in SANF and the adaptive determination strategy can remove blocks with less relation to reference block in similar block group which have little help for compression performance, and the remove of them can reduce the computational complexity of SVD. Our proposed optimization algorithm can save encoding and decoding time significantly with negligible performance loss, which achieves 70.7%, 84.4%, 80.82% and 81.95% decoding time saving with only 0.13%, 0.05%, 0.13% and 0.15% increases of BD-rate for AI, RA, LDB and LDP configurations, respectively compared to original SANF in JEM-7.0.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"375 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124688457","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Machine Learning Based Choice of Characteristics for the One-Shot Determination of the HEVC Intra Coding Tree 基于机器学习的HEVC编码树一次性确定特征选择
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456261
Alexandre Mercat, F. Arrestier, M. Pelcat, W. Hamidouche, D. Ménard
In the last few years, the Internet of Things (IoT) has become a reality. Forthcoming applications are likely to boost mobile video demand to an unprecedented level. A large number of systems are likely to integrate the latest MPEG video standard High Efficiency Video Coding (HEVC) in the long run and will particularly require energy efficiency. In this context, constraining the computational complexity of embedded HEVC encoders is a challenging task, especially in the case of software encoders. The most energy consuming part of a software intra encoder is the determination of the coding tree partitioning, i.e. the size of pixel blocks. This determination usually requires an iterative process that leads to repeating some encoding tasks. State-of-the-art studies have focused on predicting, from “easily” computed characteristics, an efficient coding tree. They have proposed and evaluated independently many characteristics for one-shot quad-tree prediction. In this paper, we present a fair comparison of these characteristics using a Machine Learning approach and a real-time HEVC encoder. Both computational complexity and information gain are considered, showing that characteristics are far from equivalent in terms of coding tree prediction performance.
在过去的几年里,物联网(IoT)已经成为现实。即将推出的应用程序可能会将移动视频需求提升到前所未有的水平。从长远来看,大量的系统可能会集成最新的MPEG视频标准HEVC (High Efficiency video Coding),并对能效提出了特别的要求。在这种情况下,限制嵌入式HEVC编码器的计算复杂度是一项具有挑战性的任务,特别是在软件编码器的情况下。软件内编码器最耗能的部分是编码树划分的确定,即像素块的大小。这种确定通常需要一个迭代过程,导致重复一些编码任务。最先进的研究集中在预测,从“容易”计算的特征,一个有效的编码树。他们提出并独立评估了单次四叉树预测的许多特征。在本文中,我们使用机器学习方法和实时HEVC编码器对这些特征进行了公平的比较。考虑了计算复杂度和信息增益,表明在编码树预测性能方面,特征远非相等。
{"title":"Machine Learning Based Choice of Characteristics for the One-Shot Determination of the HEVC Intra Coding Tree","authors":"Alexandre Mercat, F. Arrestier, M. Pelcat, W. Hamidouche, D. Ménard","doi":"10.1109/PCS.2018.8456261","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456261","url":null,"abstract":"In the last few years, the Internet of Things (IoT) has become a reality. Forthcoming applications are likely to boost mobile video demand to an unprecedented level. A large number of systems are likely to integrate the latest MPEG video standard High Efficiency Video Coding (HEVC) in the long run and will particularly require energy efficiency. In this context, constraining the computational complexity of embedded HEVC encoders is a challenging task, especially in the case of software encoders. The most energy consuming part of a software intra encoder is the determination of the coding tree partitioning, i.e. the size of pixel blocks. This determination usually requires an iterative process that leads to repeating some encoding tasks. State-of-the-art studies have focused on predicting, from “easily” computed characteristics, an efficient coding tree. They have proposed and evaluated independently many characteristics for one-shot quad-tree prediction. In this paper, we present a fair comparison of these characteristics using a Machine Learning approach and a real-time HEVC encoder. Both computational complexity and information gain are considered, showing that characteristics are far from equivalent in terms of coding tree prediction performance.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125436474","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Image-Based Rendering using Point Cloud for 2D Video Compression 使用点云进行2D视频压缩的基于图像的渲染
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456267
H. Golestani, Thibaut Meyer, M. Wien
The main idea of this paper is to extract the 3D scene geometry for the observed scene and use it for synthesizing a more precise prediction using Image-Based Rendering (IBR) for motion compensation in a hybrid coding scheme. The proposed method first extracts camera parameters using Structure from Motion (SfM). Then, a Patch-based Multi-View Stereo (PMVS) technique is employed to generate the scene Point-Cloud (PC) only from already decoded key-frames. Since the PC could be really sparse in poorly reconstructed regions, a depth expansion mechanism is also used. This 3D information helps to properly warp textures from the key-frames to the target frame. This IBR-based prediction is then used as an additional reference for motion compensation. In this way, the encoder can choose between the rendered prediction and the regular reference pictures through a rate- distortion optimization. On average, the simulation results show about 2.16% bitrate reduction compared to the reference HEVC implementation, for tested dynamic and static scene video sequences.
本文的主要思想是提取观察场景的三维场景几何形状,并使用基于图像的渲染(IBR)在混合编码方案中进行运动补偿,从而合成更精确的预测。该方法首先利用SfM (Structure from Motion)提取相机参数。然后,采用基于patch的多视点立体(PMVS)技术,仅从已解码的关键帧生成场景点云(PC)。由于PC在重建较差的区域可能非常稀疏,因此还使用了深度扩展机制。这个3D信息有助于正确地将纹理从关键帧扭曲到目标帧。这种基于ibr的预测然后用作运动补偿的附加参考。这样,编码器可以通过率失真优化在渲染的预测图像和常规参考图像之间进行选择。对于测试的动态和静态场景视频序列,仿真结果显示,与参考HEVC实现相比,平均比特率降低了2.16%。
{"title":"Image-Based Rendering using Point Cloud for 2D Video Compression","authors":"H. Golestani, Thibaut Meyer, M. Wien","doi":"10.1109/PCS.2018.8456267","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456267","url":null,"abstract":"The main idea of this paper is to extract the 3D scene geometry for the observed scene and use it for synthesizing a more precise prediction using Image-Based Rendering (IBR) for motion compensation in a hybrid coding scheme. The proposed method first extracts camera parameters using Structure from Motion (SfM). Then, a Patch-based Multi-View Stereo (PMVS) technique is employed to generate the scene Point-Cloud (PC) only from already decoded key-frames. Since the PC could be really sparse in poorly reconstructed regions, a depth expansion mechanism is also used. This 3D information helps to properly warp textures from the key-frames to the target frame. This IBR-based prediction is then used as an additional reference for motion compensation. In this way, the encoder can choose between the rendered prediction and the regular reference pictures through a rate- distortion optimization. On average, the simulation results show about 2.16% bitrate reduction compared to the reference HEVC implementation, for tested dynamic and static scene video sequences.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129131531","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Fully-Automatic Inverse Tone Mapping Preserving the Content Creator’s Artistic Intentions 全自动逆色调映射保存内容创作者的艺术意图
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456253
Gonzalo Luzardo, J. Aelterman, H. Luong, W. Philips, Daniel Ochoa, Sven Rousseaux
High Dynamic Range (HDR) displays can show images with higher color contrast levels and peak luminosities than the common Low Dynamic Range (LDR) displays. However, most existing video content is recorded and/or graded in LDR format. To show this LDR content on HDR displays, a dynamic range expansion by using an Inverse Tone Mapped Operator (iTMO) is required. In addition to requiring human intervention for tuning, most of the iTMOs don’t consider artistic intentions inherent to the HDR domain. Furthermore, the quality of their results decays with peak brightness above 1000 nits. In this paper, we propose a fully-automatic inverse tone mapping operator based on mid-level mapping. This allows expanding LDR images into HDR with peak brightness over 1000 nits, preserving the artistic intentions inherent to the HDR domain. We assessed our results using full-reference objective quality metrics as HDR- VDP-2.2 and DRIM. Experimental results demonstrate that our proposed method outperforms the current state of the art.
与普通的低动态范围(LDR)显示器相比,高动态范围(HDR)显示器可以显示具有更高色彩对比度水平和峰值亮度的图像。但是,大多数现有的视频内容都是以LDR格式录制和/或分级的。为了在HDR显示器上显示LDR内容,需要使用逆色调映射算子(iTMO)进行动态范围扩展。除了需要人为干预进行调优之外,大多数itmo都没有考虑到HDR领域固有的艺术意图。此外,当峰值亮度超过1000尼特时,其结果的质量就会下降。本文提出了一种基于中级映射的全自动逆调映射算子。这允许将LDR图像扩展为峰值亮度超过1000尼特的HDR,保留HDR域固有的艺术意图。我们使用全参考客观质量指标HDR- VDP-2.2和DRIM来评估结果。实验结果表明,我们提出的方法优于目前的技术水平。
{"title":"Fully-Automatic Inverse Tone Mapping Preserving the Content Creator’s Artistic Intentions","authors":"Gonzalo Luzardo, J. Aelterman, H. Luong, W. Philips, Daniel Ochoa, Sven Rousseaux","doi":"10.1109/PCS.2018.8456253","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456253","url":null,"abstract":"High Dynamic Range (HDR) displays can show images with higher color contrast levels and peak luminosities than the common Low Dynamic Range (LDR) displays. However, most existing video content is recorded and/or graded in LDR format. To show this LDR content on HDR displays, a dynamic range expansion by using an Inverse Tone Mapped Operator (iTMO) is required. In addition to requiring human intervention for tuning, most of the iTMOs don’t consider artistic intentions inherent to the HDR domain. Furthermore, the quality of their results decays with peak brightness above 1000 nits. In this paper, we propose a fully-automatic inverse tone mapping operator based on mid-level mapping. This allows expanding LDR images into HDR with peak brightness over 1000 nits, preserving the artistic intentions inherent to the HDR domain. We assessed our results using full-reference objective quality metrics as HDR- VDP-2.2 and DRIM. Experimental results demonstrate that our proposed method outperforms the current state of the art.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129746687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
360-Degree Video Quality Evaluation 360度视频质量评估
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456255
Philippe Hanhart, Yuwen He, Yan Ye, J. Boyce, Z. Deng, Lidong Xu
360-degree video is emerging as a new way of offering immersive visual experience. The quality evaluation of 360- degree video is more difficult compared to the quality evaluation of conventional video. However, to ensure successful development of 360-degree video coding technologies, it is essential to precisely measure both objective and subjective quality. In this paper, an overview of the 360-degree video quality evaluation framework established by the joint video exploration team (JVET) of ITU-T VCEG and ISO/IEC MPEG is provided. This framework aims at reproducing the different processes in the 360-degree video processing workflow that are related to coding. The results of different experiments conducted using the JVET framework are reported to illustrate the impact on objective and subjective quality with different projection formats and codecs.
360度视频正在成为一种提供沉浸式视觉体验的新方式。与传统视频的质量评价相比,360度视频的质量评价更加困难。然而,为了保证360度视频编码技术的成功发展,必须精确测量客观质量和主观质量。本文概述了ITU-T VCEG和ISO/IEC MPEG联合视频探索小组(JVET)建立的360度视频质量评估框架。该框架旨在再现与编码相关的360度视频处理工作流程中的不同过程。本文报告了使用JVET框架进行的不同实验的结果,以说明不同的投影格式和编解码器对客观和主观质量的影响。
{"title":"360-Degree Video Quality Evaluation","authors":"Philippe Hanhart, Yuwen He, Yan Ye, J. Boyce, Z. Deng, Lidong Xu","doi":"10.1109/PCS.2018.8456255","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456255","url":null,"abstract":"360-degree video is emerging as a new way of offering immersive visual experience. The quality evaluation of 360- degree video is more difficult compared to the quality evaluation of conventional video. However, to ensure successful development of 360-degree video coding technologies, it is essential to precisely measure both objective and subjective quality. In this paper, an overview of the 360-degree video quality evaluation framework established by the joint video exploration team (JVET) of ITU-T VCEG and ISO/IEC MPEG is provided. This framework aims at reproducing the different processes in the 360-degree video processing workflow that are related to coding. The results of different experiments conducted using the JVET framework are reported to illustrate the impact on objective and subjective quality with different projection formats and codecs.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124414602","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1