首页 > 最新文献

2018 Picture Coding Symposium (PCS)最新文献

英文 中文
Fully Connected Network for HEVC CU Split Decision equipped with Laplacian Transparent Composite Model 基于拉普拉斯透明复合模型的HEVC - CU拆分决策全连接网络
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456290
Hossam Amer, Abdullah M. Rashwan, E. Yang
High Efficiency Video Coding (HEVC) improves rate distortion (RD) performance significantly, but at the same time is computationally expensive due to the adoption of a large variety of coding unit (CU) sizes in its RD optimization. In this paper, we investigate the application of fully connected neural networks (NNs) to this time-sensitive application to improve its time complexity, while controlling the resulting bitrate loss. Specifically, four NNs are introduced with one NN for each depth of the coding tree unit. These NNs either split the current CU or terminate the CU search algorithm. Because training of NNs is time-consuming and requires large training data, we further propose a novel training strategy in which offline training and online adaptation work together to overcome this limitation. Our features are extracted from original frames based on the Laplacian Transparent Composite Model (LPTCM). Experiments carried out on all-intra configuration for HEVC reveal that our method is among the best NN methods, with an average time saving of 38% and an average controlled bitrate loss of 1.6%, compared to original HEVC.
高效视频编码(High Efficiency Video Coding, HEVC)可以显著提高码率失真(rate distortion, RD)性能,但同时由于在码率失真(rate distortion, RD)优化中采用了多种不同的编码单元(Coding unit, CU)尺寸,因此计算成本很高。在本文中,我们研究了全连接神经网络(NNs)在这种时间敏感应用中的应用,以提高其时间复杂度,同时控制由此产生的比特率损失。具体来说,引入了四个神经网络,每个神经网络对应编码树单元的每个深度。这些神经网络要么拆分当前的CU,要么终止CU搜索算法。由于神经网络的训练耗时且需要大量的训练数据,我们进一步提出了一种离线训练和在线适应相结合的新型训练策略来克服这一限制。我们的特征是基于拉普拉斯透明复合模型(LPTCM)从原始帧中提取的。在HEVC的全帧内配置上进行的实验表明,我们的方法是最好的神经网络方法之一,与原始HEVC相比,平均节省38%的时间,平均控制比特率损失为1.6%。
{"title":"Fully Connected Network for HEVC CU Split Decision equipped with Laplacian Transparent Composite Model","authors":"Hossam Amer, Abdullah M. Rashwan, E. Yang","doi":"10.1109/PCS.2018.8456290","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456290","url":null,"abstract":"High Efficiency Video Coding (HEVC) improves rate distortion (RD) performance significantly, but at the same time is computationally expensive due to the adoption of a large variety of coding unit (CU) sizes in its RD optimization. In this paper, we investigate the application of fully connected neural networks (NNs) to this time-sensitive application to improve its time complexity, while controlling the resulting bitrate loss. Specifically, four NNs are introduced with one NN for each depth of the coding tree unit. These NNs either split the current CU or terminate the CU search algorithm. Because training of NNs is time-consuming and requires large training data, we further propose a novel training strategy in which offline training and online adaptation work together to overcome this limitation. Our features are extracted from original frames based on the Laplacian Transparent Composite Model (LPTCM). Experiments carried out on all-intra configuration for HEVC reveal that our method is among the best NN methods, with an average time saving of 38% and an average controlled bitrate loss of 1.6%, compared to original HEVC.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122951301","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Content-Adaptive 360-Degree Video Coding Using Hybrid Cubemap Projection 使用混合立方体映射投影的内容自适应360度视频编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456280
Yuwen He, Xiaoyu Xiu, Philippe Hanhart, Yan Ye, Fanyi Duanmu, Yao Wang
In this paper, a novel hybrid cubemap projection (HCP) is proposed to improve the 360-degree video coding efficiency. HCP allows adaptive sampling adjustments in the horizontal and vertical directions within each cube face. HCP parameters of each cube face can be adjusted based on the input 360-degree video content characteristics for a better sampling efficiency. The HCP parameters can be updated periodically to adapt to temporal content variation. An efficient HCP parameter estimation algorithm is proposed to reduce the computational complexity of parameter estimation. Experimental results demonstrate that HCP format achieves on average luma (Y) BD-rate reduction of 11.51%, 8.0%, and 0.54% compared to equirectangular projection format, cubemap projection format, and adjusted cubemap projection format, respectively, in terms of end-to-end WS-PSNR.
为了提高360度视频编码效率,提出了一种新的混合立方体映射投影(HCP)算法。HCP允许在每个立方体表面的水平和垂直方向上进行自适应采样调整。每个立方体面的HCP参数可以根据输入的360度视频内容特性进行调整,以获得更好的采样效率。HCP参数可以定期更新以适应时间变化。为了降低参数估计的计算复杂度,提出了一种高效的HCP参数估计算法。实验结果表明,在端到端WS-PSNR方面,HCP格式与等矩形投影格式、立方体映射投影格式和调整立方体映射投影格式相比,平均亮度(Y) bd率分别降低了11.51%、8.0%和0.54%。
{"title":"Content-Adaptive 360-Degree Video Coding Using Hybrid Cubemap Projection","authors":"Yuwen He, Xiaoyu Xiu, Philippe Hanhart, Yan Ye, Fanyi Duanmu, Yao Wang","doi":"10.1109/PCS.2018.8456280","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456280","url":null,"abstract":"In this paper, a novel hybrid cubemap projection (HCP) is proposed to improve the 360-degree video coding efficiency. HCP allows adaptive sampling adjustments in the horizontal and vertical directions within each cube face. HCP parameters of each cube face can be adjusted based on the input 360-degree video content characteristics for a better sampling efficiency. The HCP parameters can be updated periodically to adapt to temporal content variation. An efficient HCP parameter estimation algorithm is proposed to reduce the computational complexity of parameter estimation. Experimental results demonstrate that HCP format achieves on average luma (Y) BD-rate reduction of 11.51%, 8.0%, and 0.54% compared to equirectangular projection format, cubemap projection format, and adjusted cubemap projection format, respectively, in terms of end-to-end WS-PSNR.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"31 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122983907","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Restricted Boltzmann Machine Image Compression 受限玻尔兹曼机图像压缩
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456279
Markus Kuchhold, Maik Simon, T. Sikora
We propose a novel lossy block-based image compression approach. Our approach builds on non-linear autoencoders that can, when properly trained, explore non-linear statistical dependencies in the image blocks for redundancy reduction. In contrast the DCT employed in JPEG is inherently restricted to exploration of linear dependencies using a second-order statistics framework. The coder is based on pre-trained class-specific Restricted Boltzmann Machines (RBM). These machines are statistical variants of neural network autoencoders that directly map pixel values in image blocks into coded bits. Decoders can be implemented with low computational complexity in a codebook design. Experimental results show that our RBM-codec outperforms JPEG at high compression rates, both in terms of PSNR, SSIM and subjective results.
提出了一种基于有损块的图像压缩方法。我们的方法建立在非线性自编码器的基础上,经过适当的训练,可以探索图像块中的非线性统计依赖关系以减少冗余。相比之下,JPEG中使用的DCT本质上仅限于使用二阶统计框架来探索线性依赖关系。编码器是基于预训练类特定的受限玻尔兹曼机(RBM)。这些机器是神经网络自动编码器的统计变体,直接将图像块中的像素值映射到编码位。解码器可以在码本设计中以较低的计算复杂度实现。实验结果表明,在高压缩率下,我们的rbm编解码器在PSNR、SSIM和主观结果方面都优于JPEG。
{"title":"Restricted Boltzmann Machine Image Compression","authors":"Markus Kuchhold, Maik Simon, T. Sikora","doi":"10.1109/PCS.2018.8456279","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456279","url":null,"abstract":"We propose a novel lossy block-based image compression approach. Our approach builds on non-linear autoencoders that can, when properly trained, explore non-linear statistical dependencies in the image blocks for redundancy reduction. In contrast the DCT employed in JPEG is inherently restricted to exploration of linear dependencies using a second-order statistics framework. The coder is based on pre-trained class-specific Restricted Boltzmann Machines (RBM). These machines are statistical variants of neural network autoencoders that directly map pixel values in image blocks into coded bits. Decoders can be implemented with low computational complexity in a codebook design. Experimental results show that our RBM-codec outperforms JPEG at high compression rates, both in terms of PSNR, SSIM and subjective results.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115516293","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Why JPEG is not JPEG — Testing a 25 years old Standard 为什么JPEG不是JPEG -测试一个25年前的标准
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456260
T. Richter, R. Clark
While ISO WG1 recently celebrated the 25th anniversary of its most successful standard, it seems to be more than surprising that up to now, no reference implementation of this standard exists. During an ongoing activity aiming at filling this gap, several observations have been made in how far the “living standard” deviates from the ISO documents. In particular, applying the official reference testing procedure of JPEG, available as ITU Recommendation T.83 or ISO/IEC 10918-2, turned out to be more a challenge than expected. This document sheds some light on the JPEG ISO standard, and our findings during reference testing a legacy, 25 year old standard.
虽然ISO WG1最近庆祝了其最成功的标准25周年,但令人惊讶的是,到目前为止,还没有这个标准的参考实施。在一项正在进行的旨在填补这一空白的活动中,对“生活标准”与ISO文件的偏离程度进行了一些观察。特别是,应用作为国际电联T.83建议书或ISO/IEC 10918-2提供的JPEG官方参考测试程序,结果比预期的更具挑战性。本文档介绍了JPEG ISO标准,以及我们在参考测试一个已有25年历史的遗留标准时的发现。
{"title":"Why JPEG is not JPEG — Testing a 25 years old Standard","authors":"T. Richter, R. Clark","doi":"10.1109/PCS.2018.8456260","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456260","url":null,"abstract":"While ISO WG1 recently celebrated the 25th anniversary of its most successful standard, it seems to be more than surprising that up to now, no reference implementation of this standard exists. During an ongoing activity aiming at filling this gap, several observations have been made in how far the “living standard” deviates from the ISO documents. In particular, applying the official reference testing procedure of JPEG, available as ITU Recommendation T.83 or ISO/IEC 10918-2, turned out to be more a challenge than expected. This document sheds some light on the JPEG ISO standard, and our findings during reference testing a legacy, 25 year old standard.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132541649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
PCS 2018 Title Page PCS 2018首页
Pub Date : 2018-06-01 DOI: 10.1109/pcs.2018.8456276
{"title":"PCS 2018 Title Page","authors":"","doi":"10.1109/pcs.2018.8456276","DOIUrl":"https://doi.org/10.1109/pcs.2018.8456276","url":null,"abstract":"","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125338905","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Hard Real-Time, Pixel-Parallel Rendering of Light Field Videos Using Steered Mixture-of-Experts 硬实时,像素并行渲染的光场视频使用操纵混合专家
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456306
Ignace P. Saenen, Ruben Verhack, Vasileios Avramelos, G. Wallendael, P. Lambert
Steered Mixture-of-Experts (SMoE) is a novel framework for the approximation, coding, and description of image modalities such as light field images and video. The future goal is to arrive at a representation for Six Degrees-of-Freedom (6DoF) image data. Previous research has shown the feasibility of real-time pixel-parallel rendering of static light field images. Each pixel is independently reconstructed by kernels that lay in its vicinity. The number of kernels involved forms the bottleneck on the achievable framerate. The goal of this paper is twofold. Firstly, we introduce pixel-level rendering of light field video, as previous work only rendered static content. Secondly, we investigate rendering using a predefined number of most significant kernels. As such, we can deliver hard real-time constraints by trading off the reconstruction quality.
导向混合专家(SMoE)是一种新的框架,用于逼近、编码和描述图像模式,如光场图像和视频。未来的目标是得到六自由度(6DoF)图像数据的表示。以往的研究已经证明了静态光场图像实时像素并行渲染的可行性。每个像素都由其附近的核独立重建。所涉及的内核数量构成了可实现帧率的瓶颈。本文的目的有两个。首先,我们介绍了光场视频的像素级渲染,因为以前的工作只渲染静态内容。其次,我们使用预定义的最重要核数来研究渲染。因此,我们可以通过权衡重建质量来交付硬实时约束。
{"title":"Hard Real-Time, Pixel-Parallel Rendering of Light Field Videos Using Steered Mixture-of-Experts","authors":"Ignace P. Saenen, Ruben Verhack, Vasileios Avramelos, G. Wallendael, P. Lambert","doi":"10.1109/PCS.2018.8456306","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456306","url":null,"abstract":"Steered Mixture-of-Experts (SMoE) is a novel framework for the approximation, coding, and description of image modalities such as light field images and video. The future goal is to arrive at a representation for Six Degrees-of-Freedom (6DoF) image data. Previous research has shown the feasibility of real-time pixel-parallel rendering of static light field images. Each pixel is independently reconstructed by kernels that lay in its vicinity. The number of kernels involved forms the bottleneck on the achievable framerate. The goal of this paper is twofold. Firstly, we introduce pixel-level rendering of light field video, as previous work only rendered static content. Secondly, we investigate rendering using a predefined number of most significant kernels. As such, we can deliver hard real-time constraints by trading off the reconstruction quality.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124054920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Two-layer Lossless HDR Coding considering Histogram Sparseness with Backward Compatibility to JPEG 考虑直方图稀疏性且向后兼容JPEG的双层无损HDR编码
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456254
Osamu Watanabe, H. Kobayashi, H. Kiya
An efficient two-layer coding method using the histogram packing technique with the backward compatibility to the legacy JPEG is proposed in this paper. The JPEG XT, which is the international standard to compress HDR images, adopts two-layer coding scheme for backward compatibility to the legacy JPEG. However, this two-layer coding structure does not give better lossless performance than the other existing single-layer coding methods for HDR images. Moreover, the JPEG XT has problems on determination of the lossless coding parameters; Finding appropriate combination of the parameter values is necessary to achieve good lossless performance. The histogram sparseness of HDR images is discussed and it is pointed out that the histogram packing technique considering the sparseness is able to improve the performance of lossless compression for HDR images and a novel two-layer coding with the histogram packing technique is proposed. The experimental results demonstrate that not only the proposed method has a better lossless compression performance than that of the JPEG XT, but also there is no need to determine image-dependent parameter values for good compression performance in spite of having the backward compatibility to the well known legacy JPEG standard.
本文提出了一种有效的双层编码方法,该方法采用直方图填充技术,并具有向后兼容传统JPEG格式的特点。JPEG XT是压缩HDR图像的国际标准,为了向后兼容传统JPEG,它采用了两层编码方案。然而,这种双层编码结构对于HDR图像的无损性能并不比现有的其他单层编码方法更好。此外,JPEG XT在无损编码参数的确定上存在问题;找到合适的参数值组合是实现良好无损性能的必要条件。讨论了HDR图像的直方图稀疏性,指出考虑稀疏性的直方图填充技术能够提高HDR图像的无损压缩性能,并提出了一种新的基于直方图填充技术的二层编码方法。实验结果表明,该方法不仅具有比JPEG XT更好的无损压缩性能,而且不需要确定与图像相关的参数值以获得良好的压缩性能,尽管它具有向后兼容众所周知的传统JPEG标准。
{"title":"Two-layer Lossless HDR Coding considering Histogram Sparseness with Backward Compatibility to JPEG","authors":"Osamu Watanabe, H. Kobayashi, H. Kiya","doi":"10.1109/PCS.2018.8456254","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456254","url":null,"abstract":"An efficient two-layer coding method using the histogram packing technique with the backward compatibility to the legacy JPEG is proposed in this paper. The JPEG XT, which is the international standard to compress HDR images, adopts two-layer coding scheme for backward compatibility to the legacy JPEG. However, this two-layer coding structure does not give better lossless performance than the other existing single-layer coding methods for HDR images. Moreover, the JPEG XT has problems on determination of the lossless coding parameters; Finding appropriate combination of the parameter values is necessary to achieve good lossless performance. The histogram sparseness of HDR images is discussed and it is pointed out that the histogram packing technique considering the sparseness is able to improve the performance of lossless compression for HDR images and a novel two-layer coding with the histogram packing technique is proposed. The experimental results demonstrate that not only the proposed method has a better lossless compression performance than that of the JPEG XT, but also there is no need to determine image-dependent parameter values for good compression performance in spite of having the backward compatibility to the well known legacy JPEG standard.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124065259","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Signal Adaptive Diffusion Filter For Video Coding 用于视频编码的信号自适应扩散滤波器
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456239
Jennifer Rasch, Jonathan Pfaff, Michael Schäfer, H. Schwarz, Martin Winken, Mischa Siekmann, D. Marpe, T. Wiegand
In this paper we combine state of the art video compression and Partial Differential Equation (PDE) based image processing methods. We introduce a new signal adaptive method to filter the predictions of a hybrid video codec using a system of PDEs describing a diffusion process. The method can be applied to intra as well as inter predictions. The filter is embedded into the framework of HEVC. The efficiency of the HEVC video codec is improved by up to −2.76% for All Intra and −3.56% for Random Access measured in Bjøntegaard delta (BD) rate. Coding gains of up to −8.76% can be observed for individual test sequences.
本文将目前最先进的视频压缩和基于偏微分方程的图像处理方法相结合。我们引入了一种新的信号自适应方法来过滤混合视频编解码器的预测,该方法使用描述扩散过程的pde系统。该方法既可用于内部预测,也可用于内部预测。该滤波器被嵌入到HEVC框架中。HEVC视频编解码器的所有Intra效率提高了- 2.76%,随机存取效率提高了- 3.56%(以Bjøntegaard delta (BD)率衡量)。对于单个测试序列,可以观察到高达- 8.76%的编码增益。
{"title":"A Signal Adaptive Diffusion Filter For Video Coding","authors":"Jennifer Rasch, Jonathan Pfaff, Michael Schäfer, H. Schwarz, Martin Winken, Mischa Siekmann, D. Marpe, T. Wiegand","doi":"10.1109/PCS.2018.8456239","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456239","url":null,"abstract":"In this paper we combine state of the art video compression and Partial Differential Equation (PDE) based image processing methods. We introduce a new signal adaptive method to filter the predictions of a hybrid video codec using a system of PDEs describing a diffusion process. The method can be applied to intra as well as inter predictions. The filter is embedded into the framework of HEVC. The efficiency of the HEVC video codec is improved by up to −2.76% for All Intra and −3.56% for Random Access measured in Bjøntegaard delta (BD) rate. Coding gains of up to −8.76% can be observed for individual test sequences.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"82 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123705502","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Object-Based Motion Estimation Using the EPD Similarity Measure 基于EPD相似度量的目标运动估计
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456287
Md. Asikuzzaman, M. Pickering
Effective motion compensated prediction plays a significant role in efficient video compression. Image registration can be used to estimate the motion of the scene in a frame by finding the geometric transformation which automatically aligns reference and target images. In the video coding literature, image registration has been applied to find the global motion in a video frame. However, if the motion of individual objects in a frame is inconsistent across time, the global motion may provide a very inefficient representation of the true motion present in the scene. In this paper we propose a motion estimation algorithm for video coding using a new similarity measure called the edge position difference (EPD). This technique estimates the motion of the individual objects based on matching the edges of objects rather than estimating the motion using the pixel values in the frame. Experimental results demonstrate that the proposed edge-based similarity measure approach achieves superior motion compensated prediction for objects in a scene when compared to the approach which only considers the pixel values of the frame.
有效的运动补偿预测是实现高效视频压缩的重要手段。图像配准可以通过寻找自动对齐参考图像和目标图像的几何变换来估计一帧内场景的运动。在视频编码文献中,图像配准已被用于寻找视频帧中的全局运动。然而,如果一个帧中单个物体的运动在时间上是不一致的,那么全局运动可能会提供一个非常低效的场景中真实运动的表示。本文提出了一种基于边缘位置差(EPD)的视频编码运动估计算法。该技术基于匹配对象的边缘来估计单个对象的运动,而不是使用帧中的像素值来估计运动。实验结果表明,与仅考虑帧像素值的方法相比,基于边缘的相似性度量方法对场景中物体的运动补偿预测效果更好。
{"title":"Object-Based Motion Estimation Using the EPD Similarity Measure","authors":"Md. Asikuzzaman, M. Pickering","doi":"10.1109/PCS.2018.8456287","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456287","url":null,"abstract":"Effective motion compensated prediction plays a significant role in efficient video compression. Image registration can be used to estimate the motion of the scene in a frame by finding the geometric transformation which automatically aligns reference and target images. In the video coding literature, image registration has been applied to find the global motion in a video frame. However, if the motion of individual objects in a frame is inconsistent across time, the global motion may provide a very inefficient representation of the true motion present in the scene. In this paper we propose a motion estimation algorithm for video coding using a new similarity measure called the edge position difference (EPD). This technique estimates the motion of the individual objects based on matching the edges of objects rather than estimating the motion using the pixel values in the frame. Experimental results demonstrate that the proposed edge-based similarity measure approach achieves superior motion compensated prediction for objects in a scene when compared to the approach which only considers the pixel values of the frame.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"37 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115628242","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Progressive Modeling of Steered Mixture-of-Experts for Light Field Video Approximation 用于光场视频逼近的导向混合专家渐进建模
Pub Date : 2018-06-01 DOI: 10.1109/PCS.2018.8456242
Ruben Verhack, G. Wallendael, Martijn Courteaux, P. Lambert, T. Sikora
Steered Mixture-of-Experts (SMoE) is a novel framework for the approximation, coding, and description of image modalities. The future goal is to arrive at a representation for Six Degrees-of-Freedom (6DoF) image data. The goal of this paper is to introduce SMoE for 4D light field videos by including the temporal dimension. However, these videos contain vast amounts of samples due to the large number of views per frame. Previous work on static light field images mitigated the problem by hard subdividing the modeling problem. However, such a hard subdivision introduces visually disturbing block artifacts on moving objects in dynamic image data. We propose a novel modeling method that does not result in block artifacts while minimizing the computational complexity and which allows for a varying spread of kernels in the spatio-temporal domain. Experiments validate that we can progressively model light field videos with increasing objective quality up to 0.97 SSIM.
导向混合专家(SMoE)是一种用于图像模态逼近、编码和描述的新框架。未来的目标是得到六自由度(6DoF)图像数据的表示。本文的目标是通过包含时间维度来引入四维光场视频的SMoE。然而,由于每帧的观看次数很多,这些视频包含了大量的样本。以前在静态光场图像上的工作通过对建模问题进行硬细分来缓解这个问题。然而,这种硬细分会在动态图像数据中对运动物体引入视觉干扰的块伪影。我们提出了一种新的建模方法,该方法不会导致块伪影,同时最大限度地降低了计算复杂性,并允许核在时空域中的不同分布。实验证明,我们可以逐步模拟光场视频,提高物镜质量,最高可达0.97 SSIM。
{"title":"Progressive Modeling of Steered Mixture-of-Experts for Light Field Video Approximation","authors":"Ruben Verhack, G. Wallendael, Martijn Courteaux, P. Lambert, T. Sikora","doi":"10.1109/PCS.2018.8456242","DOIUrl":"https://doi.org/10.1109/PCS.2018.8456242","url":null,"abstract":"Steered Mixture-of-Experts (SMoE) is a novel framework for the approximation, coding, and description of image modalities. The future goal is to arrive at a representation for Six Degrees-of-Freedom (6DoF) image data. The goal of this paper is to introduce SMoE for 4D light field videos by including the temporal dimension. However, these videos contain vast amounts of samples due to the large number of views per frame. Previous work on static light field images mitigated the problem by hard subdividing the modeling problem. However, such a hard subdivision introduces visually disturbing block artifacts on moving objects in dynamic image data. We propose a novel modeling method that does not result in block artifacts while minimizing the computational complexity and which allows for a varying spread of kernels in the spatio-temporal domain. Experiments validate that we can progressively model light field videos with increasing objective quality up to 0.97 SSIM.","PeriodicalId":433667,"journal":{"name":"2018 Picture Coding Symposium (PCS)","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128761518","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
2018 Picture Coding Symposium (PCS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1