首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Focus on visual rendering quality through content-based depth map coding 通过基于内容的深度图编码,专注于视觉渲染质量
Pub Date : 2010-12-07 DOI: 10.1109/PCS.2010.5702448
Emilie Bosc, M. Pressigout, L. Morin
Multi-view video plus depth (MVD) data is a set of multiple sequences capturing the same scene at different viewpoints, with their associated per-pixel depth value. Overcoming this large amount of data requires an effective coding framework. Yet, a simple but essential question refers to the means assessing the proposed coding methods. While the challenge in compression is the optimization of the rate-distortion ratio, a widely used objective metric to evaluate the distortion is the Peak-Signal-to-Noise-Ratio (PSNR), because of its simplicity and mathematically easiness to deal with such purposes. This paper points out the problem of reliability, concerning this metric, when estimating 3D video codec performances. We investigated the visual performances of two methods, namely H.264/MVC and Locally Adaptive Resolution (LAR) method, by encoding depth maps and reconstructing existing views from those degraded depth images. The experiments revealed that lower coding efficiency, in terms of PSNR, does not imply a lower rendering visual quality and that LAR method preserves the depth map properties correctly.
多视图视频加深度(MVD)数据是一组在不同视点捕获相同场景的多个序列,具有相关的每像素深度值。克服这些大量的数据需要一个有效的编码框架。然而,一个简单但重要的问题涉及到评估所提出的编码方法的手段。虽然压缩中的挑战是率失真比的优化,但由于峰值信噪比(PSNR)简单且在数学上易于处理,因此广泛使用客观度量来评估失真。本文指出了在评估3D视频编解码器性能时,可靠性指标的问题。研究了H.264/MVC和局部自适应分辨率(local Adaptive Resolution, LAR)两种方法的视觉性能,分别对深度图进行编码,并从退化的深度图像中重建现有视图。实验表明,较低的编码效率(就PSNR而言)并不意味着较低的渲染视觉质量,并且LAR方法正确地保留了深度图属性。
{"title":"Focus on visual rendering quality through content-based depth map coding","authors":"Emilie Bosc, M. Pressigout, L. Morin","doi":"10.1109/PCS.2010.5702448","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702448","url":null,"abstract":"Multi-view video plus depth (MVD) data is a set of multiple sequences capturing the same scene at different viewpoints, with their associated per-pixel depth value. Overcoming this large amount of data requires an effective coding framework. Yet, a simple but essential question refers to the means assessing the proposed coding methods. While the challenge in compression is the optimization of the rate-distortion ratio, a widely used objective metric to evaluate the distortion is the Peak-Signal-to-Noise-Ratio (PSNR), because of its simplicity and mathematically easiness to deal with such purposes. This paper points out the problem of reliability, concerning this metric, when estimating 3D video codec performances. We investigated the visual performances of two methods, namely H.264/MVC and Locally Adaptive Resolution (LAR) method, by encoding depth maps and reconstructing existing views from those degraded depth images. The experiments revealed that lower coding efficiency, in terms of PSNR, does not imply a lower rendering visual quality and that LAR method preserves the depth map properties correctly.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121273455","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
A novel video coding scheme for Super Hi-Vision 一种新的超高清视频编码方案
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702498
Shun-ichi Sekiguchi, Akira Minezawa, K. Sugimoto, A. Ichigaya, Kazuhisa Iguchi, Y. Shishikui
We propose a novel video coding scheme targeting Super Hi-Vision (SHV) video sources. While it takes a conventional block-based MC + Transform hybrid coding approach that is suitable for hardware implementation of a SHV video codec, the proposed scheme achieved significant coding efficiency improvement by introducing several coding tools such as intra prediction and adaptive transform. According to our experimental analysis, the proposed scheme achieves significant bit-rate saving compared to the state-of-the-art AVC/H.264 high profile.
针对超高清视频源,提出了一种新的视频编码方案。该方案采用传统的基于分块的MC + Transform混合编码方法,适合SHV视频编解码器的硬件实现,通过引入帧内预测和自适应变换等编码工具,显著提高了编码效率。根据我们的实验分析,与最先进的AVC/H相比,该方案实现了显着的比特率节省。264高调。
{"title":"A novel video coding scheme for Super Hi-Vision","authors":"Shun-ichi Sekiguchi, Akira Minezawa, K. Sugimoto, A. Ichigaya, Kazuhisa Iguchi, Y. Shishikui","doi":"10.1109/PCS.2010.5702498","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702498","url":null,"abstract":"We propose a novel video coding scheme targeting Super Hi-Vision (SHV) video sources. While it takes a conventional block-based MC + Transform hybrid coding approach that is suitable for hardware implementation of a SHV video codec, the proposed scheme achieved significant coding efficiency improvement by introducing several coding tools such as intra prediction and adaptive transform. According to our experimental analysis, the proposed scheme achieves significant bit-rate saving compared to the state-of-the-art AVC/H.264 high profile.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"91 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124422331","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Efficient free viewpoint video-on-demand scheme realizing walk-through experience 高效的免费视点视频点播方案,实现漫游体验
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702452
A. Ishikawa, Hiroshi Sankoh, S. Naito, S. Sakazawa
This paper presents an efficient video-on-demand (VOD) scheme for free viewpoint television (FTV), and proposes a data format and its data generation method to provide a walkthrough experience. We employ a hybrid rendering approach to describe a 3D scene using 3D model data for objects and textures. However, conventional hybrid rendering methods such as multi-texturing include excessive redundancy in texture data and demand a great deal of bandwidth to transmit. In this paper we propose an efficient texture data format, which removes the redundancy due to occlusion of objects by employing an orthogonal projection image for each object. The additional advantage of the data format is great simplification at the server to choose the transmitted images that correspond to the requested viewpoint. Experiments using multiview real video sequences confirm that the proposed scheme can reduce the transmission of texture data by as much as 42% compared to the conventional scheme.
本文提出了一种高效的免费视点电视(FTV)视频点播(VOD)方案,并提出了一种数据格式及其数据生成方法,以提供一种演练体验。我们采用混合渲染方法来描述一个3D场景,使用对象和纹理的3D模型数据。然而,传统的混合渲染方法(如多重纹理)在纹理数据中存在过多的冗余,并且需要大量的带宽传输。本文提出了一种有效的纹理数据格式,该格式通过对每个物体使用正交投影图像来消除由于物体遮挡造成的冗余。数据格式的另一个优点是,在服务器选择与所请求的视点相对应的传输图像时,可以大大简化。使用多视点真实视频序列的实验证明,与传统方案相比,该方案可以减少42%的纹理数据传输。
{"title":"Efficient free viewpoint video-on-demand scheme realizing walk-through experience","authors":"A. Ishikawa, Hiroshi Sankoh, S. Naito, S. Sakazawa","doi":"10.1109/PCS.2010.5702452","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702452","url":null,"abstract":"This paper presents an efficient video-on-demand (VOD) scheme for free viewpoint television (FTV), and proposes a data format and its data generation method to provide a walkthrough experience. We employ a hybrid rendering approach to describe a 3D scene using 3D model data for objects and textures. However, conventional hybrid rendering methods such as multi-texturing include excessive redundancy in texture data and demand a great deal of bandwidth to transmit. In this paper we propose an efficient texture data format, which removes the redundancy due to occlusion of objects by employing an orthogonal projection image for each object. The additional advantage of the data format is great simplification at the server to choose the transmitted images that correspond to the requested viewpoint. Experiments using multiview real video sequences confirm that the proposed scheme can reduce the transmission of texture data by as much as 42% compared to the conventional scheme.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"101 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124797228","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
An adaptive low-complexity global motion estimation algorithm 一种自适应低复杂度全局运动估计算法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702574
Md. Nazmul Haque, Moyuresh Biswas, M. Pickering, M. Frater
One important recent application of image registration has been in the estimation of global motion parameters for object-based video coding. A limitation of current global motion estimation approaches is the additional complexity of the gradient-descent optimization that is typically required to calculate the optimal set of global motion parameters. In this paper we propose a new low-complexity algorithm for global motion estimation. The complexity of the proposed algorithm is reduced by performing the majority of the operations in the gradient-descent optimization using logic operations rather than full-precision arithmetic operations. This use of logic operations means that the algorithm can be implemented much more easily in hardware platforms such as field programmable gate arrays (FPGAs). Experimental results show that the execution time for software implementations of the new algorithm is reduced by a factor of almost four when compared to existing fast implementations without any significant loss in registration accuracy.
图像配准最近的一个重要应用是基于对象的视频编码的全局运动参数估计。当前全局运动估计方法的一个局限性是梯度下降优化的额外复杂性,这通常需要计算全局运动参数的最优集。本文提出了一种新的低复杂度全局运动估计算法。本文提出的梯度下降优化算法采用逻辑运算而不是全精度算术运算,从而降低了算法的复杂度。这种逻辑运算的使用意味着该算法可以更容易地在硬件平台上实现,如现场可编程门阵列(fpga)。实验结果表明,与现有的快速算法相比,新算法的软件执行时间减少了近四倍,而配准精度没有明显下降。
{"title":"An adaptive low-complexity global motion estimation algorithm","authors":"Md. Nazmul Haque, Moyuresh Biswas, M. Pickering, M. Frater","doi":"10.1109/PCS.2010.5702574","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702574","url":null,"abstract":"One important recent application of image registration has been in the estimation of global motion parameters for object-based video coding. A limitation of current global motion estimation approaches is the additional complexity of the gradient-descent optimization that is typically required to calculate the optimal set of global motion parameters. In this paper we propose a new low-complexity algorithm for global motion estimation. The complexity of the proposed algorithm is reduced by performing the majority of the operations in the gradient-descent optimization using logic operations rather than full-precision arithmetic operations. This use of logic operations means that the algorithm can be implemented much more easily in hardware platforms such as field programmable gate arrays (FPGAs). Experimental results show that the execution time for software implementations of the new algorithm is reduced by a factor of almost four when compared to existing fast implementations without any significant loss in registration accuracy.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115034438","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Advanced inpainting-based macroblock prediction with regularized structure propagation in video compression 视频压缩中基于图像绘制的正则化结构宏块预测
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702587
Yang Xu, H. Xiong
In this paper, we propose an optimized inpainting-based macroblock (MB) prediction mode (IP-mode) in the state-of-the-art H.264/AVC video compression engine, and investigate a natural extension of structured sparsity over the ordered Belief Propagation (BP) inference in inpainting-based prediction. The IP-mode is regularized by a global spatio-temporal consistency between the predicted content and the co-located known texture, and could be adopted in both Intra and Inter frames without redundant assistant information. It is solved by an optimization problem under Markov Random Field (MRF), and the structured sparsity of the predicted macroblock region is inferred by tensor voting projected from the decoded regions to tune the priority of message scheduling in BP with a more convergent manner. Rate-distortion optimization is maintained to select the optimal mode among the inpainting-based prediction (IP-), the intra-, and inter-modes. Compared to the existing prediction modes in H.264/AVC, the proposed inpainting-based prediction scheme is validated to achieve a better R-D performance for homogeneous visual patterns and behave a more robust error resilience capability with an intrinsic probabilistic inference.
本文在最先进的H.264/AVC视频压缩引擎中提出了一种优化的基于绘画的宏块(MB)预测模式(ip模式),并研究了结构化稀疏性在基于绘画的预测中有序信念传播(BP)推理的自然扩展。ip模式通过预测内容与同一位置的已知纹理之间的全局时空一致性进行正则化,可以在帧内和帧间使用,不需要冗余的辅助信息。利用马尔可夫随机场(MRF)下的优化问题来解决该问题,并通过从解码区域投影的张量投票来推断预测宏块区域的结构化稀疏性,从而以更收敛的方式调整BP中消息调度的优先级。保持速率失真优化,以在基于涂漆的预测(IP-),内部和内部模式中选择最佳模式。与H.264/AVC中现有的预测模式相比,所提出的基于图像绘制的预测方案对同质视觉模式具有更好的R-D性能,并且具有更强的固有概率推理的容错能力。
{"title":"Advanced inpainting-based macroblock prediction with regularized structure propagation in video compression","authors":"Yang Xu, H. Xiong","doi":"10.1109/PCS.2010.5702587","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702587","url":null,"abstract":"In this paper, we propose an optimized inpainting-based macroblock (MB) prediction mode (IP-mode) in the state-of-the-art H.264/AVC video compression engine, and investigate a natural extension of structured sparsity over the ordered Belief Propagation (BP) inference in inpainting-based prediction. The IP-mode is regularized by a global spatio-temporal consistency between the predicted content and the co-located known texture, and could be adopted in both Intra and Inter frames without redundant assistant information. It is solved by an optimization problem under Markov Random Field (MRF), and the structured sparsity of the predicted macroblock region is inferred by tensor voting projected from the decoded regions to tune the priority of message scheduling in BP with a more convergent manner. Rate-distortion optimization is maintained to select the optimal mode among the inpainting-based prediction (IP-), the intra-, and inter-modes. Compared to the existing prediction modes in H.264/AVC, the proposed inpainting-based prediction scheme is validated to achieve a better R-D performance for homogeneous visual patterns and behave a more robust error resilience capability with an intrinsic probabilistic inference.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115202528","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Memory-efficient parallelization of JPEG-LS with relaxed context update 具有轻松上下文更新的JPEG-LS的内存高效并行化
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702444
S. Wahl, Zhe Wang, Chensheng Qiu, M. Wróblewski, L. Rockstroh, S. Simon
Many state-of-the-art lossless image compression standards feature adaptive error modelling. This, however, leads to data dependency loops of the compression scheme such that a parallel compression of neighboring pixels is not possible. In this paper, we propose a relaxation to the context update of JPEG-LS by delaying the update procedure, in order to achieve a guaranteed degree of parallelism with a negligible effect on the compression ratio. The lossless mode of JPEG-LS including the run-mode is considered. A descewing scheme is provided generating a bit-stream that preserves the order needed for the decoder to mimic the prediction in a consistent way. This system is memory efficient in a sense that no additional memory for the large context-set is needed.
许多最先进的无损图像压缩标准具有自适应误差建模。然而,这导致压缩方案的数据依赖循环,使得相邻像素的并行压缩是不可能的。在本文中,我们提出通过延迟更新过程来放松JPEG-LS的上下文更新,以便在对压缩比影响可以忽略不计的情况下获得保证的并行度。考虑了包括运行模式在内的JPEG-LS的无损模式。提供了一种解码方案,该方案生成的比特流保留了解码器以一致的方式模拟预测所需的顺序。从某种意义上说,该系统是内存高效的,不需要为大型上下文集提供额外的内存。
{"title":"Memory-efficient parallelization of JPEG-LS with relaxed context update","authors":"S. Wahl, Zhe Wang, Chensheng Qiu, M. Wróblewski, L. Rockstroh, S. Simon","doi":"10.1109/PCS.2010.5702444","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702444","url":null,"abstract":"Many state-of-the-art lossless image compression standards feature adaptive error modelling. This, however, leads to data dependency loops of the compression scheme such that a parallel compression of neighboring pixels is not possible. In this paper, we propose a relaxation to the context update of JPEG-LS by delaying the update procedure, in order to achieve a guaranteed degree of parallelism with a negligible effect on the compression ratio. The lossless mode of JPEG-LS including the run-mode is considered. A descewing scheme is provided generating a bit-stream that preserves the order needed for the decoder to mimic the prediction in a consistent way. This system is memory efficient in a sense that no additional memory for the large context-set is needed.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122088327","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Enhanced video compression with region-based texture models 基于区域纹理模型的增强视频压缩
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702560
Fan Zhang, D. Bull
This paper presents a region-based video compression algorithm based on texture warping and synthesis. Instead of encoding whole images or prediction residuals after translational motion estimation, this algorithm employs a perspective motion model to warp static textures and uses a texture synthesis approach to synthesise dynamic textures. Spatial and temporal artefacts are prevented by an in-loop video quality assessment module. The proposed method has been integrated into an H.264 video coding framework. The results show significant bitrate savings, up to 55%, compared with H.264, for similar visual quality.
提出了一种基于纹理变形和合成的基于区域的视频压缩算法。该算法不需要对整幅图像进行编码,也不需要对平移运动估计后的残差进行预测,而是采用透视运动模型对静态纹理进行翘曲,采用纹理合成方法对动态纹理进行合成。通过循环视频质量评估模块防止空间和时间伪影。该方法已被集成到H.264视频编码框架中。结果显示,在相同的视觉质量下,与H.264相比,比特率节省了55%。
{"title":"Enhanced video compression with region-based texture models","authors":"Fan Zhang, D. Bull","doi":"10.1109/PCS.2010.5702560","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702560","url":null,"abstract":"This paper presents a region-based video compression algorithm based on texture warping and synthesis. Instead of encoding whole images or prediction residuals after translational motion estimation, this algorithm employs a perspective motion model to warp static textures and uses a texture synthesis approach to synthesise dynamic textures. Spatial and temporal artefacts are prevented by an in-loop video quality assessment module. The proposed method has been integrated into an H.264 video coding framework. The results show significant bitrate savings, up to 55%, compared with H.264, for similar visual quality.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122223738","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Analysis of in-loop denoising in lossy transform coding 有损变换编码中的环内去噪分析
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702584
Eugen Wige, Gilbert Yammine, P. Amon, A. Hutter, André Kaup
When compressing noisy image sequences, the compression efficiency is limited by the noise amount within these image sequences as the noise part cannot be predicted. In this paper, we investigate the influence of noise within the reference frame on lossy video coding of noisy image sequences. We estimate how much noise is left within a lossy coded reference frame. Therefore we analyze the transform and quantization step inside a hybrid video coder, specifically H.264/AVC. The noise power after transform, quantization, and inverse transform is calculated analytically. We use knowledge of the noise power within the reference frame in order to improve the inter frame prediction. For noise filtering of the reference frame, we implemented a simple denoising algorithm inside the H.264/AVC reference software JM15.1. We show that the bitrate can be decreased by up to 8.1% compared to the H.264/AVC standard for high resolution noisy image sequences.
在压缩噪声图像序列时,由于噪声部分无法预测,压缩效率受到图像序列内噪声量的限制。本文研究了参考帧内噪声对含噪图像序列有损视频编码的影响。我们估计在有损编码参考帧内留下多少噪声。因此,我们分析了混合视频编码器内部的变换和量化步骤,特别是H.264/AVC。对变换、量化和逆变换后的噪声功率进行了解析计算。我们利用参考帧内噪声功率的知识来改进帧间预测。对于参考帧的噪声滤波,我们在H.264/AVC参考软件JM15.1中实现了一种简单的去噪算法。我们表明,与H.264/AVC标准相比,对于高分辨率噪声图像序列,比特率可以降低8.1%。
{"title":"Analysis of in-loop denoising in lossy transform coding","authors":"Eugen Wige, Gilbert Yammine, P. Amon, A. Hutter, André Kaup","doi":"10.1109/PCS.2010.5702584","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702584","url":null,"abstract":"When compressing noisy image sequences, the compression efficiency is limited by the noise amount within these image sequences as the noise part cannot be predicted. In this paper, we investigate the influence of noise within the reference frame on lossy video coding of noisy image sequences. We estimate how much noise is left within a lossy coded reference frame. Therefore we analyze the transform and quantization step inside a hybrid video coder, specifically H.264/AVC. The noise power after transform, quantization, and inverse transform is calculated analytically. We use knowledge of the noise power within the reference frame in order to improve the inter frame prediction. For noise filtering of the reference frame, we implemented a simple denoising algorithm inside the H.264/AVC reference software JM15.1. We show that the bitrate can be decreased by up to 8.1% compared to the H.264/AVC standard for high resolution noisy image sequences.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117314785","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Free viewpoint image generation with super resolution 免费视点图像生成与超分辨率
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702462
Norishige Fukushima, Y. Ishibashi
In this paper, we propose a method of free viewpoint image generation with super resolution. In the conventional approaches, such as nearest neighbor and linear interpolation, the synthetic image on zoomed virtual view tends to have low resolution, because the reference images do not have enough textures. To overcome this problem, we reconstruct the image with super resolution. Super resolution can generate higher image resolution than the input image one, and then we combine super resolution with free viewpoint image generation. In the experiment, we use a camera array which contains 11 × 11 aligned cameras and use 4 × 4 cameras subset per pixel to reconstruct image by means of super resolution. The experimental results show that synthesized image in the effective range has about 4.5 dB higher PSNR than ones created by the nearest neighbor and 2.5 dB higher than ones created by the linear interpolation.
本文提出了一种超分辨率的自由视点图像生成方法。在传统的最近邻插值和线性插值方法中,由于参考图像纹理不够,在放大虚拟视图上合成的图像往往分辨率较低。为了克服这个问题,我们对图像进行了超分辨率重建。超分辨率可以生成比输入图像更高的图像分辨率,然后将超分辨率与自由视点图像生成相结合。在实验中,我们使用包含11 × 11个对齐相机的相机阵列,每像素使用4 × 4个相机子集,通过超分辨率重建图像。实验结果表明,在有效范围内合成的图像比最近邻图像的PSNR高4.5 dB左右,比线性插值图像的PSNR高2.5 dB左右。
{"title":"Free viewpoint image generation with super resolution","authors":"Norishige Fukushima, Y. Ishibashi","doi":"10.1109/PCS.2010.5702462","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702462","url":null,"abstract":"In this paper, we propose a method of free viewpoint image generation with super resolution. In the conventional approaches, such as nearest neighbor and linear interpolation, the synthetic image on zoomed virtual view tends to have low resolution, because the reference images do not have enough textures. To overcome this problem, we reconstruct the image with super resolution. Super resolution can generate higher image resolution than the input image one, and then we combine super resolution with free viewpoint image generation. In the experiment, we use a camera array which contains 11 × 11 aligned cameras and use 4 × 4 cameras subset per pixel to reconstruct image by means of super resolution. The experimental results show that synthesized image in the effective range has about 4.5 dB higher PSNR than ones created by the nearest neighbor and 2.5 dB higher than ones created by the linear interpolation.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122276141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
A new hybrid parallel intra coding method based on interpolative prediction 一种新的基于插值预测的混合并行内编码方法
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702521
Cui Wang, Akira Kubota, Y. Hatori
The hybrid coding method combining the predictive coding with the orthogonal transformation and the quantization is mainly used recently. This paper proposes a new hybrid parallel Intra Coding based on interpolative prediction which uses correlations between neighboring pixels, including non-causal pixels. In order to get high prediction performance, the optimal quantizing scheme, which is used to cancel the error that expands when decoding, is used. Furthermore, a new type of block shape, which enables parallel coding, is proposed to simplify the processing of interpolative prediction. The result of comparison between proposed method and intra coding method in H.264 shows that the PSNR of proposed technique achieves 1 dB to 4 dB improvement in Luminance, especially for image with more details.
将预测编码与正交变换和量化相结合的混合编码方法是目前主要采用的编码方法。提出了一种基于插值预测的混合并行内编码方法,该方法利用相邻像素之间的相关性,包括非因果像素之间的相关性。为了获得较高的预测性能,采用了最优量化方案来抵消解码时扩大的误差。在此基础上,提出了一种能够实现并行编码的块形状,简化了插值预测的处理。将该方法与H.264的帧内编码方法进行了比较,结果表明,该方法的PSNR在亮度上提高了1 ~ 4 dB,特别是对于细节较多的图像。
{"title":"A new hybrid parallel intra coding method based on interpolative prediction","authors":"Cui Wang, Akira Kubota, Y. Hatori","doi":"10.1109/PCS.2010.5702521","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702521","url":null,"abstract":"The hybrid coding method combining the predictive coding with the orthogonal transformation and the quantization is mainly used recently. This paper proposes a new hybrid parallel Intra Coding based on interpolative prediction which uses correlations between neighboring pixels, including non-causal pixels. In order to get high prediction performance, the optimal quantizing scheme, which is used to cancel the error that expands when decoding, is used. Furthermore, a new type of block shape, which enables parallel coding, is proposed to simplify the processing of interpolative prediction. The result of comparison between proposed method and intra coding method in H.264 shows that the PSNR of proposed technique achieves 1 dB to 4 dB improvement in Luminance, especially for image with more details.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"31 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127985664","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1