首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Reducing bitrates of compressed video with enhanced view synthesis for FTV 通过增强FTV的视图合成来降低压缩视频的比特率
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702575
Lu Yang, M. O. Wildeboer, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto
View synthesis using depth maps is a well-known technique for exploiting the redundancy between multi-view videos. In this paper, we deal with the bitrates of view synthesis at the decoder side of FTV that would use compressed depth maps and views. Both inherent depth estimation error and coding distortion would degrade synthesis quality. The focus is to reduce bitrates required for generating the high-quality virtual view. We employ a reliable view synthesis method which is compared with standard MPEG view synthesis software. The experimental results show that the bitrates required for synthesizing high-quality virtual view could be reduced by utilizing our enhanced view synthesis technique to improve the PSNR at medium bitrates.
使用深度图的视图合成是一种众所周知的利用多视图视频之间冗余的技术。在本文中,我们处理在FTV解码器端使用压缩深度图和视图的视图合成的比特率。固有的深度估计误差和编码失真都会降低合成质量。其重点是降低生成高质量虚拟视图所需的比特率。与标准的MPEG视图合成软件相比,我们采用了一种可靠的视图合成方法。实验结果表明,在中等比特率下,利用增强的视图合成技术提高PSNR,可以降低合成高质量虚拟视图所需的比特率。
{"title":"Reducing bitrates of compressed video with enhanced view synthesis for FTV","authors":"Lu Yang, M. O. Wildeboer, T. Yendo, M. P. Tehrani, T. Fujii, M. Tanimoto","doi":"10.1109/PCS.2010.5702575","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702575","url":null,"abstract":"View synthesis using depth maps is a well-known technique for exploiting the redundancy between multi-view videos. In this paper, we deal with the bitrates of view synthesis at the decoder side of FTV that would use compressed depth maps and views. Both inherent depth estimation error and coding distortion would degrade synthesis quality. The focus is to reduce bitrates required for generating the high-quality virtual view. We employ a reliable view synthesis method which is compared with standard MPEG view synthesis software. The experimental results show that the bitrates required for synthesizing high-quality virtual view could be reduced by utilizing our enhanced view synthesis technique to improve the PSNR at medium bitrates.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127575055","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Improved texture compression for S3TC 改进了S3TC的纹理压缩
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702515
Yifei Jiang, Dandan Huan
Texture compression is a specialized form of still image compression employed in computer graphics systems to reduce memory bandwidth consumption. Modern texture compression schemes cannot generate satisfactory qualities for both alpha channel and color channel of texture images. We propose a novel texture compression scheme, named ImTC, based on the insight into the essential difference between transparency and color. ImTC defines new data formats and compresses the two channels flexibly. While keeping the same compression ratio as the de facto standard texture compression scheme, ImTC improves compression qualities of both channels. The average PSNR score of alpha channel is improved by about 0.2 dB, and that of color channel can be increased by 6.50 dB over a set of test images, which makes ImTC a better substitute for the standard scheme.
纹理压缩是静态图像压缩的一种特殊形式,用于计算机图形系统,以减少内存带宽消耗。现代纹理压缩方案不能同时对纹理图像的alpha通道和颜色通道产生令人满意的质量。基于对透明度和颜色之间本质区别的理解,我们提出了一种新的纹理压缩方案ImTC。ImTC定义了新的数据格式,并灵活地压缩了两个通道。在保持与事实上的标准纹理压缩方案相同的压缩比的同时,ImTC提高了两个通道的压缩质量。在一组测试图像中,α通道的平均PSNR分数提高了约0.2 dB,颜色通道的平均PSNR分数提高了6.50 dB,这使得ImTC可以更好地替代标准方案。
{"title":"Improved texture compression for S3TC","authors":"Yifei Jiang, Dandan Huan","doi":"10.1109/PCS.2010.5702515","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702515","url":null,"abstract":"Texture compression is a specialized form of still image compression employed in computer graphics systems to reduce memory bandwidth consumption. Modern texture compression schemes cannot generate satisfactory qualities for both alpha channel and color channel of texture images. We propose a novel texture compression scheme, named ImTC, based on the insight into the essential difference between transparency and color. ImTC defines new data formats and compresses the two channels flexibly. While keeping the same compression ratio as the de facto standard texture compression scheme, ImTC improves compression qualities of both channels. The average PSNR score of alpha channel is improved by about 0.2 dB, and that of color channel can be increased by 6.50 dB over a set of test images, which makes ImTC a better substitute for the standard scheme.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"32 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117262037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Medium-granularity computational complexity control for H.264/AVC H.264/AVC的中粒度计算复杂度控制
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702467
Xiang Li, M. Wien, J. Ohm
Today, video applications on handheld devices become more and more popular. Due to limited computational capability of handheld devices, complexity constrained video coding draws much attention. In this paper, a medium-granularity computational complexity control (MGCC) is proposed for H.264/AVC. First, a large dynamic range in complexity is achieved by taking 16×16 motion estimation in a single reference frame as the basic computational unit. Then a high coding efficiency is obtained by an adaptive computation allocation at MB level. Simulations show that coarse-granularity methods cannot work when the normalized complexity is below 15%. In contrast, the proposed MGCC performs well even when the complexity is reduced to 8.8%. Moreover, an average gain of 0.3 dB over coarse-granularity methods in BD-PSNR is obtained for 11 sequences when the complexity is around 20%.
如今,手持设备上的视频应用变得越来越流行。由于手持设备的计算能力有限,复杂度受限的视频编码备受关注。本文提出了一种针对H.264/AVC的中粒度计算复杂度控制(MGCC)。首先,以单个参考帧内的16×16运动估计为基本计算单元,实现了较大的动态复杂度范围;然后通过MB级的自适应计算分配获得较高的编码效率。仿真结果表明,当归一化复杂度低于15%时,粗粒度方法无法正常工作。相比之下,即使复杂性降低到8.8%,所提出的MGCC也表现良好。此外,当复杂度在20%左右时,11个序列的BD-PSNR平均增益比粗粒度方法高0.3 dB。
{"title":"Medium-granularity computational complexity control for H.264/AVC","authors":"Xiang Li, M. Wien, J. Ohm","doi":"10.1109/PCS.2010.5702467","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702467","url":null,"abstract":"Today, video applications on handheld devices become more and more popular. Due to limited computational capability of handheld devices, complexity constrained video coding draws much attention. In this paper, a medium-granularity computational complexity control (MGCC) is proposed for H.264/AVC. First, a large dynamic range in complexity is achieved by taking 16×16 motion estimation in a single reference frame as the basic computational unit. Then a high coding efficiency is obtained by an adaptive computation allocation at MB level. Simulations show that coarse-granularity methods cannot work when the normalized complexity is below 15%. In contrast, the proposed MGCC performs well even when the complexity is reduced to 8.8%. Moreover, an average gain of 0.3 dB over coarse-granularity methods in BD-PSNR is obtained for 11 sequences when the complexity is around 20%.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"229 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124532965","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Bit allocation of vertices and colors for patch-based coding in time-varying meshes 时变网格中基于补丁编码的顶点和颜色的位分配
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702449
T. Yamasaki, K. Aizawa
This paper discusses bit-rate assignments for vertices, color, reference frames, and target frames in the patch-based compression method for time-varying meshes (TVMs). TVMs are nonisomorphic 3D mesh sequences of the real-world objects generated from multiview images. Experimental results demonstrate that the bit rate for vertices greatly affects the visual quality of the rendered 3D model, whereas the bit rate for color does not contribute to quality improvement. Therefore, as many bits as possible should be assigned to vertices, with 8–10 bits per vertex (bpv) per frame being sufficient for color. For interframe coding, the visual quality is improved in proportion to the bit rate of both vertices and color. However, it is demonstrated that the use of fewer bits (5∼6 bpv) is sufficient to achieve a visual quality that matches the intraframe visual quality.
本文讨论了时变网格(tvm)基于补丁的压缩方法中顶点、颜色、参考帧和目标帧的比特率分配。tvm是由多视图图像生成的真实世界物体的非同构三维网格序列。实验结果表明,顶点的比特率对渲染3D模型的视觉质量有很大影响,而颜色的比特率对渲染3D模型的视觉质量没有影响。因此,应该为顶点分配尽可能多的位,每帧每个顶点8-10位(bpv)就足够用于颜色。对于帧间编码,视觉质量与顶点和颜色的比特率成正比。然而,研究表明,使用更少的比特(5 ~ 6 bpv)就足以获得与帧内视觉质量相匹配的视觉质量。
{"title":"Bit allocation of vertices and colors for patch-based coding in time-varying meshes","authors":"T. Yamasaki, K. Aizawa","doi":"10.1109/PCS.2010.5702449","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702449","url":null,"abstract":"This paper discusses bit-rate assignments for vertices, color, reference frames, and target frames in the patch-based compression method for time-varying meshes (TVMs). TVMs are nonisomorphic 3D mesh sequences of the real-world objects generated from multiview images. Experimental results demonstrate that the bit rate for vertices greatly affects the visual quality of the rendered 3D model, whereas the bit rate for color does not contribute to quality improvement. Therefore, as many bits as possible should be assigned to vertices, with 8–10 bits per vertex (bpv) per frame being sufficient for color. For interframe coding, the visual quality is improved in proportion to the bit rate of both vertices and color. However, it is demonstrated that the use of fewer bits (5∼6 bpv) is sufficient to achieve a visual quality that matches the intraframe visual quality.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"127 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123220500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
On the duality of rate allocation and quality indices 论费率分配与质量指标的二元性
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702484
T. Richter
In a recent work [16], the author proposed to study the performance of still image quality indices such as the SSIM by using them as objective function of rate allocation algorithms. The outcome of that work was not only a multi-scale SSIM optimal JPEG 2000 implementation, but also a first-order approximation of the MS-SSIM that is surprisingly similar to more traditional contrast-sensitivity and visual masking based approaches. It will be seen in this work that the only difference between the latter works and the MS-SSIM index is the choice of the exponent of the masking term, and furthermore, that a slight modification of the SSIM definition reproducing the traditional exponent is able to improve the performance of the index at or below the visual threshold. It is hence demonstrated that the duality of quality indices and rate allocation helps to improve both the visual performance of the compression codec and the performance of the index.
在最近的工作[16]中,作者提出将SSIM等静止图像质量指标作为速率分配算法的目标函数来研究其性能。这项工作的结果不仅是一个多尺度SSIM最优JPEG 2000实现,而且是MS-SSIM的一阶近似,它与更传统的对比度灵敏度和基于视觉掩蔽的方法惊人地相似。在这项工作中可以看到,后者与MS-SSIM指数之间的唯一区别是屏蔽项指数的选择,此外,对再现传统指数的SSIM定义进行轻微修改能够提高指数在视觉阈值或低于视觉阈值的性能。结果表明,质量指标和速率分配的对偶性有助于提高压缩编解码器的视觉性能和索引性能。
{"title":"On the duality of rate allocation and quality indices","authors":"T. Richter","doi":"10.1109/PCS.2010.5702484","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702484","url":null,"abstract":"In a recent work [16], the author proposed to study the performance of still image quality indices such as the SSIM by using them as objective function of rate allocation algorithms. The outcome of that work was not only a multi-scale SSIM optimal JPEG 2000 implementation, but also a first-order approximation of the MS-SSIM that is surprisingly similar to more traditional contrast-sensitivity and visual masking based approaches. It will be seen in this work that the only difference between the latter works and the MS-SSIM index is the choice of the exponent of the masking term, and furthermore, that a slight modification of the SSIM definition reproducing the traditional exponent is able to improve the performance of the index at or below the visual threshold. It is hence demonstrated that the duality of quality indices and rate allocation helps to improve both the visual performance of the compression codec and the performance of the index.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130777349","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A real-time system of distributed video coding 分布式视频编码的实时系统
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702557
K. Sakomizu, T. Yamasaki, Satoshi Nakagawa, T. Nishi
This paper presents a real-time system of distributed video coding (DVC). DVC is a current video compression paradigm. The decoding process of DVC is normally complex, which causes difficulty in real-time implementation. To address this problem, we propose a new configuration of DVC with three methods: simple rate control without the feedback channel, simple transmitting of dynamic range and simple bidirectional motion estimation to reduce complexity. Then we implement the system with parallelization techniques. We also develop the encoder for a low power processor. Experimental results show that the encoder on i.MX31 400 MHz could operates at about CIF 13 fps, and the decoder on Core 2 Quad 2.83 GHz operates at more than CIF 30 fps.
提出了一种分布式视频编码(DVC)实时系统。DVC是当前的视频压缩范例。DVC的解码过程通常比较复杂,这给实时实现带来了困难。为了解决这个问题,我们提出了一种新的DVC结构,采用三种方法:简单的速率控制,不带反馈通道,简单的动态范围传输和简单的双向运动估计来降低复杂度。然后采用并行化技术实现系统。我们还开发了用于低功耗处理器的编码器。实验结果表明,在i.MX31 400 MHz上编码器的工作速度约为CIF 13 fps,在Core 2 Quad 2.83 GHz上解码器的工作速度超过CIF 30 fps。
{"title":"A real-time system of distributed video coding","authors":"K. Sakomizu, T. Yamasaki, Satoshi Nakagawa, T. Nishi","doi":"10.1109/PCS.2010.5702557","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702557","url":null,"abstract":"This paper presents a real-time system of distributed video coding (DVC). DVC is a current video compression paradigm. The decoding process of DVC is normally complex, which causes difficulty in real-time implementation. To address this problem, we propose a new configuration of DVC with three methods: simple rate control without the feedback channel, simple transmitting of dynamic range and simple bidirectional motion estimation to reduce complexity. Then we implement the system with parallelization techniques. We also develop the encoder for a low power processor. Experimental results show that the encoder on i.MX31 400 MHz could operates at about CIF 13 fps, and the decoder on Core 2 Quad 2.83 GHz operates at more than CIF 30 fps.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128721470","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An improved Wyner-Ziv video coding with feedback channel 改进的带有反馈通道的Wyner-Ziv视频编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702468
Feng Ye, Aidong Men, Bo Yang, Manman Fan, Kan Chang
This paper presents an improved feedback-assisted low complexity WZVC scheme. The performance of this scheme is improved by two enhancements: an improved mode-based key frame encoding and a 3DRS-assisted (three-dimensional recursive search assisted) motion estimation algorithm for WZ encoding. Experimental results show that our coding scheme can achieve significant gain compared to state-oft he-art TDWZ codec while still low encoding complexity.
提出了一种改进的反馈辅助低复杂度WZVC方案。该方案通过改进的基于模式的关键帧编码和WZ编码的3drs辅助(三维递归搜索辅助)运动估计算法两方面的改进提高了性能。实验结果表明,与目前最先进的TDWZ编解码器相比,我们的编码方案可以获得显著的增益,同时编码复杂度也很低。
{"title":"An improved Wyner-Ziv video coding with feedback channel","authors":"Feng Ye, Aidong Men, Bo Yang, Manman Fan, Kan Chang","doi":"10.1109/PCS.2010.5702468","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702468","url":null,"abstract":"This paper presents an improved feedback-assisted low complexity WZVC scheme. The performance of this scheme is improved by two enhancements: an improved mode-based key frame encoding and a 3DRS-assisted (three-dimensional recursive search assisted) motion estimation algorithm for WZ encoding. Experimental results show that our coding scheme can achieve significant gain compared to state-oft he-art TDWZ codec while still low encoding complexity.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"49 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126833012","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Dictionary learning-based distributed compressive video sensing 基于字典学习的分布式压缩视频感知
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702466
Hung-Wei Chen, Li-Wei Kang, Chun-Shien Lu
We address an important issue of fully low-cost and low-complex video compression for use in resource-extremely limited sensors/devices. Conventional motion estimation-based video compression or distributed video coding (DVC) techniques all rely on the high-cost mechanism, namely, sensing/sampling and compression are disjointedly performed, resulting in unnecessary consumption of resources. That is, most acquired raw video data will be discarded in the (possibly) complex compression stage. In this paper, we propose a dictionary learning-based distributed compressive video sensing (DCVS) framework to “directly” acquire compressed video data. Embedded in the compressive sensing (CS)-based single-pixel camera architecture, DCVS can compressively sense each video frame in a distributed manner. At DCVS decoder, video reconstruction can be formulated as an l1-minimization problem via solving the sparse coefficients with respect to some basis functions. We investigate adaptive dictionary/basis learning for each frame based on the training samples extracted from previous reconstructed neighboring frames and argue that much better basis can be obtained to represent the frame, compared to fixed basis-based representation and recent popular “CS-based DVC” approaches without relying on dictionary learning.
我们解决了一个重要的问题,即在资源极其有限的传感器/设备中使用完全低成本和低复杂性的视频压缩。传统的基于运动估计的视频压缩或分布式视频编码(DVC)技术都依赖于高成本机制,即传感/采样和压缩分离进行,导致不必要的资源消耗。也就是说,在(可能)复杂的压缩阶段,大部分采集到的原始视频数据将被丢弃。在本文中,我们提出了一个基于字典学习的分布式压缩视频感知(DCVS)框架来“直接”获取压缩视频数据。DCVS嵌入到基于压缩感知(CS)的单像素摄像机架构中,可以以分布式的方式压缩感知每个视频帧。在DCVS解码器中,视频重构可以通过求解关于一些基函数的稀疏系数来表示为一个l1最小化问题。我们研究了基于从先前重建的相邻帧中提取的训练样本的每帧的自适应字典/基学习,并认为与基于固定基的表示和最近流行的不依赖字典学习的“基于cs的DVC”方法相比,可以获得更好的基来表示帧。
{"title":"Dictionary learning-based distributed compressive video sensing","authors":"Hung-Wei Chen, Li-Wei Kang, Chun-Shien Lu","doi":"10.1109/PCS.2010.5702466","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702466","url":null,"abstract":"We address an important issue of fully low-cost and low-complex video compression for use in resource-extremely limited sensors/devices. Conventional motion estimation-based video compression or distributed video coding (DVC) techniques all rely on the high-cost mechanism, namely, sensing/sampling and compression are disjointedly performed, resulting in unnecessary consumption of resources. That is, most acquired raw video data will be discarded in the (possibly) complex compression stage. In this paper, we propose a dictionary learning-based distributed compressive video sensing (DCVS) framework to “directly” acquire compressed video data. Embedded in the compressive sensing (CS)-based single-pixel camera architecture, DCVS can compressively sense each video frame in a distributed manner. At DCVS decoder, video reconstruction can be formulated as an l1-minimization problem via solving the sparse coefficients with respect to some basis functions. We investigate adaptive dictionary/basis learning for each frame based on the training samples extracted from previous reconstructed neighboring frames and argue that much better basis can be obtained to represent the frame, compared to fixed basis-based representation and recent popular “CS-based DVC” approaches without relying on dictionary learning.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"58 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126350243","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
3-D video coding using depth transition data 基于深度过渡数据的三维视频编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702453
Woo-Shik Kim, Antonio Ortega, Jaejoon Lee, H. Wey
The objective of this work is to develop a new 3-D video coding system which can provide better coding efficiency with improved subjective quality as compared to existing 3-D video systems. We have analyzed the distortions that occur in rendered views generated using depth image based rendering (DIBR) and classified them in order to evaluate their impact on subjective quality. As a result, we found that depth map coding distortion leads to “erosion artifacts” at object boundaries, which lead to significant degradation in perceptual quality. To solve this problem, we propose a solution in which depth transition data is encoded and transmitted to the decoder. Depth transition data for a given pixel indicates the camera position for which this pixel's depth will change. A main reason to consider transmitting explicitly this information is that it can be used to improve view interpolation at many different intermediate camera positions. Simulation results show that the subjective quality can be significantly improved by reducing the effect of erosion artifacts, using our proposed depth transition data. Maximum PSNR gains of about 0.5 dB can also be observed.
本文的目标是开发一种新的三维视频编码系统,与现有的三维视频系统相比,该系统可以提供更好的编码效率和更高的主观质量。我们分析了使用基于深度图像的渲染(DIBR)生成的渲染视图中出现的失真,并对它们进行了分类,以评估它们对主观质量的影响。因此,我们发现深度图编码失真会导致物体边界处的“侵蚀伪影”,从而导致感知质量的显著下降。为了解决这个问题,我们提出了一种将深度转换数据编码并传输到解码器的解决方案。给定像素的深度转换数据表示该像素的深度将改变的相机位置。考虑显式传输此信息的一个主要原因是,它可以用于改善许多不同中间相机位置的视图插值。仿真结果表明,采用本文提出的深度过渡数据,可以显著降低侵蚀伪影的影响,从而提高主观质量。还可以观察到约0.5 dB的最大PSNR增益。
{"title":"3-D video coding using depth transition data","authors":"Woo-Shik Kim, Antonio Ortega, Jaejoon Lee, H. Wey","doi":"10.1109/PCS.2010.5702453","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702453","url":null,"abstract":"The objective of this work is to develop a new 3-D video coding system which can provide better coding efficiency with improved subjective quality as compared to existing 3-D video systems. We have analyzed the distortions that occur in rendered views generated using depth image based rendering (DIBR) and classified them in order to evaluate their impact on subjective quality. As a result, we found that depth map coding distortion leads to “erosion artifacts” at object boundaries, which lead to significant degradation in perceptual quality. To solve this problem, we propose a solution in which depth transition data is encoded and transmitted to the decoder. Depth transition data for a given pixel indicates the camera position for which this pixel's depth will change. A main reason to consider transmitting explicitly this information is that it can be used to improve view interpolation at many different intermediate camera positions. Simulation results show that the subjective quality can be significantly improved by reducing the effect of erosion artifacts, using our proposed depth transition data. Maximum PSNR gains of about 0.5 dB can also be observed.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124260549","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
Inter prediction based on spatio-temporal adaptive localized learning model 基于时空自适应局部学习模型的内部预测
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702459
Hao Chen, R. Hu, Zhongyuan Wang, Rui Zhong
Inter prediction based on block matching motion estimation is important for video coding. But this method suffers from the additional overhead in data rate representing the motion information that needs to be transmitted to the decoder. To solve this problem, we present an improved implicit motion information inter prediction algorithm for P slice in H.264/AVC based on the spatio-temporal adaptive localized learning (STALL) model. According to 4 × 4 block transform structure in H.264/AVC, we first adaptively choose nine spatial neighbors and nine temporal neighbors, and a localized 3D casual cube is designed as training window. By using these information, the model parameters could be adaptively computed based on the Least Square Prediction (LSP) method. Finally, we add a new inter prediction mode into H.264/AVC standard for P slice. The experimental results show that our algorithm improves encoding efficiency compared with H.264/AVC standard, with relatively increases in complexity.
基于块匹配运动估计的帧间预测是视频编码的重要内容。但是这种方法在表示需要传输到解码器的运动信息的数据速率上有额外的开销。为了解决这一问题,提出了一种改进的基于时空自适应局部学习(STALL)模型的H.264/AVC中P片隐式运动信息互预测算法。根据H.264/AVC中的4 × 4块变换结构,首先自适应选择9个空间邻居和9个时间邻居,设计一个局部化的三维随机立方体作为训练窗口;利用这些信息,可以基于最小二乘预测(LSP)方法自适应计算模型参数。最后,我们在H.264/AVC标准中增加了一种新的P片间预测模式。实验结果表明,与H.264/AVC标准相比,该算法提高了编码效率,但复杂度相对增加。
{"title":"Inter prediction based on spatio-temporal adaptive localized learning model","authors":"Hao Chen, R. Hu, Zhongyuan Wang, Rui Zhong","doi":"10.1109/PCS.2010.5702459","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702459","url":null,"abstract":"Inter prediction based on block matching motion estimation is important for video coding. But this method suffers from the additional overhead in data rate representing the motion information that needs to be transmitted to the decoder. To solve this problem, we present an improved implicit motion information inter prediction algorithm for P slice in H.264/AVC based on the spatio-temporal adaptive localized learning (STALL) model. According to 4 × 4 block transform structure in H.264/AVC, we first adaptively choose nine spatial neighbors and nine temporal neighbors, and a localized 3D casual cube is designed as training window. By using these information, the model parameters could be adaptively computed based on the Least Square Prediction (LSP) method. Finally, we add a new inter prediction mode into H.264/AVC standard for P slice. The experimental results show that our algorithm improves encoding efficiency compared with H.264/AVC standard, with relatively increases in complexity.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117350256","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1