首页 > 最新文献

28th Picture Coding Symposium最新文献

英文 中文
Progress from stereoscopic to three-dimensional displays based on visual perception 基于视觉感知从立体显示到三维显示的进展
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702469
S. Yano
First, proper characteristics of stereoscopic or multi-view displays are described in the context of human visual perception. Next, the development of three-dimensional displays, matched to the function of human visual field, is described accounting for these points of stereoscopic and multi-view displays. These three-dimensional displays were developed based on the principle of light field reproduction, and the performances of each display are briefly described.
首先,在人类视觉感知的背景下,描述了立体或多视图显示的适当特征。其次,描述了与人类视野功能相匹配的三维显示的发展,说明了立体和多视图显示的这些点。这些三维显示器是基于光场再现原理开发的,并简要介绍了每种显示器的性能。
{"title":"Progress from stereoscopic to three-dimensional displays based on visual perception","authors":"S. Yano","doi":"10.1109/PCS.2010.5702469","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702469","url":null,"abstract":"First, proper characteristics of stereoscopic or multi-view displays are described in the context of human visual perception. Next, the development of three-dimensional displays, matched to the function of human visual field, is described accounting for these points of stereoscopic and multi-view displays. These three-dimensional displays were developed based on the principle of light field reproduction, and the performances of each display are briefly described.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"53 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124357804","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
View synthesis error analysis for selecting the optimal QP of depth map coding in 3D video application 三维视频应用中深度图编码最佳QP选择的视图合成误差分析
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702483
Yanwei Liu, S. Ci, Hui Tang
In 3D video communication, how to select the appropriate quantization parameter (QP) for depth map coding is very important for obtaining the optimal view synthesis quality. This paper first analyzes the depth uncertainty induced two kinds of view synthesis errors, namely the original depth error induced view synthesis error and the depth compression induced view synthesis error, and then proposes a quadratic model to characterize the relationship between the view synthesis quality and the depth quantization step size. The proposed model can find the inflexion point in the curve of the view synthesis quality with the increasing depth quantization step size. Experimental results show that, given the rate constraint for depth map, the proposed model can accurately find the optimal QP for depth map coding.
在三维视频通信中,如何选择合适的深度图编码量化参数(QP)对获得最佳的视点合成质量至关重要。本文首先分析了深度不确定性引起的两种视图综合误差,即原始深度误差引起的视图综合误差和深度压缩引起的视图综合误差,然后提出了一个二次模型来表征视图综合质量与深度量化步长之间的关系。该模型可以随着深度量化步长的增加找到视图综合质量曲线的拐点。实验结果表明,在给定深度图编码速率约束的情况下,该模型能够准确地找到深度图编码的最优QP。
{"title":"View synthesis error analysis for selecting the optimal QP of depth map coding in 3D video application","authors":"Yanwei Liu, S. Ci, Hui Tang","doi":"10.1109/PCS.2010.5702483","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702483","url":null,"abstract":"In 3D video communication, how to select the appropriate quantization parameter (QP) for depth map coding is very important for obtaining the optimal view synthesis quality. This paper first analyzes the depth uncertainty induced two kinds of view synthesis errors, namely the original depth error induced view synthesis error and the depth compression induced view synthesis error, and then proposes a quadratic model to characterize the relationship between the view synthesis quality and the depth quantization step size. The proposed model can find the inflexion point in the curve of the view synthesis quality with the increasing depth quantization step size. Experimental results show that, given the rate constraint for depth map, the proposed model can accurately find the optimal QP for depth map coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115417021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Image denoising with hard color-shrinkage and grouplet transform 采用硬色收缩和组变换对图像进行去噪
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702528
T. Saito, K. Ishikawa, Yasutaka Ueda, T. Komatsu
To remove signal-dependent noise of a digital color camera, we propose a new denoising method with our hard color-shrinkage in the tight-frame grouplet transform domain. The classic hard-shrinkage works well for monochrome-image denoising. To utilize inter-channel color dependence, a noisy image undergoes the color transformation from the RGB to the luminance-and-chrominance color space, and the luminance and the chrominance components are separately denoised; but this approach cannot cope with actual signal-dependent noise. To utilize the noise's signal-dependencies, we have constructed the hard color-shrinkage where the interchannel color dependence is directly utilized in the RGB color space. The hard color-shrinkage alleviates denoising artifacts, and improves picture quality of denoised images.
为了去除数码彩色相机的信号依赖性噪声,提出了一种在紧帧群变换域中使用硬颜色收缩的去噪方法。经典的硬收缩对单色图像去噪效果很好。为了利用通道间颜色依赖性,对噪声图像进行从RGB到亮度和色度颜色空间的颜色变换,并分别对亮度和色度分量进行去噪;但这种方法不能处理实际的信号相关噪声。为了利用噪声的信号依赖性,我们构造了硬颜色收缩,其中通道间颜色依赖性直接在RGB色彩空间中使用。硬色收缩减轻了去噪后的伪影,提高了去噪后图像的质量。
{"title":"Image denoising with hard color-shrinkage and grouplet transform","authors":"T. Saito, K. Ishikawa, Yasutaka Ueda, T. Komatsu","doi":"10.1109/PCS.2010.5702528","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702528","url":null,"abstract":"To remove signal-dependent noise of a digital color camera, we propose a new denoising method with our hard color-shrinkage in the tight-frame grouplet transform domain. The classic hard-shrinkage works well for monochrome-image denoising. To utilize inter-channel color dependence, a noisy image undergoes the color transformation from the RGB to the luminance-and-chrominance color space, and the luminance and the chrominance components are separately denoised; but this approach cannot cope with actual signal-dependent noise. To utilize the noise's signal-dependencies, we have constructed the hard color-shrinkage where the interchannel color dependence is directly utilized in the RGB color space. The hard color-shrinkage alleviates denoising artifacts, and improves picture quality of denoised images.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124906073","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Bitwise prediction error correction for Distributed Video Coding 分布式视频编码的位预测纠错
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702514
Axel Lakus-Becker, Ka-Ming Leung, Zhonghua Ma
Side information plays a key role in the performance of a Distributed Video Coding (DVC) system. However, the generation of side information often relies on complex motion estimation/interpolation operation. The correlation between the source data and the side information, sometimes referred as virtual channel, is also very difficult to model accurately. In this paper, we propose a bitwise prediction error correction method to improve the quality of the side information during Wyner-Ziv decoding. Whenever a bit error is detected in a bit plane, the less significant bits of the corresponding pixel are adjusted to match the initial prediction. The proposed method has been evaluated using a pixel-domain DVC system and delivers a better coding performance with improved decoding quality and reduced bitrate.
侧信息在分布式视频编码(DVC)系统的性能中起着关键作用。然而,侧信息的生成往往依赖于复杂的运动估计/插值操作。源数据与侧信息(有时称为虚拟信道)之间的相关性也很难准确建模。为了提高wner - ziv译码过程中边线信息的质量,本文提出了一种按位预测纠错方法。每当在位平面中检测到位错误时,将相应像素的不太重要的位进行调整以匹配初始预测。采用像素域DVC系统对该方法进行了测试,结果表明该方法具有更好的编码性能,提高了解码质量,降低了比特率。
{"title":"Bitwise prediction error correction for Distributed Video Coding","authors":"Axel Lakus-Becker, Ka-Ming Leung, Zhonghua Ma","doi":"10.1109/PCS.2010.5702514","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702514","url":null,"abstract":"Side information plays a key role in the performance of a Distributed Video Coding (DVC) system. However, the generation of side information often relies on complex motion estimation/interpolation operation. The correlation between the source data and the side information, sometimes referred as virtual channel, is also very difficult to model accurately. In this paper, we propose a bitwise prediction error correction method to improve the quality of the side information during Wyner-Ziv decoding. Whenever a bit error is detected in a bit plane, the less significant bits of the corresponding pixel are adjusted to match the initial prediction. The proposed method has been evaluated using a pixel-domain DVC system and delivers a better coding performance with improved decoding quality and reduced bitrate.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129749402","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Successive refinement of overlapped cell side quantizers for scalable multiple description coding 可扩展多重描述编码中重叠小区侧量化器的逐次细化
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702488
Muhammad Majid, G. Abhayaratne
Scalable multiple description coding (SMDC) provides reliability and facility to truncate the descriptions according to the user rate-distortion requirements. In this paper we generalize the conditions of successive refinement of the side quantizer of a multiple description scalar quantizer that has overlapped quantizer cells generated by a modified linear index assignment matrix. We propose that the split or refinement factor for each of the refinement side quantizers should be greater than the maximum side quantizer bin spread and should not be integer multiples of each other for satisfying the SMDC distortion conditions and verify through simulation results on scalable multiple description image coding.
可伸缩的多重描述编码(SMDC)提供了根据用户率失真要求截断描述的可靠性和便利性。本文推广了由改进的线性指标分配矩阵生成的具有重叠量化单元的多重描述标量量化器的侧量化器连续细化的条件。为了满足SMDC失真条件,我们提出每个细化侧量化器的分割因子或细化因子应大于最大侧量化器bin spread,而不是彼此的整数倍,并通过可扩展多重描述图像编码的仿真结果进行验证。
{"title":"Successive refinement of overlapped cell side quantizers for scalable multiple description coding","authors":"Muhammad Majid, G. Abhayaratne","doi":"10.1109/PCS.2010.5702488","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702488","url":null,"abstract":"Scalable multiple description coding (SMDC) provides reliability and facility to truncate the descriptions according to the user rate-distortion requirements. In this paper we generalize the conditions of successive refinement of the side quantizer of a multiple description scalar quantizer that has overlapped quantizer cells generated by a modified linear index assignment matrix. We propose that the split or refinement factor for each of the refinement side quantizers should be greater than the maximum side quantizer bin spread and should not be integer multiples of each other for satisfying the SMDC distortion conditions and verify through simulation results on scalable multiple description image coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128679768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Scalable video compression framework with adaptive multiresolution directional filter bank design 具有自适应多分辨率方向滤波器组设计的可扩展视频压缩框架
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702523
Lingchen Zhu, H. Xiong
Regarding orientation resolution as an isolated variable from scale, this paper proposes a dual (scale and orientation) multiresolution transform into scalable video coding (SVC) framework. By projecting 2D signals (textures and edges) onto nonuniformly divided orientation subspaces, the dual multiresolution SVC (DMSVC) can capture 2-D curve smoothness with less coefficients, and provide more flexible spatial decomposition structures than traditional wavelet-based SVC (WSVC). In the spatial decomposition module of DMSVC, the nonuniform directional distribution along scale of each frame is detected by phase congruency in the overcomplete wavelet domain. The corresponding orientational multi-resolution is achieved by nonuniform directional filter banks (NUDFB) which is fulfilled via a non-symmetric binary tree (NSBT) structured frequency division. The wavelet basis function in each scale is converted to an adaptive set of nonuniform directional basis by employing nonuniform directional filter banks. Experimental results validate a superior coding performance and visual quality over WSVC especially on those sequences full of directional edges and textures.
将方向分辨率作为独立于尺度的变量,提出了一种双(尺度和方向)多分辨率可缩放视频编码(SVC)框架。双多分辨率SVC (DMSVC)通过将二维信号(纹理和边缘)投影到非均匀划分的方向子空间上,以较少的系数捕获二维曲线的平滑性,并提供比传统小波SVC (WSVC)更灵活的空间分解结构。在DMSVC的空间分解模块中,利用过完备小波域的相位一致性检测每帧图像沿尺度的非均匀方向分布。通过非对称二叉树(NSBT)结构分频实现非均匀方向滤波器组(NUDFB)的定向多分辨率。采用非均匀方向滤波器组,将每个尺度上的小波基函数转换为一组自适应的非均匀方向基。实验结果验证了WSVC的编码性能和视觉质量,特别是在那些充满方向边缘和纹理的序列上。
{"title":"Scalable video compression framework with adaptive multiresolution directional filter bank design","authors":"Lingchen Zhu, H. Xiong","doi":"10.1109/PCS.2010.5702523","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702523","url":null,"abstract":"Regarding orientation resolution as an isolated variable from scale, this paper proposes a dual (scale and orientation) multiresolution transform into scalable video coding (SVC) framework. By projecting 2D signals (textures and edges) onto nonuniformly divided orientation subspaces, the dual multiresolution SVC (DMSVC) can capture 2-D curve smoothness with less coefficients, and provide more flexible spatial decomposition structures than traditional wavelet-based SVC (WSVC). In the spatial decomposition module of DMSVC, the nonuniform directional distribution along scale of each frame is detected by phase congruency in the overcomplete wavelet domain. The corresponding orientational multi-resolution is achieved by nonuniform directional filter banks (NUDFB) which is fulfilled via a non-symmetric binary tree (NSBT) structured frequency division. The wavelet basis function in each scale is converted to an adaptive set of nonuniform directional basis by employing nonuniform directional filter banks. Experimental results validate a superior coding performance and visual quality over WSVC especially on those sequences full of directional edges and textures.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128876624","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Content-adaptive spatial scalability for scalable video coding 可扩展视频编码的内容自适应空间可扩展性
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702471
Yongzhe Wang, N. Stefanoski, Xiangzhong Fang, A. Smolic
This paper presents an enhancement of the SVC extension of the H.264/AVC standard by content-adaptive spatial scalability (CASS). CASS introduces a novel functionality which is important for high quality content distribution. The video streams (spatial layers), which are used as input to the encoder, are created by content-adaptive and art-directable retargeting of existing high resolution video. Video is retargeted to resolutions and aspect ratios which are mainly dictated by target display devices. Thereby no content is cut off, but visually important content is preserved at the expense of a non-linear distortion of visually unimportant areas. The non-linear dependencies between such video streams are efficiently exploited by CASS for scalable coding. This is achieved by integrating warping-based non-linear texture prediction and warp coding into the SVC framework. The results indicate high prediction accuracy of non-linear predictors and high compression efficiency with limited increase in bit rate and complexity compared to the standard SVC for the case of INTRA only coding.
本文提出了一种基于内容自适应空间可扩展性(CASS)的H.264/AVC标准的SVC扩展增强方法。CASS引入了一个新的功能,这对于高质量的内容分发非常重要。视频流(空间层),用作编码器的输入,是由现有的高分辨率视频的内容自适应和艺术定向重定向创建的。视频被重新定位到主要由目标显示设备决定的分辨率和长宽比。因此,没有内容被切断,但视觉上重要的内容被保留,代价是视觉上不重要的区域的非线性扭曲。这些视频流之间的非线性依赖关系被CASS有效地用于可扩展编码。这是通过将基于翘曲的非线性纹理预测和翘曲编码集成到SVC框架中来实现的。结果表明,在仅使用INTRA编码的情况下,与标准SVC相比,非线性预测器的预测精度高,压缩效率高,比特率和复杂度的增加有限。
{"title":"Content-adaptive spatial scalability for scalable video coding","authors":"Yongzhe Wang, N. Stefanoski, Xiangzhong Fang, A. Smolic","doi":"10.1109/PCS.2010.5702471","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702471","url":null,"abstract":"This paper presents an enhancement of the SVC extension of the H.264/AVC standard by content-adaptive spatial scalability (CASS). CASS introduces a novel functionality which is important for high quality content distribution. The video streams (spatial layers), which are used as input to the encoder, are created by content-adaptive and art-directable retargeting of existing high resolution video. Video is retargeted to resolutions and aspect ratios which are mainly dictated by target display devices. Thereby no content is cut off, but visually important content is preserved at the expense of a non-linear distortion of visually unimportant areas. The non-linear dependencies between such video streams are efficiently exploited by CASS for scalable coding. This is achieved by integrating warping-based non-linear texture prediction and warp coding into the SVC framework. The results indicate high prediction accuracy of non-linear predictors and high compression efficiency with limited increase in bit rate and complexity compared to the standard SVC for the case of INTRA only coding.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127827566","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Difference detection based early mode termination for depth map coding in MVC 基于差分检测的MVC深度图编码早期模式终止
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702547
Minghui Wang, Xin Jin, S. Goto
Depth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based (ODD). A simplified ODD (sODD) strategy is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of Joint Multiview Video Coding (JMVC) version 8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have an acceptable R-D performance lost.
深度图编码是多视图视频编码(MVC)中随着深度图像渲染(DIBR)的发展而出现的一个新课题。由于深度图是单色图,纹理比彩色图少,因此快速算法是减少编码器计算负担的必要和可能。提出了基于差分检测的早模终止策略。差分检测算法分为基于重构帧(RDD)和基于原始帧(ODD)两种。提出了一种简化的ODD (sODD)策略。基于这三种DD算法的早模式终止分别在JMVC 8.0版本的参考软件中实现和评估。仿真结果表明,基于RDD的算法没有性能损失,运行时间平均减少25%。基于ODD和基于sODD的方法分别可以节省54.3%和43.6%的运行时间,并且具有可接受的R-D性能损失。
{"title":"Difference detection based early mode termination for depth map coding in MVC","authors":"Minghui Wang, Xin Jin, S. Goto","doi":"10.1109/PCS.2010.5702547","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702547","url":null,"abstract":"Depth map coding is a new topic in multiview video coding (MVC) following the development of depth-image-based rendering (DIBR). Since depth map is monochromatic and has less texture than color map, fast algorithm is necessary and possible to reduce the computation burden of the encoder. This paper proposed difference detection based early mode termination strategy. The difference detection (DD) algorithms are categorized to reconstructed frame based (RDD) and original frame based (ODD). A simplified ODD (sODD) strategy is also proposed. Early mode termination based on these three DD algorithms are implemented and evaluated in the reference software of Joint Multiview Video Coding (JMVC) version 8.0 respectively. Simulation results indicate that RDD based one has no performance lost and reduce 25% runtime on average. ODD and sODD based ones can save 54.3% and 43.6% runtime respectively and have an acceptable R-D performance lost.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"166 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121538983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Sparse representation of depth maps for efficient transform coding 深度图的稀疏表示,用于高效的变换编码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702491
Gene Cheung, Akira Kubota, Antonio Ortega
Compression of depth maps is important for “image plus depth” representation of multiview images, which enables synthesis of novel intermediate views via depth-image-based rendering (DIBR) at decoder. Previous depth map coding schemes exploit unique depth characteristics to compactly and faithfully reproduce the original signal. In contrast, given that depth maps are not directly viewed but are only used for view synthesis, in this paper we manipulate depth values themselves, without causing severe synthesized view distortion, in order to maximize sparsity in the transform domain for compression gain. We formulate the sparsity maximization problem as an l0-norm optimization. Given l0-norm optimization is hard in general, we first find a sparse representation by iteratively solving a weighted l1 minimization via linear programming (LP). We then design a heuristic to push resulting LP solution away from constraint boundaries to avoid quantization errors. Using JPEG as an example transform codec, we show that our approach gained up to 2.5dB in rate-distortion performance for the interpolated view.
深度图的压缩对于多视图图像的“图像加深度”表示非常重要,它可以在解码器上通过基于深度图像的渲染(DIBR)合成新的中间视图。以往的深度图编码方案利用独特的深度特性来紧凑、真实地再现原始信号。相比之下,考虑到深度图不是直接查看的,而是仅用于视图合成,在本文中,我们操纵深度值本身,而不会引起严重的合成视图失真,以最大化变换域中的稀疏性以获得压缩增益。我们将稀疏性最大化问题表述为一个10范数优化问题。考虑到10范数优化通常是困难的,我们首先通过线性规划(LP)迭代求解加权l1最小化来找到一个稀疏表示。然后,我们设计了一个启发式算法,将得到的LP解推离约束边界,以避免量化误差。使用JPEG作为转换编解码器的示例,我们表明我们的方法在插值视图中获得了高达2.5dB的率失真性能。
{"title":"Sparse representation of depth maps for efficient transform coding","authors":"Gene Cheung, Akira Kubota, Antonio Ortega","doi":"10.1109/PCS.2010.5702491","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702491","url":null,"abstract":"Compression of depth maps is important for “image plus depth” representation of multiview images, which enables synthesis of novel intermediate views via depth-image-based rendering (DIBR) at decoder. Previous depth map coding schemes exploit unique depth characteristics to compactly and faithfully reproduce the original signal. In contrast, given that depth maps are not directly viewed but are only used for view synthesis, in this paper we manipulate depth values themselves, without causing severe synthesized view distortion, in order to maximize sparsity in the transform domain for compression gain. We formulate the sparsity maximization problem as an l0-norm optimization. Given l0-norm optimization is hard in general, we first find a sparse representation by iteratively solving a weighted l1 minimization via linear programming (LP). We then design a heuristic to push resulting LP solution away from constraint boundaries to avoid quantization errors. Using JPEG as an example transform codec, we show that our approach gained up to 2.5dB in rate-distortion performance for the interpolated view.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"75 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121558514","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
Multiple description video transcoding with temporal drift control 具有时间漂移控制的多描述视频转码
Pub Date : 2010-12-01 DOI: 10.1109/PCS.2010.5702563
P. Correia, P. Assunção, V. Silva
This paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.
本文提出了一种多重描述(MD)转码方案,该方案能够在时间预测的运动补偿切片中通过畸变积累来防止漂移。漂移补偿是通过产生受控数量的侧信息来实现的,当描述无法到达最终用户终端时,这些侧信息将用于解码。侧信息是通过使用独立的量化参数对转码残基进行重新编码而产生的,该量化参数还控制了冗余度。设计了一种简化的体系结构,以减少处理函数数量和缓冲区需求方面的转码复杂性。实验结果表明,与经典的MD转码方案相比,临时预测的帧不会产生漂移,并且在减少冗余成本的情况下显著提高了帧的质量。
{"title":"Multiple description video transcoding with temporal drift control","authors":"P. Correia, P. Assunção, V. Silva","doi":"10.1109/PCS.2010.5702563","DOIUrl":"https://doi.org/10.1109/PCS.2010.5702563","url":null,"abstract":"This paper proposes a multiple description (MD) transcoding scheme capable of preventing drift by distortion accumulation in temporally predicted motion compensated slices. Drift compensation is achieved by generating a controlled amount of side information to be used for decoding whenever a description fails to reach the end user terminal. The side information is generated by re-encoding the transcoding residue with an independent quantisation parameter which also controls redundancy. A simplified architecture is devised to reduce transcoding complexity in regard to the number of processing functions and buffer requirements. The experimental results show that temporally predicted frames do not suffer from drift and their quality is significantly improved at reduced redundancy cost in comparison with a classic MD transcoding scheme.","PeriodicalId":255142,"journal":{"name":"28th Picture Coding Symposium","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-12-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116408155","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
28th Picture Coding Symposium
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1