首页 > 最新文献

2013 Data Compression Conference最新文献

英文 中文
Context Lossless Coding of Audio Signals 音频信号无损编码
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.102
G. Ulacha, R. Stasinski
In the paper improvements obtained for context lossless audio coding are investigated. The approach is not popular in audio compression, hence, the research concentrates on static forward predictors optimized using MMSE criterion. Two and three context algorithms are tested on 16 popular benchmark recordings. Savings due to inter-channel audio dependencies are also considered. It is shown that indeed, context approach has potential of improving data compaction properties of audio coding algorithms.
本文研究了上下文无损音频编码的改进方法。该方法在音频压缩中并不流行,因此,研究集中在使用MMSE准则优化的静态前向预测器上。在16个流行的基准记录上测试了两种和三种上下文算法。由于通道间音频依赖的节省也被考虑在内。研究结果表明,上下文方法确实具有改善音频编码算法数据压缩特性的潜力。
{"title":"Context Lossless Coding of Audio Signals","authors":"G. Ulacha, R. Stasinski","doi":"10.1109/DCC.2013.102","DOIUrl":"https://doi.org/10.1109/DCC.2013.102","url":null,"abstract":"In the paper improvements obtained for context lossless audio coding are investigated. The approach is not popular in audio compression, hence, the research concentrates on static forward predictors optimized using MMSE criterion. Two and three context algorithms are tested on 16 popular benchmark recordings. Savings due to inter-channel audio dependencies are also considered. It is shown that indeed, context approach has potential of improving data compaction properties of audio coding algorithms.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130601850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Subsampling Input Based Side Information Creation in Wyner-Ziv Video Coding Wyner-Ziv视频编码中基于子采样输入的边信息创建
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.98
Y. Shen, Ji-Ciao Luo, Ja-Ling Wu
Summary form only given. Distributed video coding (DVC) has been intensively studied in recent years. This new coding paradigm substantially differs from conventional prediction-based video codecs such as MPEG and H.26x, which are characterized by a complex encoder and simple decoder. The conventional DVC codec, e.g., DISCOVER codec, uses advanced frame interpolation techniques to create SI based on adjacent decoded reference frames. The quality of SI is a well-recognized factor in the RD performance of WZ video coding. A high SI quality implies a high correlation between the created SI and the original WZ frame, which then decreases the rate required to achieve a given decoded quality. Clearly, the performance of an SI creation process based on adjacent previously decoded frames is limited by the quality of the past and the future reference frames as well as the distance and motion behavior between them. The correlation between high-motion frames is low and vice versa. That is, SI quality in the conventional codecs depends on the temporal correlation of key frames, which affects the bitrate and PSNR of the compression process. In this work, a novel DVC architecture for dealing with the cases of high-motion and large GOP-size sequences is proposed to better the rate-distortion (RD) performance. For high-motion video sequences, the proposed architecture generates SI by using subsampled spatial information instead of interpolated temporal information. the proposed approach separates the video sequence into subsampled key frames and corresponding WZ frames, which changes the creation of SI. That is, all successive frames on the encoder side are downsized to sub-frames, which are then compressed by an H.264/AVC intra encoder. Experimental results reveal that the subsampling input based DVC codec can gain up to 1.47 dB in the RD measures and maintains the most important characteristic of the DVC codec, the encoder is lightweight, as compared with the conventional WZ codec, respectively. The novel DVC architecture evaluated in this study exploits spatial relations to create SI. The experimental results confirm that the RD performance of the proposed approach is superior to that of the conventional one for high-motion and/or large GOP-size sequences. The quality of spatial interpolation based SI is higher than that of the temporal interpolation one, which leads to a high-PSNR reconstructed WZ frame. The subsampled key frames are also decoded by LDPCA decoder to recover the information lost when H.264/AVC intra coding is used to increase PSNR gain. Since many spatial domain interpolation and super resolution schemes have been proposed for use in the fields of image processing and computer vision, the performance of the proposed DVC codec can be further enhanced by using better schemes to generate even better SI.
只提供摘要形式。分布式视频编码(DVC)近年来得到了广泛的研究。这种新的编码范式与传统的基于预测的视频编解码器(如MPEG和H.26x)有很大的不同,后者的特点是具有复杂的编码器和简单的解码器。传统的DVC编解码器,例如DISCOVER编解码器,使用先进的帧插值技术来基于相邻解码的参考帧创建SI。在WZ视频编码的RD性能中,SI质量是一个公认的因素。高SI质量意味着创建的SI和原始WZ帧之间的高相关性,从而降低了实现给定解码质量所需的速率。显然,基于相邻的先前解码帧的SI创建过程的性能受到过去和未来参考帧的质量以及它们之间的距离和运动行为的限制。高运动帧之间的相关性较低,反之亦然。也就是说,传统编解码器中的SI质量取决于关键帧的时间相关性,这影响了压缩过程的比特率和PSNR。在这项工作中,提出了一种新的DVC架构,用于处理高运动和大gop大小的序列,以提高率失真(RD)性能。对于高运动视频序列,本文提出的结构通过使用次采样的空间信息而不是插值的时间信息来生成SI。该方法将视频序列分离为下采样关键帧和相应的WZ帧,从而改变了SI的生成。也就是说,编码器侧的所有连续帧都被缩小为子帧,然后由H.264/AVC内部编码器压缩。实验结果表明,基于子采样输入的DVC编解码器在RD测量中增益高达1.47 dB,保持了DVC编解码器最重要的特性,与传统的WZ编解码器相比,编码器重量轻。本研究评估的新型DVC架构利用空间关系来创建SI。实验结果证实,对于高运动和/或大gop大小的序列,该方法的RD性能优于传统方法。基于空间插值的图像质量比基于时间插值的图像质量高,从而得到高信噪比重构的WZ帧。对采样后的关键帧进行LDPCA解码器解码,以恢复H.264/AVC帧内编码提高信噪比增益时丢失的信息。由于在图像处理和计算机视觉领域已经提出了许多空间域插值和超分辨率方案,因此所提出的DVC编解码器的性能可以通过使用更好的方案来进一步增强,从而产生更好的SI。
{"title":"Subsampling Input Based Side Information Creation in Wyner-Ziv Video Coding","authors":"Y. Shen, Ji-Ciao Luo, Ja-Ling Wu","doi":"10.1109/DCC.2013.98","DOIUrl":"https://doi.org/10.1109/DCC.2013.98","url":null,"abstract":"Summary form only given. Distributed video coding (DVC) has been intensively studied in recent years. This new coding paradigm substantially differs from conventional prediction-based video codecs such as MPEG and H.26x, which are characterized by a complex encoder and simple decoder. The conventional DVC codec, e.g., DISCOVER codec, uses advanced frame interpolation techniques to create SI based on adjacent decoded reference frames. The quality of SI is a well-recognized factor in the RD performance of WZ video coding. A high SI quality implies a high correlation between the created SI and the original WZ frame, which then decreases the rate required to achieve a given decoded quality. Clearly, the performance of an SI creation process based on adjacent previously decoded frames is limited by the quality of the past and the future reference frames as well as the distance and motion behavior between them. The correlation between high-motion frames is low and vice versa. That is, SI quality in the conventional codecs depends on the temporal correlation of key frames, which affects the bitrate and PSNR of the compression process. In this work, a novel DVC architecture for dealing with the cases of high-motion and large GOP-size sequences is proposed to better the rate-distortion (RD) performance. For high-motion video sequences, the proposed architecture generates SI by using subsampled spatial information instead of interpolated temporal information. the proposed approach separates the video sequence into subsampled key frames and corresponding WZ frames, which changes the creation of SI. That is, all successive frames on the encoder side are downsized to sub-frames, which are then compressed by an H.264/AVC intra encoder. Experimental results reveal that the subsampling input based DVC codec can gain up to 1.47 dB in the RD measures and maintains the most important characteristic of the DVC codec, the encoder is lightweight, as compared with the conventional WZ codec, respectively. The novel DVC architecture evaluated in this study exploits spatial relations to create SI. The experimental results confirm that the RD performance of the proposed approach is superior to that of the conventional one for high-motion and/or large GOP-size sequences. The quality of spatial interpolation based SI is higher than that of the temporal interpolation one, which leads to a high-PSNR reconstructed WZ frame. The subsampled key frames are also decoded by LDPCA decoder to recover the information lost when H.264/AVC intra coding is used to increase PSNR gain. Since many spatial domain interpolation and super resolution schemes have been proposed for use in the fields of image processing and computer vision, the performance of the proposed DVC codec can be further enhanced by using better schemes to generate even better SI.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"10 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131955642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Online Learning Based Face Distortion Recovery for Conversational Video Coding 会话视频编码中基于在线学习的人脸失真恢复
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.105
Xi Wang, Li Su, Qingming Huang, Guorong Li, H. Qi
In a video conversation, the participants usually remain the same. As the conversation continues, similar facial expressions of the same person would occur intermittently. However, the correlation of similar face features has not been fully used since the conventional methods only focus on independent frames. We set up a face feature database and updated it online to include new facial expressions during the whole conversation. At the receiver side, the database is used to recover the face distortion and thus improve the visual quality. Additionally, the proposed method brings small burden to update the database and is generic to various CODEC.
在视频对话中,参与者通常保持不变。随着谈话的进行,同一个人类似的面部表情会断断续续地出现。然而,由于传统的方法只关注独立的帧,因此没有充分利用相似人脸特征之间的相关性。我们建立了一个面部特征数据库,并在线更新它,以包括整个对话过程中的新面部表情。在接收端,使用数据库恢复人脸畸变,从而提高视觉质量。此外,该方法对数据库更新负担小,对各种编解码器具有通用性。
{"title":"Online Learning Based Face Distortion Recovery for Conversational Video Coding","authors":"Xi Wang, Li Su, Qingming Huang, Guorong Li, H. Qi","doi":"10.1109/DCC.2013.105","DOIUrl":"https://doi.org/10.1109/DCC.2013.105","url":null,"abstract":"In a video conversation, the participants usually remain the same. As the conversation continues, similar facial expressions of the same person would occur intermittently. However, the correlation of similar face features has not been fully used since the conventional methods only focus on independent frames. We set up a face feature database and updated it online to include new facial expressions during the whole conversation. At the receiver side, the database is used to recover the face distortion and thus improve the visual quality. Additionally, the proposed method brings small burden to update the database and is generic to various CODEC.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"34 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134412448","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Low Complexity Rate Distortion Optimization for HEVC HEVC低复杂度率失真优化
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.15
Siwei Ma, Shiqi Wang, Shanshe Wang, Liang Zhao, Qin Yu, Wen Gao
The emerging High Efficiency Video Coding (HEVC) standard has improved the coding efficiency drastically, and can provide equivalent subjective quality with more than 50% bit rate reduction compared to its predecessor H.264/AVC. As expected, the improvement on coding efficiency is obtained at the expense of more intensive computation complexity. In this paper, based on an overall analysis of computation complexity in HEVC encoder, a low complexity rate distortion optimization (RDO) coding scheme is proposed by reducing the number of available candidates for evaluation in terms of the intra prediction mode decision, reference frame selection and CU splitting. With the proposed scheme, the RDO technique of HEVC can be implemented in a low-complexity way for complexity-constrained encoders. Experimental results demonstrate that, compared with the original HEVC reference encoder implementation, the proposed algorithms can achieve about 30% reduced encoding time on average with ignorable coding performance degradation (0.8%).
新兴的高效视频编码(HEVC)标准大大提高了编码效率,与之前的H.264/AVC相比,可以提供同等的主观质量,比特率降低50%以上。正如预期的那样,编码效率的提高是以更高的计算复杂度为代价的。本文在全面分析HEVC编码器计算复杂度的基础上,从帧内预测模式的选择、参考帧的选择和帧间分割等方面,减少可评估候选帧的数量,提出了一种低复杂度率失真优化(RDO)编码方案。利用该方案,HEVC的RDO技术可以在复杂度受限的编码器中以低复杂度的方式实现。实验结果表明,与原有的HEVC参考编码器实现相比,所提算法的编码时间平均缩短了30%左右,编码性能下降可以忽略不计(0.8%)。
{"title":"Low Complexity Rate Distortion Optimization for HEVC","authors":"Siwei Ma, Shiqi Wang, Shanshe Wang, Liang Zhao, Qin Yu, Wen Gao","doi":"10.1109/DCC.2013.15","DOIUrl":"https://doi.org/10.1109/DCC.2013.15","url":null,"abstract":"The emerging High Efficiency Video Coding (HEVC) standard has improved the coding efficiency drastically, and can provide equivalent subjective quality with more than 50% bit rate reduction compared to its predecessor H.264/AVC. As expected, the improvement on coding efficiency is obtained at the expense of more intensive computation complexity. In this paper, based on an overall analysis of computation complexity in HEVC encoder, a low complexity rate distortion optimization (RDO) coding scheme is proposed by reducing the number of available candidates for evaluation in terms of the intra prediction mode decision, reference frame selection and CU splitting. With the proposed scheme, the RDO technique of HEVC can be implemented in a low-complexity way for complexity-constrained encoders. Experimental results demonstrate that, compared with the original HEVC reference encoder implementation, the proposed algorithms can achieve about 30% reduced encoding time on average with ignorable coding performance degradation (0.8%).","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128237686","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Cross Segment Decoding for Improved Quality of Experience for Video Applications 提高视频应用体验质量的交叉段解码
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.31
Jiangtao Wen, Shunyao Li, Yao Lu, Meiyuan Fang, Xuan Dong, Huiwen Chang, Pin Tao
In this paper, we present an improved algorithm for decoding live streamed or pre-encoded video bit streams with time-varying qualities. The algorithm extracts information available to the decoder from a high visual quality segment of the clip that has already been received and decoded, but was encoded independently from the current segment. The proposed decoder is capable of significantly improve the Quality of Experience of the user without incurring significant overhead to the storage and computational complexities of both the encoder and the decoder. We present simulation results using the HEVC reference encoder and standard test clips, and discuss areas of improvements to the algorithm and potential ways of incorporating the technique to a video streaming system or standards.
在本文中,我们提出了一种改进的算法,用于解码时变质量的直播或预编码视频比特流。该算法从已经接收和解码的高视觉质量片段中提取可用于解码器的信息,但独立于当前片段进行编码。所提出的解码器能够显著提高用户的体验质量,而不会对编码器和解码器的存储和计算复杂性产生显著的开销。我们展示了使用HEVC参考编码器和标准测试剪辑的仿真结果,并讨论了算法的改进领域以及将该技术纳入视频流系统或标准的潜在方法。
{"title":"Cross Segment Decoding for Improved Quality of Experience for Video Applications","authors":"Jiangtao Wen, Shunyao Li, Yao Lu, Meiyuan Fang, Xuan Dong, Huiwen Chang, Pin Tao","doi":"10.1109/DCC.2013.31","DOIUrl":"https://doi.org/10.1109/DCC.2013.31","url":null,"abstract":"In this paper, we present an improved algorithm for decoding live streamed or pre-encoded video bit streams with time-varying qualities. The algorithm extracts information available to the decoder from a high visual quality segment of the clip that has already been received and decoded, but was encoded independently from the current segment. The proposed decoder is capable of significantly improve the Quality of Experience of the user without incurring significant overhead to the storage and computational complexities of both the encoder and the decoder. We present simulation results using the HEVC reference encoder and standard test clips, and discuss areas of improvements to the algorithm and potential ways of incorporating the technique to a video streaming system or standards.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134378809","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Single-Pass Dependent Bit Allocation in Temporal Scalability Video Coding 时间可扩展性视频编码中的单通道相关位分配
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.87
Jiaying Liu, Yongjin Cho, Zongming Guo
Summary form only given. In the scalable video coding, we refer to a group-of-pictures (GOP) structure that is composed of hierarchically aligned B-pictures. It employs generalized B-pictures that can be used as a reference to following inter-coded frames. Although it introduces a structural encoding delay of one GOP size, it provides much higher coding efficiency than the conventional GOP structures [2]. Moreover, due to its natural capability of providing the temporal scalability, it is employed as a GOP structure of H.264/SVC [3]. Because of the complex inter-layer dependence of hierarchical B-pictures, the development of an efficient and effective bit allocation algorithm for H.264/SVC is a challenging task. There are several bit allocation algorithms that considered the inter-layer dependence in the literature before. Schwarz et al. proposed the QP cascading scheme that applies a fixed quantization parameter (QP) difference between adjacent temporal layers. Liu et al. introduced constant weights to temporal layers in their H.264/SVC rate control algorithm. Although these algorithms achieve superior coding efficiency, they are limited in two aspects. First, the inter-layer dependence is heuristically addressed. Second, the input video characteristics are not taken into account. For these reasons, the optimality of these bit allocation algorithms cannot be guaranteed. We propose a single-pass dependent bit allocation algorithm for scalable video coding with hierarchical B-pictures in this work. It is generally perceived that dependent bit allocation algorithms cannot be practically employed due to their extremely high complexity requirement. To develop a practical single-pass bit allocation algorithm, we use the number of skipped blocks and the ratio of the mean absolute difference (MAD) as features to measure the inter-layer signal dependence of input video signals. The proposed algorithm performs bit allocation at the target bit rate with two mechanisms: 1) the GOP based rate control and 2) adaptive temporal layer QP decision. The superior performance of the proposed algorithm is demonstrated by experimental results, which is benchmarked by two other single-pass bit allocation algorithms in the literature. The rate and the PSNR coding performance of the proposed scheme and two benchmarks at various target bit rates for GOP-4 and GOP-8, respectively. We see that the proposed rate control algorithm achieves about 0.2-0.3dB improvement in coding efficiency as compared to JSVM. Furthermore, the proposed rate control algorithm outperforms Liu's Algorithm by a significant margin.
只提供摘要形式。在可伸缩视频编码中,我们提到了一种由分层排列的b图组成的图像组(GOP)结构。它采用广义的b图,可以作为后续编码间帧的参考。虽然它引入了一个GOP大小的结构编码延迟,但它提供了比传统GOP结构[2]更高的编码效率。此外,由于其提供时间可扩展性的天然能力,它被用作H.264/SVC[3]的GOP结构。由于H.264/SVC的分层b图像具有复杂的层间依赖性,因此开发一种高效的H.264/SVC比特分配算法是一项具有挑战性的任务。文献中已有几种考虑层间依赖性的位分配算法。Schwarz等人提出了QP级联方案,该方案在相邻的时间层之间应用固定的量化参数(QP)差。Liu等人在H.264/SVC速率控制算法中对时间层引入恒定权值。这些算法虽然具有较高的编码效率,但在两个方面存在局限性。首先,启发式地解决了层间依赖。其次,没有考虑输入视频的特性。由于这些原因,不能保证这些位分配算法的最优性。在这项工作中,我们提出了一种单通道相关的位分配算法,用于具有分层b图像的可扩展视频编码。通常认为,依赖位分配算法由于其极高的复杂度要求而不能实际应用。为了开发一种实用的单通位分配算法,我们使用跳过块的数量和平均绝对差(MAD)的比率作为特征来衡量输入视频信号的层间信号依赖性。该算法通过两种机制实现目标比特率下的比特分配:1)基于GOP的速率控制和2)自适应时序QP决策。实验结果证明了该算法的优越性能,并与文献中另外两种单通比特分配算法进行了基准测试。在不同的目标比特率下,分别对GOP-4和GOP-8进行了速率和PSNR编码性能测试。我们看到,与JSVM相比,所提出的速率控制算法的编码效率提高了约0.2-0.3dB。此外,所提出的速率控制算法明显优于Liu的算法。
{"title":"Single-Pass Dependent Bit Allocation in Temporal Scalability Video Coding","authors":"Jiaying Liu, Yongjin Cho, Zongming Guo","doi":"10.1109/DCC.2013.87","DOIUrl":"https://doi.org/10.1109/DCC.2013.87","url":null,"abstract":"Summary form only given. In the scalable video coding, we refer to a group-of-pictures (GOP) structure that is composed of hierarchically aligned B-pictures. It employs generalized B-pictures that can be used as a reference to following inter-coded frames. Although it introduces a structural encoding delay of one GOP size, it provides much higher coding efficiency than the conventional GOP structures [2]. Moreover, due to its natural capability of providing the temporal scalability, it is employed as a GOP structure of H.264/SVC [3]. Because of the complex inter-layer dependence of hierarchical B-pictures, the development of an efficient and effective bit allocation algorithm for H.264/SVC is a challenging task. There are several bit allocation algorithms that considered the inter-layer dependence in the literature before. Schwarz et al. proposed the QP cascading scheme that applies a fixed quantization parameter (QP) difference between adjacent temporal layers. Liu et al. introduced constant weights to temporal layers in their H.264/SVC rate control algorithm. Although these algorithms achieve superior coding efficiency, they are limited in two aspects. First, the inter-layer dependence is heuristically addressed. Second, the input video characteristics are not taken into account. For these reasons, the optimality of these bit allocation algorithms cannot be guaranteed. We propose a single-pass dependent bit allocation algorithm for scalable video coding with hierarchical B-pictures in this work. It is generally perceived that dependent bit allocation algorithms cannot be practically employed due to their extremely high complexity requirement. To develop a practical single-pass bit allocation algorithm, we use the number of skipped blocks and the ratio of the mean absolute difference (MAD) as features to measure the inter-layer signal dependence of input video signals. The proposed algorithm performs bit allocation at the target bit rate with two mechanisms: 1) the GOP based rate control and 2) adaptive temporal layer QP decision. The superior performance of the proposed algorithm is demonstrated by experimental results, which is benchmarked by two other single-pass bit allocation algorithms in the literature. The rate and the PSNR coding performance of the proposed scheme and two benchmarks at various target bit rates for GOP-4 and GOP-8, respectively. We see that the proposed rate control algorithm achieves about 0.2-0.3dB improvement in coding efficiency as compared to JSVM. Furthermore, the proposed rate control algorithm outperforms Liu's Algorithm by a significant margin.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115591782","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Coding Tree Depth Estimation for Complexity Reduction of HEVC 降低HEVC复杂度的编码树深度估计
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.12
G. Corrêa, P. Assunção, L. Agostini, L. Cruz
The emerging HEVC standard introduces a number of tools which increase compression efficiency in comparison to its predecessors at the cost of greater computational complexity. This paper proposes a complexity control method for HEVC encoders based on dynamic adjustment of the newly proposed coding tree structures. The method improves a previous solution by adopting a strategy that takes into consideration both spatial and temporal correlation in order to decide the maximum coding tree depth allowed for each coding tree block. Complexity control capability is increased in comparison to a previous work, while compression losses are decreased by 70%. Experimental results show that the encoder computational complexity can be downscaled to 60% with an average bit rate increase around 1.3% and a PSNR decrease under 0.07 dB.
新兴的HEVC标准引入了许多工具,与之前的标准相比,这些工具以更高的计算复杂度为代价提高了压缩效率。本文提出了一种基于编码树结构动态调整的HEVC编码器复杂度控制方法。该方法改进了先前的解决方案,采用了考虑空间和时间相关性的策略,以确定每个编码树块允许的最大编码树深度。与以前的工作相比,复杂性控制能力得到了提高,而压缩损失降低了70%。实验结果表明,该编码器的计算复杂度可降低60%,平均比特率提高1.3%左右,PSNR降低0.07 dB以下。
{"title":"Coding Tree Depth Estimation for Complexity Reduction of HEVC","authors":"G. Corrêa, P. Assunção, L. Agostini, L. Cruz","doi":"10.1109/DCC.2013.12","DOIUrl":"https://doi.org/10.1109/DCC.2013.12","url":null,"abstract":"The emerging HEVC standard introduces a number of tools which increase compression efficiency in comparison to its predecessors at the cost of greater computational complexity. This paper proposes a complexity control method for HEVC encoders based on dynamic adjustment of the newly proposed coding tree structures. The method improves a previous solution by adopting a strategy that takes into consideration both spatial and temporal correlation in order to decide the maximum coding tree depth allowed for each coding tree block. Complexity control capability is increased in comparison to a previous work, while compression losses are decreased by 70%. Experimental results show that the encoder computational complexity can be downscaled to 60% with an average bit rate increase around 1.3% and a PSNR decrease under 0.07 dB.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116038730","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 49
Texture Compression 纹理压缩
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.30
Georgios Georgiadis, A. Chiuso, Stefano Soatto
We characterize ``visual textures'' as realizations of a stationary, ergodic, Markovian process, and propose using its approximate minimal sufficient statistics for compressing texture images. We propose inference algorithms for estimating the ``state'' of such process and its ``variability''. These represent the encoding stage. We also propose a non-parametric sampling scheme for decoding, by synthesizing textures from their encoding. While these are not faithful reproductions of the original textures (so they would fail a comparison test based on PSNR), they capture the statistical properties of the underlying process, as we demonstrate empirically. We also quantify the tradeoff between fidelity (measured by a proxy of a perceptual score) and complexity.
我们将“视觉纹理”描述为一个平稳的、遍历的、马尔可夫过程的实现,并建议使用其近似最小充分统计量来压缩纹理图像。我们提出了用于估计这种过程的“状态”及其“可变性”的推理算法。这些代表编码阶段。我们还提出了一种解码的非参数采样方案,通过纹理编码合成纹理。虽然这些不是原始纹理的忠实复制品(因此它们将无法通过基于PSNR的比较测试),但它们捕获了底层过程的统计特性,正如我们经验证明的那样。我们还量化了保真度(通过感知分数的代理来衡量)和复杂性之间的权衡。
{"title":"Texture Compression","authors":"Georgios Georgiadis, A. Chiuso, Stefano Soatto","doi":"10.1109/DCC.2013.30","DOIUrl":"https://doi.org/10.1109/DCC.2013.30","url":null,"abstract":"We characterize ``visual textures'' as realizations of a stationary, ergodic, Markovian process, and propose using its approximate minimal sufficient statistics for compressing texture images. We propose inference algorithms for estimating the ``state'' of such process and its ``variability''. These represent the encoding stage. We also propose a non-parametric sampling scheme for decoding, by synthesizing textures from their encoding. While these are not faithful reproductions of the original textures (so they would fail a comparison test based on PSNR), they capture the statistical properties of the underlying process, as we demonstrate empirically. We also quantify the tradeoff between fidelity (measured by a proxy of a perceptual score) and complexity.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"157 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123944816","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Image Compression via Colorization Using Semi-Regular Color Samples 使用半规则颜色样本通过着色图像压缩
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.112
Chang-jiang Zhang, Hui Fang
Summary form only given. We improves colorization-based image compression by sparsely sampling color points on a semi-regular grid and compressing them using JPEG. We generate variations of sampling locations based on extreme gray-scale values to to further improve PSNR.
只提供摘要形式。我们通过在半规则网格上稀疏采样颜色点并使用JPEG压缩它们来改进基于颜色的图像压缩。我们根据极端灰度值生成采样位置的变化,以进一步提高PSNR。
{"title":"Image Compression via Colorization Using Semi-Regular Color Samples","authors":"Chang-jiang Zhang, Hui Fang","doi":"10.1109/DCC.2013.112","DOIUrl":"https://doi.org/10.1109/DCC.2013.112","url":null,"abstract":"Summary form only given. We improves colorization-based image compression by sparsely sampling color points on a semi-regular grid and compressing them using JPEG. We generate variations of sampling locations based on extreme gray-scale values to to further improve PSNR.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124473510","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
From Run Length Encoding to LZ78 and Back Again 从运行长度编码到LZ78再返回
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.22
Yuya Tamakoshi, T. I., Shunsuke Inenaga, H. Bannai, M. Takeda
In this paper, we present efficient algorithms for interconversion between Lempel-Ziv 78 (LZ78) encoding and run length encoding (RLE). We show how, given an RLE of size n for a string S, we can compute the corresponding LZ78 encoding of size m for S in O((n + m) log σ) time, where σ is the number of distinct characters appearing in S. We also show how, given an LZ78 encoding of size m for a string S, we can compute the corresponding RLE of size n in O(n + m) time. Both algorithms use O(m) extra working space.
本文提出了LZ78 (Lempel-Ziv 78)编码与RLE编码之间的有效转换算法。我们展示了如何在给定字符串S的大小为n的RLE的情况下,在O((n + m) log σ)时间内计算出大小为m的对应的LZ78编码,其中σ是S中出现的不同字符的数量。我们还展示了如何在给定字符串S的大小为m的LZ78编码的情况下,在O(n + m)时间内计算出大小为n的对应的RLE。两种算法都使用O(m)的额外工作空间。
{"title":"From Run Length Encoding to LZ78 and Back Again","authors":"Yuya Tamakoshi, T. I., Shunsuke Inenaga, H. Bannai, M. Takeda","doi":"10.1109/DCC.2013.22","DOIUrl":"https://doi.org/10.1109/DCC.2013.22","url":null,"abstract":"In this paper, we present efficient algorithms for interconversion between Lempel-Ziv 78 (LZ78) encoding and run length encoding (RLE). We show how, given an RLE of size n for a string S, we can compute the corresponding LZ78 encoding of size m for S in O((n + m) log σ) time, where σ is the number of distinct characters appearing in S. We also show how, given an LZ78 encoding of size m for a string S, we can compute the corresponding RLE of size n in O(n + m) time. Both algorithms use O(m) extra working space.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"216 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122846541","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
期刊
2013 Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1