首页 > 最新文献

2013 Data Compression Conference最新文献

英文 中文
Progressive Image Restoration through Hybrid Graph Laplacian Regularization 基于混合图拉普拉斯正则化的渐进式图像恢复
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.18
Deming Zhai, Xianming Liu, Debin Zhao, Hong Chang, Wen Gao
In this paper, we propose a unified framework to perform progressive image restoration based on hybrid graph Laplacian regularized regression. We first construct a multi-scale representation of the target image by Laplacian pyramid, then progressively recover the degraded image in the scale space from coarse to fine so that the sharp edges and texture can be eventually recovered. On one hand, within each scale, a graph Laplacian regularization model represented by implicit kernel is learned which simultaneously minimizes the least square error on the measured samples and preserves the geometrical structure of the image data space by exploring non-local self-similarity. In this procedure, the intrinsic manifold structure is considered by using both measured and unmeasured samples. On the other hand, between two scales, the proposed model is extended to the parametric manner through explicit kernel mapping to model the inter-scale correlation, in which the local structure regularity is learned and propagated from coarser to finer scales. Experimental results on benchmark test images demonstrate that the proposed method achieves better performance than state-of-the-art image restoration algorithms.
本文提出了一种基于混合图拉普拉斯正则化回归的渐进图像恢复统一框架。首先利用拉普拉斯金字塔构造目标图像的多尺度表示,然后在尺度空间中由粗到细逐步恢复退化图像,最终恢复出锐利的边缘和纹理。一方面,在每个尺度内,学习一种以隐式核为代表的图拉普拉斯正则化模型,该模型通过探索非局部自相似性,使测量样本的最小二乘误差最小化,同时保留图像数据空间的几何结构;在这个过程中,本征流形结构被考虑使用测量和未测量的样本。另一方面,在两个尺度之间,通过显式核映射将模型扩展到参数化的方式来建模尺度间的相关性,其中局部结构规则被学习并从粗尺度传播到细尺度。在基准测试图像上的实验结果表明,该方法比现有的图像恢复算法具有更好的性能。
{"title":"Progressive Image Restoration through Hybrid Graph Laplacian Regularization","authors":"Deming Zhai, Xianming Liu, Debin Zhao, Hong Chang, Wen Gao","doi":"10.1109/DCC.2013.18","DOIUrl":"https://doi.org/10.1109/DCC.2013.18","url":null,"abstract":"In this paper, we propose a unified framework to perform progressive image restoration based on hybrid graph Laplacian regularized regression. We first construct a multi-scale representation of the target image by Laplacian pyramid, then progressively recover the degraded image in the scale space from coarse to fine so that the sharp edges and texture can be eventually recovered. On one hand, within each scale, a graph Laplacian regularization model represented by implicit kernel is learned which simultaneously minimizes the least square error on the measured samples and preserves the geometrical structure of the image data space by exploring non-local self-similarity. In this procedure, the intrinsic manifold structure is considered by using both measured and unmeasured samples. On the other hand, between two scales, the proposed model is extended to the parametric manner through explicit kernel mapping to model the inter-scale correlation, in which the local structure regularity is learned and propagated from coarser to finer scales. Experimental results on benchmark test images demonstrate that the proposed method achieves better performance than state-of-the-art image restoration algorithms.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116324178","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
A Distortion Metric for the Lossy Compression of DNA Microarray Images DNA微阵列图像有损压缩的失真度量
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.26
Miguel Hernández-Cabronero, Victor Sanchez, M. Marcellin, J. Serra-Sagristà
DNA micro arrays are state-of-the-art tools in biological and medical research. In this work, we discuss the suitability of lossy compression for DNA micro array images and highlight the necessity for a distortion metric to assess the loss of relevant information. We also propose one possible metric that considers the basic image features employed by most DNA micro array analysis techniques. Experimental results indicate that the proposed metric can identify and differentiate important and unimportant changes in DNA micro array images.
DNA微阵列是生物和医学研究中最先进的工具。在这项工作中,我们讨论了DNA微阵列图像有损压缩的适用性,并强调了使用失真度量来评估相关信息损失的必要性。我们还提出了一种可能的度量,该度量考虑了大多数DNA微阵列分析技术所采用的基本图像特征。实验结果表明,所提出的度量可以识别和区分DNA微阵列图像中重要和不重要的变化。
{"title":"A Distortion Metric for the Lossy Compression of DNA Microarray Images","authors":"Miguel Hernández-Cabronero, Victor Sanchez, M. Marcellin, J. Serra-Sagristà","doi":"10.1109/DCC.2013.26","DOIUrl":"https://doi.org/10.1109/DCC.2013.26","url":null,"abstract":"DNA micro arrays are state-of-the-art tools in biological and medical research. In this work, we discuss the suitability of lossy compression for DNA micro array images and highlight the necessity for a distortion metric to assess the loss of relevant information. We also propose one possible metric that considers the basic image features employed by most DNA micro array analysis techniques. Experimental results indicate that the proposed metric can identify and differentiate important and unimportant changes in DNA micro array images.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122063506","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Multiterminal Source Coding for Many Sensors with Entropy Coding and Gaussian Process Regression 基于熵编码和高斯过程回归的多端传感器源编码
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.62
Samuel Cheng
Summary form only given. In this paper, we take a different approach from the coding community. Instead of taking the usual route of quantization plus Slepian-Wolf coding, we do not perform any Slepian-Wolf coding on the transmitter side. We simply perform quantization on the sensor readings, compress the quantization indexes with conventional entropy coding, and send the compressed indexes to the receiver. On the decoder side, we simply perform entropy decoding and Gaussian process regression to reconstruct the joint source. To reduce the sum rate over all sensors, some sensors are censored and do not transmit anything to the decoder.
只提供摘要形式。在本文中,我们采用了与编码社区不同的方法。我们没有采用通常的量化加睡眠狼编码的方法,而是在发送端不执行任何睡眠狼编码。我们简单地对传感器读数进行量化,用传统的熵编码压缩量化指标,并将压缩后的指标发送给接收器。在解码器端,我们简单地执行熵解码和高斯过程回归来重建联合源。为了降低所有传感器的和速率,一些传感器被删减,不向解码器传输任何东西。
{"title":"Multiterminal Source Coding for Many Sensors with Entropy Coding and Gaussian Process Regression","authors":"Samuel Cheng","doi":"10.1109/DCC.2013.62","DOIUrl":"https://doi.org/10.1109/DCC.2013.62","url":null,"abstract":"Summary form only given. In this paper, we take a different approach from the coding community. Instead of taking the usual route of quantization plus Slepian-Wolf coding, we do not perform any Slepian-Wolf coding on the transmitter side. We simply perform quantization on the sensor readings, compress the quantization indexes with conventional entropy coding, and send the compressed indexes to the receiver. On the decoder side, we simply perform entropy decoding and Gaussian process regression to reconstruct the joint source. To reduce the sum rate over all sensors, some sensors are censored and do not transmit anything to the decoder.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"159 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122195954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Evaluation of Efficient Compression Properties of the Complete Oscillator Method, Part 1: Canonical Signals 完备振子方法的有效压缩特性评估,第1部分:典型信号
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.73
I. Gorodnitsky, Anton Y. Yen
The paper describes a highly efficient coding scheme based on the Complete Oscillator Method (COM). The COM has a number of powerful theoretical properties which enable it to provide very compact models for a wide range of deterministic signals. Several of these properties are studied here. The theoretical COM is shown to model and synthesize exactly all types of stationary signals, irrespective of the dimension of the system that generated it, using only one model parameter. The exact representation property independent of data dimension extends to certain amplitude-variable signals as well. The COM also reconstructs with high fidelity a number of other classes of nonstationary signals for which the exact representation cannot be guaranteed. One such class encompasses frequency-modulated signals presented here. The theoretical results obtained under idealized conditions are related to practical discrete implementations, where the COM is shown to be robust to deviations from the ideal conditions. In non-ideal conditions, increasing the order of the COM to two terms, four parameters total, delivers near exact models in many cases. The compact representation property of the COM is illustrated on several canonical waveforms, which provide representative examples for each class of signal studied here.
本文提出了一种基于完全振子法(COM)的高效编码方案。COM具有许多强大的理论特性,使其能够为大范围的确定性信号提供非常紧凑的模型。这里研究了其中的一些特性。理论COM被证明可以准确地建模和合成所有类型的平稳信号,而不考虑产生它的系统的尺寸,只使用一个模型参数。这种与数据维数无关的精确表示特性也适用于某些变幅信号。COM还以高保真度重建了许多其他类型的非平稳信号,这些信号的精确表示不能得到保证。一个这样的类包括这里介绍的调频信号。在理想条件下获得的理论结果与实际的离散实现有关,其中COM显示出对偏离理想条件的鲁棒性。在非理想条件下,将COM的顺序增加到两个项,总共四个参数,在许多情况下可以提供接近精确的模型。在几个典型波形上说明了COM的紧凑表示特性,为本文研究的每一类信号提供了代表性的例子。
{"title":"Evaluation of Efficient Compression Properties of the Complete Oscillator Method, Part 1: Canonical Signals","authors":"I. Gorodnitsky, Anton Y. Yen","doi":"10.1109/DCC.2013.73","DOIUrl":"https://doi.org/10.1109/DCC.2013.73","url":null,"abstract":"The paper describes a highly efficient coding scheme based on the Complete Oscillator Method (COM). The COM has a number of powerful theoretical properties which enable it to provide very compact models for a wide range of deterministic signals. Several of these properties are studied here. The theoretical COM is shown to model and synthesize exactly all types of stationary signals, irrespective of the dimension of the system that generated it, using only one model parameter. The exact representation property independent of data dimension extends to certain amplitude-variable signals as well. The COM also reconstructs with high fidelity a number of other classes of nonstationary signals for which the exact representation cannot be guaranteed. One such class encompasses frequency-modulated signals presented here. The theoretical results obtained under idealized conditions are related to practical discrete implementations, where the COM is shown to be robust to deviations from the ideal conditions. In non-ideal conditions, increasing the order of the COM to two terms, four parameters total, delivers near exact models in many cases. The compact representation property of the COM is illustrated on several canonical waveforms, which provide representative examples for each class of signal studied here.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116023690","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Considerations and Algorithms for Compression of Sets 集压缩的注意事项和算法
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.83
N. Larsson
We consider compression of unordered sets of distinct elements, focusing particularly on compressing sets of fixed-length bit strings in the presence of statistical information. We address previous work, and outline a novel compression algorithm that allows transparent incorporation of various estimates for probability distribution. Experiments allow the conclusion that set compression can benefit from incorporating statistics, using our method or variants of previously known techniques.
我们考虑不同元素的无序集的压缩,特别关注在存在统计信息的情况下压缩固定长度的位串集。我们解决了以前的工作,并概述了一种新的压缩算法,该算法允许透明地合并各种概率分布估计。实验得出结论,集合压缩可以受益于结合统计,使用我们的方法或变体的先前已知的技术。
{"title":"Considerations and Algorithms for Compression of Sets","authors":"N. Larsson","doi":"10.1109/DCC.2013.83","DOIUrl":"https://doi.org/10.1109/DCC.2013.83","url":null,"abstract":"We consider compression of unordered sets of distinct elements, focusing particularly on compressing sets of fixed-length bit strings in the presence of statistical information. We address previous work, and outline a novel compression algorithm that allows transparent incorporation of various estimates for probability distribution. Experiments allow the conclusion that set compression can benefit from incorporating statistics, using our method or variants of previously known techniques.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116082610","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Context-Based Algorithms for the List-Update Problem under Alternative Cost Models 备选成本模型下基于上下文的列表更新问题算法
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.44
Shahin Kamali, Susana Ladra, A. López-Ortiz, Diego Seco
The List-Update Problem is a well studied online problem with direct applications in data compression. Although the model proposed by Sleator & Tarjan has become the standard in the field for the problem, its applicability in some domains, and in particular for compression purposes, has been questioned. In this paper, we focus on two alternative models for the problem that arguably have more practical significance than the standard model. We provide new algorithms for these models, and show that these algorithms outperform all classical algorithms under the discussed models. This is done via an empirical study of the performance of these algorithms on the reference data set for the list-update problem. The presented algorithms make use of the context-based strategies for compression, which have not been considered before in the context of the list-update problem and lead to improved compression algorithms. In addition, we study the adaptability of these algorithms to different measures of locality of reference and compressibility.
列表更新问题是一个在线问题,在数据压缩中有直接的应用。尽管Sleator & Tarjan提出的模型已经成为该问题领域的标准,但其在某些领域的适用性,特别是在压缩目的方面,仍受到质疑。在本文中,我们重点讨论了两种替代模型,这些模型可以说比标准模型更具有实际意义。我们为这些模型提供了新的算法,并证明这些算法在所讨论的模型下优于所有经典算法。这是通过对这些算法在列表更新问题的参考数据集上的性能进行实证研究来完成的。本文提出的算法利用了基于上下文的压缩策略,这在之前的列表更新问题中没有被考虑过,并导致了改进的压缩算法。此外,我们还研究了这些算法对不同参考局部性和可压缩性度量的适应性。
{"title":"Context-Based Algorithms for the List-Update Problem under Alternative Cost Models","authors":"Shahin Kamali, Susana Ladra, A. López-Ortiz, Diego Seco","doi":"10.1109/DCC.2013.44","DOIUrl":"https://doi.org/10.1109/DCC.2013.44","url":null,"abstract":"The List-Update Problem is a well studied online problem with direct applications in data compression. Although the model proposed by Sleator & Tarjan has become the standard in the field for the problem, its applicability in some domains, and in particular for compression purposes, has been questioned. In this paper, we focus on two alternative models for the problem that arguably have more practical significance than the standard model. We provide new algorithms for these models, and show that these algorithms outperform all classical algorithms under the discussed models. This is done via an empirical study of the performance of these algorithms on the reference data set for the list-update problem. The presented algorithms make use of the context-based strategies for compression, which have not been considered before in the context of the list-update problem and lead to improved compression algorithms. In addition, we study the adaptability of these algorithms to different measures of locality of reference and compressibility.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"89 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120901267","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Structural Group Sparse Representation for Image Compressive Sensing Recovery 图像压缩感知恢复的结构群稀疏表示
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.41
Jian Zhang, Debin Zhao, F. Jiang, Wen Gao
Compressive Sensing (CS) theory shows that a signal can be decoded from many fewer measurements than suggested by the Nyquist sampling theory, when the signal is sparse in some domain. Most of conventional CS recovery approaches, however, exploited a set of fixed bases (e.g. DCT, wavelet, contour let and gradient domain) for the entirety of a signal, which are irrespective of the nonstationarity of natural signals and cannot achieve high enough degree of sparsity, thus resulting in poor rate-distortion performance. In this paper, we propose a new framework for image compressive sensing recovery via structural group sparse representation (SGSR) modeling, which enforces image sparsity and self-similarity simultaneously under a unified framework in an adaptive group domain, thus greatly confining the CS solution space. In addition, an efficient iterative shrinkage/thresholding algorithm based technique is developed to solve the above optimization problem. Experimental results demonstrate that the novel CS recovery strategy achieves significant performance improvements over the current state-of-the-art schemes and exhibits nice convergence.
压缩感知(CS)理论表明,当信号在某些域中稀疏时,可以从比奈奎斯特采样理论所建议的更少的测量中解码信号。然而,传统的CS恢复方法大多对整个信号使用一组固定基(如DCT、小波、轮廓let和梯度域),不考虑自然信号的非平稳性,不能达到足够高的稀疏度,导致率失真性能较差。本文提出了一种基于结构群稀疏表示(structural group sparse representation, SGSR)建模的图像压缩感知恢复新框架,该框架在自适应群域的统一框架下同时增强了图像的稀疏性和自相似性,从而极大地限制了CS解空间。此外,本文还提出了一种基于迭代收缩/阈值算法的优化方法。实验结果表明,与现有方案相比,新的CS恢复策略取得了显著的性能改进,并表现出良好的收敛性。
{"title":"Structural Group Sparse Representation for Image Compressive Sensing Recovery","authors":"Jian Zhang, Debin Zhao, F. Jiang, Wen Gao","doi":"10.1109/DCC.2013.41","DOIUrl":"https://doi.org/10.1109/DCC.2013.41","url":null,"abstract":"Compressive Sensing (CS) theory shows that a signal can be decoded from many fewer measurements than suggested by the Nyquist sampling theory, when the signal is sparse in some domain. Most of conventional CS recovery approaches, however, exploited a set of fixed bases (e.g. DCT, wavelet, contour let and gradient domain) for the entirety of a signal, which are irrespective of the nonstationarity of natural signals and cannot achieve high enough degree of sparsity, thus resulting in poor rate-distortion performance. In this paper, we propose a new framework for image compressive sensing recovery via structural group sparse representation (SGSR) modeling, which enforces image sparsity and self-similarity simultaneously under a unified framework in an adaptive group domain, thus greatly confining the CS solution space. In addition, an efficient iterative shrinkage/thresholding algorithm based technique is developed to solve the above optimization problem. Experimental results demonstrate that the novel CS recovery strategy achieves significant performance improvements over the current state-of-the-art schemes and exhibits nice convergence.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"15 3","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"120928649","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 77
Evaluation of Efficient Compression Properties of the Complete Oscillator Method, Part 2: Speech Coding 完整振荡器方法的有效压缩特性评估,第2部分:语音编码
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.110
Anton Y. Yen, I. Gorodnitsky
Summary form only given. This paper examines the performance of the recently proposed Complete Oscillator Method (COM) in the context of coding speech. The COM is shown to provide several advantages over traditional predictive coding techniques. Unlike the cascaded method employed by codecs such as Adaptive Multi-Rate (AMR), the COM encodes short and long-term data features jointly using a single, flexible representation. Joint approaches have previously been shown to yield efficiency gains [1]. Furthermore, the COM does not always require an explicit encoding of the residual error to reconstruct the signal. As AMR can allocate as much as 85% of its coding budget towards encoding the residual, there is substantial motivation for finding alternatives to source-filter coding methods. The first part of the paper compares the synthesis of speech frames using the COM versus a combination of linear predictor and adaptive codebook (LPAC) in order to assess the deterministic modeling capabilities of the COM relative to linear predictive codes. With both approaches optimized by minimizing the perceptually-weighted error (PWE) between the original and reconstructed speech, the COM is shown to achieve lower PWE on average than LPAC as implemented in the AMR standard for several types of speech. The COM improved PWE in 78.20% of voiced frames yielding a 2.02 dB PWE gain on average. For voiced to unvoiced transitions, the COM improved PWE in 76.75% of the frames with a 1.26 dB average gain. For unvoiced speech, the COM consistently improved PWE but the average gain was not significant. Only for unvoiced to voiced transitions did the COM not produce gains in average PWE. The second part of the paper compares the synthesis of speech frames using the COM at several bit rates to standard AMR and Speex codecs to show that the COM can produce comparable quality speech in a significant percentage of frames. Using weighted spectral slope distance (WSS) as a metric, a 5.5 kbps COM was seen to outperform 12.2 kbps AMR in 24.12% of speech frames. These results are not intended to demonstrate the workings of a COM-only speech coder, but rather to suggest how existing codecs can achieve lower bit rates by using the COM to encode some subset of frames. For example, by using the COM in the lowest bit rate mode sufficient to achieve a similar WSS as 12.2 kbps AMR, the average bit rate can potentially be reduced to 9.16 kbps.
只提供摘要形式。本文研究了最近提出的完全振荡器方法(COM)在语音编码中的性能。与传统的预测编码技术相比,COM具有许多优点。与自适应多速率(AMR)等编解码器采用的级联方法不同,COM使用单一、灵活的表示方式联合编码短期和长期数据特征。联合方法先前已被证明可以提高效率[1]。此外,COM并不总是需要对残差进行显式编码来重建信号。由于AMR可以将高达85%的编码预算分配给残差编码,因此寻找源滤波器编码方法的替代方法具有很大的动机。本文的第一部分比较了使用COM的语音帧合成与线性预测器和自适应码本(LPAC)的组合,以评估COM相对于线性预测码的确定性建模能力。两种方法都通过最小化原始语音和重建语音之间的感知加权误差(PWE)进行了优化,结果表明,对于几种类型的语音,COM的平均PWE比AMR标准中实现的LPAC要低。COM提高了78.20%的浊音帧的PWE,平均PWE增益为2.02 dB。对于浊音到非浊音转换,COM在76.75%的帧中提高了PWE,平均增益为1.26 dB。对于不发音的语音,COM持续提高PWE,但平均增益并不显著。只有在不发音到发音的转换中,COM不会产生平均PWE的增益。论文的第二部分比较了使用COM在几个比特率下与标准AMR和Speex编解码器合成语音帧的情况,以表明COM可以在很大比例的帧中产生相当质量的语音。使用加权频谱斜率距离(WSS)作为度量,在24.12%的语音帧中,5.5 kbps的COM表现优于12.2 kbps的AMR。这些结果并不是为了演示纯COM语音编码器的工作原理,而是建议现有的编解码器如何通过使用COM对帧的某些子集进行编码来实现更低的比特率。例如,通过在最低比特率模式下使用COM,足以实现与12.2 kbps AMR相似的WSS,平均比特率可能会降低到9.16 kbps。
{"title":"Evaluation of Efficient Compression Properties of the Complete Oscillator Method, Part 2: Speech Coding","authors":"Anton Y. Yen, I. Gorodnitsky","doi":"10.1109/DCC.2013.110","DOIUrl":"https://doi.org/10.1109/DCC.2013.110","url":null,"abstract":"Summary form only given. This paper examines the performance of the recently proposed Complete Oscillator Method (COM) in the context of coding speech. The COM is shown to provide several advantages over traditional predictive coding techniques. Unlike the cascaded method employed by codecs such as Adaptive Multi-Rate (AMR), the COM encodes short and long-term data features jointly using a single, flexible representation. Joint approaches have previously been shown to yield efficiency gains [1]. Furthermore, the COM does not always require an explicit encoding of the residual error to reconstruct the signal. As AMR can allocate as much as 85% of its coding budget towards encoding the residual, there is substantial motivation for finding alternatives to source-filter coding methods. The first part of the paper compares the synthesis of speech frames using the COM versus a combination of linear predictor and adaptive codebook (LPAC) in order to assess the deterministic modeling capabilities of the COM relative to linear predictive codes. With both approaches optimized by minimizing the perceptually-weighted error (PWE) between the original and reconstructed speech, the COM is shown to achieve lower PWE on average than LPAC as implemented in the AMR standard for several types of speech. The COM improved PWE in 78.20% of voiced frames yielding a 2.02 dB PWE gain on average. For voiced to unvoiced transitions, the COM improved PWE in 76.75% of the frames with a 1.26 dB average gain. For unvoiced speech, the COM consistently improved PWE but the average gain was not significant. Only for unvoiced to voiced transitions did the COM not produce gains in average PWE. The second part of the paper compares the synthesis of speech frames using the COM at several bit rates to standard AMR and Speex codecs to show that the COM can produce comparable quality speech in a significant percentage of frames. Using weighted spectral slope distance (WSS) as a metric, a 5.5 kbps COM was seen to outperform 12.2 kbps AMR in 24.12% of speech frames. These results are not intended to demonstrate the workings of a COM-only speech coder, but rather to suggest how existing codecs can achieve lower bit rates by using the COM to encode some subset of frames. For example, by using the COM in the lowest bit rate mode sufficient to achieve a similar WSS as 12.2 kbps AMR, the average bit rate can potentially be reduced to 9.16 kbps.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127005446","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast Coding Unit Depth Decision Algorithm for Interframe Coding in HEVC HEVC帧间编码的快速编码单元深度判定算法
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.13
Yongfei Zhang, Haibo Wang, Zhe Li
As the next generation standard of video coding, the High Efficiency Video Coding (HEVC) achieves significantly better coding efficiency than all existing video coding standards. A Coding Unit (CU) quad tree concept is introduced to HEVC to improve the coding efficiency. Each CU node in quad tree will be traversed by depth first search process to find the best Coding Tree Unit (CTU) partition. Although this quad tree search process can obtain the best CTU partition, it is very time consuming, especially in interframe coding. To alleviate the encoder computation load in interframe coding, a fast CU depth decision method is proposed by reducing the depth search range. Based on the depth information correlation between spatio-temporal adjacent CTUs and the current CTU, some depths can be adaptively excluded from the depth search process in advance. Experimental results show that the proposed scheme provides almost 30% encoder time savings on average compared to the default encoding scheme in HM8.0 with only 0.38% bit rate increment in coding performance.
高效视频编码(High Efficiency video coding, HEVC)作为下一代视频编码标准,其编码效率明显高于现有的所有视频编码标准。为了提高编码效率,在HEVC中引入了编码单元四叉树的概念。通过深度优先搜索遍历四叉树中的每个CU节点,找到最佳的编码树单元分区。虽然这种四叉树搜索过程可以获得最佳的CTU分区,但非常耗时,特别是在帧间编码时。为了减轻帧间编码时编码器的计算负担,提出了一种通过减小深度搜索范围来快速确定帧间编码深度的方法。基于时空相邻CTU与当前CTU之间的深度信息相关性,可以提前自适应地从深度搜索过程中排除某些深度。实验结果表明,与HM8.0的默认编码方案相比,该方案平均节省了近30%的编码器时间,编码性能仅提高0.38%。
{"title":"Fast Coding Unit Depth Decision Algorithm for Interframe Coding in HEVC","authors":"Yongfei Zhang, Haibo Wang, Zhe Li","doi":"10.1109/DCC.2013.13","DOIUrl":"https://doi.org/10.1109/DCC.2013.13","url":null,"abstract":"As the next generation standard of video coding, the High Efficiency Video Coding (HEVC) achieves significantly better coding efficiency than all existing video coding standards. A Coding Unit (CU) quad tree concept is introduced to HEVC to improve the coding efficiency. Each CU node in quad tree will be traversed by depth first search process to find the best Coding Tree Unit (CTU) partition. Although this quad tree search process can obtain the best CTU partition, it is very time consuming, especially in interframe coding. To alleviate the encoder computation load in interframe coding, a fast CU depth decision method is proposed by reducing the depth search range. Based on the depth information correlation between spatio-temporal adjacent CTUs and the current CTU, some depths can be adaptively excluded from the depth search process in advance. Experimental results show that the proposed scheme provides almost 30% encoder time savings on average compared to the default encoding scheme in HM8.0 with only 0.38% bit rate increment in coding performance.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"123 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124185378","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 68
Real-Time Compression of Intra-Cerebral EEG Using Eigendecomposition with Dynamic Dictionary 基于动态字典特征分解的脑电实时压缩
Pub Date : 2013-03-20 DOI: 10.1109/DCC.2013.68
H. Daou, F. Labeau
A novel technique for Intra-cerebral Electroencephalogram (iEEG) compression in real-time is proposed in this article. This technique uses eigendecomposition and dynamic dictionary update to reduce the EEG channels to only one decor related channel or eigenchannel. Experimental results show that this technique is able to provide low distortion values at very low bit rates (BRs). In addition, performance results of this method show to be better and more stable than JPEG2000. Results do not vary a lot both in time and between different patients which proves the stability of the method.
本文提出了一种新的实时脑内脑电图(iEEG)压缩技术。该技术使用特征分解和动态字典更新将EEG通道减少到只有一个与装饰相关的通道或特征通道。实验结果表明,该技术能够在非常低的比特率(BRs)下提供低失真值。此外,该方法的性能结果表明,该方法比JPEG2000更好,更稳定。结果在时间和不同患者之间变化不大,证明了该方法的稳定性。
{"title":"Real-Time Compression of Intra-Cerebral EEG Using Eigendecomposition with Dynamic Dictionary","authors":"H. Daou, F. Labeau","doi":"10.1109/DCC.2013.68","DOIUrl":"https://doi.org/10.1109/DCC.2013.68","url":null,"abstract":"A novel technique for Intra-cerebral Electroencephalogram (iEEG) compression in real-time is proposed in this article. This technique uses eigendecomposition and dynamic dictionary update to reduce the EEG channels to only one decor related channel or eigenchannel. Experimental results show that this technique is able to provide low distortion values at very low bit rates (BRs). In addition, performance results of this method show to be better and more stable than JPEG2000. Results do not vary a lot both in time and between different patients which proves the stability of the method.","PeriodicalId":388717,"journal":{"name":"2013 Data Compression Conference","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2013-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128051896","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2013 Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1