首页 > 最新文献

Proceedings. Data Compression Conference最新文献

英文 中文
Parameter analysis for the generalized LZ compression of audio 音频广义LZ压缩的参数分析
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.70
D. Kirovski, Zeph Landau
Summary form only given. We introduced (Kirovski and Landau (2004)) a memory-based model of the source signal, which explores multimedia repetitiveness to improve upon compression rates achieved by classic memoryless or simple prediction-based audio compression algorithms such as MP3. The representation error is masked using a psycho-acoustic filter. The goal of the masking function is to set the error such that reconstruction of audible samples is exact whereas the reconstruction of inaudible samples is such that the absolute magnitude of the error is minimized. We compute the entropy of the quantized pointers to all blocks, the quantized pointers to the applied transforms, the quantized scalars used to create the linear combination of transformed blocks, and the error vector returned.
只提供摘要形式。我们介绍了(Kirovski和Landau(2004))一种基于记忆的源信号模型,该模型探索了多媒体重复性,以提高经典的无记忆或简单的基于预测的音频压缩算法(如MP3)所达到的压缩率。使用心理-声学滤波器来掩盖表征误差。掩蔽函数的目标是设置误差,使可听样本的重建是精确的,而不可听样本的重建是使误差的绝对幅度最小化。我们计算指向所有块的量化指针的熵,指向应用变换的量化指针,用于创建转换块的线性组合的量化标量,以及返回的误差向量。
{"title":"Parameter analysis for the generalized LZ compression of audio","authors":"D. Kirovski, Zeph Landau","doi":"10.1109/DCC.2005.70","DOIUrl":"https://doi.org/10.1109/DCC.2005.70","url":null,"abstract":"Summary form only given. We introduced (Kirovski and Landau (2004)) a memory-based model of the source signal, which explores multimedia repetitiveness to improve upon compression rates achieved by classic memoryless or simple prediction-based audio compression algorithms such as MP3. The representation error is masked using a psycho-acoustic filter. The goal of the masking function is to set the error such that reconstruction of audible samples is exact whereas the reconstruction of inaudible samples is such that the absolute magnitude of the error is minimized. We compute the entropy of the quantized pointers to all blocks, the quantized pointers to the applied transforms, the quantized scalars used to create the linear combination of transformed blocks, and the error vector returned.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"40 1","pages":"465-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80069132","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Real-time traversal in grammar-based compressed files 实时遍历基于语法的压缩文件
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.78
L. Gąsieniec, R. Kolpakov, I. Potapov, P. Sant
Summary form only given. In text compression applications, it is important to be able to process compressed data without requiring (complete) decompression. In this context it is crucial to study compression methods that allow time/space efficient access to any fragment of a compressed file without being forced to perform complete decompression. We study here the real-time recovery of consecutive symbols from compressed files, in the context of grammar-based compression. In this setting, a compressed text is represented as a small (a few Kb) dictionary D (containing a set of code words), and a very long (a few Mb) string based on symbols drawn from the dictionary D. The space efficiency of this kind of compression is comparable with standard compression methods based on the Lempel-Ziv approach. We show, that one can visit consecutive symbols of the original text, moving from one symbol to another in constant time and extra O(|D|) space. This algorithm is an improvement of the on-line linear (amortised) time algorithm presented in (L. Gasieniec et al, Proc. 13th Int. Symp. on Fund. of Comp. Theo., LNCS, vol.2138, p.138-152, 2001).
只提供摘要形式。在文本压缩应用程序中,能够处理压缩数据而不需要(完全)解压缩是很重要的。在这种情况下,研究压缩方法是至关重要的,这些方法允许时间/空间有效地访问压缩文件的任何片段,而不必强制执行完全解压缩。本文研究了基于语法压缩的压缩文件中连续符号的实时恢复。在这种情况下,压缩文本被表示为一个小的(几Kb)字典D(包含一组码字)和一个非常长的(几Mb)字符串(基于从字典D中绘制的符号)。这种压缩的空间效率与基于Lempel-Ziv方法的标准压缩方法相当。我们证明,一个人可以访问原始文本的连续符号,在恒定的时间和额外的O(|D|)空间内从一个符号移动到另一个符号。该算法是对(L. Gasieniec et al ., Proc. 13 Int)中提出的在线线性(摊平)时间算法的改进。计算机协会。在基金。西奥公司。生物医学工程学报,vol.2138, p.138-152, 2001)。
{"title":"Real-time traversal in grammar-based compressed files","authors":"L. Gąsieniec, R. Kolpakov, I. Potapov, P. Sant","doi":"10.1109/DCC.2005.78","DOIUrl":"https://doi.org/10.1109/DCC.2005.78","url":null,"abstract":"Summary form only given. In text compression applications, it is important to be able to process compressed data without requiring (complete) decompression. In this context it is crucial to study compression methods that allow time/space efficient access to any fragment of a compressed file without being forced to perform complete decompression. We study here the real-time recovery of consecutive symbols from compressed files, in the context of grammar-based compression. In this setting, a compressed text is represented as a small (a few Kb) dictionary D (containing a set of code words), and a very long (a few Mb) string based on symbols drawn from the dictionary D. The space efficiency of this kind of compression is comparable with standard compression methods based on the Lempel-Ziv approach. We show, that one can visit consecutive symbols of the original text, moving from one symbol to another in constant time and extra O(|D|) space. This algorithm is an improvement of the on-line linear (amortised) time algorithm presented in (L. Gasieniec et al, Proc. 13th Int. Symp. on Fund. of Comp. Theo., LNCS, vol.2138, p.138-152, 2001).","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"9 1","pages":"458-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"81949172","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
Minimum distortion color image retrieval based on Lloyd-clustered Gauss mixtures 基于劳埃德聚类高斯混合的最小失真彩色图像检索
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.52
Sangoh Jeong, R. Gray
We consider image retrieval based on minimum distortion selection of features of color images modelled by Gauss mixtures. The proposed algorithm retrieves the image in a database having minimum distortion when the query image is encoded by a separate Gauss mixture codebook representing each image in the database. We use Gauss mixture vector quantization (GMVQ) for clustering Gauss mixtures, instead of the conventional expectation-maximization (EM) algorithm. Experimental comparison shows that the simpler GMVQ and the EM algorithms have close Gauss mixture parameters with similar convergence speeds. We also provide a new color-interleaving method, reducing the dimension of feature vectors and the size of covariance matrices, thereby reducing computation. This method shows a slightly better retrieval performance than the usual color-interleaving method in HSV color space. Our proposed minimum distortion image retrieval performs better than probabilistic image retrieval.
我们考虑基于高斯混合模型的彩色图像特征的最小失真选择的图像检索。当查询图像由表示数据库中每个图像的单独高斯混合码本编码时,该算法检索数据库中具有最小失真的图像。我们使用高斯混合矢量量化(GMVQ)来代替传统的期望最大化(EM)算法对高斯混合进行聚类。实验对比表明,较简单的GMVQ算法和EM算法具有相近的高斯混合参数和相似的收敛速度。我们还提出了一种新的颜色交织方法,减少了特征向量的维数和协方差矩阵的大小,从而减少了计算量。在HSV色彩空间中,该方法的检索性能略好于通常的颜色交错法。我们提出的最小失真图像检索比概率图像检索性能更好。
{"title":"Minimum distortion color image retrieval based on Lloyd-clustered Gauss mixtures","authors":"Sangoh Jeong, R. Gray","doi":"10.1109/DCC.2005.52","DOIUrl":"https://doi.org/10.1109/DCC.2005.52","url":null,"abstract":"We consider image retrieval based on minimum distortion selection of features of color images modelled by Gauss mixtures. The proposed algorithm retrieves the image in a database having minimum distortion when the query image is encoded by a separate Gauss mixture codebook representing each image in the database. We use Gauss mixture vector quantization (GMVQ) for clustering Gauss mixtures, instead of the conventional expectation-maximization (EM) algorithm. Experimental comparison shows that the simpler GMVQ and the EM algorithms have close Gauss mixture parameters with similar convergence speeds. We also provide a new color-interleaving method, reducing the dimension of feature vectors and the size of covariance matrices, thereby reducing computation. This method shows a slightly better retrieval performance than the usual color-interleaving method in HSV color space. Our proposed minimum distortion image retrieval performs better than probabilistic image retrieval.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"112 1","pages":"279-288"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85782650","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Generalization of the rate-distortion function for Wyner-Ziv coding of noisy sources in the quadratic-Gaussian case 二次高斯情况下噪声源Wyner-Ziv编码的率失真函数推广
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.6
D. Rebollo-Monedero, B. Girod
We extend the rate-distortion function for Wyner-Ziv coding of noisy sources with quadratic distortion, in the jointly Gaussian case, to more general statistics. It suffices that the noisy observation Z be the sum of a function of the side information Y and independent Gaussian noise, while the source data X must be the sum of a function of Y, a linear function of Z, and a random variable N such that the conditional expectation of N given Y and Z is zero, almost surely. Furthermore, the side information Y may be arbitrarily distributed in any alphabet, discrete or continuous. Under these general conditions, we prove that no rate loss is incurred due to the unavailability of the side information at the encoder. In the noiseless Wyner-Ziv case, i.e., when the source data is directly observed, the assumptions are still less restrictive than those recently established in the literature. We confirm, theoretically and experimentally, the consistency of this analysis with some of the main results on high-rate Wyner-Ziv quantization of noisy sources.
我们将二次失真噪声源的wner - ziv编码的速率失真函数扩展到更一般的统计量。噪声观测值Z是侧信息Y的函数和独立高斯噪声的和就足够了,而源数据X必须是Y的函数、Z的线性函数和随机变量N的和,使得给定Y和Z的N的条件期望几乎肯定为零。此外,边信息Y可以任意分布在任何字母中,离散或连续。在这些一般条件下,我们证明了由于编码器侧信息不可用而不会产生速率损失。在无噪声的Wyner-Ziv案例中,即当源数据被直接观察到时,假设仍然比最近在文献中建立的假设限制性更小。我们从理论上和实验上证实了这一分析与噪声源的高速率Wyner-Ziv量化的一些主要结果的一致性。
{"title":"Generalization of the rate-distortion function for Wyner-Ziv coding of noisy sources in the quadratic-Gaussian case","authors":"D. Rebollo-Monedero, B. Girod","doi":"10.1109/DCC.2005.6","DOIUrl":"https://doi.org/10.1109/DCC.2005.6","url":null,"abstract":"We extend the rate-distortion function for Wyner-Ziv coding of noisy sources with quadratic distortion, in the jointly Gaussian case, to more general statistics. It suffices that the noisy observation Z be the sum of a function of the side information Y and independent Gaussian noise, while the source data X must be the sum of a function of Y, a linear function of Z, and a random variable N such that the conditional expectation of N given Y and Z is zero, almost surely. Furthermore, the side information Y may be arbitrarily distributed in any alphabet, discrete or continuous. Under these general conditions, we prove that no rate loss is incurred due to the unavailability of the side information at the encoder. In the noiseless Wyner-Ziv case, i.e., when the source data is directly observed, the assumptions are still less restrictive than those recently established in the literature. We confirm, theoretically and experimentally, the consistency of this analysis with some of the main results on high-rate Wyner-Ziv quantization of noisy sources.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"2 1","pages":"23-32"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"86028489","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
On the performance of linear Slepian-Wolf codes for correlated stationary memoryless sources 相关平稳无记忆源下线性睡眠-狼码的性能研究
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.65
Shengtian Yang, Peiliang Qiu
We derive an upper bound on the average MAP decoding error probability of random linear SW codes for arbitrary correlated stationary memoryless sources defined on Galois fields. By using this tool, we analyze the performance of SW codes based on LDPC codes and random permutations, and show that under some conditions, all but a diminishingly small proportion of LDPC encoders and permutations are good enough for the design of practical SW systems when the coding length is very large.
我们推导了在伽罗瓦场上定义的任意相关平稳无记忆源的随机线性SW码的MAP平均译码错误概率的上界。利用该工具,我们分析了基于LDPC编码和随机排列的软件编码的性能,并表明在某些条件下,当编码长度非常大时,除了越来越少的LDPC编码器和排列外,所有的LDPC编码器和排列都足以满足实际软件系统的设计。
{"title":"On the performance of linear Slepian-Wolf codes for correlated stationary memoryless sources","authors":"Shengtian Yang, Peiliang Qiu","doi":"10.1109/DCC.2005.65","DOIUrl":"https://doi.org/10.1109/DCC.2005.65","url":null,"abstract":"We derive an upper bound on the average MAP decoding error probability of random linear SW codes for arbitrary correlated stationary memoryless sources defined on Galois fields. By using this tool, we analyze the performance of SW codes based on LDPC codes and random permutations, and show that under some conditions, all but a diminishingly small proportion of LDPC encoders and permutations are good enough for the design of practical SW systems when the coding length is very large.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"18 1","pages":"53-62"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"77994994","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
An extension of the Burrows Wheeler transform to k words Burrows Wheeler变换的扩展到k个单词
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.13
S. Mantaci, A. Restivo, M. Sciortino
Summary form only given. We introduce an extension of the Burrows-Wheeler transform to a multiset of primitive words. Primitiveness is not actually a restrictive hypothesis, since in practice almost all the processed texts are primitive (or become primitive by adding an end-of-string symbol). We prove that such a transformation as the BWT is reversible. We show how to use the transformation as a preprocessing for the simultaneous compression of different texts.
只提供摘要形式。我们将Burrows-Wheeler变换扩展到原始词的多集。原始性实际上并不是一个限制性假设,因为在实践中几乎所有处理过的文本都是原始性的(或者通过添加字符串结束符号而变得原始性)。我们证明了像BWT这样的变换是可逆的。我们展示了如何使用转换作为同时压缩不同文本的预处理。
{"title":"An extension of the Burrows Wheeler transform to k words","authors":"S. Mantaci, A. Restivo, M. Sciortino","doi":"10.1109/DCC.2005.13","DOIUrl":"https://doi.org/10.1109/DCC.2005.13","url":null,"abstract":"Summary form only given. We introduce an extension of the Burrows-Wheeler transform to a multiset of primitive words. Primitiveness is not actually a restrictive hypothesis, since in practice almost all the processed texts are primitive (or become primitive by adding an end-of-string symbol). We prove that such a transformation as the BWT is reversible. We show how to use the transformation as a preprocessing for the simultaneous compression of different texts.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"15 1","pages":"469-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73149064","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Compression algorithm for infrared hyperspectral sounder data 红外高光谱测深数据压缩算法
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.27
I. Gladkova, L. Roytman, M. Goldberg
Summary form only given. The research is undertaken by NOAA/NESDIS, for its GOES-R Earth observation satellite series, to be launched in the 2013 time frame, to enable greater distribution of its scientific data, both within the US and internationally. We have developed a new lossless algorithm for compression of the signals from NOAA's environmental satellites using current spacecraft to simulate data from the upcoming GOES-R instrument, and focusing on Aqua Spacecraft's AIRS (atmospheric infrared sounder) instrument in our case study. The AIRS is a high resolution instrument which measures infrared radiances at 2378 wavelengths ranging from 3.74-15.4 /spl mu/m. The AIRS takes 90 measurements as it scans 48.95 degrees perpendicular to the satellite's orbit every 2.667 seconds. We use Level 1A digital count data granules, which represent 6 minutes (or 135 scans) of measurements. Therefore, our data set consists of a 90/spl times/135/spl times/1502 cube of integers ranging from 12-14 bits. Our compression algorithm consists of the following steps: 1) channel partitioning; 2) adaptive clustering; 3) projection onto principal directions; 4) entropy coding of the residuals.
只提供摘要形式。这项研究是由NOAA/NESDIS为其将于2013年发射的GOES-R地球观测卫星系列进行的,以便在美国国内和国际上更好地分发其科学数据。我们开发了一种新的无损算法,用于压缩来自NOAA环境卫星的信号,使用现有航天器模拟即将推出的GOES-R仪器的数据,并在我们的案例研究中重点关注Aqua航天器的AIRS(大气红外探测仪)仪器。AIRS是一种高分辨率仪器,可测量2378个波长的红外辐射,范围为3.74-15.4 /spl mu/m。AIRS每2.667秒扫描与卫星轨道垂直48.95度的卫星,进行90次测量。我们使用1A级数字计数数据颗粒,代表6分钟(或135次扫描)的测量。因此,我们的数据集由一个90/spl乘以/135/spl乘以/1502的整数立方组成,范围从12-14位。我们的压缩算法包括以下步骤:1)通道划分;2)自适应聚类;3)主方向投影;4)残差熵编码。
{"title":"Compression algorithm for infrared hyperspectral sounder data","authors":"I. Gladkova, L. Roytman, M. Goldberg","doi":"10.1109/DCC.2005.27","DOIUrl":"https://doi.org/10.1109/DCC.2005.27","url":null,"abstract":"Summary form only given. The research is undertaken by NOAA/NESDIS, for its GOES-R Earth observation satellite series, to be launched in the 2013 time frame, to enable greater distribution of its scientific data, both within the US and internationally. We have developed a new lossless algorithm for compression of the signals from NOAA's environmental satellites using current spacecraft to simulate data from the upcoming GOES-R instrument, and focusing on Aqua Spacecraft's AIRS (atmospheric infrared sounder) instrument in our case study. The AIRS is a high resolution instrument which measures infrared radiances at 2378 wavelengths ranging from 3.74-15.4 /spl mu/m. The AIRS takes 90 measurements as it scans 48.95 degrees perpendicular to the satellite's orbit every 2.667 seconds. We use Level 1A digital count data granules, which represent 6 minutes (or 135 scans) of measurements. Therefore, our data set consists of a 90/spl times/135/spl times/1502 cube of integers ranging from 12-14 bits. Our compression algorithm consists of the following steps: 1) channel partitioning; 2) adaptive clustering; 3) projection onto principal directions; 4) entropy coding of the residuals.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"12 1","pages":"460-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82354247","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Design of VQ-based hybrid digital-analog joint source-channel codes for image communication 基于vq的图像通信混合数模联合源信道码设计
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.30
Yadong Wang, F. Alajaji, T. Linder
A joint source-channel coding system for image communication over an additive white Gaussian noise channel is presented. It employs vector quantization based hybrid digital-analog modulation techniques with bandwidth compression and expansion for transmitting and reconstructing the wavelet coefficients of an image. The main advantage of the proposed system is that it achieves good performance at the design channel signal-to-noise ratio (CSNR), while still maintaining a "graceful improvement" characteristic at higher CSNR. Comparisons are made with two purely digital systems and two purely analog systems. Simulation shows that the proposed system is superior to the other investigated systems for a wide range of CSNR.
提出了一种用于加性高斯白噪声信道上图像通信的联合信源信道编码系统。它采用基于矢量量化的混合数模调制技术,通过带宽压缩和扩展来传输和重建图像的小波系数。该系统的主要优点是在设计信道信噪比(CSNR)下具有良好的性能,同时在较高的CSNR下仍然保持“优雅改进”特性。对两个纯数字系统和两个纯模拟系统进行了比较。仿真结果表明,该系统在较宽的信噪比范围内优于其他研究过的系统。
{"title":"Design of VQ-based hybrid digital-analog joint source-channel codes for image communication","authors":"Yadong Wang, F. Alajaji, T. Linder","doi":"10.1109/DCC.2005.30","DOIUrl":"https://doi.org/10.1109/DCC.2005.30","url":null,"abstract":"A joint source-channel coding system for image communication over an additive white Gaussian noise channel is presented. It employs vector quantization based hybrid digital-analog modulation techniques with bandwidth compression and expansion for transmitting and reconstructing the wavelet coefficients of an image. The main advantage of the proposed system is that it achieves good performance at the design channel signal-to-noise ratio (CSNR), while still maintaining a \"graceful improvement\" characteristic at higher CSNR. Comparisons are made with two purely digital systems and two purely analog systems. Simulation shows that the proposed system is superior to the other investigated systems for a wide range of CSNR.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"33 1","pages":"193-202"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76237579","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 14
AXECHOP: a grammar-based compressor for XML AXECHOP:基于语法的XML压缩器
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.20
G. Leighton, Jim Diamond, T. Müldner
Summary form only given. XML is gaining widespread acceptance as a standard for storing and transmitting structured data. One of the drawbacks of XML is that it is quite verbose: an XML representation of a set of data can easily be ten times as large as a more economical representation of the data. To overcome this limitation, we present a compression scheme tailored specifically to XML named AXECHOP. The compression strategy used in AXECHOP begins by dividing the source XML document into structural and data segments. The former is represented using a byte tokenization scheme that preserves the original structure of the document (i.e. it maintains the proper nesting and ordering of elements, attributes, and data values). The MPM compression algorithm is used to generate a context-free grammar capable of deriving this original structure, and the grammar is passed through an adaptive arithmetic coder before being written to the compressed file. The document's data is organized into a series of containers (where container membership is determined by the identity of the XML element or attribute that encloses the data) and then the Burrows-Wheeler transform (BWT) is applied to the contents of each dictionary, with the results being appended to the compressed file.
只提供摘要形式。XML作为存储和传输结构化数据的标准正在获得广泛的接受。XML的缺点之一是它非常冗长:一组数据的XML表示很容易比更经济的数据表示大十倍。为了克服这个限制,我们提出了一个专门为XML定制的压缩方案,名为AXECHOP。AXECHOP中使用的压缩策略首先将源XML文档划分为结构段和数据段。前者使用保留文档原始结构的字节标记化方案来表示(即,它维护元素、属性和数据值的适当嵌套和顺序)。MPM压缩算法用于生成能够派生此原始结构的上下文无关语法,该语法在写入压缩文件之前经过自适应算术编码器。文档的数据被组织到一系列容器中(其中容器的成员关系由包含数据的XML元素或属性的标识决定),然后将Burrows-Wheeler转换(BWT)应用于每个字典的内容,并将结果附加到压缩文件中。
{"title":"AXECHOP: a grammar-based compressor for XML","authors":"G. Leighton, Jim Diamond, T. Müldner","doi":"10.1109/DCC.2005.20","DOIUrl":"https://doi.org/10.1109/DCC.2005.20","url":null,"abstract":"Summary form only given. XML is gaining widespread acceptance as a standard for storing and transmitting structured data. One of the drawbacks of XML is that it is quite verbose: an XML representation of a set of data can easily be ten times as large as a more economical representation of the data. To overcome this limitation, we present a compression scheme tailored specifically to XML named AXECHOP. The compression strategy used in AXECHOP begins by dividing the source XML document into structural and data segments. The former is represented using a byte tokenization scheme that preserves the original structure of the document (i.e. it maintains the proper nesting and ordering of elements, attributes, and data values). The MPM compression algorithm is used to generate a context-free grammar capable of deriving this original structure, and the grammar is passed through an adaptive arithmetic coder before being written to the compressed file. The document's data is organized into a series of containers (where container membership is determined by the identity of the XML element or attribute that encloses the data) and then the Burrows-Wheeler transform (BWT) is applied to the contents of each dictionary, with the results being appended to the compressed file.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"44 1","pages":"467-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76094702","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
A flexible compressed text retrieval system using a modified LZW algorithm 一个使用改进LZW算法的灵活压缩文本检索系统
Pub Date : 2005-03-29 DOI: 10.1109/DCC.2005.5
Nan Zhang, Tao Tao, R. Satya, A. Mukherjee
Summary form only given. With an increasing amount of text data being stored in compressed format, being able to access the compressed data randomly and decode it partially is highly desirable for efficient retrieval in many applications. The efficiency of these operations depends on the compression method used. We present a modified LZW algorithm that supports efficient indexing and searching on compressed files. Our method performs in a sublinear complexity, since we only decode a small portion of the file. The proposed approach not only provides the flexibility for dynamic indexing in different text granularities, but also provides the possibility for parallel processing in both encoding and decoding sides, independent of the number of processors available. It also provides good error resilience. The compression ratio is improved using the proposed modified LZW algorithm. Test results show that our public trie method has a compression ratio of 0.34 for the TREC corpus and 0.32 with text preprocessing using a star transform with an optimal static dictionary; this is very close to the efficient word Huffman and phrase based word Huffman schemes, but has a more flexible random access ability.
只提供摘要形式。随着越来越多的文本数据以压缩格式存储,能够随机访问压缩数据并对其进行部分解码对于许多应用程序中的高效检索是非常理想的。这些操作的效率取决于所使用的压缩方法。提出了一种改进的LZW算法,支持对压缩文件的高效索引和搜索。我们的方法执行的复杂度是次线性的,因为我们只解码文件的一小部分。所提出的方法不仅提供了在不同文本粒度下动态索引的灵活性,而且还提供了在编码和解码方面并行处理的可能性,而不受可用处理器数量的影响。它还提供了良好的错误恢复能力。采用改进后的LZW算法提高了压缩比。测试结果表明,我们的公共trie方法对TREC语料库的压缩比为0.34,对使用最优静态字典进行星形变换的文本预处理的压缩比为0.32;它非常接近高效的词霍夫曼和基于短语的词霍夫曼方案,但具有更灵活的随机访问能力。
{"title":"A flexible compressed text retrieval system using a modified LZW algorithm","authors":"Nan Zhang, Tao Tao, R. Satya, A. Mukherjee","doi":"10.1109/DCC.2005.5","DOIUrl":"https://doi.org/10.1109/DCC.2005.5","url":null,"abstract":"Summary form only given. With an increasing amount of text data being stored in compressed format, being able to access the compressed data randomly and decode it partially is highly desirable for efficient retrieval in many applications. The efficiency of these operations depends on the compression method used. We present a modified LZW algorithm that supports efficient indexing and searching on compressed files. Our method performs in a sublinear complexity, since we only decode a small portion of the file. The proposed approach not only provides the flexibility for dynamic indexing in different text granularities, but also provides the possibility for parallel processing in both encoding and decoding sides, independent of the number of processors available. It also provides good error resilience. The compression ratio is improved using the proposed modified LZW algorithm. Test results show that our public trie method has a compression ratio of 0.34 for the TREC corpus and 0.32 with text preprocessing using a star transform with an optimal static dictionary; this is very close to the efficient word Huffman and phrase based word Huffman schemes, but has a more flexible random access ability.","PeriodicalId":91161,"journal":{"name":"Proceedings. Data Compression Conference","volume":"43 1","pages":"493-"},"PeriodicalIF":0.0,"publicationDate":"2005-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76321207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
期刊
Proceedings. Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1