首页 > 最新文献

Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)最新文献

英文 中文
Embedded post-processing for enhancement of compressed images 用于增强压缩图像的嵌入式后处理
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755655
Aria Nosratinia
This paper presents a simple and effective post-processing method for compressed images. This work focuses on the cyclic time-variance introduced by block-based and subband transform coders. We propose an algorithm to (almost) restore stationarity to the cyclo-stationary output of the conventional transform coders. Despite a simple, non-iterative structure, this method outperforms other methods of image enhancement known to us, e.g. linear and nonlinear filtering, projection on convex sets (POCS), wavelet-based and optimization-based methods. In particular, the proposed method performs very well in suppressing both blocking and ringing artifacts. Furthermore, it admits a solution with successive approximation. The resulting embeddedness is very useful for multimedia applications such as image browsing on the World Wide Web.
提出了一种简单有效的压缩图像后处理方法。本文主要研究基于分块和子带变换编码器引入的循环时变。我们提出了一种算法(几乎)恢复平稳性的循环平稳输出的传统变换编码器。尽管结构简单,非迭代,但该方法优于我们已知的其他图像增强方法,例如线性和非线性滤波,凸集投影(POCS),基于小波和基于优化的方法。特别是,该方法在抑制阻塞和振铃伪影方面表现良好。进一步地,它允许一个逐次逼近的解。由此产生的嵌入性对于多媒体应用非常有用,例如在万维网上浏览图像。
{"title":"Embedded post-processing for enhancement of compressed images","authors":"Aria Nosratinia","doi":"10.1109/DCC.1999.755655","DOIUrl":"https://doi.org/10.1109/DCC.1999.755655","url":null,"abstract":"This paper presents a simple and effective post-processing method for compressed images. This work focuses on the cyclic time-variance introduced by block-based and subband transform coders. We propose an algorithm to (almost) restore stationarity to the cyclo-stationary output of the conventional transform coders. Despite a simple, non-iterative structure, this method outperforms other methods of image enhancement known to us, e.g. linear and nonlinear filtering, projection on convex sets (POCS), wavelet-based and optimization-based methods. In particular, the proposed method performs very well in suppressing both blocking and ringing artifacts. Furthermore, it admits a solution with successive approximation. The resulting embeddedness is very useful for multimedia applications such as image browsing on the World Wide Web.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"298 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132627938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 23
Modifications of the Burrows and Wheeler data compression algorithm 对Burrows和Wheeler数据压缩算法的改进
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755668
B. Balkenhol, S. Kurtz, Y. Shtarkov
We improve upon previous results on the Burrows and Wheeler (BW)-algorithm. Based on the context tree model, we consider the specific statistical properties of the data at the output of the BWT. We describe six important properties, three of which have not been described elsewhere. These considerations lead to modifications of the coding method, which in turn improve the coding efficiency. We briefly describe how to compute the BWT with low complexity in time and space, using suffix trees in two different representations. Finally, we present experimental results about the compression rate and running time of our method, and compare these results to previous achievements.
我们改进了Burrows and Wheeler (BW)算法的先前结果。基于上下文树模型,我们考虑了BWT输出时数据的特定统计属性。我们描述了六个重要的性质,其中三个没有在其他地方描述。这些考虑导致了对编码方法的修改,从而提高了编码效率。我们简要描述了如何使用后缀树在两种不同的表示中计算低时间和空间复杂度的BWT。最后,给出了该方法的压缩率和运行时间的实验结果,并与前人的成果进行了比较。
{"title":"Modifications of the Burrows and Wheeler data compression algorithm","authors":"B. Balkenhol, S. Kurtz, Y. Shtarkov","doi":"10.1109/DCC.1999.755668","DOIUrl":"https://doi.org/10.1109/DCC.1999.755668","url":null,"abstract":"We improve upon previous results on the Burrows and Wheeler (BW)-algorithm. Based on the context tree model, we consider the specific statistical properties of the data at the output of the BWT. We describe six important properties, three of which have not been described elsewhere. These considerations lead to modifications of the coding method, which in turn improve the coding efficiency. We briefly describe how to compute the BWT with low complexity in time and space, using suffix trees in two different representations. Finally, we present experimental results about the compression rate and running time of our method, and compare these results to previous achievements.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"1048 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131797779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
Performance of quantizers on noisy channels using structured families of codes 使用结构化编码族的噪声信道上量化器的性能
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755697
A. Méhes, K. Zeger
Achievable distortion bounds are derived for the cascade of structured families of binary linear channel codes and binary lattice vector quantizers. It is known that for the cascade of asymptotically good channel codes and asymptotically good vector quantizers the end-to-end distortion decays to zero exponentially fast as a function of the overall transmission rate, and is achieved by choosing a channel code rate that is independent of the overall transmission rate. We show that for certain families of practical channel codes and binary lattice vector quantizers, the overall distortion can still be made to decay to zero exponentially fast as the transmission rate grows, although the exponent is a sub-linear function of the transmission rate. This is achieved by carefully choosing a channel code rate that decays to zero as the transmission rate grows. Explicit channel code rate schedules are obtained for several well-known families of channel codes.
推导了二元线性信道码和二元点阵矢量量化器的结构族级联可实现的失真界。众所周知,对于渐近良好信道码和渐近良好矢量量化器的级联,端到端失真作为总体传输速率的函数呈指数级快速衰减到零,并且通过选择与总体传输速率无关的信道码率来实现。我们表明,对于某些实用的信道码族和二进制点阵矢量量化器,尽管指数是传输速率的次线性函数,但随着传输速率的增长,总体失真仍然可以以指数速度衰减到零。这是通过仔细选择一个随传输速率增长而衰减到零的信道码率来实现的。对几种已知的信道码族,给出了明确的信道码率表。
{"title":"Performance of quantizers on noisy channels using structured families of codes","authors":"A. Méhes, K. Zeger","doi":"10.1109/DCC.1999.755697","DOIUrl":"https://doi.org/10.1109/DCC.1999.755697","url":null,"abstract":"Achievable distortion bounds are derived for the cascade of structured families of binary linear channel codes and binary lattice vector quantizers. It is known that for the cascade of asymptotically good channel codes and asymptotically good vector quantizers the end-to-end distortion decays to zero exponentially fast as a function of the overall transmission rate, and is achieved by choosing a channel code rate that is independent of the overall transmission rate. We show that for certain families of practical channel codes and binary lattice vector quantizers, the overall distortion can still be made to decay to zero exponentially fast as the transmission rate grows, although the exponent is a sub-linear function of the transmission rate. This is achieved by carefully choosing a channel code rate that decays to zero as the transmission rate grows. Explicit channel code rate schedules are obtained for several well-known families of channel codes.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"38 ","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114059027","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Protein is incompressible 蛋白质是不可压缩的
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755675
C. Nevill-Manning, I. Witten
Life is based on two polymers, DNA and protein, whose properties can be described in a simple text file. It is natural to expect that standard text compression techniques would work on biological sequences as they do on English text. But biological sequences have a fundamentally different structure from linguistic ones, and standard compression schemes exhibit disappointing performance on them. We describe a new approach to compression that takes account of the underlying biochemical principles. This gives rise to a generalization of blending for statistical compressors where every context is used, weighted by its similarity to the current context. Results support what research in bioinformatics has shown, that there is little Markov dependency in protein. This cripples data compression schemes and reduces them to order zero models.
生命基于两种聚合物,DNA和蛋白质,它们的特性可以用一个简单的文本文件来描述。我们很自然地期望标准文本压缩技术能够像处理英语文本一样处理生物序列。但是生物序列与语言序列有着根本不同的结构,标准的压缩方案在它们上面表现得令人失望。我们描述了一种新的压缩方法,考虑到潜在的生化原理。这导致了统计压缩器混合的泛化,其中使用每个上下文,并根据其与当前上下文的相似性进行加权。结果支持生物信息学研究表明,蛋白质中几乎没有马尔可夫依赖性。这削弱了数据压缩方案,并将其减少到零阶模型。
{"title":"Protein is incompressible","authors":"C. Nevill-Manning, I. Witten","doi":"10.1109/DCC.1999.755675","DOIUrl":"https://doi.org/10.1109/DCC.1999.755675","url":null,"abstract":"Life is based on two polymers, DNA and protein, whose properties can be described in a simple text file. It is natural to expect that standard text compression techniques would work on biological sequences as they do on English text. But biological sequences have a fundamentally different structure from linguistic ones, and standard compression schemes exhibit disappointing performance on them. We describe a new approach to compression that takes account of the underlying biochemical principles. This gives rise to a generalization of blending for statistical compressors where every context is used, weighted by its similarity to the current context. Results support what research in bioinformatics has shown, that there is little Markov dependency in protein. This cripples data compression schemes and reduces them to order zero models.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114190945","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 96
Distributed source coding using syndromes (DISCUS): design and construction 使用综合征的分布式源编码(DISCUS):设计和构建
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755665
S. Pradhan, K. Ramchandran
We address the problem of distributed source coding, i.e. compression of correlated sources that are not co-located and/or cannot communicate with each other to minimize their joint description cost. In this work we tackle the related problem of compressing a source that is correlated with another source which is available only at the decoder. In contrast to prior information-theoretic approaches, we introduce a new construction and practical framework for tackling the problem based on the judicious incorporation of channel coding principles into this source coding problem. We dub our approach as distributed source coding using syndromes (DISCUS). We focus in this paper on trellis-structured constructions of the framework to illustrate its utility. Simulation results confirm the power of DISCUS, opening up a new and exciting constructive playing-ground for the distributed source coding problem. For the distributed coding of correlated i.i.d. Gaussian sources that are noisy versions of each other with "correlation-SNR" in the range of 12 to 20 dB, the DISCUS method attains gains of 7-15 dB in SNR over the Shannon-bound using "naive" independent coding of the sources.
我们解决了分布式源编码的问题,即压缩不位于同一位置和/或不能相互通信的相关源,以最小化它们的联合描述成本。在这项工作中,我们解决了压缩与仅在解码器上可用的另一个源相关的源的相关问题。与先前的信息论方法相比,我们引入了一种新的结构和实用框架来解决基于明智地将信道编码原则纳入该源编码问题的问题。我们将我们的方法称为使用综合征的分布式源代码编码(DISCUS)。在本文中,我们将重点放在框架的网格结构结构上,以说明其实用性。仿真结果证实了DISCUS的强大功能,为分布式源编码问题开辟了一个新的、令人兴奋的建设性领域。对于相互具有“相关信噪比”(correlation-SNR)在12 ~ 20 dB范围内的相关i.i.d高斯源的噪声版本的分布式编码,DISCUS方法使用源的“朴素”独立编码,比香农边界获得7 ~ 15 dB的信噪比增益。
{"title":"Distributed source coding using syndromes (DISCUS): design and construction","authors":"S. Pradhan, K. Ramchandran","doi":"10.1109/DCC.1999.755665","DOIUrl":"https://doi.org/10.1109/DCC.1999.755665","url":null,"abstract":"We address the problem of distributed source coding, i.e. compression of correlated sources that are not co-located and/or cannot communicate with each other to minimize their joint description cost. In this work we tackle the related problem of compressing a source that is correlated with another source which is available only at the decoder. In contrast to prior information-theoretic approaches, we introduce a new construction and practical framework for tackling the problem based on the judicious incorporation of channel coding principles into this source coding problem. We dub our approach as distributed source coding using syndromes (DISCUS). We focus in this paper on trellis-structured constructions of the framework to illustrate its utility. Simulation results confirm the power of DISCUS, opening up a new and exciting constructive playing-ground for the distributed source coding problem. For the distributed coding of correlated i.i.d. Gaussian sources that are noisy versions of each other with \"correlation-SNR\" in the range of 12 to 20 dB, the DISCUS method attains gains of 7-15 dB in SNR over the Shannon-bound using \"naive\" independent coding of the sources.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114374689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1072
A simple technique for bounding the pointwise redundancy of the 1978 Lempel-Ziv algorithm 1978 Lempel-Ziv算法的点向冗余边界的简单技术
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755693
J. Kieffer, E. Yang
If x is a string of finite length over a finite alphabet A, let LZ(x) denote the length of the binary codeword assigned to x by the 1978 version of the Lempel-Ziv data compression algorithm, let t(x) be the number of phrases in the Lempel-Ziv parsing of x, and let /spl mu/(x) be the probability assigned to x by a memoryless source model. Using a very simple technique, we probe the pointwise redundancy bound LZ(x)+log/sub 2//spl mu/(x)/spl les/8t(x)max{-log/sub 2//spl mu/(a):a/spl isin/A}.
如果x是有限字母a上的有限长度的字符串,设LZ(x)表示1978年版本的Lempel-Ziv数据压缩算法分配给x的二进制码字的长度,设t(x)是x在Lempel-Ziv解析中的短语数,设/spl mu/(x)是无内存源模型分配给x的概率。使用一个非常简单的技术,我们探测了点向冗余界LZ(x)+log/sub 2//spl mu/(x)/spl les/8t(x)max{-log/sub 2//spl mu/(a):a/spl isin/ a}。
{"title":"A simple technique for bounding the pointwise redundancy of the 1978 Lempel-Ziv algorithm","authors":"J. Kieffer, E. Yang","doi":"10.1109/DCC.1999.755693","DOIUrl":"https://doi.org/10.1109/DCC.1999.755693","url":null,"abstract":"If x is a string of finite length over a finite alphabet A, let LZ(x) denote the length of the binary codeword assigned to x by the 1978 version of the Lempel-Ziv data compression algorithm, let t(x) be the number of phrases in the Lempel-Ziv parsing of x, and let /spl mu/(x) be the probability assigned to x by a memoryless source model. Using a very simple technique, we probe the pointwise redundancy bound LZ(x)+log/sub 2//spl mu/(x)/spl les/8t(x)max{-log/sub 2//spl mu/(a):a/spl isin/A}.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117316459","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Fast, modified Z-coding of wavelet pyramids 快速,修改的z编码的小波金字塔
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785695
W. Lynch, K. Kolarov, Bill Arrighi
[Summary form only given]. This paper describes a fast, low-complexity, entropy efficient coder for wavelet pyramids. This coder approaches the entropy-limited coding rate of video wavelet pyramids, is fast in both hardware and software implementations, and has low complexity for use in ASICs. It consists of a modified Z-coder used to code the zero/non-zero significance function without adaptation. The wavelet pyramid is further sharpened by scaling to match the characteristics of the human visual system (HVS). We derive the statistical characteristics of quantized wavelet pyramids from NTSC video viewed under standard conditions. These video pyramids have substantial runs of zeros and also substantial runs of non-zeros. To explore these we developed a modification of the Z-coder and explored an application of it to code zero versus non-zero. Z-codecs have the advantage of a simple (no multipliers) and fast implementation combined with coding performance approximating that of an arithmetic codec. Our experiments showed that this coder compares favorably with straight arithmetic coding. Our encoder has significant speed advantage due to low cost implementation.
[仅提供摘要形式]。本文描述了一种快速、低复杂度、熵高效的小波金字塔编码器。该编码器接近视频小波金字塔的熵限编码率,硬件和软件实现速度快,并且在asic中使用的复杂性低。它由一个修改的z编码器组成,用于编码零/非零显著性函数而不自适应。小波金字塔通过缩放进一步锐化,以匹配人类视觉系统(HVS)的特征。我们从标准条件下观看的NTSC视频中导出了量化小波金字塔的统计特征。这些视频金字塔有大量的零运行和大量的非零运行。为了探索这些,我们开发了z编码器的修改,并探索了它对编码零与非零的应用。z编解码器具有简单(无乘法器)和快速实现的优点,其编码性能接近算术编解码器。我们的实验表明,该编码器优于直接算术编码。由于实现成本低,我们的编码器具有显著的速度优势。
{"title":"Fast, modified Z-coding of wavelet pyramids","authors":"W. Lynch, K. Kolarov, Bill Arrighi","doi":"10.1109/DCC.1999.785695","DOIUrl":"https://doi.org/10.1109/DCC.1999.785695","url":null,"abstract":"[Summary form only given]. This paper describes a fast, low-complexity, entropy efficient coder for wavelet pyramids. This coder approaches the entropy-limited coding rate of video wavelet pyramids, is fast in both hardware and software implementations, and has low complexity for use in ASICs. It consists of a modified Z-coder used to code the zero/non-zero significance function without adaptation. The wavelet pyramid is further sharpened by scaling to match the characteristics of the human visual system (HVS). We derive the statistical characteristics of quantized wavelet pyramids from NTSC video viewed under standard conditions. These video pyramids have substantial runs of zeros and also substantial runs of non-zeros. To explore these we developed a modification of the Z-coder and explored an application of it to code zero versus non-zero. Z-codecs have the advantage of a simple (no multipliers) and fast implementation combined with coding performance approximating that of an arithmetic codec. Our experiments showed that this coder compares favorably with straight arithmetic coding. Our encoder has significant speed advantage due to low cost implementation.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"417 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116181774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fast multi-match Lempel-Ziv 快速多匹配Lempel-Ziv
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785702
M. Pinho, W. Finamore, W. Pearlman
Summary form only given. One of the most popular encoders in the literature is the LZ78, which was proposed by Ziv and Lempel (1978). We establish a recursive way to find the longest m-tuple match. We prove the following theorem that shows how to obtain a longest (m+1)-tuple match from the longest m-tuple match. It shows that a (m+1)-tuple match is the concatenation of the first (m-1) words of the m-tuple match with the next longest double match. Therefore, the longest (m+1)-tuple match can be found using the m-tuple match and a procedure to compute the longest double match. Our theorem is as follows. Let A be a source alphabet, let A* be the set of all finite strings of A, and D/spl sub/A*, such that if x/spl isin/D then all prefixes of x belong to D. Let u denote a one-sided infinite sequence. If b/sub 1//sup m/ is the longest m-tuple match in u, with respect to D, then there is a longest (m+1)-tuple match b/spl circ//sub 1//sup m+1/, such that b/spl circ//sub i/=b/sub i/,/spl forall/i/spl isin/{1,...m-1}. We implemented the fast mmLZ and the results show a improvement in compression of around 5% over the LZW, in the Canterbury Corpus (Arnold and Bell, 1997) with little extra computational cost.
只提供摘要形式。文献中最流行的编码器之一是LZ78,它是由Ziv和Lempel(1978)提出的。我们建立了一种递归的方法来找到最长的m元组匹配。我们证明了以下定理,该定理展示了如何从最长的m元组匹配中获得最长的(m+1)元组匹配。它表明(m+1)元组匹配是m元组匹配的第一个(m-1)个单词与下一个最长的双匹配的连接。因此,可以使用m元组匹配和计算最长双匹配的过程来找到最长(m+1)元组匹配。我们的定理如下。设A是一个源字母,设A*是A的所有有限字符串的集合,D/spl sub/A*,使得如果x/spl是in/D,则x的所有前缀都属于D。设u表示一个单侧无限序列。如果b/sub 1//sup m/是u中最长的m元组匹配,相对于D,则存在最长的(m+1)-元组匹配b/spl circ//sub 1//sup m+1/,使得b/spl circ//sub i/=b/sub i/,/spl forall/i/spl isin/{1,…m-1}。我们实现了快速mmLZ,结果显示在坎特伯雷语料库(Arnold and Bell, 1997)中,压缩比LZW提高了约5%,并且几乎没有额外的计算成本。
{"title":"Fast multi-match Lempel-Ziv","authors":"M. Pinho, W. Finamore, W. Pearlman","doi":"10.1109/DCC.1999.785702","DOIUrl":"https://doi.org/10.1109/DCC.1999.785702","url":null,"abstract":"Summary form only given. One of the most popular encoders in the literature is the LZ78, which was proposed by Ziv and Lempel (1978). We establish a recursive way to find the longest m-tuple match. We prove the following theorem that shows how to obtain a longest (m+1)-tuple match from the longest m-tuple match. It shows that a (m+1)-tuple match is the concatenation of the first (m-1) words of the m-tuple match with the next longest double match. Therefore, the longest (m+1)-tuple match can be found using the m-tuple match and a procedure to compute the longest double match. Our theorem is as follows. Let A be a source alphabet, let A* be the set of all finite strings of A, and D/spl sub/A*, such that if x/spl isin/D then all prefixes of x belong to D. Let u denote a one-sided infinite sequence. If b/sub 1//sup m/ is the longest m-tuple match in u, with respect to D, then there is a longest (m+1)-tuple match b/spl circ//sub 1//sup m+1/, such that b/spl circ//sub i/=b/sub i/,/spl forall/i/spl isin/{1,...m-1}. We implemented the fast mmLZ and the results show a improvement in compression of around 5% over the LZW, in the Canterbury Corpus (Arnold and Bell, 1997) with little extra computational cost.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123920298","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A general-purpose compression scheme for databases 用于数据库的通用压缩方案
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785676
A. Cannane, H. Williams, J. Zobel
Summary form only given. Current adaptive compression schemes such as GZIP and COMPRESS are impractical for database compression as they do not allow random access to individual records. A compression algorithm for general-purpose database systems must address the problem of randomly accessing and individually decompressing records, while maintaining compact storage of data. The SEQUITUR algorithm of Nevill-Manning et al., (1994, 1996, 1997) also adaptively compresses data, achieving excellent compression but with significant main-memory requirements. A preliminary version of SEQUITUR used a semi-static modelling approach to achieve slightly worse compression than the adaptive approach. We describe a new variant of the semi-static SEQUITUR algorithm, RAY, that reduces main-memory use and allows random-access to databases. RAY models repetition in sequences by progressively constructing a hierarchical grammar with multiple passes through the data. The multiple pass approach of RAY uses statistics on character pair repetition, or digram frequency, to create rules in the grammar. While our preliminary implementation is not especially fast, the multi-pass approach permits reductions in compression time, at the cost of affecting compression performance, by limiting the number of passes. We have found that RAY has practicable main-memory requirements and achieves better compression than an efficient Huffmann scheme and popular adaptive compression techniques. Moreover, our scheme allows random access to data and is not restricted to databases of text.
只提供摘要形式。当前的自适应压缩方案(如GZIP和COMPRESS)对于数据库压缩是不切实际的,因为它们不允许随机访问单个记录。通用数据库系统的压缩算法必须解决随机访问和单独解压缩记录的问题,同时保持数据的紧凑存储。neville - manning等人(1994,1996,1997)的SEQUITUR算法也自适应压缩数据,虽然压缩效果很好,但对主存的要求很高。SEQUITUR的初步版本使用半静态建模方法来实现比自适应方法稍差的压缩。我们描述了半静态SEQUITUR算法的一个新变体,RAY,它减少了主内存的使用,并允许随机访问数据库。RAY通过逐步构建具有多次遍历数据的分层语法来对序列中的重复进行建模。RAY的多次传递方法使用有关字符对重复或图频率的统计信息来创建语法中的规则。虽然我们的初步实现不是特别快,但通过限制传递次数,以影响压缩性能为代价,多通道方法可以减少压缩时间。我们发现RAY具有实际的主存需求,并且比高效的霍夫曼方案和流行的自适应压缩技术实现更好的压缩。此外,我们的方案允许随机访问数据,而不局限于文本数据库。
{"title":"A general-purpose compression scheme for databases","authors":"A. Cannane, H. Williams, J. Zobel","doi":"10.1109/DCC.1999.785676","DOIUrl":"https://doi.org/10.1109/DCC.1999.785676","url":null,"abstract":"Summary form only given. Current adaptive compression schemes such as GZIP and COMPRESS are impractical for database compression as they do not allow random access to individual records. A compression algorithm for general-purpose database systems must address the problem of randomly accessing and individually decompressing records, while maintaining compact storage of data. The SEQUITUR algorithm of Nevill-Manning et al., (1994, 1996, 1997) also adaptively compresses data, achieving excellent compression but with significant main-memory requirements. A preliminary version of SEQUITUR used a semi-static modelling approach to achieve slightly worse compression than the adaptive approach. We describe a new variant of the semi-static SEQUITUR algorithm, RAY, that reduces main-memory use and allows random-access to databases. RAY models repetition in sequences by progressively constructing a hierarchical grammar with multiple passes through the data. The multiple pass approach of RAY uses statistics on character pair repetition, or digram frequency, to create rules in the grammar. While our preliminary implementation is not especially fast, the multi-pass approach permits reductions in compression time, at the cost of affecting compression performance, by limiting the number of passes. We have found that RAY has practicable main-memory requirements and achieves better compression than an efficient Huffmann scheme and popular adaptive compression techniques. Moreover, our scheme allows random access to data and is not restricted to databases of text.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129286146","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Experiments with single-pass adaptive vector quantization 单次自适应矢量量化实验
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785703
F. Rizzo, J. Storer, B. Carpentieri
Summary form only given. Constantinescu and Storer (1994) introduced an adaptive vector quantization algorithm (AVQ) that combines adaptive dictionary techniques with vector quantization (VQ). The algorithm typically equals or exceeds the compression of the JPEG standard on different classes of images and it often outperforms traditional trained VQ. We show how it is possible to improve AVQ on the class of images on which JPEG does best (i.e., "magazine photographs"). The improvement is possible by exploring the similarities in the dictionary built by AVQ. This is achieved by transforming the input vectors in a way similar to the one used in mean-shape-gain VQ (Oehler and Gray, 1993). In MSGVQ each vector x~/spl isin/R/sup n/ is decomposed as x~=g/spl middot/s~+E/sub x//spl middot/1~, where g=/spl par/x~-E/sub x//spl middot/1~/spl par/ and s~=(x~-E/sub x//spl middot/1~)/g; mean, gain and shape are quantized separately. We apply this idea to AVQ, changing the match heuristic: let and be respectively the of the dictionary block b and of the one anchored in p. The entry b is the best match if d(x~/sub p/,x/spl circ/)/spl les/T (x/spl circ/=g/sub p//spl middot/s~/sub b/+E/sub p//spl middot/1~) and its size is maximum. The triple is entropy coded and sent to the decoder. This simple modification of the match heuristic allows AVQ to improve the compression ratio on many images. In some cases this improvement is as high as 60%. Along with the better compression results, there is also an improvement in the overall visual quality of the decoded image, especially at high compression rate.
只提供摘要形式。Constantinescu和stover(1994)提出了一种自适应矢量量化算法(AVQ),该算法将自适应字典技术与矢量量化(VQ)相结合。该算法在不同类别的图像上通常等于或超过JPEG标准的压缩,并且通常优于传统的训练VQ。我们展示了如何在JPEG最擅长的图像类别(即“杂志照片”)上提高AVQ。这种改进可以通过探索AVQ构建的词典中的相似性来实现。这是通过以类似于平均形状增益VQ中使用的方式变换输入向量来实现的(Oehler和Gray, 1993)。在MSGVQ中,各向量x~/spl isin/R/sup n/分解为x~=g/spl middot/s~+E/sub x//spl middot/1~,其中g=/spl par/x~-E/sub x//spl middot/1~/spl par/, s~=(x~-E/sub x//spl middot/1~)/g;均值、增益和形状分别量化。我们将这一思想应用到AVQ中,改变匹配启发式:让和分别为字典块b和锚定在p中的块。如果d(x~/sub p/,x/spl circ/)/spl les/T (x/spl circ/=g/sub p//spl middot/s~/sub b/+E/sub p//spl middot/1~)且其大小最大,则条目b是最佳匹配。这三个是熵编码并发送到解码器。这种对匹配启发式的简单修改允许AVQ提高许多图像的压缩比。在某些情况下,这种改善高达60%。随着压缩效果的改善,解码图像的整体视觉质量也有所提高,特别是在高压缩率下。
{"title":"Experiments with single-pass adaptive vector quantization","authors":"F. Rizzo, J. Storer, B. Carpentieri","doi":"10.1109/DCC.1999.785703","DOIUrl":"https://doi.org/10.1109/DCC.1999.785703","url":null,"abstract":"Summary form only given. Constantinescu and Storer (1994) introduced an adaptive vector quantization algorithm (AVQ) that combines adaptive dictionary techniques with vector quantization (VQ). The algorithm typically equals or exceeds the compression of the JPEG standard on different classes of images and it often outperforms traditional trained VQ. We show how it is possible to improve AVQ on the class of images on which JPEG does best (i.e., \"magazine photographs\"). The improvement is possible by exploring the similarities in the dictionary built by AVQ. This is achieved by transforming the input vectors in a way similar to the one used in mean-shape-gain VQ (Oehler and Gray, 1993). In MSGVQ each vector x~/spl isin/R/sup n/ is decomposed as x~=g/spl middot/s~+E/sub x//spl middot/1~, where g=/spl par/x~-E/sub x//spl middot/1~/spl par/ and s~=(x~-E/sub x//spl middot/1~)/g; mean, gain and shape are quantized separately. We apply this idea to AVQ, changing the match heuristic: let and be respectively the of the dictionary block b and of the one anchored in p. The entry b is the best match if d(x~/sub p/,x/spl circ/)/spl les/T (x/spl circ/=g/sub p//spl middot/s~/sub b/+E/sub p//spl middot/1~) and its size is maximum. The triple is entropy coded and sent to the decoder. This simple modification of the match heuristic allows AVQ to improve the compression ratio on many images. In some cases this improvement is as high as 60%. Along with the better compression results, there is also an improvement in the overall visual quality of the decoded image, especially at high compression rate.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122731764","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
期刊
Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1