首页 > 最新文献

Proceedings DCC '97. Data Compression Conference最新文献

英文 中文
A new trellis vector residual quantizer with applications to speech and image coding 一种新的栅格矢量残差量化器,应用于语音和图像编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582083
B. Carpentieri, G. Motta
Summary form only given. We present a new trellis coded vector residual quantizer (TCVRQ) that combines trellis coding and vector residual quantization. Our TCVRQ is a general-purpose sub-optimal vector quantizer with low computational costs and small memory requirement that permits high memory savings when compared to traditional quantizers. Our experiments confirm that TCVRQ is a good compromise between memory/speed requirements and quality and that it is not sensitive to codebook design errors. We propose a method for computing quantization levels and experimentally analyze the performance of our TCVRQ when applied to speech coding at very low bit rates and to direct image coding. We employed our TCVRQ in a linear prediction based speech codec for the quantization of the LP parameters. Several experiments were performed using both SNR and a perceptive measure of distortion known as cepstral distance. The results obtained and some informal listening tests show that nearly transparent quantization can be performed at a rate of 1.9 bits per parameter. The experiments in image coding were performed encoding some 256 gray levels, 512/spl times/512 pixel images using blocks of 3/spl times/3 pixels. Our TCVRQ were compared, on the same training and test sets, to an exhaustive search vector quantizer (built using the generalized Lloyd algorithm) and to a tree quantizer for different coding rates ranging from 3 to 10 bits per block.
只提供摘要形式。提出了一种结合网格编码和矢量残差量化的网格编码矢量残差量化器(TCVRQ)。我们的TCVRQ是一种通用的次优矢量量化器,具有低计算成本和小内存需求,与传统量化器相比,可以节省大量内存。我们的实验证实TCVRQ是内存/速度要求和质量之间的一个很好的折衷,并且它对码本设计错误不敏感。我们提出了一种量化水平的计算方法,并实验分析了TCVRQ应用于极低比特率的语音编码和直接图像编码时的性能。我们将TCVRQ用于基于线性预测的语音编解码器,用于LP参数的量化。几个实验进行了使用信噪比和感知测量失真称为倒谱距离。得到的结果和一些非正式的听力测试表明,几乎透明的量化可以在每个参数1.9比特的速率下进行。图像编码实验采用3/spl次/3像素块对256个灰度、512/spl次/512像素的图像进行编码。在相同的训练集和测试集上,我们的TCVRQ与穷举搜索向量量化器(使用广义Lloyd算法构建)和树量化器进行了比较,用于不同的编码速率,从3到10位/块。
{"title":"A new trellis vector residual quantizer with applications to speech and image coding","authors":"B. Carpentieri, G. Motta","doi":"10.1109/DCC.1997.582083","DOIUrl":"https://doi.org/10.1109/DCC.1997.582083","url":null,"abstract":"Summary form only given. We present a new trellis coded vector residual quantizer (TCVRQ) that combines trellis coding and vector residual quantization. Our TCVRQ is a general-purpose sub-optimal vector quantizer with low computational costs and small memory requirement that permits high memory savings when compared to traditional quantizers. Our experiments confirm that TCVRQ is a good compromise between memory/speed requirements and quality and that it is not sensitive to codebook design errors. We propose a method for computing quantization levels and experimentally analyze the performance of our TCVRQ when applied to speech coding at very low bit rates and to direct image coding. We employed our TCVRQ in a linear prediction based speech codec for the quantization of the LP parameters. Several experiments were performed using both SNR and a perceptive measure of distortion known as cepstral distance. The results obtained and some informal listening tests show that nearly transparent quantization can be performed at a rate of 1.9 bits per parameter. The experiments in image coding were performed encoding some 256 gray levels, 512/spl times/512 pixel images using blocks of 3/spl times/3 pixels. Our TCVRQ were compared, on the same training and test sets, to an exhaustive search vector quantizer (built using the generalized Lloyd algorithm) and to a tree quantizer for different coding rates ranging from 3 to 10 bits per block.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124525476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generalized node splitting and bilevel image compression 广义节点分割与双层图像压缩
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582102
H. Helfgott, J. Storer
Summary form only given. Among the methods for lossless compression of bilevel images, algorithms that do node splitting on context pixels obtain the highest compression ratios. For the most part, these methods use binary variables to do the splitting. Variables that can adopt more than two values are sometimes used, but each possible value of the variable always determines a separate child of a node. We put forward the use of splitting variables that can adopt a very large number of values, including intervals over the reals. At the same time, the number of children per node is kept small as needed. We use a greedy algorithm to repeatedly divide the range of the splitting variable so as to maximize entropy reduction at each step. Both non-local information, e.g., position, and functions on neighborhood pixels can go into tree-building. The resulting compression ratios are higher than those of traditional node-splitting methods. We also show that a context-based codebook, i.e. a function from the set of all possible contexts to the real interval [0,1], can be composed with the inverse of a function from the set of all possible contexts to the reals, such as a function based on Grey coding of the context bitstring, to produce a function from the reals to [0,1] that is very amenable to moderately lossy compression. Even though compression of the codebook is lossy, compression of the image itself is lossless.
只提供摘要形式。在对二层图像进行无损压缩的方法中,对上下文像素进行节点分割的算法压缩比最高。在大多数情况下,这些方法使用二进制变量进行拆分。有时使用可以采用两个以上值的变量,但是变量的每个可能值总是决定节点的一个单独的子节点。我们提出了拆分变量的用法,它可以采用非常多的值,包括实数上的区间。同时,每个节点的子节点数量根据需要保持较小。我们使用贪心算法对分割变量的范围进行重复分割,使每一步的熵降最大化。非局部信息(例如位置)和邻域像素上的函数都可以用于树的构建。所得压缩比高于传统的节点分裂方法。我们还证明了一个基于上下文的码本,即一个从所有可能上下文的集合到实数区间[0,1]的函数,可以与一个从所有可能上下文的集合到实数的函数的逆组成,例如一个基于上下文位串的灰色编码的函数,以产生一个从实数到[0,1]的函数,这个函数非常适合适度的有损压缩。尽管码本的压缩是有损的,但图像本身的压缩是无损的。
{"title":"Generalized node splitting and bilevel image compression","authors":"H. Helfgott, J. Storer","doi":"10.1109/DCC.1997.582102","DOIUrl":"https://doi.org/10.1109/DCC.1997.582102","url":null,"abstract":"Summary form only given. Among the methods for lossless compression of bilevel images, algorithms that do node splitting on context pixels obtain the highest compression ratios. For the most part, these methods use binary variables to do the splitting. Variables that can adopt more than two values are sometimes used, but each possible value of the variable always determines a separate child of a node. We put forward the use of splitting variables that can adopt a very large number of values, including intervals over the reals. At the same time, the number of children per node is kept small as needed. We use a greedy algorithm to repeatedly divide the range of the splitting variable so as to maximize entropy reduction at each step. Both non-local information, e.g., position, and functions on neighborhood pixels can go into tree-building. The resulting compression ratios are higher than those of traditional node-splitting methods. We also show that a context-based codebook, i.e. a function from the set of all possible contexts to the real interval [0,1], can be composed with the inverse of a function from the set of all possible contexts to the reals, such as a function based on Grey coding of the context bitstring, to produce a function from the reals to [0,1] that is very amenable to moderately lossy compression. Even though compression of the codebook is lossy, compression of the image itself is lossless.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"85 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121387774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Efficient storage compression for 3D regions 有效的3D区域存储压缩
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582128
G. Panagopoulou, S. Sirmakessis, A. Tsakalidis
Summary form only given. We present the results of the comparison of heuristics algorithms for efficient storage compression for 3D regions. We have implemented five different algorithms. We present the experimental results of the comparison of these five algorithms; the first one is a simple, space consuming, approach that works as the upper bound for the storage requirements of the other four algorithms. It actually groups cubes into larger parallelepipeds. The second algorithm is an invariant of the algorithm of Franzblau-Kleitman (1984). We produced an invariant of their algorithm for 3D regions. Our contribution is the development of the other three algorithms that have less storage requirements than the algorithm of Franzblau-Kleitman. The algorithms have been tested in practice. We used them on files containing 3D regions. Every file contained cubes, described with triples of coordinates. We calculated the number of rectangles that every algorithm generates. A small number of rectangles indicates good performance.
只提供摘要形式。我们提出了对有效的三维区域存储压缩的启发式算法的比较结果。我们已经实现了五种不同的算法。给出了这五种算法的对比实验结果;第一种是一种简单的、消耗空间的方法,它是其他四种算法存储需求的上限。它实际上把立方体分成了更大的平行六面体。第二种算法是Franzblau-Kleitman(1984)算法的不变量。我们为3D区域生成了他们算法的不变量。我们的贡献是开发了其他三种比Franzblau-Kleitman算法存储需求更少的算法。这些算法已经在实践中得到了验证。我们在包含3D区域的文件上使用它们。每个文件都包含用坐标三元组描述的多维数据集。我们计算了每个算法生成的矩形的数量。矩形数量少表示性能好。
{"title":"Efficient storage compression for 3D regions","authors":"G. Panagopoulou, S. Sirmakessis, A. Tsakalidis","doi":"10.1109/DCC.1997.582128","DOIUrl":"https://doi.org/10.1109/DCC.1997.582128","url":null,"abstract":"Summary form only given. We present the results of the comparison of heuristics algorithms for efficient storage compression for 3D regions. We have implemented five different algorithms. We present the experimental results of the comparison of these five algorithms; the first one is a simple, space consuming, approach that works as the upper bound for the storage requirements of the other four algorithms. It actually groups cubes into larger parallelepipeds. The second algorithm is an invariant of the algorithm of Franzblau-Kleitman (1984). We produced an invariant of their algorithm for 3D regions. Our contribution is the development of the other three algorithms that have less storage requirements than the algorithm of Franzblau-Kleitman. The algorithms have been tested in practice. We used them on files containing 3D regions. Every file contained cubes, described with triples of coordinates. We calculated the number of rectangles that every algorithm generates. A small number of rectangles indicates good performance.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132066462","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Symbol ranking text compressors 符号排序文本压缩器
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582093
P. Fenwick
Summary form only given. In 1951 Shannon estimated the entropy of English text by giving human subjects a sample of text and asking them to guess the next letters. He found, in one example, that 79% of the attempts were correct at the first try, 8% needed two attempts and 3% needed 3 attempts. By regarding the number of attempts as an information source he could estimate the language entropy. Shannon also stated that an "identical twin" to the original predictor could recover the original text and these ideas are developed here to provide a new taxonomy of text compressors. In all cases these compressors recode the input into "rankings" of "most probable symbol", "next most probable symbol", and so on. The rankings have a very skew distribution (low entropy) and are processed by a conventional statistical compressor. Several "symbol ranking" compressors have appeared in the literature, though seldom with that name or even reference to Shannon's work. The author has developed a compressor which uses constant-order contexts and is based on a set-associative cache with LRU update. A software implementation has run at about 1 Mbyte/s with an average compression of 3.6 bits/byte on the Calgary Corpus.
只提供摘要形式。1951年,香农通过给人类受试者一份文本样本,并让他们猜测下一个字母,估计了英语文本的熵。他发现,在一个例子中,79%的尝试在第一次尝试时是正确的,8%需要两次尝试,3%需要三次尝试。通过将尝试次数作为信息源,他可以估计语言熵。香农还表示,原始预测器的“同卵双胞胎”可以恢复原始文本,这些想法在这里得到发展,以提供一种新的文本压缩器分类法。在所有情况下,这些压缩器将输入重新编码为“最可能符号”、“下一个最可能符号”的“排名”,以此类推。排名具有非常不均匀的分布(低熵),并由传统的统计压缩器处理。几个“符号排序”压缩器已经出现在文献中,虽然很少有这个名字,甚至参考香农的工作。作者开发了一种使用常序上下文的压缩器,该压缩器基于集关联缓存和LRU更新。在卡尔加里语料库上,软件实现的运行速度约为1mbyte /s,平均压缩率为3.6 bits/byte。
{"title":"Symbol ranking text compressors","authors":"P. Fenwick","doi":"10.1109/DCC.1997.582093","DOIUrl":"https://doi.org/10.1109/DCC.1997.582093","url":null,"abstract":"Summary form only given. In 1951 Shannon estimated the entropy of English text by giving human subjects a sample of text and asking them to guess the next letters. He found, in one example, that 79% of the attempts were correct at the first try, 8% needed two attempts and 3% needed 3 attempts. By regarding the number of attempts as an information source he could estimate the language entropy. Shannon also stated that an \"identical twin\" to the original predictor could recover the original text and these ideas are developed here to provide a new taxonomy of text compressors. In all cases these compressors recode the input into \"rankings\" of \"most probable symbol\", \"next most probable symbol\", and so on. The rankings have a very skew distribution (low entropy) and are processed by a conventional statistical compressor. Several \"symbol ranking\" compressors have appeared in the literature, though seldom with that name or even reference to Shannon's work. The author has developed a compressor which uses constant-order contexts and is based on a set-associative cache with LRU update. A software implementation has run at about 1 Mbyte/s with an average compression of 3.6 bits/byte on the Calgary Corpus.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128469294","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Orthonormal sets of filters obtained by modulations and rotations of a prototype 由原型的调制和旋转得到的标准正交滤波器集
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582079
R. Bernardini, J. Kovacevic
Summary form only given. In the past decade the field of image processing has grown considerably, and although various successful techniques have been developed for tasks such as image compression, understanding and segmentation, one final piece is missing. Bearing in mind that ultimately, an image is evaluated by a human observer, it is obvious that the usual mean-square error is not appropriate and thus we still sorely lack subjective measures of image quality. Some physiological results point out that the early stage of the human visual system (HVS) works like a filter bank on the retinal image. The filters in such a filter bank can be seen as being obtained by rotation and modulation from an original prototype filter. This work is just the first step and concentrates on the design of local bases obtained by unitary transformations of a single (or more than one) prototype filter.
只提供摘要形式。在过去的十年里,图像处理领域有了很大的发展,尽管各种成功的技术已经被开发出来,用于图像压缩、理解和分割等任务,但最后一点是缺失的。记住,图像最终是由人类观察者评估的,很明显,通常的均方误差是不合适的,因此我们仍然严重缺乏对图像质量的主观衡量。一些生理学结果指出,人类视觉系统(HVS)的早期阶段就像视网膜图像的滤光器。这种滤波器组中的滤波器可以看作是由原始原型滤波器的旋转和调制获得的。这项工作只是第一步,重点是设计通过单个(或多个)原型滤波器的幺正变换获得的局部基。
{"title":"Orthonormal sets of filters obtained by modulations and rotations of a prototype","authors":"R. Bernardini, J. Kovacevic","doi":"10.1109/DCC.1997.582079","DOIUrl":"https://doi.org/10.1109/DCC.1997.582079","url":null,"abstract":"Summary form only given. In the past decade the field of image processing has grown considerably, and although various successful techniques have been developed for tasks such as image compression, understanding and segmentation, one final piece is missing. Bearing in mind that ultimately, an image is evaluated by a human observer, it is obvious that the usual mean-square error is not appropriate and thus we still sorely lack subjective measures of image quality. Some physiological results point out that the early stage of the human visual system (HVS) works like a filter bank on the retinal image. The filters in such a filter bank can be seen as being obtained by rotation and modulation from an original prototype filter. This work is just the first step and concentrates on the design of local bases obtained by unitary transformations of a single (or more than one) prototype filter.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123695229","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Calligraphic character boundary coding with rational B-spline based on energy minimization using genetic algorithm 基于遗传算法能量最小化的有理b样条书法字符边界编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582076
P. Bao, S. Lam
Summary form only given. Traditional salient point based approaches fail in coding calligraphic characters as noisy boundaries make the extraction of the salient points a difficult task. We propose an alternative solution based on genetic algorithm which searches through the space of possible parameter values until a global optimal solution is found. The objective function we employed is a modified version of the total energy function found in the active contour literature.
只提供摘要形式。传统的基于显著点的方法无法对书法字符进行编码,因为边界噪声使得显著点的提取变得困难。我们提出了一种基于遗传算法的替代解决方案,该算法在可能的参数值空间中搜索,直到找到全局最优解。我们采用的目标函数是在活动轮廓文献中发现的总能量函数的修改版本。
{"title":"Calligraphic character boundary coding with rational B-spline based on energy minimization using genetic algorithm","authors":"P. Bao, S. Lam","doi":"10.1109/DCC.1997.582076","DOIUrl":"https://doi.org/10.1109/DCC.1997.582076","url":null,"abstract":"Summary form only given. Traditional salient point based approaches fail in coding calligraphic characters as noisy boundaries make the extraction of the salient points a difficult task. We propose an alternative solution based on genetic algorithm which searches through the space of possible parameter values until a global optimal solution is found. The objective function we employed is a modified version of the total energy function found in the active contour literature.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123807401","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Facsimile-images of the future 未来的传真图像
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582143
M. J. Turner, K. Halton
Summary form only given. Fax is immensely successful with over 100 million machines sold world-wide. It is successful because a fax machine is simple and easy to use as a stand-alone device, but it is also very versatile when combined with PC technology. This trend highlights the divergence between PCs and fax machines in image resolution. Many PC applications require resolutions as high as 600 dpi, for example desktop publishing, graphic design, while the highest resolution commonly supported by fax machines is 200 lines/inch. Clearly, transmission of the sort of images which can be generated by a PC places new demands on facsimile. There are at present three main standards for coding facsimile images that have been introduced within the last few years, of which the vast majority of machines use the simplest and oldest technique. This study compares the current standards with a fourth technique developed at De Montfort University called the contour tree format. This new format is a strict two dimensional representation of regions and offers some intrinsic advantages. A set of criteria was investigated with emphasis regarding the compression ratio using many different input conditions.
只提供摘要形式。Fax取得了巨大的成功,在全球销售了超过1亿台机器。它之所以成功,是因为传真机作为一个独立的设备简单易用,但当它与个人电脑技术结合使用时,它也非常通用。这一趋势凸显了个人电脑和传真机在图像分辨率上的差异。许多PC应用程序需要高达600 dpi的分辨率,例如桌面出版,图形设计,而传真机通常支持的最高分辨率是200线/英寸。显然,这种可以由个人电脑生成的图像的传输对传真提出了新的要求。目前有三种主要的传真图像编码标准,这些标准是最近几年引进的,其中绝大多数机器使用的是最简单和最古老的技术。这项研究将目前的标准与德蒙福特大学开发的第四种技术——轮廓树格式进行了比较。这种新格式是严格的二维区域表示,并提供了一些内在的优势。研究了一套标准,重点研究了不同输入条件下的压缩比。
{"title":"Facsimile-images of the future","authors":"M. J. Turner, K. Halton","doi":"10.1109/DCC.1997.582143","DOIUrl":"https://doi.org/10.1109/DCC.1997.582143","url":null,"abstract":"Summary form only given. Fax is immensely successful with over 100 million machines sold world-wide. It is successful because a fax machine is simple and easy to use as a stand-alone device, but it is also very versatile when combined with PC technology. This trend highlights the divergence between PCs and fax machines in image resolution. Many PC applications require resolutions as high as 600 dpi, for example desktop publishing, graphic design, while the highest resolution commonly supported by fax machines is 200 lines/inch. Clearly, transmission of the sort of images which can be generated by a PC places new demands on facsimile. There are at present three main standards for coding facsimile images that have been introduced within the last few years, of which the vast majority of machines use the simplest and oldest technique. This study compares the current standards with a fourth technique developed at De Montfort University called the contour tree format. This new format is a strict two dimensional representation of regions and offers some intrinsic advantages. A set of criteria was investigated with emphasis regarding the compression ratio using many different input conditions.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122053010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Arithmetic coding with improved solution for the carry-over problem 算术编码与结转问题的改进解
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582146
Xiaohui Xue, Wen Gao
Summary form only given. The carry-over problem is inherent in arithmetic coding as a result of using finite precision arithmetic. As far as we know, the currently widely used solution for this problem is the bit-stuffing technique, which was proposed by Rissanen and Langdon (198l). However, this technique is not completely satisfactory. The stuffed-bit affects the coding efficiency slightly. The code stream inserted with several stuffed-bits loses its concept as a real number. This conflicts with the principle that arithmetic coding maps an input stream to an interval on the real line, which is neither perfect nor convenient for analysis. We present our solution for the carry-over problem, the carry-trap technique, which works without the deliberately inserted stuffed-bit. We also present a concise termination method, named the medium termination technique. Both are proved rigorously.
只提供摘要形式。由于使用有限精度算法,算术编码固有的结转问题。据我们所知,目前广泛使用的解决这个问题的方法是位填充技术,由Rissanen和Langdon (198l)提出。然而,这种技术并不完全令人满意。填充位对编码效率影响较小。插入几个填充位的码流失去了作为实数的概念。这与算术编码将输入流映射到实线上的间隔的原则相冲突,这既不完美,也不便于分析。我们提出了一种解决结转问题的方法,即不需要故意插入填充钻头的结转陷阱技术。我们还提出了一种简洁的终止方法,称为介质终止技术。两者都得到了严格的证明。
{"title":"Arithmetic coding with improved solution for the carry-over problem","authors":"Xiaohui Xue, Wen Gao","doi":"10.1109/DCC.1997.582146","DOIUrl":"https://doi.org/10.1109/DCC.1997.582146","url":null,"abstract":"Summary form only given. The carry-over problem is inherent in arithmetic coding as a result of using finite precision arithmetic. As far as we know, the currently widely used solution for this problem is the bit-stuffing technique, which was proposed by Rissanen and Langdon (198l). However, this technique is not completely satisfactory. The stuffed-bit affects the coding efficiency slightly. The code stream inserted with several stuffed-bits loses its concept as a real number. This conflicts with the principle that arithmetic coding maps an input stream to an interval on the real line, which is neither perfect nor convenient for analysis. We present our solution for the carry-over problem, the carry-trap technique, which works without the deliberately inserted stuffed-bit. We also present a concise termination method, named the medium termination technique. Both are proved rigorously.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130053612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A percolating state selector for suffix-tree context models 用于后缀树上下文模型的渗透状态选择器
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.581957
S. Bunton
This paper introduces into practice and empirically evaluates a set of techniques for information-theoretic state selection that have been developing in asymptotic results state selection, which actually implements the selection of an entire model from among a set of competing models, is performed at least trivially by all of the suffix-tree FSMs used for on-line probability estimation. The set of state-selection techniques presented combines orthogonally with the other sets of design options covered in the companion paper of Bunton (Proceedings Data Compression Conference, p.42, 1997).
本文在实践中引入并经验地评价了一组在渐近结果状态选择中发展起来的信息论状态选择技术,这些技术实际上实现了从一组竞争模型中选择整个模型,至少可以通过用于在线概率估计的所有词尾树fsm来实现。所提出的状态选择技术集与Bunton的论文(Proceedings Data Compression Conference,第42页,1997)中涵盖的其他设计选项集正交结合。
{"title":"A percolating state selector for suffix-tree context models","authors":"S. Bunton","doi":"10.1109/DCC.1997.581957","DOIUrl":"https://doi.org/10.1109/DCC.1997.581957","url":null,"abstract":"This paper introduces into practice and empirically evaluates a set of techniques for information-theoretic state selection that have been developing in asymptotic results state selection, which actually implements the selection of an entire model from among a set of competing models, is performed at least trivially by all of the suffix-tree FSMs used for on-line probability estimation. The set of state-selection techniques presented combines orthogonally with the other sets of design options covered in the companion paper of Bunton (Proceedings Data Compression Conference, p.42, 1997).","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121219450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
A corpus for the evaluation of lossless compression algorithms 一个评价无损压缩算法的语料库
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582019
R. Arnold, T. Bell
A number of authors have used the Calgary corpus of texts to provide empirical results for lossless compression algorithms. This corpus was collected in 1987, although it was not published until 1990. The advances with compression algorithms have been achieving relatively small improvements in compression, measured using the Calgary corpus. There is a concern that algorithms are being fine-tuned to this corpus, and that small improvements measured in this way may not apply to other files. Furthermore, the corpus is almost ten years old, and over this period there have been changes in the kinds of files that are compressed, particularly with the development of the Internet, and the rapid growth of high-capacity secondary storage for personal computers. We explore the issues raised above, and develop a principled technique for collecting a corpus of test data for compression methods. A corpus, called the Canterbury corpus, is developed using this technique, and we report the performance of a collection of compression methods using the new corpus.
许多作者使用卡尔加里文本语料库为无损压缩算法提供经验结果。这个语料库于1987年收集,但直到1990年才出版。压缩算法的进步在压缩方面取得了相对较小的改进,使用卡尔加里语料库进行了测量。有一种担忧是,算法正在针对这个语料库进行微调,以这种方式测量的小改进可能不适用于其他文件。此外,语料库已经有将近十年的历史了,在这段时间里,压缩文件的种类发生了变化,特别是随着互联网的发展,以及个人计算机高容量二级存储的迅速增长。我们探讨了上面提出的问题,并开发了一种原则性的技术来收集压缩方法的测试数据库。使用这种技术开发了一个名为坎特伯雷语料库的语料库,我们报告了使用新语料库的压缩方法集合的性能。
{"title":"A corpus for the evaluation of lossless compression algorithms","authors":"R. Arnold, T. Bell","doi":"10.1109/DCC.1997.582019","DOIUrl":"https://doi.org/10.1109/DCC.1997.582019","url":null,"abstract":"A number of authors have used the Calgary corpus of texts to provide empirical results for lossless compression algorithms. This corpus was collected in 1987, although it was not published until 1990. The advances with compression algorithms have been achieving relatively small improvements in compression, measured using the Calgary corpus. There is a concern that algorithms are being fine-tuned to this corpus, and that small improvements measured in this way may not apply to other files. Furthermore, the corpus is almost ten years old, and over this period there have been changes in the kinds of files that are compressed, particularly with the development of the Internet, and the rapid growth of high-capacity secondary storage for personal computers. We explore the issues raised above, and develop a principled technique for collecting a corpus of test data for compression methods. A corpus, called the Canterbury corpus, is developed using this technique, and we report the performance of a collection of compression methods using the new corpus.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"09 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127216853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 217
期刊
Proceedings DCC '97. Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1