首页 > 最新文献

Proceedings DCC '97. Data Compression Conference最新文献

英文 中文
Progressive Ziv-Lempel encoding of synthetic images 合成图像的渐进式Ziv-Lempel编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582099
Derek Greene, M. Vishwanath, F. Yao, Tong Zhang
Summary form only given. We describe an algorithm that gives a progression of compressed versions of a single image. Each stage of the progression is a lossy compression of the image, with the distortion decreasing in each stage, until the last image is losslessly compressed. Both compressor and decompressor make use of earlier stages to significantly improve the compression of later stages of the progression. Our algorithm uses vector quantization to improve the distortion at the beginning of the progression, and adapts Ziv and Lempel's algorithm to make it efficient for progressive encoding.
只提供摘要形式。我们描述了一种算法,该算法给出了单个图像的压缩版本的级数。级数的每一阶段都是图像的有损压缩,每一阶段的失真都在减少,直到最后一个图像被无损压缩。压气机和减压机都利用较早的阶段来显著改善后期阶段的压缩。我们的算法采用矢量量化来改善序列开始时的失真,并采用Ziv和Lempel算法来提高序列编码的效率。
{"title":"Progressive Ziv-Lempel encoding of synthetic images","authors":"Derek Greene, M. Vishwanath, F. Yao, Tong Zhang","doi":"10.1109/DCC.1997.582099","DOIUrl":"https://doi.org/10.1109/DCC.1997.582099","url":null,"abstract":"Summary form only given. We describe an algorithm that gives a progression of compressed versions of a single image. Each stage of the progression is a lossy compression of the image, with the distortion decreasing in each stage, until the last image is losslessly compressed. Both compressor and decompressor make use of earlier stages to significantly improve the compression of later stages of the progression. Our algorithm uses vector quantization to improve the distortion at the beginning of the progression, and adapts Ziv and Lempel's algorithm to make it efficient for progressive encoding.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"158 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114470169","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Encoding of intervals with conditional coding 区间的条件编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582097
U. Graf
Summary form only given. With conditional coding a new technique is presented that encodes equally likely symbols of an input alphabet A (|A|=m) efficiently. The code consists of bitstrings with size n=[log/sub 2/(m)] and (n+1) and is a prefix code. The encoding needs only one comparison, one shift, and one addition per encoded symbol. Compared to the theoretical limit the method loses only at most 0.086071... bits per encoding and 0.057304... bits in average. Opposed to radix conversion (which achieves the theoretical limit) the algorithm works without multiplication and division and does not need a single-bit writing loop or bitstring arithmetic in the encoding step. Therefore it works a lot faster than radix conversion and can easily be implemented in hardware. The decoding step has the same properties. Encoding and decoding can be exchanged for better adaption to the code alphabet size.
只提供摘要形式。利用条件编码,提出了一种有效编码输入字母a (| a |=m)等可能符号的新方法。该码由大小为n=[log/sub 2/(m)]和(n+1)的位串组成,是前缀码。编码只需要对每个编码符号进行一次比较、一次移位和一次加法。与理论极限相比,该方法最多损失0.086071…每编码位和0.057304…平均位数。与基数转换(达到理论极限)相反,该算法不需要乘法和除法,也不需要在编码步骤中进行单位写入循环或位串算术。因此,它比基数转换快得多,并且可以很容易地在硬件中实现。解码步骤具有相同的属性。编码和解码可以交换,以更好地适应代码字母表的大小。
{"title":"Encoding of intervals with conditional coding","authors":"U. Graf","doi":"10.1109/DCC.1997.582097","DOIUrl":"https://doi.org/10.1109/DCC.1997.582097","url":null,"abstract":"Summary form only given. With conditional coding a new technique is presented that encodes equally likely symbols of an input alphabet A (|A|=m) efficiently. The code consists of bitstrings with size n=[log/sub 2/(m)] and (n+1) and is a prefix code. The encoding needs only one comparison, one shift, and one addition per encoded symbol. Compared to the theoretical limit the method loses only at most 0.086071... bits per encoding and 0.057304... bits in average. Opposed to radix conversion (which achieves the theoretical limit) the algorithm works without multiplication and division and does not need a single-bit writing loop or bitstring arithmetic in the encoding step. Therefore it works a lot faster than radix conversion and can easily be implemented in hardware. The decoding step has the same properties. Encoding and decoding can be exchanged for better adaption to the code alphabet size.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116050276","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A codebook generation algorithm for document image compression 一个用于文档图像压缩的码本生成算法
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582053
Qin Zhang, J. Danskin, N. Young
Pattern-matching based document compression systems rely on finding a small set of patterns that can be used to represent all of the ink in the document. Finding an optimal set of patterns is NP-hard; previous compression schemes have resorted to heuristics. We extend the cross-entropy approach, used previously for measuring pattern similarity, to this problem. Using this approach we reduce the problem to the fixed-cost k-median problem, for which we present a new algorithm with a good provable performance guarantee. We test our new algorithm in place of the previous heuristics (First Fit, with and without generalized Lloyd's (k-means) postprocessing steps). The new algorithm generates a better codebook, resulting in an overall improvement in compression performance of almost 17%.
基于模式匹配的文档压缩系统依赖于找到一小组可用于表示文档中所有墨水的模式。找到一组最优模式是np困难的;以前的压缩方案采用了启发式方法。我们将以前用于测量模式相似性的交叉熵方法扩展到这个问题。利用这种方法,我们将问题简化为固定代价的k-中值问题,并提出了一种具有良好可证明性能保证的新算法。我们测试了我们的新算法,取代了之前的启发式算法(第一次拟合,有和没有广义劳埃德(k-means)后处理步骤)。新算法生成了更好的码本,压缩性能整体提升了近17%。
{"title":"A codebook generation algorithm for document image compression","authors":"Qin Zhang, J. Danskin, N. Young","doi":"10.1109/DCC.1997.582053","DOIUrl":"https://doi.org/10.1109/DCC.1997.582053","url":null,"abstract":"Pattern-matching based document compression systems rely on finding a small set of patterns that can be used to represent all of the ink in the document. Finding an optimal set of patterns is NP-hard; previous compression schemes have resorted to heuristics. We extend the cross-entropy approach, used previously for measuring pattern similarity, to this problem. Using this approach we reduce the problem to the fixed-cost k-median problem, for which we present a new algorithm with a good provable performance guarantee. We test our new algorithm in place of the previous heuristics (First Fit, with and without generalized Lloyd's (k-means) postprocessing steps). The new algorithm generates a better codebook, resulting in an overall improvement in compression performance of almost 17%.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121320020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
A fixed-rate quantizer using block-based entropy-constrained quantization and run-length coding 使用基于块的熵约束量化和运行长度编码的固定速率量化器
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582054
Dongchang Yu, M. Marcellin
A fast and efficient quantization technique is described. It is fixed-length, robust to bit errors, and compatible with most current compression standards. It is based on entropy-constrained quantization and uses the well-known and efficient Viterbi algorithm to force the coded sequence to be fixed-rate. Run-length coding techniques are used to improve the performance at low encoding rates. Simulation results show that it can achieve performance comparable to that of Huffman coded entropy-constrained scalar quantization with computational complexity increasing only linearly in block length.
描述了一种快速有效的量化技术。它是固定长度的,对比特错误具有鲁棒性,并且与大多数当前的压缩标准兼容。它基于熵约束量化,并使用众所周知的高效Viterbi算法强制编码序列固定速率。运行长度编码技术用于提高低编码率下的性能。仿真结果表明,该方法可以达到与霍夫曼编码熵约束标量量化相当的性能,且计算复杂度仅随数据块长度线性增加。
{"title":"A fixed-rate quantizer using block-based entropy-constrained quantization and run-length coding","authors":"Dongchang Yu, M. Marcellin","doi":"10.1109/DCC.1997.582054","DOIUrl":"https://doi.org/10.1109/DCC.1997.582054","url":null,"abstract":"A fast and efficient quantization technique is described. It is fixed-length, robust to bit errors, and compatible with most current compression standards. It is based on entropy-constrained quantization and uses the well-known and efficient Viterbi algorithm to force the coded sequence to be fixed-rate. Run-length coding techniques are used to improve the performance at low encoding rates. Simulation results show that it can achieve performance comparable to that of Huffman coded entropy-constrained scalar quantization with computational complexity increasing only linearly in block length.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123896360","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
On adaptive strategies for an extended family of Golomb-type codes 一类扩展的golomb型码族的自适应策略
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.581993
G. Seroussi, M. Weinberger
Off-centered, two-sided geometric distributions of the integers are often encountered in lossless image compression applications, as probabilistic models for prediction residuals. Based on a recent characterization of the family of optimal prefix codes for these distributions, which is an extension of the Golomb (1966) codes, we investigate adaptive strategies for their symbol-by-symbol prefix coding, as opposed to arithmetic coding. Our strategies allow for adaptive coding of prediction residuals at very low complexity. They provide a theoretical framework for the heuristic approximations frequently used when modifying the Golomb code, originally designed for one-sided geometric distributions of non-negative integers, so as to apply to the encoding of any integer.
作为预测残差的概率模型,在无损图像压缩应用中经常遇到整数的偏心、双面几何分布。基于最近对这些分布的最优前缀码族的表征,这是Golomb(1966)码的扩展,我们研究了它们的符号对符号前缀编码的自适应策略,而不是算术编码。我们的策略允许在非常低的复杂度下对预测残差进行自适应编码。它们为启发式近似提供了一个理论框架,在修改最初为非负整数的单侧几何分布设计的Golomb编码时经常使用,以便适用于任何整数的编码。
{"title":"On adaptive strategies for an extended family of Golomb-type codes","authors":"G. Seroussi, M. Weinberger","doi":"10.1109/DCC.1997.581993","DOIUrl":"https://doi.org/10.1109/DCC.1997.581993","url":null,"abstract":"Off-centered, two-sided geometric distributions of the integers are often encountered in lossless image compression applications, as probabilistic models for prediction residuals. Based on a recent characterization of the family of optimal prefix codes for these distributions, which is an extension of the Golomb (1966) codes, we investigate adaptive strategies for their symbol-by-symbol prefix coding, as opposed to arithmetic coding. Our strategies allow for adaptive coding of prediction residuals at very low complexity. They provide a theoretical framework for the heuristic approximations frequently used when modifying the Golomb code, originally designed for one-sided geometric distributions of non-negative integers, so as to apply to the encoding of any integer.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127664618","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 33
Conditional entropy coding of VQ indexes for image compression 图像压缩中VQ索引的条件熵编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582058
Xiaolin Wu, Jiang Wen, W. H. Wong
Vector quantization (VQ) is a source coding methodology with provable rate-distortion optimality. However, despite more than two decades of intensive research, VQ theoretical promise is yet to be fully realized in image compression practice. Restricted by high VQ complexity in dimensions and due to high-order sample correlations in images, block sizes of practical VQ image coders are hardly large enough to achieve the rate-distortion optimality. Among the large number of VQ variants in the literature, a technique called address VQ (A-VQ) by Nasrabadi and Feng (1990) achieved the best rate-distortion performance so far to the best of our knowledge. The essence of A-VQ is to effectively increase VQ dimensions by a lossless coding of a group of 16-dimensional VQ codewords that are spatially adjacent. From a different perspective, we can consider a signal source that is coded by memoryless basic VQ to be just another signal source whose samples are the indices of the memoryless VQ codewords, and then induce the problem of lossless compression of the VQ-coded source. If the memoryless VQ is not rate-distortion optimal (often the case in practice), then there must exist hidden structures between the samples of VQ-coded source (VQ codewords). Therefore, an alternative way of approaching the rate-distortion optimality is to model and utilize these inter-codewords structures or correlations by context modeling and conditional entropy coding of VQ indexes.
矢量量化(VQ)是一种具有可证明的率失真最优性的源编码方法。然而,尽管经过了二十多年的深入研究,VQ的理论前景尚未在图像压缩实践中得到充分实现。由于VQ在维度上的高复杂度和图像中的高阶样本相关性的限制,实际VQ图像编码器的块大小很难大到足以实现率失真的最优性。在文献中大量的VQ变体中,Nasrabadi和Feng(1990)的一种称为地址VQ (a -VQ)的技术达到了迄今为止我们所知的最佳速率失真性能。a -VQ的本质是通过对一组空间相邻的16维VQ码字进行无损编码,有效地增加VQ的维数。从另一个角度来看,我们可以把一个由无记忆基本VQ编码的信号源看作是另一个信号源,其样本是无记忆基本VQ码字的索引,从而引出VQ编码的信号源的无损压缩问题。如果无记忆VQ不是速率失真最优(在实践中经常出现这种情况),那么在VQ编码源(VQ码字)的样本之间一定存在隐藏结构。因此,接近率失真最优性的另一种方法是通过上下文建模和VQ索引的条件熵编码来建模和利用这些码字间结构或相关性。
{"title":"Conditional entropy coding of VQ indexes for image compression","authors":"Xiaolin Wu, Jiang Wen, W. H. Wong","doi":"10.1109/DCC.1997.582058","DOIUrl":"https://doi.org/10.1109/DCC.1997.582058","url":null,"abstract":"Vector quantization (VQ) is a source coding methodology with provable rate-distortion optimality. However, despite more than two decades of intensive research, VQ theoretical promise is yet to be fully realized in image compression practice. Restricted by high VQ complexity in dimensions and due to high-order sample correlations in images, block sizes of practical VQ image coders are hardly large enough to achieve the rate-distortion optimality. Among the large number of VQ variants in the literature, a technique called address VQ (A-VQ) by Nasrabadi and Feng (1990) achieved the best rate-distortion performance so far to the best of our knowledge. The essence of A-VQ is to effectively increase VQ dimensions by a lossless coding of a group of 16-dimensional VQ codewords that are spatially adjacent. From a different perspective, we can consider a signal source that is coded by memoryless basic VQ to be just another signal source whose samples are the indices of the memoryless VQ codewords, and then induce the problem of lossless compression of the VQ-coded source. If the memoryless VQ is not rate-distortion optimal (often the case in practice), then there must exist hidden structures between the samples of VQ-coded source (VQ codewords). Therefore, an alternative way of approaching the rate-distortion optimality is to model and utilize these inter-codewords structures or correlations by context modeling and conditional entropy coding of VQ indexes.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"59 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130441020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Perceptual rate control algorithms for fax-based video compression 基于传真的视频压缩的感知速率控制算法
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582086
Yi-Jen Chin, T. Berger
Summary form only given. Video samples usually are predicted from coded versions of nearby samples sent either earlier in the same frame or in the previous frame. Analysis of the human vision system (HVS) suggests that we may not need to correct values of residuals that do not exceed a perceptual threshold sometimes referred to in the literature of perception as the just-noticeable-distortion (JND). The ideal JND provides each pixel being coded with a threshold level below which discrepancies are perceptually distortion-free. Also of interest is the rate control analysis of noticeable, above threshold distortions that inevitably result at low bit rates. Because facsimile-based video compression (FBVC) processing is done in the spatio-temporal pixel domain, we can exploit the local psycho-perceptual properties of the HVS. Our proposed rate control algorithms are distinguished by being computationally economical, transform-free, devoid of block-based artifacts, and capable of easily providing a constant bit rate video stream.
只提供摘要形式。视频样本通常是根据同一帧或前一帧发送的附近样本的编码版本来预测的。对人类视觉系统(HVS)的分析表明,我们可能不需要校正不超过感知阈值的残差值,有时在感知文献中称为just- visible -distortion (JND)。理想的JND为每个像素提供编码的阈值水平,低于该阈值水平,差异在感知上不会失真。同样令人感兴趣的是对明显的、高于阈值的失真的速率控制分析,这些失真不可避免地会导致低比特率。由于基于传真的视频压缩(FBVC)处理是在时空像素域进行的,因此我们可以利用HVS的局部心理感知特性。我们提出的速率控制算法的特点是计算经济,无变换,没有基于块的伪影,并且能够很容易地提供恒定的比特率视频流。
{"title":"Perceptual rate control algorithms for fax-based video compression","authors":"Yi-Jen Chin, T. Berger","doi":"10.1109/DCC.1997.582086","DOIUrl":"https://doi.org/10.1109/DCC.1997.582086","url":null,"abstract":"Summary form only given. Video samples usually are predicted from coded versions of nearby samples sent either earlier in the same frame or in the previous frame. Analysis of the human vision system (HVS) suggests that we may not need to correct values of residuals that do not exceed a perceptual threshold sometimes referred to in the literature of perception as the just-noticeable-distortion (JND). The ideal JND provides each pixel being coded with a threshold level below which discrepancies are perceptually distortion-free. Also of interest is the rate control analysis of noticeable, above threshold distortions that inevitably result at low bit rates. Because facsimile-based video compression (FBVC) processing is done in the spatio-temporal pixel domain, we can exploit the local psycho-perceptual properties of the HVS. Our proposed rate control algorithms are distinguished by being computationally economical, transform-free, devoid of block-based artifacts, and capable of easily providing a constant bit rate video stream.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"60 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125912440","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Progressive image transmission: an adaptive quadtree-pruning approach 渐进图像传输:一种自适应四叉树修剪方法
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582075
C. Bajaj, Guozhong Zhuang
Summary form only given. Progressive, adaptive and hierarchical modes are desirable image coding features. This paper presents a quadtree-pruning pyramid coding scheme satisfying all these objectives. Pyramid coding is an approach suitable for progressive image transmission, where the original image is divided into different levels that correspond to successive approximants of the original one. Starting from the original image, a sequence of reduced-size images is formed by averaging intensity values over 2/spl times/2-pixel blocks. This sequence, called the mean pyramid, ends with an image with only one pixel. Then another sequence of images, called the difference pyramid which can be further encoded via vector quantization, is formed by taking the difference of two consecutive images in the mean pyramid. Our quadtree-pruning approach uses only the mean pyramid. Experiments show that the quadtree-pruning pyramid method is quite efficient for lossy compression. Our approach can also be used for lossless compression by simply setting the threshold function to be zero.
只提供摘要形式。渐进、自适应和分层模式是理想的图像编码特征。本文提出了一种满足上述要求的四叉树剪枝金字塔编码方案。金字塔编码是一种适用于渐进图像传输的方法,它将原始图像分成不同的层次,对应于原始图像的连续近似值。从原始图像开始,通过平均2/spl倍/2像素块上的强度值,形成一系列减小尺寸的图像。这个序列被称为平均金字塔,以只有一个像素的图像结束。然后在平均金字塔中取两幅连续图像的差,形成另一序列图像,称为差金字塔,可以通过矢量量化进一步编码。我们的四叉树修剪方法只使用平均金字塔。实验结果表明,四叉树剪枝金字塔方法对有损压缩是非常有效的。我们的方法也可以用于无损压缩,只需将阈值函数设置为零。
{"title":"Progressive image transmission: an adaptive quadtree-pruning approach","authors":"C. Bajaj, Guozhong Zhuang","doi":"10.1109/DCC.1997.582075","DOIUrl":"https://doi.org/10.1109/DCC.1997.582075","url":null,"abstract":"Summary form only given. Progressive, adaptive and hierarchical modes are desirable image coding features. This paper presents a quadtree-pruning pyramid coding scheme satisfying all these objectives. Pyramid coding is an approach suitable for progressive image transmission, where the original image is divided into different levels that correspond to successive approximants of the original one. Starting from the original image, a sequence of reduced-size images is formed by averaging intensity values over 2/spl times/2-pixel blocks. This sequence, called the mean pyramid, ends with an image with only one pixel. Then another sequence of images, called the difference pyramid which can be further encoded via vector quantization, is formed by taking the difference of two consecutive images in the mean pyramid. Our quadtree-pruning approach uses only the mean pyramid. Experiments show that the quadtree-pruning pyramid method is quite efficient for lossy compression. Our approach can also be used for lossless compression by simply setting the threshold function to be zero.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126873208","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Universal transform coding based on backward adaptation 基于后向自适应的通用变换编码
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582046
Vivek K Goyal, Jun Zhuang, M. Vetterli
The method for universal transform coding based on backward adaptation introduced by Goyal et al. (see IEEE Int. Conf. Image Proc., vol.II, p.365-8, 1996) is reviewed and further analyzed. This algorithm uses a linear transform which is periodically updated based on a local Karhunen-Loeve transform (KLT) estimate. The KLT estimate is derived purely from quantized data, so the decoder can track the encoder state without any side information. The effect of estimating only from quantized data is quantitatively analyzed. Two convergence results which hold in the absence of estimation noise are presented. The first applies for any vector dimension but does not preclude the necessity of a sequence of quantization step sizes that goes to zero. The second applies only in the two-dimensional case, but shows local convergence for a fixed, sufficiently small quantization step size. Refinements which reduce the storage and computational requirements of the algorithm are suggested.
Goyal等人提出了一种基于后向自适应的通用变换编码方法。对《Conf. Image Proc, vol.II, p.365-8, 1996》进行了审查和进一步分析。该算法使用基于局部Karhunen-Loeve变换(KLT)估计的周期性更新的线性变换。KLT估计纯粹来自量化数据,因此解码器可以跟踪编码器状态而不需要任何侧信息。定量分析了仅从量化数据进行估计的效果。给出了在没有估计噪声情况下的两个收敛结果。第一个适用于任何向量维度,但并不排除量化步长序列趋近于零的必要性。第二种方法仅适用于二维情况,但对于一个固定的、足够小的量化步长显示出局部收敛性。提出了减少算法存储和计算需求的改进方法。
{"title":"Universal transform coding based on backward adaptation","authors":"Vivek K Goyal, Jun Zhuang, M. Vetterli","doi":"10.1109/DCC.1997.582046","DOIUrl":"https://doi.org/10.1109/DCC.1997.582046","url":null,"abstract":"The method for universal transform coding based on backward adaptation introduced by Goyal et al. (see IEEE Int. Conf. Image Proc., vol.II, p.365-8, 1996) is reviewed and further analyzed. This algorithm uses a linear transform which is periodically updated based on a local Karhunen-Loeve transform (KLT) estimate. The KLT estimate is derived purely from quantized data, so the decoder can track the encoder state without any side information. The effect of estimating only from quantized data is quantitatively analyzed. Two convergence results which hold in the absence of estimation noise are presented. The first applies for any vector dimension but does not preclude the necessity of a sequence of quantization step sizes that goes to zero. The second applies only in the two-dimensional case, but shows local convergence for a fixed, sufficiently small quantization step size. Refinements which reduce the storage and computational requirements of the algorithm are suggested.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128473848","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Video compression with weighted finite automata 加权有限自动机的视频压缩
Pub Date : 1997-03-25 DOI: 10.1109/DCC.1997.582071
J. Albert, S. Frank, U. Hafner, M. Unger
Summary form only given. Weighted finite automata (WFA) exploit self-similarities within single images and also video streams to remove spatial and temporal redundancies. The WFA image codec combines techniques from fractal image compression and vector-quantization to achieve performance results for low bit-rates which can be put on a par with state-of-the-art codecs like embedded zerotree wavelet coding. Moreover, frame regeneration of WFA encoded video streams is faster than that of wavelet coded video streams due to the simple mathematical structure of WFA. Therefore, WFA were chosen as a starting point for a fractal-like video compression with hierarchical motion-compensation. Video streams are structured as proposed by the MPEG standards: the entire video is subdivided into several groups of pictures which are made up of one I-frame and a given number of predicted B- or P-frames. The macro block concept of the MPEG standard is replaced by a hierarchical and adaptive image partitioning. We integrated motion compensation with variable block sizes into the WFA coder to exploit the inter-frame redundancy. The general concept of the WFA compression was retained since it already provides a hierarchical subdivision of the image. The video stream is encoded frame by frame with an improved version of the WFA inference algorithm.
只提供摘要形式。加权有限自动机(WFA)利用单个图像和视频流中的自相似性来消除空间和时间冗余。WFA图像编解码器结合了分形图像压缩和矢量量化技术,实现了低比特率的性能结果,可以与最先进的编解码器(如嵌入式零树小波编码)相提并论。此外,由于WFA编码的视频流数学结构简单,其帧再生速度比小波编码的视频流快。因此,选择WFA作为具有分层运动补偿的类分形视频压缩的起点。视频流按照MPEG标准提出的结构进行构建:整个视频被细分为几组图像,这些图像由一个i帧和给定数量的预测B帧或p帧组成。MPEG标准的宏块概念被分层自适应图像分割所取代。我们将可变块大小的运动补偿集成到WFA编码器中,以利用帧间冗余。保留了WFA压缩的一般概念,因为它已经提供了图像的分层细分。视频流是用改进的WFA推理算法逐帧编码的。
{"title":"Video compression with weighted finite automata","authors":"J. Albert, S. Frank, U. Hafner, M. Unger","doi":"10.1109/DCC.1997.582071","DOIUrl":"https://doi.org/10.1109/DCC.1997.582071","url":null,"abstract":"Summary form only given. Weighted finite automata (WFA) exploit self-similarities within single images and also video streams to remove spatial and temporal redundancies. The WFA image codec combines techniques from fractal image compression and vector-quantization to achieve performance results for low bit-rates which can be put on a par with state-of-the-art codecs like embedded zerotree wavelet coding. Moreover, frame regeneration of WFA encoded video streams is faster than that of wavelet coded video streams due to the simple mathematical structure of WFA. Therefore, WFA were chosen as a starting point for a fractal-like video compression with hierarchical motion-compensation. Video streams are structured as proposed by the MPEG standards: the entire video is subdivided into several groups of pictures which are made up of one I-frame and a given number of predicted B- or P-frames. The macro block concept of the MPEG standard is replaced by a hierarchical and adaptive image partitioning. We integrated motion compensation with variable block sizes into the WFA coder to exploit the inter-frame redundancy. The general concept of the WFA compression was retained since it already provides a hierarchical subdivision of the image. The video stream is encoded frame by frame with an improved version of the WFA inference algorithm.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128506193","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
Proceedings DCC '97. Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1