Universal lossy compression of a data sequence can be obtained by fitting to the source sequence a "simple" reconstruction sequence that can be encoded efficiently and yet be within a tolerable distortion from the given source sequence. We develop iterative algorithms to find such a reconstruction sequence, for a given source sequence, using different criteria of simplicity for the reconstruction sequence. As a result we obtain a practical universal lossy compression method. The proposed method can be applied to source sequences defined over finite or continuous alphabets. We discuss the relation between our method and quantization techniques like entropy coded vector quantization (ECVQ) and trellis coded quantization (TCQ).
{"title":"An iterative technique for universal lossy compression of individual sequences","authors":"Daniel Manor, M. Feder","doi":"10.1109/DCC.1997.581995","DOIUrl":"https://doi.org/10.1109/DCC.1997.581995","url":null,"abstract":"Universal lossy compression of a data sequence can be obtained by fitting to the source sequence a \"simple\" reconstruction sequence that can be encoded efficiently and yet be within a tolerable distortion from the given source sequence. We develop iterative algorithms to find such a reconstruction sequence, for a given source sequence, using different criteria of simplicity for the reconstruction sequence. As a result we obtain a practical universal lossy compression method. The proposed method can be applied to source sequences defined over finite or continuous alphabets. We discuss the relation between our method and quantization techniques like entropy coded vector quantization (ECVQ) and trellis coded quantization (TCQ).","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131174903","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We study high-fidelity image compression with a given tight bound on the maximum error magnitude. We propose some practical adaptive context modeling techniques to correct prediction biases caused by quantizing prediction residues, a problem common to the current DPCM like predictive nearly-lossless image coders. By incorporating the proposed techniques into the nearly-lossless version of CALIC, we were able to increase its PSNR by 1 dB or more and/or reduce its bit rate by ten per cent or more. More encouragingly, at bit rates around 1.25 bpp our method obtained competitive PSNR results against the best wavelet coders, while obtaining much smaller maximum error magnitude.
{"title":"L/sub /spl infin//-constrained high-fidelity image compression via adaptive context modeling","authors":"Xiaolin Wu, W. K. Choi, P. Bao","doi":"10.1109/DCC.1997.581978","DOIUrl":"https://doi.org/10.1109/DCC.1997.581978","url":null,"abstract":"We study high-fidelity image compression with a given tight bound on the maximum error magnitude. We propose some practical adaptive context modeling techniques to correct prediction biases caused by quantizing prediction residues, a problem common to the current DPCM like predictive nearly-lossless image coders. By incorporating the proposed techniques into the nearly-lossless version of CALIC, we were able to increase its PSNR by 1 dB or more and/or reduce its bit rate by ten per cent or more. More encouragingly, at bit rates around 1.25 bpp our method obtained competitive PSNR results against the best wavelet coders, while obtaining much smaller maximum error magnitude.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131131968","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Summary form only given. Although rate-distortion theory establishes optimal coding properties for vector quantization (VQ) of stationary sources, the fact that real sources are, in actuality, nonstationary has led to the proposal of adaptive-VQ (AVQ) algorithms that compensate for changing source statistics. Because of the scarcity of rate-distortion results for nonstationary sources, proposed AVQ algorithms have been mostly heuristically, rather than analytically, motivated. As a result, there has been, to date, little attempt to develop a general model of AVQ or to compare the performance associated with existing AVQ algorithms. We summarize observations resulting from detailed studies of a number of previously published AVQ algorithms. To our knowledge, the observations represent the first attempt to define and describe AVQ in a general framework. We begin by proposing a mathematical definition of AVQ. Because of the large variety of algorithms that have purported to be AVQ, it is unclear from prior literature precisely what is meant by this term. Any resulting confusion is likely due to a certain imprecise, and sometimes ambiguous, use of the word "adaptive" in VQ literature. However, common to a large part of this literature is the notion that AVQ properly refers to techniques that dynamically vary the contents of a VQ codebook as coding progresses. Our definition of AVQ captures this idea of progressive codebook updating in a general mathematical framework.
{"title":"Adaptive vector quantization-Part I: a unifying structure","authors":"J. Fowler","doi":"10.1109/DCC.1997.582094","DOIUrl":"https://doi.org/10.1109/DCC.1997.582094","url":null,"abstract":"Summary form only given. Although rate-distortion theory establishes optimal coding properties for vector quantization (VQ) of stationary sources, the fact that real sources are, in actuality, nonstationary has led to the proposal of adaptive-VQ (AVQ) algorithms that compensate for changing source statistics. Because of the scarcity of rate-distortion results for nonstationary sources, proposed AVQ algorithms have been mostly heuristically, rather than analytically, motivated. As a result, there has been, to date, little attempt to develop a general model of AVQ or to compare the performance associated with existing AVQ algorithms. We summarize observations resulting from detailed studies of a number of previously published AVQ algorithms. To our knowledge, the observations represent the first attempt to define and describe AVQ in a general framework. We begin by proposing a mathematical definition of AVQ. Because of the large variety of algorithms that have purported to be AVQ, it is unclear from prior literature precisely what is meant by this term. Any resulting confusion is likely due to a certain imprecise, and sometimes ambiguous, use of the word \"adaptive\" in VQ literature. However, common to a large part of this literature is the notion that AVQ properly refers to techniques that dynamically vary the contents of a VQ codebook as coding progresses. Our definition of AVQ captures this idea of progressive codebook updating in a general mathematical framework.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132748021","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Lossless textual image compression based on pattern matching classically includes a "residue" coding step that refines an initially lossy reconstructed image to its lossless original form. This step is typically accomplished by arithmetically coding the predicted value for each lossless image pixel, based on the values of previously reconstructed nearby pixels in both the lossless image and its precursor lossy image. Our contribution describes background typical prediction (TPR-B), a fast method for residue coding based on "typical prediction" which permits the skipping of pixels to be arithmetically encoded; and non-symbol typical prediction (TPR-NS), an improved compression method for residue coding also based on "typical prediction". Experimental results are reported based on the residue coding method proposed in Howard's (see Proc. of '96 Data Compression Conf., Snowbird, Utah, p.210-19, 1996) SPM algorithm and the lossy images it generates when applied to eight CCITT bi-level test images. These results demonstrate that after lossy image coding, 88% of the lossless image pixels in the test set can be predicted using TPR-B and need not be residue coded at all. In terms of saved SPM arithmetic coding operations while residue coding, TPR-B achieves an average coding speed increase of 8 times. Using TPR-NS together with TPR-B increases the SPM residue coding compression ratios by an average of 11%.
基于模式匹配的无损文本图像压缩通常包括一个“残差”编码步骤,该步骤将初始有损重构图像细化到其无损原始形式。这一步通常是通过对每个无损图像像素的预测值进行算术编码来完成的,这是基于在无损图像及其前体有损图像中先前重建的附近像素的值。我们的贡献描述了背景典型预测(TPR-B),这是一种基于“典型预测”的快速残差编码方法,它允许对像素的跳过进行算术编码;非符号典型预测(TPR-NS),也是基于“典型预测”改进的残基编码压缩方法。基于Howard提出的残差编码方法(参见Proc. of '96 Data Compression Conf., Snowbird, Utah, p.210- 19,1996) SPM算法及其对8张CCITT双电平测试图像产生的有损图像,报告了实验结果。这些结果表明,在有损图像编码后,使用TPR-B可以预测测试集中88%的无损图像像素,根本不需要进行残差编码。TPR-B在剩余编码时节省的SPM算术编码操作方面,平均编码速度提高了8倍。将TPR-NS与TPR-B结合使用,SPM残基编码压缩比平均提高11%。
{"title":"Fast residue coding for lossless textual image compression","authors":"C. Constantinescu, R. Arps","doi":"10.1109/DCC.1997.582065","DOIUrl":"https://doi.org/10.1109/DCC.1997.582065","url":null,"abstract":"Lossless textual image compression based on pattern matching classically includes a \"residue\" coding step that refines an initially lossy reconstructed image to its lossless original form. This step is typically accomplished by arithmetically coding the predicted value for each lossless image pixel, based on the values of previously reconstructed nearby pixels in both the lossless image and its precursor lossy image. Our contribution describes background typical prediction (TPR-B), a fast method for residue coding based on \"typical prediction\" which permits the skipping of pixels to be arithmetically encoded; and non-symbol typical prediction (TPR-NS), an improved compression method for residue coding also based on \"typical prediction\". Experimental results are reported based on the residue coding method proposed in Howard's (see Proc. of '96 Data Compression Conf., Snowbird, Utah, p.210-19, 1996) SPM algorithm and the lossy images it generates when applied to eight CCITT bi-level test images. These results demonstrate that after lossy image coding, 88% of the lossless image pixels in the test set can be predicted using TPR-B and need not be residue coded at all. In terms of saved SPM arithmetic coding operations while residue coding, TPR-B achieves an average coding speed increase of 8 times. Using TPR-NS together with TPR-B increases the SPM residue coding compression ratios by an average of 11%.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124385847","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
We report the performance of the embedded zerotree wavelet (EZW) using successive-approximation quantization and an adaptive arithmetic coding for effective reduction in bit rates while maintaining high visual quality of reconstructed color images. For 24 bit color images, excellent visual quality is maintained upto a bit rate reduction to approximately 0.48 bpp by EZW yielding a compression ratio (CR) of 50:1. Further bit rate reduction to 0.375 bpp results in a visible degradation by EZW, as is the case when using the adaptive vector quantizer AFLC-VQ. However, the bit rate reduction by AFLC-VQ was computed from the quantizer output and did not include any subsequent entropy coding. Therefore entropy coding of the multi-resolution codebooks generated by adaptive vector quantization of the wavelet coefficients in the AFLC-VQ scheme should reduce the bit rate to at least 0.36 bpp (CR 67:1) at the desired quality currently obtainable at 0.48 bpp by EZW.
{"title":"Low bit rate color image coding with adaptive encoding of wavelet coefficients","authors":"S. Meadows, S. Mitra","doi":"10.1109/DCC.1997.582118","DOIUrl":"https://doi.org/10.1109/DCC.1997.582118","url":null,"abstract":"We report the performance of the embedded zerotree wavelet (EZW) using successive-approximation quantization and an adaptive arithmetic coding for effective reduction in bit rates while maintaining high visual quality of reconstructed color images. For 24 bit color images, excellent visual quality is maintained upto a bit rate reduction to approximately 0.48 bpp by EZW yielding a compression ratio (CR) of 50:1. Further bit rate reduction to 0.375 bpp results in a visible degradation by EZW, as is the case when using the adaptive vector quantizer AFLC-VQ. However, the bit rate reduction by AFLC-VQ was computed from the quantizer output and did not include any subsequent entropy coding. Therefore entropy coding of the multi-resolution codebooks generated by adaptive vector quantization of the wavelet coefficients in the AFLC-VQ scheme should reduce the bit rate to at least 0.36 bpp (CR 67:1) at the desired quality currently obtainable at 0.48 bpp by EZW.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123506803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency.
{"title":"An overhead reduction technique for mega-state compression schemes","authors":"A. Bookstein, S. T. Klein, T. Raita","doi":"10.1109/DCC.1997.582061","DOIUrl":"https://doi.org/10.1109/DCC.1997.582061","url":null,"abstract":"Many of the most effective compression methods involve complicated models. Unfortunately, as model complexity increases, so does the cost of storing the model itself. This paper examines a method to reduce the amount of storage needed to represent a Markov model with an extended alphabet, by applying a clustering scheme that brings together similar states. Experiments run on a variety of large natural language texts show that much of the overhead of storing the model can be saved at the cost of a very small loss of compression efficiency.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129909918","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Summary form only given. This article introduces a non-linear statistical approach to the interframe video coding assuming a priori that the source is non-Gaussian. To this end generalised Gaussian (GG) modelling and high-order statistics are used and a new optimal coding problem is identified as a simultaneous diagonalisation of 2nd and 4th order cumulant tensors. This problem, named the high-order Karhunen-Loeve transform (HOKLT), is an independent component analysis (ICA) method. Using the available linear techniques for cumulant tensor diagonalisation the HOKLT problem cannot be, in general, solved exactly. Considering the impossibility of solving HOKLT problem within the linear group, a non-linear methodology named non-linear independent components analysis (NLICA) that solves the HOKLT problem was introduced. The structure of the analysis operator produced by NLICA is a linear-nonlinear-linear transformation where the first linear stage is an isoentropic ICA operator and the last linear stage is a principal components analysis (PCA) operator. The non-linear stage is diagonal and it converts marginal densities to Gaussianity conserving marginal entropies. Considering the three basic coding modes within DPCM video coders and the three colour components there are nine different sources. Fitting this sources to GG family, done in this work, has shown how far from Gaussianity these sources are and supports the GG modelling effectiveness.
{"title":"Compression of generalised Gaussian sources","authors":"A. Puga, A. P. Alves","doi":"10.1109/DCC.1997.582131","DOIUrl":"https://doi.org/10.1109/DCC.1997.582131","url":null,"abstract":"Summary form only given. This article introduces a non-linear statistical approach to the interframe video coding assuming a priori that the source is non-Gaussian. To this end generalised Gaussian (GG) modelling and high-order statistics are used and a new optimal coding problem is identified as a simultaneous diagonalisation of 2nd and 4th order cumulant tensors. This problem, named the high-order Karhunen-Loeve transform (HOKLT), is an independent component analysis (ICA) method. Using the available linear techniques for cumulant tensor diagonalisation the HOKLT problem cannot be, in general, solved exactly. Considering the impossibility of solving HOKLT problem within the linear group, a non-linear methodology named non-linear independent components analysis (NLICA) that solves the HOKLT problem was introduced. The structure of the analysis operator produced by NLICA is a linear-nonlinear-linear transformation where the first linear stage is an isoentropic ICA operator and the last linear stage is a principal components analysis (PCA) operator. The non-linear stage is diagonal and it converts marginal densities to Gaussianity conserving marginal entropies. Considering the three basic coding modes within DPCM video coders and the three colour components there are nine different sources. Fitting this sources to GG family, done in this work, has shown how far from Gaussianity these sources are and supports the GG modelling effectiveness.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"1994 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125550350","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Summary form only given. This paper presents an algorithm for perceptually lossless image compression. The approach utilizes properties of the human visual system in the form of a perceptual threshold function (PTF) model. The PTF model determines the amount of distortion that can be introduced at each location of the image. Thus, constraining all quantization errors to levels below the PTF results in perceptually lossless image compression. The system employs a modified form of the embedded zerotree wavelet (EZW) coding algorithm that limits the quantization errors of the wavelet transform coefficients to levels below those specified by the model of the perceptual threshold function. Experimental results demonstrate perceptually lossless compression of monochrome images at bit rates ranging from 0.4 to 1.2 bits per pixel at a viewing distance of six times the image height and at bit rates from 0.2 to 0.5 bits per pixel at a viewing distance of twelve times the image height.
{"title":"Perceptually lossless image compression","authors":"Peter J. Hahn, V., John Mathews","doi":"10.1109/DCC.1997.582100","DOIUrl":"https://doi.org/10.1109/DCC.1997.582100","url":null,"abstract":"Summary form only given. This paper presents an algorithm for perceptually lossless image compression. The approach utilizes properties of the human visual system in the form of a perceptual threshold function (PTF) model. The PTF model determines the amount of distortion that can be introduced at each location of the image. Thus, constraining all quantization errors to levels below the PTF results in perceptually lossless image compression. The system employs a modified form of the embedded zerotree wavelet (EZW) coding algorithm that limits the quantization errors of the wavelet transform coefficients to levels below those specified by the model of the perceptual threshold function. Experimental results demonstrate perceptually lossless compression of monochrome images at bit rates ranging from 0.4 to 1.2 bits per pixel at a viewing distance of six times the image height and at bit rates from 0.2 to 0.5 bits per pixel at a viewing distance of twelve times the image height.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"103 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126371180","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Y. Okada, M. Tokuyo, S. Yoshida, N. Okayasu, H. Shimoi
Summary form only given. Lossless data compression is commonly used on personal computers to increase their storage capacity. For example, we can get twice the normal capacity by using lossless data compression algorithms. However, it is necessary to locate compressed data of variable sizes in a fixed-size block with as little fragmentation as possible. This can be accomplished by compressed data management (CDM). The amount of storage capacity provided by data compression depends on the ability of CDM. If CDM does not eliminate fragmentation sufficiently, one cannot attain the storage capacity corresponding to the compression ratio. We present an efficient CDM using a new packed file system (PFS). We confirmed that the PFS achieves and maintains 95% of high space efficiency by using only 1/1000 of the table size needed for the entire storage capacity without employing garbage collection.
{"title":"Effective management of compressed data with packed file systems","authors":"Y. Okada, M. Tokuyo, S. Yoshida, N. Okayasu, H. Shimoi","doi":"10.1109/DCC.1997.582127","DOIUrl":"https://doi.org/10.1109/DCC.1997.582127","url":null,"abstract":"Summary form only given. Lossless data compression is commonly used on personal computers to increase their storage capacity. For example, we can get twice the normal capacity by using lossless data compression algorithms. However, it is necessary to locate compressed data of variable sizes in a fixed-size block with as little fragmentation as possible. This can be accomplished by compressed data management (CDM). The amount of storage capacity provided by data compression depends on the ability of CDM. If CDM does not eliminate fragmentation sufficiently, one cannot attain the storage capacity corresponding to the compression ratio. We present an efficient CDM using a new packed file system (PFS). We confirmed that the PFS achieves and maintains 95% of high space efficiency by using only 1/1000 of the table size needed for the entire storage capacity without employing garbage collection.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"86 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116219927","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
The SPIHT (set partitioning in hierarchical trees) algorithm by Said and Pearlman (see IEEE Trans. on Circuits and Systems for Video Technology, no.6, p.243-250, 1996) is known to have produced some of the best results in still image coding. It is a fully embedded wavelet coding algorithm with precise rate control and low complexity. We present an application of the SPIHT algorithm to video sequences, using three-dimensional (3D) wavelet decompositions and 3D spatio-temporal dependence trees. A full 3D-SPIHT encoder/decoder is implemented in software and is compared against MPEG-2 in parallel simulations. Although there is no motion estimation or compensation in the 3D SPIHT, it performs measurably and visually better than MPEG-2, which employs complicated motion estimation and compensation.
{"title":"An embedded wavelet video coder using three-dimensional set partitioning in hierarchical trees (SPIHT)","authors":"Beong-Jo Kim, W. Pearlman","doi":"10.1109/DCC.1997.582048","DOIUrl":"https://doi.org/10.1109/DCC.1997.582048","url":null,"abstract":"The SPIHT (set partitioning in hierarchical trees) algorithm by Said and Pearlman (see IEEE Trans. on Circuits and Systems for Video Technology, no.6, p.243-250, 1996) is known to have produced some of the best results in still image coding. It is a fully embedded wavelet coding algorithm with precise rate control and low complexity. We present an application of the SPIHT algorithm to video sequences, using three-dimensional (3D) wavelet decompositions and 3D spatio-temporal dependence trees. A full 3D-SPIHT encoder/decoder is implemented in software and is compared against MPEG-2 in parallel simulations. Although there is no motion estimation or compensation in the 3D SPIHT, it performs measurably and visually better than MPEG-2, which employs complicated motion estimation and compensation.","PeriodicalId":403990,"journal":{"name":"Proceedings DCC '97. Data Compression Conference","volume":"113 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1997-03-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127311303","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}