首页 > 最新文献

Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)最新文献

英文 中文
Data compression using long common strings 使用长公共字符串进行数据压缩
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755678
J. Bentley
We describe a precompression algorithm that effectively represents any long common strings that appear in a file. The algorithm interacts well with standard compression algorithms that represent shorter strings that are near in the input text. Our experiments show that some real data sets do indeed contain many long common strings. We extend the fingerprint mechanisms of our algorithm to a program that identifies long common strings in an input file. This program gives interesting insights into the structure of real data files that contain long common strings.
我们描述了一种预压缩算法,它可以有效地表示文件中出现的任何长公共字符串。该算法与表示接近输入文本的较短字符串的标准压缩算法很好地交互。我们的实验表明,一些真实的数据集确实包含许多长公共字符串。我们将算法的指纹机制扩展到一个程序,该程序可以识别输入文件中的长公共字符串。这个程序对包含长公共字符串的实际数据文件的结构提供了有趣的见解。
{"title":"Data compression using long common strings","authors":"J. Bentley","doi":"10.1109/DCC.1999.755678","DOIUrl":"https://doi.org/10.1109/DCC.1999.755678","url":null,"abstract":"We describe a precompression algorithm that effectively represents any long common strings that appear in a file. The algorithm interacts well with standard compression algorithms that represent shorter strings that are near in the input text. Our experiments show that some real data sets do indeed contain many long common strings. We extend the fingerprint mechanisms of our algorithm to a program that identifies long common strings in an input file. This program gives interesting insights into the structure of real data files that contain long common strings.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"356 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116241451","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 72
Modified SPIHT encoding for SAR image data 修改了SAR图像数据的SPIHT编码
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785719
Z. Zeng, I. Cumming
Summary form only given. We developed a wavelet-based SAR image compression algorithm which combines tree-structured texture analysis, soft-thresholding speckle reduction, quadtree homogeneous decomposition, and a modified zero-tree coding scheme. First, the tree-structured wavelet transform is applied to the SAR image. The decomposition is no longer simply applied to the low-scale subsignals recursively but to the output of any filter. The measurement of the decomposition is the energy of the image. If the energy of a subimage is significantly smaller than others, we stop the decomposition in this region since it contains less information. The texture factors are created after this step, which represents the amount of texture information. Second, quadtree decomposition is used to split the components in the lowest scale component into two sets, a homogeneous set and a target set. The homogeneous set consists of the relatively homogeneous regions. The target set consists of those non-homogeneous regions which have been further decomposed into single component regions. A conventional soft-threshold is applied to reduce speckle noise on all the wavelet coefficients except those of the lowest scale. The feature factor is used to set the threshold. Finally, the conventional SPIHT methods are modified based on the result from the tree-structured decomposition and the quadtree decomposition. In the encoder, the amount of speckle reduction is chosen based on the requirements of the user. Different coding schemes are applied to the homogeneous set and the target set. The skewed distribution of the residuals makes arithmetic coding the best choice for lossless compression.
只提供摘要形式。我们开发了一种基于小波的SAR图像压缩算法,该算法结合了树结构纹理分析、软阈值斑点消减、四叉树均匀分解和改进的零树编码方案。首先,对SAR图像进行树结构小波变换。分解不再简单地递归地应用于低尺度子信号,而是应用于任何滤波器的输出。分解的度量是图像的能量。如果一个子图像的能量明显小于其他子图像,我们停止在该区域分解,因为它包含的信息较少。纹理因子是在这一步之后创建的,它代表了纹理信息的数量。其次,利用四叉树分解将最小尺度分量中的分量分成两个集合,即齐次集合和目标集合。齐次集合由相对齐次的区域组成。目标集由非均匀区域组成,这些非均匀区域被进一步分解为单个分量区域。采用传统的软阈值法去除除最低尺度外的所有小波系数上的散斑噪声。特征因子用于设置阈值。最后,根据树结构分解和四叉树分解的结果,对传统的SPIHT方法进行了改进。在编码器中,根据用户的要求选择斑点减少量。对齐次集和目标集采用了不同的编码方案。残差的偏态分布使得算术编码成为无损压缩的最佳选择。
{"title":"Modified SPIHT encoding for SAR image data","authors":"Z. Zeng, I. Cumming","doi":"10.1109/DCC.1999.785719","DOIUrl":"https://doi.org/10.1109/DCC.1999.785719","url":null,"abstract":"Summary form only given. We developed a wavelet-based SAR image compression algorithm which combines tree-structured texture analysis, soft-thresholding speckle reduction, quadtree homogeneous decomposition, and a modified zero-tree coding scheme. First, the tree-structured wavelet transform is applied to the SAR image. The decomposition is no longer simply applied to the low-scale subsignals recursively but to the output of any filter. The measurement of the decomposition is the energy of the image. If the energy of a subimage is significantly smaller than others, we stop the decomposition in this region since it contains less information. The texture factors are created after this step, which represents the amount of texture information. Second, quadtree decomposition is used to split the components in the lowest scale component into two sets, a homogeneous set and a target set. The homogeneous set consists of the relatively homogeneous regions. The target set consists of those non-homogeneous regions which have been further decomposed into single component regions. A conventional soft-threshold is applied to reduce speckle noise on all the wavelet coefficients except those of the lowest scale. The feature factor is used to set the threshold. Finally, the conventional SPIHT methods are modified based on the result from the tree-structured decomposition and the quadtree decomposition. In the encoder, the amount of speckle reduction is chosen based on the requirements of the user. Different coding schemes are applied to the homogeneous set and the target set. The skewed distribution of the residuals makes arithmetic coding the best choice for lossless compression.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129473334","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Rate-distortion analysis of spike processes 尖峰过程的速率失真分析
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755657
C. Weidmann, M. Vetterli
Recent rate-distortion analyses of image transform coders are based on a trade-off between the lossless coding of coefficient positions versus the lossy coding of the coefficient values. We propose spike processes as a tool that allows a more fundamental trade-off, namely between lossy position coding and lossy value coding. We investigate the Hamming distortion case and give analytic results for single and multiple spikes. We then consider upper bounds for a single Gaussian spike with squared error distortion. The obtained results show a rate distortion behavior which switches from linear at low rates to exponential at high rates.
最近的图像变换编码器的率失真分析是基于系数位置的无损编码与系数值的有损编码之间的权衡。我们提出尖峰过程作为一种工具,允许更基本的权衡,即有损位置编码和有损值编码之间。我们研究了汉明失真的情况,并给出了单尖峰和多尖峰的分析结果。然后,我们考虑具有平方误差失真的单个高斯尖峰的上界。得到的结果表明,在低速率下由线性转换为指数速率的速率畸变行为。
{"title":"Rate-distortion analysis of spike processes","authors":"C. Weidmann, M. Vetterli","doi":"10.1109/DCC.1999.755657","DOIUrl":"https://doi.org/10.1109/DCC.1999.755657","url":null,"abstract":"Recent rate-distortion analyses of image transform coders are based on a trade-off between the lossless coding of coefficient positions versus the lossy coding of the coefficient values. We propose spike processes as a tool that allows a more fundamental trade-off, namely between lossy position coding and lossy value coding. We investigate the Hamming distortion case and give analytic results for single and multiple spikes. We then consider upper bounds for a single Gaussian spike with squared error distortion. The obtained results show a rate distortion behavior which switches from linear at low rates to exponential at high rates.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125838046","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 29
A perceptual-based video coder for error resilience 基于感知的纠错视频编码器
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785678
Yi-jen Chiu
Summary form only given. Error resilience is an important requirement when errors occur during video transmission. The video transmitted over the Internet is usually a packetized stream and thus the common errors for the Internet video are due to packet loss, caused by buffer overflows in routers, late arrival of packets, and bit errors in the network. This loss results in single or multiple macroblock losses in the decoding process and causes severe degradation in perceived quality and error propagation. We present a perceptual preprocessor based on the insensitivity of the human visual system to the mild changes in pixel intensity in order to segment video into regions according to perceptibility of picture changes. With the information of segmentation, we determine which macroblocks require motion estimation and then which macroblocks need to be included in the second layer. The second layer contains the coarse (less quantized) version of the most perceptually-critical picture information to provide redundancy used to reconstruct lost coding blocks. This information is transmitted in a separate packet, which provides path and time diversities when packet losses are uncorrelated. This combination of methods provides a significant improvement in received quality when losses occur, without significantly degrading the video in a low-bit-rate video channel. Our proposed scheme is easily scalable to various data bitrates, picture quality, and computational complexity for use on different platforms. Because the data in our layered video stream is standards-compliant, our proposed schemes require no extra non-standard device to encode/decode the video and they are easily integrated into the current video standards such as H.261/263, MPEG1/MPEG2 and the forthcoming MPEG4.
只提供摘要形式。当视频传输过程中出现错误时,错误恢复能力是一个重要的要求。在Internet上传输的视频通常是分组流,因此Internet视频的常见错误是由于丢包、路由器缓冲区溢出、数据包延迟到达和网络中的比特错误造成的。这种丢失会导致解码过程中单个或多个宏块丢失,并导致感知质量和错误传播的严重下降。基于人类视觉系统对像素强度轻微变化的不敏感性,提出了一种感知预处理器,根据图像变化的可感知性将视频分割成多个区域。根据分割的信息,我们确定哪些宏块需要运动估计,然后哪些宏块需要包含在第二层中。第二层包含最感知关键的图像信息的粗糙(较少量化)版本,以提供用于重建丢失的编码块的冗余。该信息在一个单独的数据包中传输,当数据包丢失不相关时,它提供了路径和时间的多样性。当出现丢失时,这种方法的组合可以显著提高接收质量,而不会显著降低低比特率视频通道中的视频质量。我们提出的方案可以很容易地扩展到不同的数据比特率、图像质量和计算复杂性,以便在不同的平台上使用。由于我们的分层视频流中的数据是符合标准的,我们提出的方案不需要额外的非标准设备来编码/解码视频,并且它们很容易集成到当前的视频标准中,如H.261/263, MPEG1/MPEG2和即将推出的MPEG4。
{"title":"A perceptual-based video coder for error resilience","authors":"Yi-jen Chiu","doi":"10.1109/DCC.1999.785678","DOIUrl":"https://doi.org/10.1109/DCC.1999.785678","url":null,"abstract":"Summary form only given. Error resilience is an important requirement when errors occur during video transmission. The video transmitted over the Internet is usually a packetized stream and thus the common errors for the Internet video are due to packet loss, caused by buffer overflows in routers, late arrival of packets, and bit errors in the network. This loss results in single or multiple macroblock losses in the decoding process and causes severe degradation in perceived quality and error propagation. We present a perceptual preprocessor based on the insensitivity of the human visual system to the mild changes in pixel intensity in order to segment video into regions according to perceptibility of picture changes. With the information of segmentation, we determine which macroblocks require motion estimation and then which macroblocks need to be included in the second layer. The second layer contains the coarse (less quantized) version of the most perceptually-critical picture information to provide redundancy used to reconstruct lost coding blocks. This information is transmitted in a separate packet, which provides path and time diversities when packet losses are uncorrelated. This combination of methods provides a significant improvement in received quality when losses occur, without significantly degrading the video in a low-bit-rate video channel. Our proposed scheme is easily scalable to various data bitrates, picture quality, and computational complexity for use on different platforms. Because the data in our layered video stream is standards-compliant, our proposed schemes require no extra non-standard device to encode/decode the video and they are easily integrated into the current video standards such as H.261/263, MPEG1/MPEG2 and the forthcoming MPEG4.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127526337","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Binary pseudowavelets and applications to bilevel image processing 二值伪小波及其在二值图像处理中的应用
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755686
S. Pigeon, Yoshua Bengio
This paper shows the existence of binary pseudowavelets, bases on the binary domain that exhibit some of the properties of wavelets, such as multiresolution reconstruction and compact support. The binary pseudowavelets are defined on B/sup n/ (binary vectors of length n) and are operated upon with the binary operators logical and, and exclusive or. The forward transform, or analysis, is the decomposition of a binary vector into its constituent binary pseudowavelets. Binary pseudowavelets allow multiresolution, progressive reconstruction of binary vectors by using progressively more coefficients in the inverse transform. Binary pseudowavelets bases, being sparse matrices, also provide for fast transforms; moreover pseudowavelets rely on hardware-friendly operations for efficient software and hardware implementation.
本文基于二值域证明了二值伪小波的存在性,这些伪小波具有小波的一些特性,如多分辨率重构和紧凑支持。二元伪小波定义在B/sup n/(长度为n的二进制向量)上,并使用二元算子逻辑与和异或进行运算。前向变换或分析是将一个二值向量分解成其组成的二值伪小波。二值伪小波允许多分辨率,通过在反变换中使用逐步增加的系数来逐步重建二值矢量。二值伪小波基作为稀疏矩阵,也提供了快速变换;此外,伪小波依赖于硬件友好的操作,以实现高效的软件和硬件实现。
{"title":"Binary pseudowavelets and applications to bilevel image processing","authors":"S. Pigeon, Yoshua Bengio","doi":"10.1109/DCC.1999.755686","DOIUrl":"https://doi.org/10.1109/DCC.1999.755686","url":null,"abstract":"This paper shows the existence of binary pseudowavelets, bases on the binary domain that exhibit some of the properties of wavelets, such as multiresolution reconstruction and compact support. The binary pseudowavelets are defined on B/sup n/ (binary vectors of length n) and are operated upon with the binary operators logical and, and exclusive or. The forward transform, or analysis, is the decomposition of a binary vector into its constituent binary pseudowavelets. Binary pseudowavelets allow multiresolution, progressive reconstruction of binary vectors by using progressively more coefficients in the inverse transform. Binary pseudowavelets bases, being sparse matrices, also provide for fast transforms; moreover pseudowavelets rely on hardware-friendly operations for efficient software and hardware implementation.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129891432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Reduced comparison search for the exact GLA 减少了精确GLA的比较搜索
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755651
T. Kaukoranta, P. Fränti, O. Nevalainen
This paper introduces a new method for reducing the number of distance calculations in the generalized Lloyd algorithm (GLA), which is a widely used method to construct a codebook in vector quantization. The reduced comparison search detects the activity of the code vectors and utilizes it on the classification of the training vectors. For training vectors whose current code vector has not been modified, we calculate distances only to the active code vectors. A large proportion of the distance calculations can be omitted without sacrificing the optimality of the partition. The new method is included in several fast GLA variants reducing their running times over 50% on average.
本文介绍了一种减少距离计算次数的新方法——广义劳埃德算法(GLA),它是矢量量化中广泛使用的构造码本的方法。简化比较搜索检测代码向量的活动并将其用于训练向量的分类。对于当前代码向量没有被修改的训练向量,我们只计算到活动代码向量的距离。在不牺牲分区的最优性的情况下,可以省去大部分的距离计算。新方法包含在几个快速GLA变体中,平均减少了50%以上的运行时间。
{"title":"Reduced comparison search for the exact GLA","authors":"T. Kaukoranta, P. Fränti, O. Nevalainen","doi":"10.1109/DCC.1999.755651","DOIUrl":"https://doi.org/10.1109/DCC.1999.755651","url":null,"abstract":"This paper introduces a new method for reducing the number of distance calculations in the generalized Lloyd algorithm (GLA), which is a widely used method to construct a codebook in vector quantization. The reduced comparison search detects the activity of the code vectors and utilizes it on the classification of the training vectors. For training vectors whose current code vector has not been modified, we calculate distances only to the active code vectors. A large proportion of the distance calculations can be omitted without sacrificing the optimality of the partition. The new method is included in several fast GLA variants reducing their running times over 50% on average.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130475499","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Two space-economical algorithms for calculating minimum redundancy prefix codes 计算最小冗余前缀码的两种空间经济算法
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755676
R. Milidiú, A. Pessoa, E. Laber
The minimum redundancy prefix code problem is to determine, for a given list W=[w/sub 1/,...,w/sub n/] of n positive symbol weights, a list L=[l/sub 1/,...,l/sub n/] of n corresponding integer codeword lengths such that /spl Sigma//sub i=1//sup n/2/sup -li//spl les/1 and /spl Sigma//sub i=1//sup n/w/sub i/l/sub i/ is minimized. Let us consider the case where W is already sorted. In this case, the output list L can be represented by a list M=[m/sub 1/,...,m/sub H/], where m(l/sub 1/), for l=1,...,H, denotes the multiplicity of the codeword length l in L and H is the length of the greatest codeword. Fortunately, H is proved to be O(min{log(1/(p/sub 1/)),n}), where p/sub 1/ is the smallest symbol probability, given by w/sub 1///spl Sigma//sub i=1//sup n/w/sub i/. We present the F-LazyHuff and the E-LazyHuff algorithms. F-LazyHuff runs in O(n) time but requires O(min{H/sup 2/,n}) additional space. On the other hand, E-LazyHuff runs in O(nlog(n/H)) time, requiring only O(H) additional space. Finally, since our two algorithms have the advantage of not writing at the input buffer during the code calculation, we discuss some applications where this feature is very useful.
最小冗余前缀码问题是确定,对于给定列表W=[W /sub 1/,…],w/下标n/]的n个正符号权值,一个列表L=[L /下标1/,…],l/下标n/]的n个对应的整数码字长度,使得/spl Sigma//下标i=1//sup n/2/sup -li//spl les/1和/spl Sigma//下标i=1//sup n/w/下标i/l/下标i/最小。让我们考虑W已经排序的情况。在这种情况下,输出列表L可以表示为列表M=[M /sub 1/,…],m/下标H/],其中m(l/下标1/),对于l=1,…,H为码字长度l在l中的多重数,H为最大码字的长度。幸运的是,H被证明为O(min{log(1/(p/下标1/)),n}),其中p/下标1/是最小的符号概率,由w/下标1///spl Sigma//下标i=1//sup n/w/下标i/给出。提出了F-LazyHuff算法和E-LazyHuff算法。F-LazyHuff运行时间为O(n),但需要O(min{H/sup 2/,n})额外空间。另一方面,E-LazyHuff在O(nlog(n/H))时间内运行,只需要O(H)额外空间。最后,由于我们的两种算法具有在代码计算期间不写入输入缓冲区的优点,因此我们讨论了一些应用程序,其中该特性非常有用。
{"title":"Two space-economical algorithms for calculating minimum redundancy prefix codes","authors":"R. Milidiú, A. Pessoa, E. Laber","doi":"10.1109/DCC.1999.755676","DOIUrl":"https://doi.org/10.1109/DCC.1999.755676","url":null,"abstract":"The minimum redundancy prefix code problem is to determine, for a given list W=[w/sub 1/,...,w/sub n/] of n positive symbol weights, a list L=[l/sub 1/,...,l/sub n/] of n corresponding integer codeword lengths such that /spl Sigma//sub i=1//sup n/2/sup -li//spl les/1 and /spl Sigma//sub i=1//sup n/w/sub i/l/sub i/ is minimized. Let us consider the case where W is already sorted. In this case, the output list L can be represented by a list M=[m/sub 1/,...,m/sub H/], where m(l/sub 1/), for l=1,...,H, denotes the multiplicity of the codeword length l in L and H is the length of the greatest codeword. Fortunately, H is proved to be O(min{log(1/(p/sub 1/)),n}), where p/sub 1/ is the smallest symbol probability, given by w/sub 1///spl Sigma//sub i=1//sup n/w/sub i/. We present the F-LazyHuff and the E-LazyHuff algorithms. F-LazyHuff runs in O(n) time but requires O(min{H/sup 2/,n}) additional space. On the other hand, E-LazyHuff runs in O(nlog(n/H)) time, requiring only O(H) additional space. Finally, since our two algorithms have the advantage of not writing at the input buffer during the code calculation, we discuss some applications where this feature is very useful.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131888574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
The effect of flexible parsing for dynamic dictionary-based data compression 灵活解析对基于动态字典的数据压缩的影响
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755673
Yossi Matias, N. Rajpoot, S. C. Sahinalp
We report on the performance evaluation of greedy parsing with a single-step lookahead, denoted as flexible parsing. We also introduce a new fingerprint-based data structure which enables efficient linear-time implementation.
我们报告了单步前瞻性贪婪解析的性能评估,称为灵活解析。我们还介绍了一种新的基于指纹的数据结构,它可以实现高效的线性时间实现。
{"title":"The effect of flexible parsing for dynamic dictionary-based data compression","authors":"Yossi Matias, N. Rajpoot, S. C. Sahinalp","doi":"10.1109/DCC.1999.755673","DOIUrl":"https://doi.org/10.1109/DCC.1999.755673","url":null,"abstract":"We report on the performance evaluation of greedy parsing with a single-step lookahead, denoted as flexible parsing. We also introduce a new fingerprint-based data structure which enables efficient linear-time implementation.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"80 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126037652","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
On taking advantage of similarities between parameters in lossless sequential coding 无损序列编码中参数间相似性的利用
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785670
J. Åberg
Summary form only given. In sequential lossless data compression algorithms the data stream is often transformed into short subsequences that are modeled as memoryless. Then it is desirable to use any information that each sequence might provide about the behaviour of other sequences that can be expected to have similar properties. Here we examine one such situation, as follows. We want to encode, using arithmetic coding with a sequential estimator, an M-ary memoryless source with unknown parameters /spl theta/, from which we have encoded already a sequence x/sup n/. In addition, both the encoder and the decoder have observed a sequence y/sup n/ that is generated independently by another source with unknown parameters /spl theta//spl tilde/ that are known to be "similar" to /spl theta/ by a pseudodistance /spl delta/(/spl theta/,/spl theta//spl tilde/) that is approximately equal to the relative entropy. Known to both sides is also a number d such that /spl delta/(/spl theta/,/spl theta//spl tilde/)/spl les/d. For a stand-alone memoryless source, the worst-case average redundancy of the (n+1)-th encoding is lower bounded by 0.5(M-1)/n+O(1/n/sup 2/), and the Dirichlet estimator is close to optimal for this case. We show that this bound holds also for the case with side information as described above, meaning that we can improve, at best, the O(1/n/sup 2/)-term. We define a frequency weighted estimator for this. Application of the frequency weighted estimator to to the PPM algorithm (Bell et al., 1989) by weighting order-4 statistics into an order-5 model, with d estimated during encoding, yields improvements that are consistent with the bounds above, which means that in practice we improve the performance by about 0.5 bits per active state of the model, making a gain of approximately 20000 bits on the Calgary Corpus.
只提供摘要形式。在顺序无损数据压缩算法中,数据流通常被转换成短的子序列,这些子序列被建模为无内存的。然后,需要使用每个序列可能提供的关于可以预期具有类似属性的其他序列的行为的任何信息。在这里,我们考察这样一种情况,如下所示。我们想要编码,使用算术编码与序列估计器,一个M-ary无记忆源与未知参数/spl θ /,从中我们已经编码了序列x/sup n/。此外,编码器和解码器都观察到一个序列y/sup n/,该序列由另一个具有未知参数的源独立产生/spl theta//spl tilde/,已知与/spl theta/“相似”,通过伪距离/spl delta/(/spl theta/,/spl theta//spl tilde/),该序列近似等于相对熵。双方都知道一个数字d,使得/spl delta/(/spl theta/,/spl theta//spl波浪/)/spl les/d。对于独立无内存源,(n+1)次编码的最坏情况平均冗余下界为0.5(M-1)/n+O(1/n/sup 2/), Dirichlet估计器在这种情况下接近最优。我们证明了这个界也适用于上面描述的边信息的情况,这意味着我们最多可以改进O(1/n/sup 2/)项。我们为此定义了一个频率加权估计器。将频率加权估计器应用于PPM算法(Bell et al., 1989),将阶-4统计量加权到阶-5模型中,在编码期间估计d,产生与上述边界一致的改进,这意味着在实践中,我们将模型的每个活动状态的性能提高了约0.5比特,在卡尔加里语料库上获得了约20000比特的增益。
{"title":"On taking advantage of similarities between parameters in lossless sequential coding","authors":"J. Åberg","doi":"10.1109/DCC.1999.785670","DOIUrl":"https://doi.org/10.1109/DCC.1999.785670","url":null,"abstract":"Summary form only given. In sequential lossless data compression algorithms the data stream is often transformed into short subsequences that are modeled as memoryless. Then it is desirable to use any information that each sequence might provide about the behaviour of other sequences that can be expected to have similar properties. Here we examine one such situation, as follows. We want to encode, using arithmetic coding with a sequential estimator, an M-ary memoryless source with unknown parameters /spl theta/, from which we have encoded already a sequence x/sup n/. In addition, both the encoder and the decoder have observed a sequence y/sup n/ that is generated independently by another source with unknown parameters /spl theta//spl tilde/ that are known to be \"similar\" to /spl theta/ by a pseudodistance /spl delta/(/spl theta/,/spl theta//spl tilde/) that is approximately equal to the relative entropy. Known to both sides is also a number d such that /spl delta/(/spl theta/,/spl theta//spl tilde/)/spl les/d. For a stand-alone memoryless source, the worst-case average redundancy of the (n+1)-th encoding is lower bounded by 0.5(M-1)/n+O(1/n/sup 2/), and the Dirichlet estimator is close to optimal for this case. We show that this bound holds also for the case with side information as described above, meaning that we can improve, at best, the O(1/n/sup 2/)-term. We define a frequency weighted estimator for this. Application of the frequency weighted estimator to to the PPM algorithm (Bell et al., 1989) by weighting order-4 statistics into an order-5 model, with d estimated during encoding, yields improvements that are consistent with the bounds above, which means that in practice we improve the performance by about 0.5 bits per active state of the model, making a gain of approximately 20000 bits on the Calgary Corpus.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114462094","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Finite automata and regularized edge-preserving wavelet transform scheme 有限自动机和正则化保边小波变换方案
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785687
Sung-Wai Hong, P. Bao
Summary form only given. We present an edge-preserving image compression technique based on the wavelet transform and iterative constrained least square regularization. This approach treats image reconstruction from lossy image compression as the process of image restoration. It utilizes the edge information detected from the source image as a priori knowledge for the subsequent reconstruction. Image restoration refers to the problem of estimating the source image from its degraded version. The reconstruction of DWT-coded images is formulated as a regularized image recovery problem and makes use of the edge information as the a priori knowledge about the source image to recover the details, as well as to reduce the ringing artifact of the DWT-coded image. To compromise the rate of edge information and DWT-coded image data, a scheme based on generalized finite automata (GFA) is used. GFA is used instead of vector quantization in order to achieve adaptive encoding of the edge image.
只提供摘要形式。提出了一种基于小波变换和迭代约束最小二乘正则化的图像边缘保持压缩技术。该方法将有损图像压缩后的图像重建过程视为图像恢复过程。它利用从源图像中检测到的边缘信息作为后续重建的先验知识。图像恢复是指从降级图像中估计源图像的问题。将dwt编码图像的重建表述为一个正则化的图像恢复问题,利用边缘信息作为源图像的先验知识来恢复细节,并减少dwt编码图像的振铃伪影。为了折衷边缘信息和dwt编码图像数据的速率,采用了一种基于广义有限自动机(GFA)的方案。为了实现边缘图像的自适应编码,采用梯度分解法代替矢量量化。
{"title":"Finite automata and regularized edge-preserving wavelet transform scheme","authors":"Sung-Wai Hong, P. Bao","doi":"10.1109/DCC.1999.785687","DOIUrl":"https://doi.org/10.1109/DCC.1999.785687","url":null,"abstract":"Summary form only given. We present an edge-preserving image compression technique based on the wavelet transform and iterative constrained least square regularization. This approach treats image reconstruction from lossy image compression as the process of image restoration. It utilizes the edge information detected from the source image as a priori knowledge for the subsequent reconstruction. Image restoration refers to the problem of estimating the source image from its degraded version. The reconstruction of DWT-coded images is formulated as a regularized image recovery problem and makes use of the edge information as the a priori knowledge about the source image to recover the details, as well as to reduce the ringing artifact of the DWT-coded image. To compromise the rate of edge information and DWT-coded image data, a scheme based on generalized finite automata (GFA) is used. GFA is used instead of vector quantization in order to achieve adaptive encoding of the edge image.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"69 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124225516","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1