首页 > 最新文献

Proceedings DCC '95 Data Compression Conference最新文献

英文 中文
A multi-dimensional measure for image quality 图像质量的多维度量
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515579
A. Eskicioglu
Summary form only given. It is necessary to develop a quality measure that is capable of determining (1) the amount of degradation, (2) the type of degradation, and (3) the impact of compression on different frequency ranges, in a reconstructed image. We discuss the development of a new graphical measure based on three criteria. To be able to make a local error analysis, we first divide a given image (the original or a degraded) into areas with certain activity levels using, as in the case of Hosaka plots, a quadtree decomposition. The largest and smallest block sizes in our decomposition scheme are 16 and 2, respectively. This gives us 4 classes of blocks having the same size. Class i represents the collection of i/spl times/i blocks; a higher value of i denotes a lower frequency area of the image. After obtaining the quadtree decomposition for a specified value of the variance threshold, we compute three values for each class i (i=2,4,8,16), and normalize them according to: (1) the number of pixels/the number of pixels in the entire image; (2) the number of distinct pixel values/the number of possible pixel values; and (3) the average of the standard deviations in the blocks/a preset maximum standard deviation. The essential characteristics of the image are then displayed in a normalized bar chart. This lays the foundations for designing optimized image coders.
只提供摘要形式。有必要开发一种质量测量方法,能够确定(1)退化的数量,(2)退化的类型,以及(3)压缩对重建图像中不同频率范围的影响。我们讨论了基于三个准则的一种新的图形度量的发展。为了能够进行局部误差分析,我们首先使用四叉树分解将给定图像(原始图像或降级图像)划分为具有特定活动水平的区域,就像在Hosaka地块的情况下一样。我们的分解方案中最大和最小的块大小分别为16和2。这给了我们4类具有相同大小的块。类i表示i/spl次/i块的集合;I值越大,表示图像的频率区域越低。在获得方差阈值指定值的四叉树分解后,我们对每一类i (i=2,4,8,16)计算三个值,并根据:(1)像素数/整个图像的像素数进行归一化;(2)不同像素值的个数/可能的像素值的个数;(3)块内标准差的平均值/预设的最大标准差。然后将图像的基本特征显示在规范化的条形图中。这为设计优化的图像编码器奠定了基础。
{"title":"A multi-dimensional measure for image quality","authors":"A. Eskicioglu","doi":"10.1109/DCC.1995.515579","DOIUrl":"https://doi.org/10.1109/DCC.1995.515579","url":null,"abstract":"Summary form only given. It is necessary to develop a quality measure that is capable of determining (1) the amount of degradation, (2) the type of degradation, and (3) the impact of compression on different frequency ranges, in a reconstructed image. We discuss the development of a new graphical measure based on three criteria. To be able to make a local error analysis, we first divide a given image (the original or a degraded) into areas with certain activity levels using, as in the case of Hosaka plots, a quadtree decomposition. The largest and smallest block sizes in our decomposition scheme are 16 and 2, respectively. This gives us 4 classes of blocks having the same size. Class i represents the collection of i/spl times/i blocks; a higher value of i denotes a lower frequency area of the image. After obtaining the quadtree decomposition for a specified value of the variance threshold, we compute three values for each class i (i=2,4,8,16), and normalize them according to: (1) the number of pixels/the number of pixels in the entire image; (2) the number of distinct pixel values/the number of possible pixel values; and (3) the average of the standard deviations in the blocks/a preset maximum standard deviation. The essential characteristics of the image are then displayed in a normalized bar chart. This lays the foundations for designing optimized image coders.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133608207","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Quantization of overcomplete expansions 过完全展开的量化
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515491
Vivek K Goyal, M. Vetterli, N. T. Thao
We present a method that represents a signal with respect to an overcomplete set of vectors which we call a dictionary. The use of overcomplete sets of vectors (redundant bases or frames) together with quantization is explored as an alternative to transform coding for signal compression. The goal is to retain the computational simplicity of transform coding while adding flexibility like adaptation to signal statistics. We show results using both fixed quantization in frames and greedy quantization using matching pursuit. An MSE slope of -6 dB/octave of frame redundancy is shown for a particular tight frame and is verified experimentally for another frame.
我们提出了一种表示信号相对于我们称之为字典的过完备向量集的方法。利用过完备的向量集(冗余基或帧)与量化一起作为信号压缩变换编码的替代方法。目标是保持变换编码的计算简单性,同时增加灵活性,如适应信号统计。我们展示了在帧中使用固定量化和使用匹配追踪的贪婪量化的结果。对于一个特定的紧框架,显示了帧冗余的MSE斜率为-6 dB/倍频,并在另一个框架上进行了实验验证。
{"title":"Quantization of overcomplete expansions","authors":"Vivek K Goyal, M. Vetterli, N. T. Thao","doi":"10.1109/DCC.1995.515491","DOIUrl":"https://doi.org/10.1109/DCC.1995.515491","url":null,"abstract":"We present a method that represents a signal with respect to an overcomplete set of vectors which we call a dictionary. The use of overcomplete sets of vectors (redundant bases or frames) together with quantization is explored as an alternative to transform coding for signal compression. The goal is to retain the computational simplicity of transform coding while adding flexibility like adaptation to signal statistics. We show results using both fixed quantization in frames and greedy quantization using matching pursuit. An MSE slope of -6 dB/octave of frame redundancy is shown for a particular tight frame and is verified experimentally for another frame.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"77 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115067141","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Context models in the MDL framework MDL框架中的上下文模型
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515496
E. Ristad, Robert G. Thomas
Current approaches to speech and handwriting recognition demand a strong language model with a small number of states and an even smaller number of parameters. We introduce four new techniques for statistical language models: multicontextual modeling, nonmonotonic contexts, implicit context growth, and the divergence heuristic. Together these techniques result in language models that have few states, even fewer parameters, and low message entropies. For example, our techniques achieve a message entropy of 2.16 bits/char on the Brown corpus using only 19374 contexts and 54621 parameters. Multicontextual modeling and nonmonotonic contexts, are generalizations of the traditional context model. Implicit context growth ensures that the state transition probabilities of a variable-length Markov process are estimated accurately. This technique is generally applicable to any variable-length Markov process whose state transition probabilities are estimated from string frequencies. In our case, each state in the Markov process represents a context, and implicit context growth conditions the shorter contexts on the fact that the longer contexts did not occur. In a traditional unicontext model, this technique reduces the message entropy of typical English text by 0.1 bits/char. The divergence heuristic, is a heuristic estimation algorithm based on Rissanen's (1978, 1983) minimum description length (MDL) principle and universal data compression algorithm.
当前的语音和手写识别方法需要一个具有少量状态和更少参数的强语言模型。我们介绍了统计语言模型的四种新技术:多上下文建模、非单调上下文、隐含上下文增长和发散启发式。这些技术结合在一起,产生了状态更少、参数更少、消息熵更低的语言模型。例如,我们的技术仅使用19374个上下文和54621个参数在Brown语料库上实现了2.16位/字符的消息熵。多上下文建模和非单调上下文是传统上下文模型的推广。隐式上下文增长保证了变长马尔可夫过程状态转移概率的准确估计。该方法一般适用于任何由弦频率估计状态转移概率的变长马尔可夫过程。在我们的例子中,马尔可夫过程中的每个状态都代表一个上下文,隐式上下文的增长以较长的上下文没有发生为前提,为较短的上下文提供条件。在传统的单上下文模型中,该技术将典型英语文本的消息熵降低了0.1 bits/char。散度启发式算法是一种基于Rissanen(1978, 1983)最小描述长度(MDL)原理和通用数据压缩算法的启发式估计算法。
{"title":"Context models in the MDL framework","authors":"E. Ristad, Robert G. Thomas","doi":"10.1109/DCC.1995.515496","DOIUrl":"https://doi.org/10.1109/DCC.1995.515496","url":null,"abstract":"Current approaches to speech and handwriting recognition demand a strong language model with a small number of states and an even smaller number of parameters. We introduce four new techniques for statistical language models: multicontextual modeling, nonmonotonic contexts, implicit context growth, and the divergence heuristic. Together these techniques result in language models that have few states, even fewer parameters, and low message entropies. For example, our techniques achieve a message entropy of 2.16 bits/char on the Brown corpus using only 19374 contexts and 54621 parameters. Multicontextual modeling and nonmonotonic contexts, are generalizations of the traditional context model. Implicit context growth ensures that the state transition probabilities of a variable-length Markov process are estimated accurately. This technique is generally applicable to any variable-length Markov process whose state transition probabilities are estimated from string frequencies. In our case, each state in the Markov process represents a context, and implicit context growth conditions the shorter contexts on the fact that the longer contexts did not occur. In a traditional unicontext model, this technique reduces the message entropy of typical English text by 0.1 bits/char. The divergence heuristic, is a heuristic estimation algorithm based on Rissanen's (1978, 1983) minimum description length (MDL) principle and universal data compression algorithm.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129747861","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Bitgroup modeling of signal data for image compression 图像压缩信号数据的位组建模
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515576
J. Vaisey, Mark Trumbo
Summary form only given. Binary variable order adaptive algorithms like the UMC of Rissanen (1986) and JBIG can be used to losslessly compress non-binary data by splitting the data into planes, each of 1 bit resolution, and passing each plane to a separate instance of the algorithm. The UMC algorithm operated in this way is the most powerful lossless signal data compressor the authors are aware of. We attempt to develop an understanding of why this approach is so effective. We investigate the common technique of Gray coding the data before splitting it into single-bit planes and passing to the modeler and coder, and compare it to a simple weighted binary coding. We then propose a non-binary pseudo-Gray code as a method of generating planes of resolution greater than or equal to 1 bit, and compare it with the other conventional methods. The algorithm to generate the pseudo-Gray code is much the same as that for the construction of a binary Gray code, except that instead of minimizing the Hamming distance between neighboring bit planes, we instead minimize the Euclidean distance between adjacent groups of bit planes.
只提供摘要形式。二进制可变顺序自适应算法,如Rissanen的UMC(1986)和JBIG,可用于无损压缩非二进制数据,将数据分成平面,每个平面为1位分辨率,并将每个平面传递给算法的单独实例。以这种方式运行的UMC算法是作者所知道的最强大的无损信号数据压缩器。我们试图理解为什么这种方法如此有效。我们研究了在将数据分割成单位平面并传递给建模器和编码器之前对数据进行灰色编码的常用技术,并将其与简单的加权二进制编码进行比较。然后,我们提出了一种非二进制伪格雷码作为生成分辨率大于或等于1位的平面的方法,并将其与其他传统方法进行了比较。生成伪格雷码的算法与构造二进制格雷码的算法非常相似,不同之处在于,我们不是最小化相邻位平面之间的汉明距离,而是最小化相邻位平面组之间的欧几里得距离。
{"title":"Bitgroup modeling of signal data for image compression","authors":"J. Vaisey, Mark Trumbo","doi":"10.1109/DCC.1995.515576","DOIUrl":"https://doi.org/10.1109/DCC.1995.515576","url":null,"abstract":"Summary form only given. Binary variable order adaptive algorithms like the UMC of Rissanen (1986) and JBIG can be used to losslessly compress non-binary data by splitting the data into planes, each of 1 bit resolution, and passing each plane to a separate instance of the algorithm. The UMC algorithm operated in this way is the most powerful lossless signal data compressor the authors are aware of. We attempt to develop an understanding of why this approach is so effective. We investigate the common technique of Gray coding the data before splitting it into single-bit planes and passing to the modeler and coder, and compare it to a simple weighted binary coding. We then propose a non-binary pseudo-Gray code as a method of generating planes of resolution greater than or equal to 1 bit, and compare it with the other conventional methods. The algorithm to generate the pseudo-Gray code is much the same as that for the construction of a binary Gray code, except that instead of minimizing the Hamming distance between neighboring bit planes, we instead minimize the Euclidean distance between adjacent groups of bit planes.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"56 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130010762","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A tree based binary encoding of text using LZW algorithm 使用LZW算法的基于树的文本二进制编码
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515573
T. Acharya, A. Mukherjee
Summary form only given. The most popular adaptive dictionary coding scheme used for text compression is the LZW algorithm. In the LZW algorithm, a changing dictionary contains common strings that have been encountered so far in the text. The dictionary can be represented by a dynamic trie. The input text is examined character by character and the longest substring (called a prefix string) of the text which already exists in the trie, is replaced by a pointer to a node in the trie which represents the prefix string. Motivation of our research is to explore a variation of the LZW algorithm for variable-length binary encoding of text (we call it the LZWA algorithm) and to develop a memory-based VLSI architecture for text compression. We proposed a new methodology to represent the trie in the form of a binary tree (we call it a binary trie) to maintain the dictionary used in the LZW scheme. This binary tree maintains all the properties of the trie and can easily be mapped into memory. As a result, the common substrings can be encoded using variable length prefix binary codes. The prefix codes enable us to uniquely decode the text in its original form. The algorithm outperforms the usual LZW scheme when the size of the text is small (usually less than 5 K). Depending upon the characteristics of the text, the improvement of the compression ratio has been achieved around 10-30% compared to the LZW scheme. But its performance degrades for larger size texts.
只提供摘要形式。用于文本压缩的最流行的自适应字典编码方案是LZW算法。在LZW算法中,不断变化的字典包含到目前为止在文本中遇到的常见字符串。字典可以用一个动态树表示。输入文本将一个字符一个字符地检查,并且已经存在于树中的文本的最长子字符串(称为前缀字符串)将被指向树中代表前缀字符串的节点的指针所替换。我们的研究动机是探索用于文本变长二进制编码的LZW算法的一种变体(我们称之为LZWA算法),并开发用于文本压缩的基于内存的VLSI架构。我们提出了一种新的方法,以二叉树的形式表示树(我们称之为二叉树),以维护LZW方案中使用的字典。这个二叉树维护了树的所有属性,并且可以很容易地映射到内存中。因此,可以使用可变长度前缀二进制代码对公共子字符串进行编码。前缀代码使我们能够以其原始形式唯一地解码文本。当文本大小较小(通常小于5 K)时,该算法优于通常的LZW方案。根据文本的特征,与LZW方案相比,压缩比的提高约为10-30%。但是对于较大的文本,其性能会下降。
{"title":"A tree based binary encoding of text using LZW algorithm","authors":"T. Acharya, A. Mukherjee","doi":"10.1109/DCC.1995.515573","DOIUrl":"https://doi.org/10.1109/DCC.1995.515573","url":null,"abstract":"Summary form only given. The most popular adaptive dictionary coding scheme used for text compression is the LZW algorithm. In the LZW algorithm, a changing dictionary contains common strings that have been encountered so far in the text. The dictionary can be represented by a dynamic trie. The input text is examined character by character and the longest substring (called a prefix string) of the text which already exists in the trie, is replaced by a pointer to a node in the trie which represents the prefix string. Motivation of our research is to explore a variation of the LZW algorithm for variable-length binary encoding of text (we call it the LZWA algorithm) and to develop a memory-based VLSI architecture for text compression. We proposed a new methodology to represent the trie in the form of a binary tree (we call it a binary trie) to maintain the dictionary used in the LZW scheme. This binary tree maintains all the properties of the trie and can easily be mapped into memory. As a result, the common substrings can be encoded using variable length prefix binary codes. The prefix codes enable us to uniquely decode the text in its original form. The algorithm outperforms the usual LZW scheme when the size of the text is small (usually less than 5 K). Depending upon the characteristics of the text, the improvement of the compression ratio has been achieved around 10-30% compared to the LZW scheme. But its performance degrades for larger size texts.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"160 11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128972985","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Operations on compressed image data 对压缩图像数据的操作
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515542
J. Kanai, S. Latifi, G. Rajarathinam, G. Nagy, H. Bunke
A formal framework of directly processing the encoded data is presented. Image operations which can be directly and efficiently applied on run-length encoded data are identified. The FSM and attributed FSM models are used to describe these operations.
提出了一种直接处理编码数据的形式化框架。识别出可以直接有效地应用于游程编码数据的图像操作。FSM模型和FSM属性模型用于描述这些操作。
{"title":"Operations on compressed image data","authors":"J. Kanai, S. Latifi, G. Rajarathinam, G. Nagy, H. Bunke","doi":"10.1109/DCC.1995.515542","DOIUrl":"https://doi.org/10.1109/DCC.1995.515542","url":null,"abstract":"A formal framework of directly processing the encoded data is presented. Image operations which can be directly and efficiently applied on run-length encoded data are identified. The FSM and attributed FSM models are used to describe these operations.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"23 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125686828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Vector quantization for lossless textual data compression 矢量量化无损文本数据压缩
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515584
W. K. Ng, C. Ravishankar
Summary form only given. Vector quantisation (VQ) may be adapted for lossless data compression if the data exhibit vector structures, such as in textural relational databases. Lossless VQ is discussed and it is demonstrated that a relation of tuples may be encoded and allocated to physical disk blocks such that standard database operations such as access, insertion, deletion, and update may be fully supported.
只提供摘要形式。如果数据表现出向量结构,例如在纹理关系数据库中,矢量量化(VQ)可以适用于无损数据压缩。讨论了无损VQ,并证明了元组的关系可以被编码并分配到物理磁盘块,从而可以完全支持标准的数据库操作,如访问、插入、删除和更新。
{"title":"Vector quantization for lossless textual data compression","authors":"W. K. Ng, C. Ravishankar","doi":"10.1109/DCC.1995.515584","DOIUrl":"https://doi.org/10.1109/DCC.1995.515584","url":null,"abstract":"Summary form only given. Vector quantisation (VQ) may be adapted for lossless data compression if the data exhibit vector structures, such as in textural relational databases. Lossless VQ is discussed and it is demonstrated that a relation of tuples may be encoded and allocated to physical disk blocks such that standard database operations such as access, insertion, deletion, and update may be fully supported.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130761452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Optimal linear prediction for the lossless compression of volume data 体数据无损压缩的最优线性预测
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515568
J. Fowler, R. Yagel
Summary form only given. Data in volume form consumes an extraordinary amount of storage space. For efficient storage and transmission of such data, compression algorithms are imperative. However, most volumetric data sets are used in biomedicine and other scientific applications where lossy compression is unacceptable. We present a lossless data compression algorithm which uses optimal linear prediction to exploit correlations in all three dimensions. Our algorithm is a combination of differential pulse-code modulation (DPCM) and Huffman coding and results in compression of around 50% for a set of volume data files. The compression algorithm was run with each of the different predictors on a set of volumes consisting of MRI images, CT images, and electron-density map data.
只提供摘要形式。卷形式的数据消耗了大量的存储空间。为了有效地存储和传输这些数据,压缩算法是必不可少的。然而,大多数体积数据集用于生物医学和其他科学应用,在这些应用中有损压缩是不可接受的。我们提出了一种无损数据压缩算法,该算法使用最优线性预测来利用所有三个维度的相关性。我们的算法是差分脉冲编码调制(DPCM)和霍夫曼编码的组合,对一组体积数据文件压缩了大约50%。在一组由MRI图像、CT图像和电子密度图数据组成的数据集上,使用每种不同的预测器运行压缩算法。
{"title":"Optimal linear prediction for the lossless compression of volume data","authors":"J. Fowler, R. Yagel","doi":"10.1109/DCC.1995.515568","DOIUrl":"https://doi.org/10.1109/DCC.1995.515568","url":null,"abstract":"Summary form only given. Data in volume form consumes an extraordinary amount of storage space. For efficient storage and transmission of such data, compression algorithms are imperative. However, most volumetric data sets are used in biomedicine and other scientific applications where lossy compression is unacceptable. We present a lossless data compression algorithm which uses optimal linear prediction to exploit correlations in all three dimensions. Our algorithm is a combination of differential pulse-code modulation (DPCM) and Huffman coding and results in compression of around 50% for a set of volume data files. The compression algorithm was run with each of the different predictors on a set of volumes consisting of MRI images, CT images, and electron-density map data.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129264430","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Experiments on the zero frequency problem 零频率问题的实验研究
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515590
J. Cleary, W. Teahan
Summary form only given. A fundamental problem in the construction of statistical techniques for data compression of sequential text is the generation of probabilities from counts of previous occurrences. Each context used in the statistical model accumulates counts of the number of times each symbol has occurred in that context. So in a binary alphabet there will be two counts C/sub 0/ and C/sub 1/ (the number of times a 0 or 1 has occurred). The problem then is to take the counts and generate from them a probability that the next character will be a 0 or 1. A naive estimate of the probability of character i could be obtained by the ratio p/sub i/=C/sub i//(C/sub 0/+C/sub i/). A fundamental problem with this is that it will generate a zero probability if C/sub 0/ or C/sub 1/ is zero. Unfortunately, a zero probability prevents coding from working correctly as the "optimum" code length in this case is infinite. Consequently any estimate of the probabilities must be non-zero even in the presence of zero counts. This problem is called the zero frequency problem . A well known solution to the problem was formulated by Laplace and is known as Laplace's law of succession. We have investigated the correctness of Laplace's law by experiment.
只提供摘要形式。构建序列文本数据压缩的统计技术中的一个基本问题是从先前出现的计数中生成概率。统计模型中使用的每个上下文中都会累积每个符号在该上下文中出现的次数。因此,在二进制字母表中,有两个计数C/下标0/和C/下标1/(0或1出现的次数)。接下来的问题是获取计数并从中生成下一个字符为0或1的概率。对字符i的概率的朴素估计可以通过比值p/下标i/=C/下标i//(C/下标0/+C/下标i/)得到。一个基本的问题是,如果C/下标0/或C/下标1/为零,它将产生零概率。不幸的是,零概率会阻止编码正常工作,因为在这种情况下“最佳”代码长度是无限的。因此,即使存在零计数,对概率的任何估计也必须是非零的。这个问题被称为零频率问题。这个问题的一个众所周知的解决方案是由拉普拉斯公式化的,被称为拉普拉斯演替定律。我们用实验研究了拉普拉斯定律的正确性。
{"title":"Experiments on the zero frequency problem","authors":"J. Cleary, W. Teahan","doi":"10.1109/DCC.1995.515590","DOIUrl":"https://doi.org/10.1109/DCC.1995.515590","url":null,"abstract":"Summary form only given. A fundamental problem in the construction of statistical techniques for data compression of sequential text is the generation of probabilities from counts of previous occurrences. Each context used in the statistical model accumulates counts of the number of times each symbol has occurred in that context. So in a binary alphabet there will be two counts C/sub 0/ and C/sub 1/ (the number of times a 0 or 1 has occurred). The problem then is to take the counts and generate from them a probability that the next character will be a 0 or 1. A naive estimate of the probability of character i could be obtained by the ratio p/sub i/=C/sub i//(C/sub 0/+C/sub i/). A fundamental problem with this is that it will generate a zero probability if C/sub 0/ or C/sub 1/ is zero. Unfortunately, a zero probability prevents coding from working correctly as the \"optimum\" code length in this case is infinite. Consequently any estimate of the probabilities must be non-zero even in the presence of zero counts. This problem is called the zero frequency problem . A well known solution to the problem was formulated by Laplace and is known as Laplace's law of succession. We have investigated the correctness of Laplace's law by experiment.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125588622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 35
Adaptive image quantization based on learning classifier systems 基于学习分类器系统的自适应图像量化
Pub Date : 1995-03-28 DOI: 10.1109/DCC.1995.515587
Jianhua Lin
Summary form only given. The performance of a quantizer depends primarily on the selection of a codebook. Most of the quantization techniques used in the past are based on a static codebook which stays unchanged for the entire input. As already demonstrated successfully in lossless data compression, adaptation can be very beneficial in the compression of typically changing input data. Adaptive quantization has been difficult to accomplish because of its lossy nature. We present a model for distribution-free adaptive image quantization based on learning classifier systems which have been used successfully in machine learning. A basic learning classifier system is a special type of message-processing, rule-based system that produces output according to its input environment. Probabilistic learning mechanisms are used to dynamically direct the behavior of the system to adapt to its environment. The adaptiveness of a learning classifier system seems very appropriate for the quantization problem. A learning classifier system based adaptive quantizer consists of the input data, a codebook, and the output. When an input can not be matched, a new codebook entry is constructed to match the input. Such an algorithm allows us not only to deal with the changing environment, but also to control the quality of the quantized output. The adaptive quantizers presented can be applied to both scalar quantization and vector quantization. Experimental results for each case in image quantization are very promising.
只提供摘要形式。量化器的性能主要取决于码本的选择。过去使用的大多数量化技术都是基于静态码本,在整个输入过程中保持不变。正如已经在无损数据压缩中成功证明的那样,自适应在压缩通常变化的输入数据时非常有益。自适应量化由于其有损耗性而难以实现。我们提出了一种基于学习分类器系统的无分布自适应图像量化模型,该模型已成功应用于机器学习。基本学习分类器系统是一种特殊类型的消息处理,基于规则的系统,根据其输入环境产生输出。概率学习机制用于动态地指导系统的行为以适应其环境。学习分类器系统的自适应似乎非常适合于量化问题。基于自适应量化器的学习分类器系统由输入数据、码本和输出组成。当无法匹配输入时,将构造一个新的码本条目来匹配该输入。这样的算法不仅可以处理不断变化的环境,还可以控制量化输出的质量。所提出的自适应量化器既可以用于标量量化,也可以用于矢量量化。每种情况下的图像量化实验结果都很有希望。
{"title":"Adaptive image quantization based on learning classifier systems","authors":"Jianhua Lin","doi":"10.1109/DCC.1995.515587","DOIUrl":"https://doi.org/10.1109/DCC.1995.515587","url":null,"abstract":"Summary form only given. The performance of a quantizer depends primarily on the selection of a codebook. Most of the quantization techniques used in the past are based on a static codebook which stays unchanged for the entire input. As already demonstrated successfully in lossless data compression, adaptation can be very beneficial in the compression of typically changing input data. Adaptive quantization has been difficult to accomplish because of its lossy nature. We present a model for distribution-free adaptive image quantization based on learning classifier systems which have been used successfully in machine learning. A basic learning classifier system is a special type of message-processing, rule-based system that produces output according to its input environment. Probabilistic learning mechanisms are used to dynamically direct the behavior of the system to adapt to its environment. The adaptiveness of a learning classifier system seems very appropriate for the quantization problem. A learning classifier system based adaptive quantizer consists of the input data, a codebook, and the output. When an input can not be matched, a new codebook entry is constructed to match the input. Such an algorithm allows us not only to deal with the changing environment, but also to control the quality of the quantized output. The adaptive quantizers presented can be applied to both scalar quantization and vector quantization. Experimental results for each case in image quantization are very promising.","PeriodicalId":107017,"journal":{"name":"Proceedings DCC '95 Data Compression Conference","volume":"66 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1995-03-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123829818","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings DCC '95 Data Compression Conference
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1