首页 > 最新文献

Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)最新文献

英文 中文
Sorted sliding window compression 分类滑动窗压缩
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785684
U. Graf
Sorted Sliding Window Compression (SSWC) uses a new model (Sorted Sliding Window Model | SSWM) to encode strings e cient, which appear again while encoding a symbol sequence. The SSWM holds statistics of all strings up to certain length k in a sliding window of size n (the sliding window is de ned like in lz77). The compression program can use the SSWM to determine if the string of the next symbols are already contained in the sliding window and returns the length of match. SSWM gives directly statistics (borders of subinterval in an interval) for use in entropy encoding methods like Arithmetic Coding or Dense Coding [Gra97]. For a given number in an interval and the string length the SSWM gives back the corresponding string which is used in decompressing. After an encoding (decoding) step the model is updated with the just encoded (decoded) characters. The Model sorts all string starting points in the sliding window lexicographically. A simple way to implement the SSWM is by exhaustive search in the sliding window. An implementation with a B-tree together with special binary searches is used here. SSWC is a simple compression scheme, which uses this new model to evaluate its properties. It looks on the next characters to encode and determines the longest match with the SSWM. If the match is smaller than 2, the character is encoded. Otherwise the length and the subinterval of the string are encoded. The length values are encoded together with the single characters by using the same adaptive frequency model. Additionally some rules are used to reduce the matching length if the code length get worse. Encoding of frequencies and intervals is done with Dense Coding. SSWC is in average better than gzip [Gai93] on the Calgary corpus: 0:2 0:5 bits-per-byte better on most les and at most 0:03 bits-per-byte worse (progc and progl). This proves the quality and gives con dence in the usability of SSWM as a new building block in models for compression. SSWM has O(log k) computing complexity on all operations and needs O(n) space. SSWM can be used to implement PPM or Markov models in limited space environments because it holds all necessary informations.
只提供摘要形式。排序滑动窗口压缩(SSWC)采用一种新的模型(排序滑动窗口模型,SSWM)对字符串进行高效编码,使字符串在编码符号序列时再次出现。SSWM在大小为n的滑动窗口中保存所有长度为k的字符串的统计信息。压缩程序可以使用SSWM来确定下一个符号的字符串是否已经包含在滑动窗口中,并返回匹配的长度。ssm直接给出统计信息(区间内子区间的边界),用于熵编码方法,如算术编码或密集编码。对于间隔和字符串长度中的给定数字,ssswm会返回用于解压缩的相应字符串。在编码(解码)步骤之后,使用刚刚编码(解码)的字符更新模型。该模型按字典顺序对滑动窗口中的所有字符串起始点进行排序。实现SSWM的一种简单方法是在滑动窗口中进行穷举搜索。这里使用了b树和特殊二叉搜索的实现。SSWC是一种简单的压缩方案,它使用这个新模型来评估其性能。它查找下一个要编码的字符,并确定与ssm的最长匹配。如果匹配小于2,则对该字符进行编码。否则将对字符串的长度和子间隔进行编码。使用相同的自适应频率模型将长度值与单个字符一起编码。此外,如果代码长度变差,则使用一些规则来减少匹配长度。频率和间隔的编码是用密集编码完成的。
{"title":"Sorted sliding window compression","authors":"U. Graf","doi":"10.1109/DCC.1999.785684","DOIUrl":"https://doi.org/10.1109/DCC.1999.785684","url":null,"abstract":"Sorted Sliding Window Compression (SSWC) uses a new model (Sorted Sliding Window Model | SSWM) to encode strings e cient, which appear again while encoding a symbol sequence. The SSWM holds statistics of all strings up to certain length k in a sliding window of size n (the sliding window is de ned like in lz77). The compression program can use the SSWM to determine if the string of the next symbols are already contained in the sliding window and returns the length of match. SSWM gives directly statistics (borders of subinterval in an interval) for use in entropy encoding methods like Arithmetic Coding or Dense Coding [Gra97]. For a given number in an interval and the string length the SSWM gives back the corresponding string which is used in decompressing. After an encoding (decoding) step the model is updated with the just encoded (decoded) characters. The Model sorts all string starting points in the sliding window lexicographically. A simple way to implement the SSWM is by exhaustive search in the sliding window. An implementation with a B-tree together with special binary searches is used here. SSWC is a simple compression scheme, which uses this new model to evaluate its properties. It looks on the next characters to encode and determines the longest match with the SSWM. If the match is smaller than 2, the character is encoded. Otherwise the length and the subinterval of the string are encoded. The length values are encoded together with the single characters by using the same adaptive frequency model. Additionally some rules are used to reduce the matching length if the code length get worse. Encoding of frequencies and intervals is done with Dense Coding. SSWC is in average better than gzip [Gai93] on the Calgary corpus: 0:2 0:5 bits-per-byte better on most les and at most 0:03 bits-per-byte worse (progc and progl). This proves the quality and gives con dence in the usability of SSWM as a new building block in models for compression. SSWM has O(log k) computing complexity on all operations and needs O(n) space. SSWM can be used to implement PPM or Markov models in limited space environments because it holds all necessary informations.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114568823","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Three-dimensional wavelet coding of video with global motion compensation 基于全局运动补偿的视频三维小波编码
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755690
Albert Wang, Zixiang Xiong, P. Chou, S. Mehrotra
Three-dimensional (2D+T) wavelet coding of video using SPIHT has been shown to outperform standard predictive video coders on complex high-motion sequences, and is competitive with standard predictive video coders on simple low-motion sequences. However, on a number of typical moderate-motion sequences characterized by largely rigid motions, 3D SPIHT performs several dB worse than motion-compensated predictive coders, because it is does not take advantage of the real physical motion underlying the scene. We introduce global motion compensation for 3D subband video coders, and find 0.5 to 2 dB gain on sequences with dominant background motion. Our approach is a hybrid of video coding based on sprites, or mosaics, and subband coding.
使用SPIHT的视频三维(2D+T)小波编码在复杂的高运动序列上优于标准预测视频编码器,在简单的低运动序列上与标准预测视频编码器竞争。然而,在一些典型的以刚性运动为特征的中等运动序列上,3D SPIHT的性能比运动补偿预测编码器差几个dB,因为它没有利用场景底层的真实物理运动。我们引入了三维子带视频编码器的全局运动补偿,并找到了具有主导背景运动的序列的0.5到2db增益。我们的方法是基于精灵或马赛克的视频编码和子带编码的混合。
{"title":"Three-dimensional wavelet coding of video with global motion compensation","authors":"Albert Wang, Zixiang Xiong, P. Chou, S. Mehrotra","doi":"10.1109/DCC.1999.755690","DOIUrl":"https://doi.org/10.1109/DCC.1999.755690","url":null,"abstract":"Three-dimensional (2D+T) wavelet coding of video using SPIHT has been shown to outperform standard predictive video coders on complex high-motion sequences, and is competitive with standard predictive video coders on simple low-motion sequences. However, on a number of typical moderate-motion sequences characterized by largely rigid motions, 3D SPIHT performs several dB worse than motion-compensated predictive coders, because it is does not take advantage of the real physical motion underlying the scene. We introduce global motion compensation for 3D subband video coders, and find 0.5 to 2 dB gain on sequences with dominant background motion. Our approach is a hybrid of video coding based on sprites, or mosaics, and subband coding.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130260346","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 63
Text mining: a new frontier for lossless compression 文本挖掘:无损压缩的新前沿
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755669
I. Witten, Zane Bray, M. Mahoui, W. Teahan
Data mining, a burgeoning new technology, is about looking for patterns in data. Likewise, text mining is about looking for patterns in text. Text mining is possible because you do not have to understand text in order to extract useful information from it. Here are four examples. First, if only names could be identified, links could be inserted automatically to other places that mention the same name, links that are "dynamically evaluated" by calling upon a search engine to bind them at click time. Second, actions can be associated with different types of data, using either explicit programming or programming-by-demonstration techniques. A day/time specification appearing anywhere within one's E-mail could be associated with diary actions such as updating a personal organizer or creating an automatic reminder, and each mention of a day/time in the text could raise a popup menu of calendar-based actions. Third, text could be mined for data in tabular format, allowing databases to be created from formatted tables such as stock-market information on Web pages. Fourth, an agent could monitor incoming newswire stories for company names and collect documents that mention them, an automated press clipping service. This paper aims to promote text compression as a key technology for text mining.
数据挖掘是一项新兴的新技术,主要是在数据中寻找模式。同样,文本挖掘就是在文本中寻找模式。文本挖掘是可能的,因为您不必为了从中提取有用的信息而理解文本。这里有四个例子。首先,如果只有名字可以识别,链接就可以自动插入到提到相同名字的其他地方,这些链接通过调用搜索引擎在点击时绑定它们来“动态评估”。其次,可以使用显式编程或演示编程技术将操作与不同类型的数据关联起来。在电子邮件中出现的日期/时间规范可以与日记操作相关联,例如更新个人组织者或创建自动提醒,并且每次在文本中提到日期/时间都可以弹出一个基于日历的操作菜单。第三,可以从文本中挖掘表格格式的数据,从而允许从格式化的表(如Web页面上的股票市场信息)创建数据库。第四,代理可以监控收到的新闻专线报道中的公司名称,并收集提到它们的文件,这是一种自动剪报服务。本文旨在推广文本压缩作为文本挖掘的一项关键技术。
{"title":"Text mining: a new frontier for lossless compression","authors":"I. Witten, Zane Bray, M. Mahoui, W. Teahan","doi":"10.1109/DCC.1999.755669","DOIUrl":"https://doi.org/10.1109/DCC.1999.755669","url":null,"abstract":"Data mining, a burgeoning new technology, is about looking for patterns in data. Likewise, text mining is about looking for patterns in text. Text mining is possible because you do not have to understand text in order to extract useful information from it. Here are four examples. First, if only names could be identified, links could be inserted automatically to other places that mention the same name, links that are \"dynamically evaluated\" by calling upon a search engine to bind them at click time. Second, actions can be associated with different types of data, using either explicit programming or programming-by-demonstration techniques. A day/time specification appearing anywhere within one's E-mail could be associated with diary actions such as updating a personal organizer or creating an automatic reminder, and each mention of a day/time in the text could raise a popup menu of calendar-based actions. Third, text could be mined for data in tabular format, allowing databases to be created from formatted tables such as stock-market information on Web pages. Fourth, an agent could monitor incoming newswire stories for company names and collect documents that mention them, an automated press clipping service. This paper aims to promote text compression as a key technology for text mining.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"106 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130849680","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Source coding with quantized redundant expansions: accuracy and reconstruction 量化冗余展开的源编码:精度与重构
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755684
Z. Cvetković
Signal representations based on low-resolution quantization of redundant expansions is an interesting source coding paradigm, the most important practical case of which is oversampled A/D conversion. Signal reconstruction from quantized coefficients of a redundant expansion and accuracy of representations of this kind are problems which are still not well understood and these are studied in this paper in finite dimensional spaces. It has been previously proven that accuracy of signal representations based on quantized redundant expansions, measured as the squared Euclidean norm of the reconstruction error, cannot be better than O(1/(r/sup 2/)), where r is the expansion redundancy. We give some general conditions under which 1/(r/sup 2/) accuracy can be attained. We also suggest a form of structure for overcomplete families which facilitates reconstruction, and which enables efficient encoding of quantized coefficients with a logarithmic increase of the bit-rate in redundancy.
基于冗余展开的低分辨率量化的信号表示是一种有趣的源编码范式,其中最重要的实际案例是过采样A/D转换。从冗余展开的量化系数重构信号及其表示的准确性是目前尚不清楚的问题,本文在有限维空间中对这些问题进行了研究。以前已经证明,基于量化冗余展开的信号表示的精度,以重构误差的平方欧氏范数来衡量,不能优于O(1/(r/sup 2/)),其中r为展开冗余。我们给出了可以达到1/(r/sup 2/)精度的一般条件。我们还提出了一种结构形式的过完备族,这有利于重建,并使量化系数的有效编码与冗余比特率的对数增长。
{"title":"Source coding with quantized redundant expansions: accuracy and reconstruction","authors":"Z. Cvetković","doi":"10.1109/DCC.1999.755684","DOIUrl":"https://doi.org/10.1109/DCC.1999.755684","url":null,"abstract":"Signal representations based on low-resolution quantization of redundant expansions is an interesting source coding paradigm, the most important practical case of which is oversampled A/D conversion. Signal reconstruction from quantized coefficients of a redundant expansion and accuracy of representations of this kind are problems which are still not well understood and these are studied in this paper in finite dimensional spaces. It has been previously proven that accuracy of signal representations based on quantized redundant expansions, measured as the squared Euclidean norm of the reconstruction error, cannot be better than O(1/(r/sup 2/)), where r is the expansion redundancy. We give some general conditions under which 1/(r/sup 2/) accuracy can be attained. We also suggest a form of structure for overcomplete families which facilitates reconstruction, and which enables efficient encoding of quantized coefficients with a logarithmic increase of the bit-rate in redundancy.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128812773","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Universal lossless source coding with the Burrows Wheeler transform 通用无损源代码编码与Burrows惠勒变换
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755667
M. Effros, Karthik Venkat Ramanan, S. R. Kulkarni, S. Verdú
We here consider a theoretical evaluation of data compression algorithms based on the Burrows Wheeler transform (BWT). The main contributions include a variety of very simple new techniques for BWT-based universal lossless source coding on finite-memory sources and a set of new rate of convergence results for BWT-based source codes. The result is a theoretical validation and quantification of the earlier experimental observation that BWT-based lossless source codes give performance better than that of Ziv-Lempel-style codes and almost as good as that of prediction by partial mapping (PPM) algorithms.
本文考虑了基于Burrows Wheeler变换(BWT)的数据压缩算法的理论评价。主要贡献包括各种非常简单的新技术,用于在有限内存源上基于bwt的通用无损源代码编码,以及一组新的基于bwt的源代码收敛速度结果。结果是对早期实验观察的理论验证和量化,即基于bwt的无损源代码的性能优于Ziv-Lempel-style代码,几乎与部分映射(PPM)算法的预测一样好。
{"title":"Universal lossless source coding with the Burrows Wheeler transform","authors":"M. Effros, Karthik Venkat Ramanan, S. R. Kulkarni, S. Verdú","doi":"10.1109/DCC.1999.755667","DOIUrl":"https://doi.org/10.1109/DCC.1999.755667","url":null,"abstract":"We here consider a theoretical evaluation of data compression algorithms based on the Burrows Wheeler transform (BWT). The main contributions include a variety of very simple new techniques for BWT-based universal lossless source coding on finite-memory sources and a set of new rate of convergence results for BWT-based source codes. The result is a theoretical validation and quantification of the earlier experimental observation that BWT-based lossless source codes give performance better than that of Ziv-Lempel-style codes and almost as good as that of prediction by partial mapping (PPM) algorithms.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132678213","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 134
Edge-adaptive prediction for lossless image coding 无损图像编码的边缘自适应预测
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755698
Wee Sun Lee
We design an edge-adaptive predictor for lossless image coding. The predictor adaptively weights a four-directional predictor together with an adaptive linear predictor based on information from neighbouring pixels. Although conceptually simple, the performance of the resulting coder is comparable to state-of-the-art image coders when a simple context-based coder is used to encode the prediction errors.
设计了一种用于无损图像编码的边缘自适应预测器。该预测器自适应地加权一个四向预测器和一个基于邻近像素信息的自适应线性预测器。虽然概念上很简单,但当使用简单的基于上下文的编码器对预测错误进行编码时,所得到的编码器的性能与最先进的图像编码器相当。
{"title":"Edge-adaptive prediction for lossless image coding","authors":"Wee Sun Lee","doi":"10.1109/DCC.1999.755698","DOIUrl":"https://doi.org/10.1109/DCC.1999.755698","url":null,"abstract":"We design an edge-adaptive predictor for lossless image coding. The predictor adaptively weights a four-directional predictor together with an adaptive linear predictor based on information from neighbouring pixels. Although conceptually simple, the performance of the resulting coder is comparable to state-of-the-art image coders when a simple context-based coder is used to encode the prediction errors.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"73 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134257490","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
A novel dual-path architecture for HDTV video decoding 一种新的HDTV视频解码双路径架构
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785714
N. Wang, N. Ling
Summary form only given. We present an architecture for digital HDTV video decoding (MPEG-2 MP@HL), based on dual decoding data paths controlled in a block layer synchronization manner and an efficient write back scheme. Our fixed schedule controller synchronizes the baseline units on a block basis in both data-paths. This scheme reduces embedded buffer sizes within the decoder and eliminates a lot of external memory bus contentions. In our write back scheme, the display DRAM is physically separated from the anchor picture DRAM, and is added to the display engine, not to the bus. The slight increase in overall DRAM size is acceptable due to the low DRAM cost today. This improves the parallelism in accessing anchor and display pictures and saves about 80 clock cycles per macroblock (based on a 81 MHz clock). Compared to the other decoding approaches such as the slice bar decoding method and the crossing-divided method, this scheme reduces memory access contentions and the amount of embedded local memory required. Our simulations show that with a relatively low speed 81 MHz clock, our architecture uses fewer than the 332 cycles (required real-time decoding upper bound), to decode each macroblock, without a high cost in overall chip area.
只提供摘要形式。我们提出了一种基于块层同步方式控制的双解码数据路径和高效回写方案的数字HDTV视频解码架构(MPEG-2 MP@HL)。我们的固定调度控制器在两个数据路径上以块为基础同步基线单元。该方案减少了解码器内的嵌入式缓冲区大小,并消除了许多外部内存总线争用。在我们的回写方案中,显示DRAM在物理上与锚点图像DRAM分离,并添加到显示引擎中,而不是添加到总线中。由于DRAM成本较低,整体DRAM尺寸的小幅增加是可以接受的。这提高了访问锚和显示图片的并行性,并为每个宏块节省了大约80个时钟周期(基于81 MHz时钟)。与其他解码方法(如切片条解码方法和交叉分割方法)相比,该方案减少了内存访问争用和所需的嵌入式本地内存量。我们的模拟表明,在相对低速的81 MHz时钟下,我们的架构使用少于332个周期(所需的实时解码上限)来解码每个宏块,而不会在整个芯片面积上造成高成本。
{"title":"A novel dual-path architecture for HDTV video decoding","authors":"N. Wang, N. Ling","doi":"10.1109/DCC.1999.785714","DOIUrl":"https://doi.org/10.1109/DCC.1999.785714","url":null,"abstract":"Summary form only given. We present an architecture for digital HDTV video decoding (MPEG-2 MP@HL), based on dual decoding data paths controlled in a block layer synchronization manner and an efficient write back scheme. Our fixed schedule controller synchronizes the baseline units on a block basis in both data-paths. This scheme reduces embedded buffer sizes within the decoder and eliminates a lot of external memory bus contentions. In our write back scheme, the display DRAM is physically separated from the anchor picture DRAM, and is added to the display engine, not to the bus. The slight increase in overall DRAM size is acceptable due to the low DRAM cost today. This improves the parallelism in accessing anchor and display pictures and saves about 80 clock cycles per macroblock (based on a 81 MHz clock). Compared to the other decoding approaches such as the slice bar decoding method and the crossing-divided method, this scheme reduces memory access contentions and the amount of embedded local memory required. Our simulations show that with a relatively low speed 81 MHz clock, our architecture uses fewer than the 332 cycles (required real-time decoding upper bound), to decode each macroblock, without a high cost in overall chip area.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"2007 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127300052","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Lossless JBIG2 coding performance 无损JBIG2编码性能
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785710
D. Tompkins, F. Kossentini
Summary form only given. The Joint Bi-Level Expert Group (JBIG), an international study group affiliated with the ISO/IEC and ITU-T, has recently completed a committee draft of the JBIG2 standard for lossy and lossless bi-level image compression. We study design considerations for a purely lossless encoder. First, we outline the JBIG2 bitstream, focusing on the options and parameters available to an encoder. Then, we present numerous lossless encoder design strategies, including lossy to lossless coding approaches. For each strategy, we determine the compression performance, and the execution times for both encoding and decoding. The strategy that achieved the highest compression performance in our experiment used a double dictionary approach, with a residue cleanup. In this strategy, small and unique symbols were coded as a generic region residue. Only repeated symbols or those used as a basis for soft matches were added to a dictionary, with the remaining symbols embedded as refinements in the symbol region segment. The second dictionary was encoded as a refinement-aggregate dictionary, where dictionary symbols were encoded as refinements of symbols from the first dictionary, or previous entries in the second dictionary. With all other bitstream parameters optimized, this strategy can easily achieve an additional 30% compression over simpler symbol dictionary approaches. Next, we continue the experiment with an evaluation of each of the bitstream options and configuration parameters, and their impact on complexity and compression. We also demonstrate the consequences of choosing incorrect parameters. We conclude with a summary of our compression results, and general recommendations for encoder designers.
只提供摘要形式。联合双级专家组(JBIG)是一个隶属于ISO/IEC和ITU-T的国际研究小组,最近完成了有损和无损双级图像压缩JBIG2标准的委员会草案。我们研究了纯无损编码器的设计考虑。首先,我们概述了JBIG2比特流,重点介绍了编码器可用的选项和参数。然后,我们提出了许多无损编码器的设计策略,包括有损到无损的编码方法。对于每种策略,我们确定了压缩性能,以及编码和解码的执行时间。在我们的实验中,实现最高压缩性能的策略使用了双字典方法,并进行了残留清理。在该策略中,小而独特的符号被编码为通用区域残基。只将重复的符号或作为软匹配基础的符号添加到字典中,其余的符号作为改进嵌入符号区域段。第二个字典被编码为一个精化聚合字典,其中字典符号被编码为第一个字典中符号的精化,或者第二个字典中以前的条目。通过优化所有其他比特流参数,该策略可以轻松实现比更简单的符号字典方法额外30%的压缩。接下来,我们继续实验,评估每个比特流选项和配置参数,以及它们对复杂性和压缩的影响。我们还演示了选择不正确参数的后果。最后,我们总结了我们的压缩结果,并为编码器设计者提供了一般建议。
{"title":"Lossless JBIG2 coding performance","authors":"D. Tompkins, F. Kossentini","doi":"10.1109/DCC.1999.785710","DOIUrl":"https://doi.org/10.1109/DCC.1999.785710","url":null,"abstract":"Summary form only given. The Joint Bi-Level Expert Group (JBIG), an international study group affiliated with the ISO/IEC and ITU-T, has recently completed a committee draft of the JBIG2 standard for lossy and lossless bi-level image compression. We study design considerations for a purely lossless encoder. First, we outline the JBIG2 bitstream, focusing on the options and parameters available to an encoder. Then, we present numerous lossless encoder design strategies, including lossy to lossless coding approaches. For each strategy, we determine the compression performance, and the execution times for both encoding and decoding. The strategy that achieved the highest compression performance in our experiment used a double dictionary approach, with a residue cleanup. In this strategy, small and unique symbols were coded as a generic region residue. Only repeated symbols or those used as a basis for soft matches were added to a dictionary, with the remaining symbols embedded as refinements in the symbol region segment. The second dictionary was encoded as a refinement-aggregate dictionary, where dictionary symbols were encoded as refinements of symbols from the first dictionary, or previous entries in the second dictionary. With all other bitstream parameters optimized, this strategy can easily achieve an additional 30% compression over simpler symbol dictionary approaches. Next, we continue the experiment with an evaluation of each of the bitstream options and configuration parameters, and their impact on complexity and compression. We also demonstrate the consequences of choosing incorrect parameters. We conclude with a summary of our compression results, and general recommendations for encoder designers.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"70 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123037559","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Progressive joint source-channel coding in feedback channels 反馈信道中渐进联合源信道编码
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.755663
Jin Lu, Aria Nosratinia, B. Aazhang
It is well known that Shannon's separation result does not hold under finite computation or finite delay constraints, thus joint source-channel coding is of great interest for practical reasons. For progressive source-channel coding systems, efficient codes have been proposed for feedforward channels and the important problem of rate allocation between the source and channel codes has been solved. For memoryless channels with feedback, the rate allocation problem was studied by Chande et al. (1998). In this paper, we consider the case of fading channels with feedback. Feedback routes are provided in many existing standard wireless channels, making rate allocation with feedback a problem of considerable practical importance. We address the question of rate allocation between the source and channel codes in the forward channel, in the presence of feedback information and under a distortion cost function. We show that the presence of feedback shifts the optimal rate allocation point, resulting in higher rates for error-correcting codes and smaller overall distortion. Simulations on both memoryless and fading channels show that the presence of feedback allows up to 1 dB improvement in PSNR compared to the similarly optimized feedforward scheme.
众所周知,在有限计算或有限延迟约束下,香农分离结果不成立,因此联合信源信道编码在实际应用中具有重要意义。对于渐进式信源信道编码系统,提出了前馈信道的有效码,解决了信源码和信道码之间速率分配的重要问题。对于有反馈的无记忆信道,Chande等(1998)研究了速率分配问题。本文考虑了带反馈的衰落信道的情况。在现有的许多标准无线信道中都提供了反馈路由,使得带反馈的速率分配问题具有相当重要的实际意义。我们解决了前向信道中存在反馈信息和失真代价函数的源码和信道码之间的速率分配问题。我们表明,反馈的存在移动了最佳速率分配点,导致更高的纠错码速率和更小的整体失真。在无记忆信道和衰落信道上的仿真表明,与类似优化的前馈方案相比,反馈的存在使PSNR提高了1db。
{"title":"Progressive joint source-channel coding in feedback channels","authors":"Jin Lu, Aria Nosratinia, B. Aazhang","doi":"10.1109/DCC.1999.755663","DOIUrl":"https://doi.org/10.1109/DCC.1999.755663","url":null,"abstract":"It is well known that Shannon's separation result does not hold under finite computation or finite delay constraints, thus joint source-channel coding is of great interest for practical reasons. For progressive source-channel coding systems, efficient codes have been proposed for feedforward channels and the important problem of rate allocation between the source and channel codes has been solved. For memoryless channels with feedback, the rate allocation problem was studied by Chande et al. (1998). In this paper, we consider the case of fading channels with feedback. Feedback routes are provided in many existing standard wireless channels, making rate allocation with feedback a problem of considerable practical importance. We address the question of rate allocation between the source and channel codes in the forward channel, in the presence of feedback information and under a distortion cost function. We show that the presence of feedback shifts the optimal rate allocation point, resulting in higher rates for error-correcting codes and smaller overall distortion. Simulations on both memoryless and fading channels show that the presence of feedback allows up to 1 dB improvement in PSNR compared to the similarly optimized feedforward scheme.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"159-160 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122126653","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 16
Towards a calibrated corpus for compression testing 迈向压缩测试的校准语料库
Pub Date : 1999-03-29 DOI: 10.1109/DCC.1999.785711
M. Titchener, P. Fenwick, M. C. Chen
Summary form only given. A mini-corpus of twelve 'calibrated' binary-data files have been produced for systematic evaluation of compression algorithms. These are generated within the framework of a deterministic theory of string complexity. Here the T-complexity of a string x (measured in taugs) is defined as C/sub T/(x/sub i/)=/spl Sigma//sub i/log/sub 2/(k/sub i/+1), where the positive integers k/sub i/ are the T-expansion parameters for the corresponding string production process. C/sub T/(x) is observed to be the logarithmic integral of the total information content I/sub x/ of x (measured in nats), i.e., C/sub T/(x)=li(I/sub x/). The average entropy is H~/sub x/=I/sub x//|x|, i.e., the total information content divided by the length of x. Thus C/sub T/(x)=li(H~/sub x//spl times/|x|). Alternatively, the information rate along a string may be described by an entropy function H/sub x/(n),0/spl les/n/spl les/|x| for the string. Assuming that H/sub x/(n) is continuously integrable along the length of the x, then I/sub x/=/spl int//sub 0//sup |/x|H/sub x/(n)/spl delta/n. Thus C/sub T/(x)=li(/spl int//sub 0//sup |/x|H/sub x/(n)/spl delta/n). Solving for H/sub x/(n): that is differentiating both sides and rearranging, we get: H/sub x/(n)=(/spl delta/C/sub T/(x|n)//spl delta/n)/spl times/log/sub e/(li/sup -1/(C/sub T/(x|/sub n/))). With x being in fact discrete, and the T-complexity function being computed in terms of the discrete T-augmentation steps, we may accordingly re-express the equation in terms of the T-prefix increments: /spl delta/n/spl ap//spl Delta//sub i/|x|=k/sub i/|p/sub i/|; and from the definition of C/sub T/(x): /spl delta/C/sub T/(x) is replaced by /spl Delta//sub i/C/sub T/(x)=log/sub 2/(k/sub i/+1). The average slope over the i-th T-prefix p/sub i/ increment is then simply (/spl Delta//sub i/C/sub T/(x))/(/spl Delta//sub i/|x|)=(log/sub 2/(k/sub i/+1))/(k/sub i/|p/sub i/|). The entropy function is now replaced by a discrete approximation.
只提供摘要形式。12个“校准”二进制数据文件的迷你语料库已经产生了压缩算法的系统评估。这些都是在弦复杂性的确定性理论框架内生成的。这里,字符串x的T-复杂度(以标签为单位)定义为C/下标T/(x/下标i/)=/spl Sigma//下标i/log/下标2/(k/下标i/+1),其中正整数k/下标i/是对应的字符串生产过程的T-展开参数。观察到C/ T/(x)是x(以纳特为单位)的总信息量I/下标x/的对数积分,即C/ T/(x)=li(I/下标x/)。平均熵为H~/sub x/=I/sub x//|x|,即总信息量除以x的长度,因此C/sub T/(x)=li(H~/sub x//spl乘以/|x|)。或者,沿着字符串的信息速率可以用熵函数H/sub x/(n)来描述,对于字符串,0/spl les/n/spl les/|x|。假设H/下标x/(n)沿x的长度连续可积,则I/下标x/=/spl int//下标0//sup |/x|H/下标x/(n)/spl /n。因此C / sub T /李(x) = (spl int / / sub x | 0 | / /晚餐/ H / sub x / (n) / splδ/ n)。求解H/下标x/(n)也就是两边求导并重新排列,我们得到H/下标x/(n)=(/spl /C/ T/(x|n)//spl /(n) /spl乘以/log/ e/(li/sup -1/(C/下标T/(x|/下标n/)))由于x实际上是离散的,并且t -复杂度函数是用离散的t增积步骤来计算的,因此我们可以用t前缀增量来重新表示方程:/spl delta/n/spl ap//spl delta/ /下标i/|x|=k/下标i/|p/下标i/|;由C/ T/(x)的定义:/spl /C/ T/(x)被/spl //下标i/C/下标T/(x)=log/下标2/(k/下标i/+1)所取代。第i个T前缀p/下标i/增量的平均斜率为(/spl Delta//下标i/C/下标T/(x))/(/spl Delta//下标i/|x|)=(log/下标2/(k/下标i/+1) /(k/下标i/|p/下标i/|))。熵函数现在被一个离散的近似代替了。
{"title":"Towards a calibrated corpus for compression testing","authors":"M. Titchener, P. Fenwick, M. C. Chen","doi":"10.1109/DCC.1999.785711","DOIUrl":"https://doi.org/10.1109/DCC.1999.785711","url":null,"abstract":"Summary form only given. A mini-corpus of twelve 'calibrated' binary-data files have been produced for systematic evaluation of compression algorithms. These are generated within the framework of a deterministic theory of string complexity. Here the T-complexity of a string x (measured in taugs) is defined as C/sub T/(x/sub i/)=/spl Sigma//sub i/log/sub 2/(k/sub i/+1), where the positive integers k/sub i/ are the T-expansion parameters for the corresponding string production process. C/sub T/(x) is observed to be the logarithmic integral of the total information content I/sub x/ of x (measured in nats), i.e., C/sub T/(x)=li(I/sub x/). The average entropy is H~/sub x/=I/sub x//|x|, i.e., the total information content divided by the length of x. Thus C/sub T/(x)=li(H~/sub x//spl times/|x|). Alternatively, the information rate along a string may be described by an entropy function H/sub x/(n),0/spl les/n/spl les/|x| for the string. Assuming that H/sub x/(n) is continuously integrable along the length of the x, then I/sub x/=/spl int//sub 0//sup |/x|H/sub x/(n)/spl delta/n. Thus C/sub T/(x)=li(/spl int//sub 0//sup |/x|H/sub x/(n)/spl delta/n). Solving for H/sub x/(n): that is differentiating both sides and rearranging, we get: H/sub x/(n)=(/spl delta/C/sub T/(x|n)//spl delta/n)/spl times/log/sub e/(li/sup -1/(C/sub T/(x|/sub n/))). With x being in fact discrete, and the T-complexity function being computed in terms of the discrete T-augmentation steps, we may accordingly re-express the equation in terms of the T-prefix increments: /spl delta/n/spl ap//spl Delta//sub i/|x|=k/sub i/|p/sub i/|; and from the definition of C/sub T/(x): /spl delta/C/sub T/(x) is replaced by /spl Delta//sub i/C/sub T/(x)=log/sub 2/(k/sub i/+1). The average slope over the i-th T-prefix p/sub i/ increment is then simply (/spl Delta//sub i/C/sub T/(x))/(/spl Delta//sub i/|x|)=(log/sub 2/(k/sub i/+1))/(k/sub i/|p/sub i/|). The entropy function is now replaced by a discrete approximation.","PeriodicalId":103598,"journal":{"name":"Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"1999-03-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125258810","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
期刊
Proceedings DCC'99 Data Compression Conference (Cat. No. PR00096)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1