{"title":"Worst-case Analysis of the Low-complexity Symbol Grouping Coding Technique","authors":"A. Said","doi":"10.1109/ISIT.2006.262028","DOIUrl":null,"url":null,"abstract":"The symbol grouping technique is widely used in practice because it allows great reductions on the complexity of entropy coding symbols from large alphabets, at the expense of small losses in compression. While it has been used mostly in an ad hoc manner, it is not known how general this technique is, i.e., in exactly what type of data sources it can be effective. We try to answer this question by searching for worst-case data sources, measuring the performance, and trying to identify trends. We show that finding the worst-case source is a very challenging optimization problem, and propose some solution methods that can be used in alphabets of moderate size. The numerical results provide evidence confirming the hypotheses that all data sources with large number of symbols can be more efficiently coded, with very small loss, using symbol grouping","PeriodicalId":115298,"journal":{"name":"2006 IEEE International Symposium on Information Theory","volume":"62 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2006-07-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2006 IEEE International Symposium on Information Theory","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISIT.2006.262028","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 1
Abstract
The symbol grouping technique is widely used in practice because it allows great reductions on the complexity of entropy coding symbols from large alphabets, at the expense of small losses in compression. While it has been used mostly in an ad hoc manner, it is not known how general this technique is, i.e., in exactly what type of data sources it can be effective. We try to answer this question by searching for worst-case data sources, measuring the performance, and trying to identify trends. We show that finding the worst-case source is a very challenging optimization problem, and propose some solution methods that can be used in alphabets of moderate size. The numerical results provide evidence confirming the hypotheses that all data sources with large number of symbols can be more efficiently coded, with very small loss, using symbol grouping