Arithmetic N-gram: an efficient data compression technique

IF 1.7 3区 计算机科学 Q3 COMPUTER SCIENCE, INFORMATION SYSTEMS Information Retrieval Journal Pub Date : 2024-03-13 DOI:10.1007/s10791-024-09431-y
Ali Hassan, Sadaf Javed, Sajjad Hussain, Rizwan Ahmad, Shams Qazi
{"title":"Arithmetic N-gram: an efficient data compression technique","authors":"Ali Hassan, Sadaf Javed, Sajjad Hussain, Rizwan Ahmad, Shams Qazi","doi":"10.1007/s10791-024-09431-y","DOIUrl":null,"url":null,"abstract":"<p>Due to the increase in the growth of data in this era of the digital world and limited resources, there is a need for more efficient data compression techniques for storing and transmitting data. Data compression can significantly reduce the amount of storage space and transmission time to store and transmit given data. More specifically, text compression has got more attention for effectively managing and processing data due to the increased use of the internet, digital devices, data transfer, etc. Over the years, various algorithms have been used for text compression such as Huffman coding, Lempel-Ziv-Welch (LZW) coding, arithmetic coding, etc. However, these methods have a limited compression ratio specifically for data storage applications where a considerable amount of data must be compressed to use storage resources efficiently. They consider individual characters to compress data. It can be more advantageous to consider words or sequences of words rather than individual characters to get a better compression ratio. Compressing individual characters results in a sizeable compressed representation due to their less repetition and structure in the data. In this paper, we proposed the ArthNgram model, in which the N-gram language model coupled with arithmetic coding is used to compress data more efficiently for data storage applications. The performance of the proposed model is evaluated based on compression ratio and compression speed. Results show that the proposed model performs better than traditional techniques.</p>","PeriodicalId":54352,"journal":{"name":"Information Retrieval Journal","volume":"126 1","pages":""},"PeriodicalIF":1.7000,"publicationDate":"2024-03-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Retrieval Journal","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1007/s10791-024-09431-y","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"COMPUTER SCIENCE, INFORMATION SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

Due to the increase in the growth of data in this era of the digital world and limited resources, there is a need for more efficient data compression techniques for storing and transmitting data. Data compression can significantly reduce the amount of storage space and transmission time to store and transmit given data. More specifically, text compression has got more attention for effectively managing and processing data due to the increased use of the internet, digital devices, data transfer, etc. Over the years, various algorithms have been used for text compression such as Huffman coding, Lempel-Ziv-Welch (LZW) coding, arithmetic coding, etc. However, these methods have a limited compression ratio specifically for data storage applications where a considerable amount of data must be compressed to use storage resources efficiently. They consider individual characters to compress data. It can be more advantageous to consider words or sequences of words rather than individual characters to get a better compression ratio. Compressing individual characters results in a sizeable compressed representation due to their less repetition and structure in the data. In this paper, we proposed the ArthNgram model, in which the N-gram language model coupled with arithmetic coding is used to compress data more efficiently for data storage applications. The performance of the proposed model is evaluated based on compression ratio and compression speed. Results show that the proposed model performs better than traditional techniques.

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
算术 N 图:一种高效的数据压缩技术
在这个数字世界的时代,由于数据的增长和资源的有限,人们需要更有效的数据压缩技术来存储和传输数据。数据压缩可以大大减少存储和传输给定数据的存储空间和传输时间。更具体地说,由于互联网、数字设备和数据传输等的使用越来越多,文本压缩在有效管理和处理数据方面得到了更多关注。多年来,人们使用了各种算法来进行文本压缩,如 Huffman 编码、Lempel-Ziv-Welch(LZW)编码、算术编码等。然而,这些方法的压缩率有限,特别是在数据存储应用中,必须压缩大量数据才能有效利用存储资源。它们只考虑单个字符来压缩数据。如果考虑单词或单词序列,而不是单个字符,可能会获得更好的压缩率。由于单个字符在数据中的重复性和结构性较低,因此压缩单个字符可获得较大的压缩表示。在本文中,我们提出了 ArthNgram 模型,在该模型中,N-gram 语言模型与算术编码相结合,可更有效地压缩数据,用于数据存储应用。我们根据压缩率和压缩速度对所提模型的性能进行了评估。结果表明,所提模型的性能优于传统技术。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Information Retrieval Journal
Information Retrieval Journal 工程技术-计算机:信息系统
CiteScore
6.20
自引率
0.00%
发文量
17
审稿时长
13.5 months
期刊介绍: The journal provides an international forum for the publication of theory, algorithms, analysis and experiments across the broad area of information retrieval. Topics of interest include search, indexing, analysis, and evaluation for applications such as the web, social and streaming media, recommender systems, and text archives. This includes research on human factors in search, bridging artificial intelligence and information retrieval, and domain-specific search applications.
期刊最新文献
Searching rooms with top-k passenger flows using indoor trajectories An innovative approach for PCO morphology segmentation using a novel MOT-SF technique A graph residual generation network for node classification based on multi-information aggregation Similarity-based ranking of videos from fixed-size one-dimensional video signature The accessibility of digital technologies for people with visual impairment and blindness: a scoping review
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1