云存储的重复数据删除技术综述

Ayad Hasan Adhab, Naseer Ali Hussien
{"title":"云存储的重复数据删除技术综述","authors":"Ayad Hasan Adhab, Naseer Ali Hussien","doi":"10.31695/ijerat.2022.8.4.2","DOIUrl":null,"url":null,"abstract":"With the rapid advancement of information technology and network, it's becoming increasingly difficult to keep up, as well as the rapid expansion of data center size, energy consumption as a percentage of IT investment is increasing. As the amount of digital data grows, so does the need for greater storage space, which drives up the cost and performance of backups. Traditional backup solutions don't have any built-in protection against duplicate data being saved up. Duplicate data backups severely lengthen backup times and consume needless resources. Data deduplication is critical for removing redundant data and lowering storage costs. Data deduplication is a new technique of compressing data that helps with storage efficiency while also proving to be a more efficient technique of dealing with duplicate data. Deduplication enables a single data copy to be uploaded to storage and subsequent copies to be provided with a pointer to the original stored copy. This paper consists of extensive literature survey and summarizes numerous storage approaches, concepts, and categories that are used in data reduplication. Also in this paper, the researchers carried out the survey for chunk based data deduplication techniques in detail.","PeriodicalId":424923,"journal":{"name":"International Journal of Engineering Research and Advanced Technology","volume":"35 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Techniques of Data Deduplication for Cloud Storage: A Review\",\"authors\":\"Ayad Hasan Adhab, Naseer Ali Hussien\",\"doi\":\"10.31695/ijerat.2022.8.4.2\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"With the rapid advancement of information technology and network, it's becoming increasingly difficult to keep up, as well as the rapid expansion of data center size, energy consumption as a percentage of IT investment is increasing. As the amount of digital data grows, so does the need for greater storage space, which drives up the cost and performance of backups. Traditional backup solutions don't have any built-in protection against duplicate data being saved up. Duplicate data backups severely lengthen backup times and consume needless resources. Data deduplication is critical for removing redundant data and lowering storage costs. Data deduplication is a new technique of compressing data that helps with storage efficiency while also proving to be a more efficient technique of dealing with duplicate data. Deduplication enables a single data copy to be uploaded to storage and subsequent copies to be provided with a pointer to the original stored copy. This paper consists of extensive literature survey and summarizes numerous storage approaches, concepts, and categories that are used in data reduplication. Also in this paper, the researchers carried out the survey for chunk based data deduplication techniques in detail.\",\"PeriodicalId\":424923,\"journal\":{\"name\":\"International Journal of Engineering Research and Advanced Technology\",\"volume\":\"35 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"International Journal of Engineering Research and Advanced Technology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.31695/ijerat.2022.8.4.2\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"International Journal of Engineering Research and Advanced Technology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.31695/ijerat.2022.8.4.2","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

随着信息技术和网络的飞速发展,数据中心的规模也在迅速扩大,能源消耗在it投资中所占的比例也在不断增加。随着数字数据量的增长,对更大存储空间的需求也在增长,这就推高了备份的成本和性能。传统的备份解决方案没有任何内置的保护措施来防止重复数据的保存。重复备份会严重延长备份时间,消耗不必要的资源。重复数据删除对于消除冗余数据和降低存储成本至关重要。重复数据删除是一种压缩数据的新技术,它有助于提高存储效率,同时也是处理重复数据的一种更有效的技术。重复数据删除可以将单个数据副本上传到存储中,并为后续副本提供指向原始存储副本的指针。本文包括广泛的文献综述,并总结了用于数据重复的许多存储方法、概念和类别。本文还对基于块的重复数据删除技术进行了详细的研究。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Techniques of Data Deduplication for Cloud Storage: A Review
With the rapid advancement of information technology and network, it's becoming increasingly difficult to keep up, as well as the rapid expansion of data center size, energy consumption as a percentage of IT investment is increasing. As the amount of digital data grows, so does the need for greater storage space, which drives up the cost and performance of backups. Traditional backup solutions don't have any built-in protection against duplicate data being saved up. Duplicate data backups severely lengthen backup times and consume needless resources. Data deduplication is critical for removing redundant data and lowering storage costs. Data deduplication is a new technique of compressing data that helps with storage efficiency while also proving to be a more efficient technique of dealing with duplicate data. Deduplication enables a single data copy to be uploaded to storage and subsequent copies to be provided with a pointer to the original stored copy. This paper consists of extensive literature survey and summarizes numerous storage approaches, concepts, and categories that are used in data reduplication. Also in this paper, the researchers carried out the survey for chunk based data deduplication techniques in detail.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Object Boundaries Detection Based Center Of Gravity Optimization of Interval Type-2 Fuzzy Logic System for Software Reliability Prediction Analysis of Quality Control for Automatic Transmission Products MCVT Model By Six Sigma Approach at PT HPPM, Indonesia A New wireless sensor networks Routing Algorithm Based on SPIN Protocols and Circumference Technique Optimization of Production process in Sewing Work for Ethiopia Cultural Clothe Apparel Production
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1