Compressed incremental checkpointing for efficient replicated key-value stores

Berkin Guler, Öznur Özkasap
{"title":"Compressed incremental checkpointing for efficient replicated key-value stores","authors":"Berkin Guler, Öznur Özkasap","doi":"10.1109/ISCC.2017.8024508","DOIUrl":null,"url":null,"abstract":"The prominent cloud services rely on geographically distributed nodes running replication and other fault-tolerance mechanisms so as to provide flawless availability and dependability. In this paper, we address the communication cost of the well known primary-backup replication protocol, and propose compressed periodic incremental checkpoint algorithms to achieve improved throughput. We set up a replicated key-value store on geographically distributed nodes of the PlanetLab platform, and developed compressed incremental checkpointing algorithms to support primary-backup replication. By considering performance metrics of interest including blocking time, checkpointing time, compression ratio, compression/ decompression times, we conducted a comprehensive analysis. We used the well-known benchmarking tool YCSB and established different sample workloads to test where each workload represents diverse plots. Our findings indicate that Zstd is the most competent compression method under all scenarios and through comparing with an uncompressed approach we point out that compressing the communication data disseminated from the primary replica coupled with the periodic incremental checkpointing algorithm not only decreases the average blocking time up to 5% but it also improves the overall system throughput by 4% compared to the no compression case.","PeriodicalId":106141,"journal":{"name":"2017 IEEE Symposium on Computers and Communications (ISCC)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2017-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"4","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2017 IEEE Symposium on Computers and Communications (ISCC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISCC.2017.8024508","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 4

Abstract

The prominent cloud services rely on geographically distributed nodes running replication and other fault-tolerance mechanisms so as to provide flawless availability and dependability. In this paper, we address the communication cost of the well known primary-backup replication protocol, and propose compressed periodic incremental checkpoint algorithms to achieve improved throughput. We set up a replicated key-value store on geographically distributed nodes of the PlanetLab platform, and developed compressed incremental checkpointing algorithms to support primary-backup replication. By considering performance metrics of interest including blocking time, checkpointing time, compression ratio, compression/ decompression times, we conducted a comprehensive analysis. We used the well-known benchmarking tool YCSB and established different sample workloads to test where each workload represents diverse plots. Our findings indicate that Zstd is the most competent compression method under all scenarios and through comparing with an uncompressed approach we point out that compressing the communication data disseminated from the primary replica coupled with the periodic incremental checkpointing algorithm not only decreases the average blocking time up to 5% but it also improves the overall system throughput by 4% compared to the no compression case.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
压缩增量检查点,用于高效复制键值存储
杰出的云服务依赖于地理上分布的节点来运行复制和其他容错机制,从而提供完美的可用性和可靠性。在本文中,我们解决了众所周知的主备份复制协议的通信成本,并提出了压缩周期性增量检查点算法来提高吞吐量。我们在PlanetLab平台的地理分布节点上建立了一个复制的键值存储,并开发了压缩增量检查点算法来支持主备份复制。通过考虑感兴趣的性能指标,包括阻塞时间、检查点时间、压缩比、压缩/解压时间,我们进行了全面的分析。我们使用了著名的基准测试工具YCSB,并建立了不同的样本工作负载进行测试,其中每个工作负载代表不同的图。我们的研究结果表明,Zstd是所有场景下最有效的压缩方法,通过与未压缩方法的比较,我们指出压缩从主副本传播的通信数据并结合周期性增量检查点算法不仅减少了平均阻塞时间高达5%,而且与未压缩情况相比,它还提高了4%的整体系统吞吐量。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Uncertainty-driven ensemble forecasting of QoS in Software Defined Networks A Reactive Security Framework for operational wind parks using Service Function Chaining The use of predictive models in dynamic treatment planning Link quality and path based clustering in IEEE 802.15.4-2015 TSCH networks Using repeated game for maximizing high priority data trustworthiness in Wireless Sensor Networks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1