首页 > 最新文献

ACM Transactions on Storage (TOS)最新文献

英文 中文
HIL
Pub Date : 2018-12-04 DOI: 10.1145/3281030
Jin-yong Choi, E. Nam, Yoon-Jae Seong, Jinhyuk Yoon, Sookwan Lee, Hongseok Kim, Jeongsu Park, Yeong-Jae Woo, Sheayun Lee, S. Min
We present a framework called Hierarchically Interacting Logs (HIL) for constructing Flash Translation Layers (FTLs). The main goal of the HIL framework is to heal the Achilles heel —the crash recovery—of FTLs (hence, its name). Nonetheless, the framework itself is general enough to encompass not only block-mapped and page-mapped FTLs but also many of their variants, including hybrid ones, because of its compositional nature. Crash recovery within the HIL framework proceeds in two phases: structural recovery and functional recovery. During the structural recovery, residual effects due to program operations ongoing at the time of the crash are eliminated in an atomic manner using shadow paging. During the functional recovery, operations that would have been performed if there had been no crash are replayed in a redo-only fashion. Both phases operate in an idempotent manner, preventing repeated crashes during recovery from causing any additional problems. We demonstrate the practicality of the proposed HIL framework by implementing a prototype and showing that its performance during normal execution and also during crash recovery is at least as good as those of state-of-the-art SSDs.
我们提出了一个名为层次交互日志(HIL)的框架,用于构建Flash翻译层(ftl)。HIL框架的主要目标是治愈超光速飞行的致命弱点——崩溃恢复(因此得名)。尽管如此,这个框架本身是足够通用的,不仅可以包含块映射和页面映射的ftl,还可以包含它们的许多变体,包括混合的,因为它的组成性质。HIL框架内的崩溃恢复分两个阶段进行:结构恢复和功能恢复。在结构恢复期间,由于在崩溃时正在进行的程序操作而产生的残余影响将使用影子分页以原子的方式消除。在功能恢复期间,将以仅重做的方式重播在没有崩溃的情况下执行的操作。这两个阶段都以幂等的方式运行,防止在恢复过程中重复崩溃导致任何其他问题。我们通过实现一个原型来证明所提出的HIL框架的实用性,并表明其在正常执行和崩溃恢复期间的性能至少与最先进的ssd一样好。
{"title":"HIL","authors":"Jin-yong Choi, E. Nam, Yoon-Jae Seong, Jinhyuk Yoon, Sookwan Lee, Hongseok Kim, Jeongsu Park, Yeong-Jae Woo, Sheayun Lee, S. Min","doi":"10.1145/3281030","DOIUrl":"https://doi.org/10.1145/3281030","url":null,"abstract":"We present a framework called Hierarchically Interacting Logs (HIL) for constructing Flash Translation Layers (FTLs). The main goal of the HIL framework is to heal the Achilles heel —the crash recovery—of FTLs (hence, its name). Nonetheless, the framework itself is general enough to encompass not only block-mapped and page-mapped FTLs but also many of their variants, including hybrid ones, because of its compositional nature. Crash recovery within the HIL framework proceeds in two phases: structural recovery and functional recovery. During the structural recovery, residual effects due to program operations ongoing at the time of the crash are eliminated in an atomic manner using shadow paging. During the functional recovery, operations that would have been performed if there had been no crash are replayed in a redo-only fashion. Both phases operate in an idempotent manner, preventing repeated crashes during recovery from causing any additional problems. We demonstrate the practicality of the proposed HIL framework by implementing a prototype and showing that its performance during normal execution and also during crash recovery is at least as good as those of state-of-the-art SSDs.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128584964","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards Robust File System Checkers 迈向健壮的文件系统检查器
Pub Date : 2018-12-04 DOI: 10.1145/3281031
Om Rameshwar Gatla, Muhammad Hameed, Mai Zheng, Viacheslav Dubeyko, A. Manzanares, F. Blagojevic, Cyril Guyot, R. Mateescu
File systems may become corrupted for many reasons despite various protection techniques. Therefore, most file systems come with a checker to recover the file system to a consistent state. However, existing checkers are commonly assumed to be able to complete the repair without interruption, which may not be true in practice. In this work, we demonstrate via fault injection experiments that checkers of widely used file systems (EXT4, XFS, BtrFS, and F2FS) may leave the file system in an uncorrectable state if the repair procedure is interrupted unexpectedly. To address the problem, we first fix the ordering issue in the undo logging of e2fsck and then build a general logging library (i.e., rfsck-lib) for strengthening checkers. To demonstrate the practicality, we integrate rfsck-lib with existing checkers and create two new checkers: rfsck-ext, a robust checker for Ext-family file systems, and rfsck-xfs, a robust checker for XFS file systems, both of which require only tens of lines of modification to the original versions. Both rfsck-ext and rfsck-xfs are resilient to faults in our experiments. Also, both checkers incur reasonable performance overhead (i.e., up to 12%) compared to the original unreliable versions. Moreover, rfsck-ext outperforms the patched e2fsck by up to nine times while achieving the same level of robustness.
尽管采用了各种保护技术,但文件系统可能由于多种原因而损坏。因此,大多数文件系统都带有检查器来将文件系统恢复到一致状态。然而,现有的检查器通常被认为能够不间断地完成修复,这在实践中可能不是真的。在这项工作中,我们通过故障注入实验证明,如果修复过程意外中断,广泛使用的文件系统(EXT4、XFS、BtrFS和F2FS)的检查器可能会使文件系统处于不可纠正的状态。为了解决这个问题,我们首先修复e2fsck的undo日志中的顺序问题,然后构建一个通用的日志库(即rfsck-lib)来加强检查器。为了演示其实用性,我们将rfsck-lib与现有的检查器集成,并创建了两个新的检查器:rfsck-ext (ext系列文件系统的健壮检查器)和rfsck-xfs (XFS文件系统的健壮检查器),这两个检查器都只需要对原始版本进行数十行修改。在我们的实验中,rfsck-ext和rfsck-xfs都对故障具有弹性。此外,与原始的不可靠版本相比,两个检查器都会产生合理的性能开销(即高达12%)。此外,rfsck-ext的性能比打过补丁的e2fsck高出9倍,同时达到了相同的健壮性水平。
{"title":"Towards Robust File System Checkers","authors":"Om Rameshwar Gatla, Muhammad Hameed, Mai Zheng, Viacheslav Dubeyko, A. Manzanares, F. Blagojevic, Cyril Guyot, R. Mateescu","doi":"10.1145/3281031","DOIUrl":"https://doi.org/10.1145/3281031","url":null,"abstract":"File systems may become corrupted for many reasons despite various protection techniques. Therefore, most file systems come with a checker to recover the file system to a consistent state. However, existing checkers are commonly assumed to be able to complete the repair without interruption, which may not be true in practice. In this work, we demonstrate via fault injection experiments that checkers of widely used file systems (EXT4, XFS, BtrFS, and F2FS) may leave the file system in an uncorrectable state if the repair procedure is interrupted unexpectedly. To address the problem, we first fix the ordering issue in the undo logging of e2fsck and then build a general logging library (i.e., rfsck-lib) for strengthening checkers. To demonstrate the practicality, we integrate rfsck-lib with existing checkers and create two new checkers: rfsck-ext, a robust checker for Ext-family file systems, and rfsck-xfs, a robust checker for XFS file systems, both of which require only tens of lines of modification to the original versions. Both rfsck-ext and rfsck-xfs are resilient to faults in our experiments. Also, both checkers incur reasonable performance overhead (i.e., up to 12%) compared to the original unreliable versions. Moreover, rfsck-ext outperforms the patched e2fsck by up to nine times while achieving the same level of robustness.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"530 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124148803","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Management of Next-Generation NAND Flash to Achieve Enterprise-Level Endurance and Latency Targets 下一代NAND闪存的管理,以实现企业级的持久和延迟目标
Pub Date : 2018-12-04 DOI: 10.1145/3241060
R. Pletka, Ioannis Koltsidas, Nikolas Ioannou, Sasa Tomic, N. Papandreou, Thomas Parnell, H. Pozidis, Aaron Fry, T. Fisher
Despite its widespread use in consumer devices and enterprise storage systems, NAND flash faces a growing number of challenges. While technology advances have helped to increase the storage density and reduce costs, they have also led to reduced endurance and larger block variations, which cannot be compensated solely by stronger ECC or read-retry schemes but have to be addressed holistically. Our goal is to enable low-cost NAND flash in enterprise storage for cost efficiency. We present novel flash-management approaches that reduce write amplification, achieve better wear leveling, and enhance endurance without sacrificing performance. We introduce block calibration, a technique to determine optimal read-threshold voltage levels that minimize error rates, and novel garbage-collection as well as data-placement schemes that alleviate the effects of block health variability and show how these techniques complement one another and thereby achieve enterprise storage requirements. By combining the proposed schemes, we improve endurance by up to 15× compared to the baseline endurance of NAND flash without using a stronger ECC scheme. The flash-management algorithms presented herein were designed and implemented in simulators, hardware test platforms, and eventually in the flash controllers of production enterprise all-flash arrays. Their effectiveness has been validated across thousands of customer deployments since 2015.
尽管NAND闪存在消费设备和企业存储系统中得到了广泛的应用,但它面临着越来越多的挑战。虽然技术进步有助于提高存储密度和降低成本,但它们也导致了耐久性降低和更大的块变化,这不能仅仅通过更强的ECC或读重试方案来补偿,而是必须全面解决。我们的目标是在企业存储中实现低成本的NAND闪存,以提高成本效率。我们提出了新颖的闪存管理方法,可以减少写入放大,实现更好的磨损均衡,并在不牺牲性能的情况下提高耐用性。我们介绍了块校准,这是一种确定最佳读阈值电压水平的技术,可将错误率降至最低,还介绍了新的垃圾收集和数据放置方案,这些方案可减轻块健康变异性的影响,并展示了这些技术如何相互补充,从而实现企业存储需求。通过结合所提出的方案,我们在不使用更强的ECC方案的情况下,将NAND闪存的基线续航时间提高了15倍。本文提出的闪存管理算法在仿真器、硬件测试平台上进行了设计和实现,并最终应用于生产企业全闪存阵列的闪存控制器中。自2015年以来,其有效性已在数千个客户部署中得到验证。
{"title":"Management of Next-Generation NAND Flash to Achieve Enterprise-Level Endurance and Latency Targets","authors":"R. Pletka, Ioannis Koltsidas, Nikolas Ioannou, Sasa Tomic, N. Papandreou, Thomas Parnell, H. Pozidis, Aaron Fry, T. Fisher","doi":"10.1145/3241060","DOIUrl":"https://doi.org/10.1145/3241060","url":null,"abstract":"Despite its widespread use in consumer devices and enterprise storage systems, NAND flash faces a growing number of challenges. While technology advances have helped to increase the storage density and reduce costs, they have also led to reduced endurance and larger block variations, which cannot be compensated solely by stronger ECC or read-retry schemes but have to be addressed holistically. Our goal is to enable low-cost NAND flash in enterprise storage for cost efficiency. We present novel flash-management approaches that reduce write amplification, achieve better wear leveling, and enhance endurance without sacrificing performance. We introduce block calibration, a technique to determine optimal read-threshold voltage levels that minimize error rates, and novel garbage-collection as well as data-placement schemes that alleviate the effects of block health variability and show how these techniques complement one another and thereby achieve enterprise storage requirements. By combining the proposed schemes, we improve endurance by up to 15× compared to the baseline endurance of NAND flash without using a stronger ECC scheme. The flash-management algorithms presented herein were designed and implemented in simulators, hardware test platforms, and eventually in the flash controllers of production enterprise all-flash arrays. Their effectiveness has been validated across thousands of customer deployments since 2015.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"201 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115700192","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
LibPM
Pub Date : 2018-12-04 DOI: 10.1145/3278141
L. Mármol, M. Chowdhury, R. Rangaswami
Persistent Memory devices present properties that are uniquely different from prior technologies for which applications have been built. Unfortunately, the conventional approach to building applications fail to either efficiently utilize these new devices or provide programmers a seamless development experience. We have built LibPM, a Persistent Memory Library that implements an easy-to-use container abstraction for consuming PM. LibPM’s containers are data hosting units that can store arbitrarily complex data types while preserving their integrity and consistency. Consequently, LibPM’s containers provide a generic interface to applications, allowing applications to store and manipulate arbitrarily structured data with strong durability and consistency properties, all without having to navigate all the myriad pitfalls of programming PM directly. By providing a simple and high-performing transactional update mechanism, LibPM allows applications to manipulate persistent data at the speed of memory. The container abstraction and automatic persistent data discovery mechanisms within LibPM also simplify porting legacy applications to PM. From a performance perspective, LibPM closely matches and often exceeds the performance of state-of-the-art application libraries for PM. For instance, LibPM ’s performance is 195× better for write intensive workloads and 2.6× better for read intensive workloads when compared with the state-of-the-art Pmem.IO persistent memory library.
持久内存设备提供的属性与构建应用程序所用的先前技术唯一不同。不幸的是,构建应用程序的传统方法既不能有效地利用这些新设备,也不能为程序员提供无缝的开发体验。我们已经构建了LibPM,一个持久内存库,它实现了一个易于使用的容器抽象,用于消费PM。LibPM的容器是数据托管单元,可以存储任意复杂的数据类型,同时保持其完整性和一致性。因此,LibPM的容器为应用程序提供了一个通用接口,允许应用程序存储和操作具有强持久性和一致性属性的任意结构化数据,而无需直接导航编程PM的所有无数陷阱。通过提供简单而高性能的事务性更新机制,LibPM允许应用程序以内存的速度操作持久数据。LibPM中的容器抽象和自动持久数据发现机制也简化了将遗留应用程序移植到PM的过程。从性能的角度来看,LibPM的性能与最先进的PM应用程序库非常接近,甚至经常超过它们。例如,与最先进的Pmem相比,LibPM在写密集型工作负载方面的性能提高了195倍,在读密集型工作负载方面的性能提高了2.6倍。IO持久内存库。
{"title":"LibPM","authors":"L. Mármol, M. Chowdhury, R. Rangaswami","doi":"10.1145/3278141","DOIUrl":"https://doi.org/10.1145/3278141","url":null,"abstract":"Persistent Memory devices present properties that are uniquely different from prior technologies for which applications have been built. Unfortunately, the conventional approach to building applications fail to either efficiently utilize these new devices or provide programmers a seamless development experience. We have built LibPM, a Persistent Memory Library that implements an easy-to-use container abstraction for consuming PM. LibPM’s containers are data hosting units that can store arbitrarily complex data types while preserving their integrity and consistency. Consequently, LibPM’s containers provide a generic interface to applications, allowing applications to store and manipulate arbitrarily structured data with strong durability and consistency properties, all without having to navigate all the myriad pitfalls of programming PM directly. By providing a simple and high-performing transactional update mechanism, LibPM allows applications to manipulate persistent data at the speed of memory. The container abstraction and automatic persistent data discovery mechanisms within LibPM also simplify porting legacy applications to PM. From a performance perspective, LibPM closely matches and often exceeds the performance of state-of-the-art application libraries for PM. For instance, LibPM ’s performance is 195× better for write intensive workloads and 2.6× better for read intensive workloads when compared with the state-of-the-art Pmem.IO persistent memory library.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"87 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126295612","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
Efficient Directory Mutations in a Full-Path-Indexed File System 全路径索引文件系统中的有效目录突变
Pub Date : 2018-11-26 DOI: 10.1145/3241061
Yang Zhan, Yizheng Jiao, Donald E. Porter, Alex Conway, Eric Knorr, Martín Farach-Colton, M. A. Bender, Jun Yuan, William Jannen, Rob Johnson
Full-path indexing can improve I/O efficiency for workloads that operate on data organized using traditional, hierarchical directories, because data is placed on persistent storage in scan order. Prior results indicate, however, that renames in a local file system with full-path indexing are prohibitively expensive. This article shows how to use full-path indexing in a file system to realize fast directory scans, writes, and renames. The article introduces a range-rename mechanism for efficient key-space changes in a write-optimized dictionary. This mechanism is encapsulated in the key-value Application Programming Interface (API) and simplifies the overall file system design. We implemented this mechanism in B&egr;-trees File System (BetrFS), an in-kernel, local file system for Linux. This new version, BetrFS 0.4, performs recursive greps 1.5x faster and random writes 1.2x faster than BetrFS 0.3, but renames are competitive with indirection-based file systems for a range of sizes. BetrFS 0.4 outperforms BetrFS 0.3, as well as traditional file systems, such as ext4, Extents File System (XFS), and Z File System (ZFS), across a variety of workloads.
对于使用传统的分层目录组织数据的工作负载,全路径索引可以提高I/O效率,因为数据是按扫描顺序放在持久存储上的。但是,先前的结果表明,使用全路径索引的本地文件系统中的重命名代价非常高。本文展示了如何在文件系统中使用全路径索引来实现快速的目录扫描、写入和重命名。本文介绍了一种范围重命名机制,用于在写优化字典中有效地更改键空间。这种机制被封装在键值应用程序编程接口(API)中,简化了整个文件系统的设计。我们在B&egr; trees文件系统(BetrFS)中实现了这种机制,这是Linux的内核内本地文件系统。这个新版本,即BetrFS 0.4,执行递归greps的速度比BetrFS 0.3快1.5倍,随机写速度比BetrFS 0.3快1.2倍,但是对于各种大小的文件系统,重命名与基于间接的文件系统竞争。在各种工作负载中,BetrFS 0.4的性能优于BetrFS 0.3,也优于传统的文件系统,如ext4、Extents file System (XFS)和Z file System (ZFS)。
{"title":"Efficient Directory Mutations in a Full-Path-Indexed File System","authors":"Yang Zhan, Yizheng Jiao, Donald E. Porter, Alex Conway, Eric Knorr, Martín Farach-Colton, M. A. Bender, Jun Yuan, William Jannen, Rob Johnson","doi":"10.1145/3241061","DOIUrl":"https://doi.org/10.1145/3241061","url":null,"abstract":"Full-path indexing can improve I/O efficiency for workloads that operate on data organized using traditional, hierarchical directories, because data is placed on persistent storage in scan order. Prior results indicate, however, that renames in a local file system with full-path indexing are prohibitively expensive. This article shows how to use full-path indexing in a file system to realize fast directory scans, writes, and renames. The article introduces a range-rename mechanism for efficient key-space changes in a write-optimized dictionary. This mechanism is encapsulated in the key-value Application Programming Interface (API) and simplifies the overall file system design. We implemented this mechanism in B&egr;-trees File System (BetrFS), an in-kernel, local file system for Linux. This new version, BetrFS 0.4, performs recursive greps 1.5x faster and random writes 1.2x faster than BetrFS 0.3, but renames are competitive with indirection-based file systems for a range of sizes. BetrFS 0.4 outperforms BetrFS 0.3, as well as traditional file systems, such as ext4, Extents File System (XFS), and Z File System (ZFS), across a variety of workloads.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123042450","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Write Energy Reduction for PCM via Pumping Efficiency Improvement 通过提高泵送效率来降低PCM的写入能耗
Pub Date : 2018-11-26 DOI: 10.1145/3200139
Huizhang Luo, Qing Liu, J. Hu, Qiao Li, Liang Shi, Qingfeng Zhuge, E. Sha
The emerging Phase Change Memory (PCM) is considered to be a promising candidate to replace DRAM as the next generation main memory due to its higher scalability and lower leakage power. However, the high write power consumption has become a major challenge in adopting PCM as main memory. In addition to the fact that writing to PCM cells requires high write current and voltage, current loss in the charge pumps also contributes a large percentage of high power consumption. The pumping efficiency of a PCM chip is a concave function of the write current. Leveraging the characteristics of the concave function, the overall pumping efficiency can be improved if the write current is uniform. In this article, we propose a peak-to-average (PTA) write scheme, which smooths the write current fluctuation by regrouping write units. In particular, we calculate the current requirements for each write unit by their values when they are evicted from the last level cache (LLC). When the write units are waiting in the memory controller, we regroup the write units by LLC-assisted PTA to reach the current-uniform goal. Experimental results show that LLC-assisted PTA achieved 13.4% of overall energy saving compared to the baseline.
新兴的相变存储器(PCM)由于具有更高的可扩展性和更低的泄漏功率,被认为是取代DRAM作为下一代主存储器的有希望的候选人。然而,高写入功耗已成为采用PCM作主存的主要挑战。除了写入PCM电池需要高写入电流和电压之外,电荷泵中的电流损耗也占高功耗的很大比例。PCM芯片的泵浦效率是写电流的凹函数。利用凹函数的特性,如果写入电流均匀,则可以提高整体泵浦效率。在本文中,我们提出了一种峰平均(PTA)写入方案,该方案通过重新分组写入单元来平滑写入电流波动。特别是,我们根据每个写单元从最后一级缓存(LLC)中被驱逐时的值来计算当前对它们的需求。当写单元在内存控制器中等待时,我们通过plc辅助PTA对写单元进行重组,以达到电流均匀的目标。实验结果表明,与基线相比,llc辅助PTA实现了13.4%的总体节能。
{"title":"Write Energy Reduction for PCM via Pumping Efficiency Improvement","authors":"Huizhang Luo, Qing Liu, J. Hu, Qiao Li, Liang Shi, Qingfeng Zhuge, E. Sha","doi":"10.1145/3200139","DOIUrl":"https://doi.org/10.1145/3200139","url":null,"abstract":"The emerging Phase Change Memory (PCM) is considered to be a promising candidate to replace DRAM as the next generation main memory due to its higher scalability and lower leakage power. However, the high write power consumption has become a major challenge in adopting PCM as main memory. In addition to the fact that writing to PCM cells requires high write current and voltage, current loss in the charge pumps also contributes a large percentage of high power consumption. The pumping efficiency of a PCM chip is a concave function of the write current. Leveraging the characteristics of the concave function, the overall pumping efficiency can be improved if the write current is uniform. In this article, we propose a peak-to-average (PTA) write scheme, which smooths the write current fluctuation by regrouping write units. In particular, we calculate the current requirements for each write unit by their values when they are evicted from the last level cache (LLC). When the write units are waiting in the memory controller, we regroup the write units by LLC-assisted PTA to reach the current-uniform goal. Experimental results show that LLC-assisted PTA achieved 13.4% of overall energy saving compared to the baseline.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131807670","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Enabling Efficient Updates in KV Storage via Hashing 通过哈希在KV存储中实现高效更新
Pub Date : 2018-11-25 DOI: 10.1145/3340287
Yongkun Li, H. Chan, P. Lee, Yinlong Xu
Persistent key-value (KV) stores mostly build on the Log-Structured Merge (LSM) tree for high write performance, yet the LSM-tree suffers from the inherently high I/O amplification. KV separation mitigates I/O amplification by storing only keys in the LSM-tree and values in separate storage. However, the current KV separation design remains inefficient under update-intensive workloads due to its high garbage collection (GC) overhead in value storage. We propose HashKV, which aims for high update performance atop KV separation under update-intensive workloads. HashKV uses hash-based data grouping, which deterministically maps values to storage space to make both updates and GC efficient. We further relax the restriction of such deterministic mappings via simple but useful design extensions. We extensively evaluate various design aspects of HashKV. We show that HashKV achieves 4.6× update throughput and 53.4% less write traffic compared to the current KV separation design. In addition, we demonstrate that we can integrate the design of HashKV with state-of-the-art KV stores and improve their respective performance.
持久键值(KV)存储主要构建在日志结构合并(LSM)树上,以获得较高的写性能,但是LSM树本身就存在高I/O放大的问题。KV分离通过仅在lsm树中存储键和在单独存储中存储值来减轻I/O放大。然而,当前的KV分离设计在更新密集型工作负载下仍然效率低下,因为它在值存储中的垃圾收集(GC)开销很高。我们提出HashKV,它的目标是在更新密集型工作负载下,在KV分离的基础上实现高更新性能。HashKV使用基于哈希的数据分组,它确定地将值映射到存储空间,以提高更新和GC效率。我们通过简单但有用的设计扩展进一步放宽了这种确定性映射的限制。我们广泛评估了HashKV的各个设计方面。我们表明,与当前的KV分离设计相比,HashKV实现了4.6倍的更新吞吐量和53.4%的写流量。此外,我们证明了我们可以将HashKV的设计与最先进的KV存储集成在一起,并提高它们各自的性能。
{"title":"Enabling Efficient Updates in KV Storage via Hashing","authors":"Yongkun Li, H. Chan, P. Lee, Yinlong Xu","doi":"10.1145/3340287","DOIUrl":"https://doi.org/10.1145/3340287","url":null,"abstract":"Persistent key-value (KV) stores mostly build on the Log-Structured Merge (LSM) tree for high write performance, yet the LSM-tree suffers from the inherently high I/O amplification. KV separation mitigates I/O amplification by storing only keys in the LSM-tree and values in separate storage. However, the current KV separation design remains inefficient under update-intensive workloads due to its high garbage collection (GC) overhead in value storage. We propose HashKV, which aims for high update performance atop KV separation under update-intensive workloads. HashKV uses hash-based data grouping, which deterministically maps values to storage space to make both updates and GC efficient. We further relax the restriction of such deterministic mappings via simple but useful design extensions. We extensively evaluate various design aspects of HashKV. We show that HashKV achieves 4.6× update throughput and 53.4% less write traffic compared to the current KV separation design. In addition, we demonstrate that we can integrate the design of HashKV with state-of-the-art KV stores and improve their respective performance.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"19 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125911951","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 64
Introduction to the Special Issue on USENIX FAST 2018 USENIX FAST 2018特刊简介
Pub Date : 2018-11-09 DOI: 10.1145/3242152
Nitin Agrawal, R. Rangaswami
This special issue of the ACM Transactions on Storage (TOS) presents some of the highlights of the 16th USENIX Conference on File and Storage Technologies (FAST’18). Over the years, FAST has evolved into a community of researchers and practitioners working on a diverse and expanding set of research topics; the conference represents some of the latest and best work being done, and this year was no different. FAST’18 received a record number of 139 submissions on topics ranging from non-volatile memory; distributed, cloud, and data center storage; and performance and scalability to experiences with deployed systems. Of these, we selected five high-quality articles for publication in this special issue of ACM TOS. The first article, which was also selected as one of the best papers at the conference, is “Protocol-Aware Recovery for Consensus-based Storage” by Ramnatthan Alagappan, Aishwarya Ganesan, Eric Lee, Aws Albarghouthi, Vijay Chidambaram, Andrea Arpaci-Dusseau, and Remzi Arpaci-Dusseau. Distributed storage systems are in widespread use today. The authors demonstrate how storage faults can significantly affect recovery in distributed storage systems that are based on replicated state machines, including ones in widespread use today. They then propose corruption-tolerant replication as a solution that can ensure safe recovery. The second article is “Efficient Directory Mutations in a Full-Path Indexed File System” by Yang Zhan, Alex Conway, Yizheng Jiao, Eric Knorr, Michael A. Bender, Martin Farach-Colton, William Jannen, Rob Johnson, Donald E. Porter, and Jun Yuan. BetrFS is a file system that offers dramatically faster execution times for common modern-day file-system operations. In this significant update to the design of BetrFS, the authors tackle the last stronghold of performance challenges, rename, with a new “range-rename” mechanism. The third article is “Fail-Slow at Scale: Evidence of Hardware Performance Faults in Large Production Systems” by Haryadi S. Gunawi, Riza O. Suminto, Russell Sears, Casey Golliher, Swaminathan Sundararaman, Xing Lin, Tim Emami, Weiguang Sheng, Nematollah Bidokhti, Caitie McCaffrey, Gary Grider, Parks M. Fields, Kevin Harms, Robert B. Ross, Andree Jacobson, Robert Ricci, Kirk Webb, Peter Alvaro, H. Birali Runesha, Mingzhe Hao, and Huaicheng Li. Mysterious storage faults are legends within the computer industry and increasingly more so as the scale of deployed systems grows rapidly; this article presents a lively discussion of one such class of faults, namely fail-slow, that has significant impact. The authors draw from a large-scale study based on significant documented and anecdotal evidence obtained from 101 reports of such incidents sourced from 12 different institutions. The fourth article, which was also selected as one of the best papers at the conference, is “Bringing Order to Chaos: Barrier-Enabled I/O Stack for Flash Storage” by Youjip Won, Joontaek Oh, Jaemin Jung, Gyeongyeol Choi, Seongbae Son, Jo
本期ACM存储交易(TOS)特刊介绍了第16届USENIX文件和存储技术会议(FAST’18)的一些亮点。多年来,FAST已经发展成为一个研究人员和实践者的社区,致力于多样化和不断扩大的研究主题;会议代表了一些最新和最好的工作,今年也不例外。FAST’18收到了创纪录的139份提交,主题从非易失性存储器;分布式、云、数据中心存储;以及部署系统的性能和可伸缩性。其中,我们选择了5篇高质量的文章发表在本期ACM TOS特刊上。第一篇文章是Ramnatthan Alagappan、Aishwarya Ganesan、Eric Lee、Aws Albarghouthi、Vijay Chidambaram、Andrea Arpaci-Dusseau和Remzi Arpaci-Dusseau撰写的“基于共识的存储的协议感知恢复”,也是会议上最好的论文之一。分布式存储系统在今天得到了广泛的应用。作者演示了存储故障如何显著影响基于复制状态机的分布式存储系统的恢复,包括目前广泛使用的存储系统。然后,他们提出了一种可以确保安全恢复的容错复制解决方案。第二篇文章是《全路径索引文件系统中的有效目录突变》,作者是詹杨、Alex Conway、焦奕正、Eric Knorr、Michael a . Bender、Martin Farach-Colton、William Jannen、Rob Johnson、Donald E. Porter和Yuan Jun。BetrFS是一种文件系统,它为常见的现代文件系统操作提供了更快的执行时间。在这个对BetrFS设计的重大更新中,作者使用一种新的“range-rename”机制解决了性能挑战的最后一个据点——重命名。第三篇文章是Haryadi S. Gunawi、Riza O. Suminto、Russell Sears、Casey Golliher、Swaminathan Sundararaman、林星、Tim Emami、沈伟光、Nematollah Bidokhti、Caitie McCaffrey、Gary Grider、Parks M. Fields、Kevin Harms、andrew Jacobson、Robert Ricci、Kirk Webb、Peter Alvaro、H. Birali Runesha、Mingzhe Hao和Huaicheng Li。神秘的存储故障是计算机行业的传奇,随着部署系统规模的迅速增长,这种传奇越来越多;本文生动地讨论了其中一类具有重大影响的故障,即慢速故障。作者从一项大规模研究中得出结论,该研究基于从12个不同机构获得的101份此类事件报告中获得的重要文献和轶事证据。被选为本次大会最佳论文的第4篇论文是《将秩序带入混乱:基于障碍的闪存I/O堆栈》,作者是元酉杰、吴俊泽、郑在民、崔景烈、孙成培、黄炯荣、赵尚渊。现代存储I/O堆栈极其复杂;造成这种复杂性的主要原因是分层和跨层的“阻抗不匹配”。本文的作者重新审视了这个久已成熟的领域,并做出了令人惊讶的原创贡献,它不仅功能强大,而且从根本上简单,能够安全地从高性能存储中提取最大的数据。
{"title":"Introduction to the Special Issue on USENIX FAST 2018","authors":"Nitin Agrawal, R. Rangaswami","doi":"10.1145/3242152","DOIUrl":"https://doi.org/10.1145/3242152","url":null,"abstract":"This special issue of the ACM Transactions on Storage (TOS) presents some of the highlights of the 16th USENIX Conference on File and Storage Technologies (FAST’18). Over the years, FAST has evolved into a community of researchers and practitioners working on a diverse and expanding set of research topics; the conference represents some of the latest and best work being done, and this year was no different. FAST’18 received a record number of 139 submissions on topics ranging from non-volatile memory; distributed, cloud, and data center storage; and performance and scalability to experiences with deployed systems. Of these, we selected five high-quality articles for publication in this special issue of ACM TOS. The first article, which was also selected as one of the best papers at the conference, is “Protocol-Aware Recovery for Consensus-based Storage” by Ramnatthan Alagappan, Aishwarya Ganesan, Eric Lee, Aws Albarghouthi, Vijay Chidambaram, Andrea Arpaci-Dusseau, and Remzi Arpaci-Dusseau. Distributed storage systems are in widespread use today. The authors demonstrate how storage faults can significantly affect recovery in distributed storage systems that are based on replicated state machines, including ones in widespread use today. They then propose corruption-tolerant replication as a solution that can ensure safe recovery. The second article is “Efficient Directory Mutations in a Full-Path Indexed File System” by Yang Zhan, Alex Conway, Yizheng Jiao, Eric Knorr, Michael A. Bender, Martin Farach-Colton, William Jannen, Rob Johnson, Donald E. Porter, and Jun Yuan. BetrFS is a file system that offers dramatically faster execution times for common modern-day file-system operations. In this significant update to the design of BetrFS, the authors tackle the last stronghold of performance challenges, rename, with a new “range-rename” mechanism. The third article is “Fail-Slow at Scale: Evidence of Hardware Performance Faults in Large Production Systems” by Haryadi S. Gunawi, Riza O. Suminto, Russell Sears, Casey Golliher, Swaminathan Sundararaman, Xing Lin, Tim Emami, Weiguang Sheng, Nematollah Bidokhti, Caitie McCaffrey, Gary Grider, Parks M. Fields, Kevin Harms, Robert B. Ross, Andree Jacobson, Robert Ricci, Kirk Webb, Peter Alvaro, H. Birali Runesha, Mingzhe Hao, and Huaicheng Li. Mysterious storage faults are legends within the computer industry and increasingly more so as the scale of deployed systems grows rapidly; this article presents a lively discussion of one such class of faults, namely fail-slow, that has significant impact. The authors draw from a large-scale study based on significant documented and anecdotal evidence obtained from 101 reports of such incidents sourced from 12 different institutions. The fourth article, which was also selected as one of the best papers at the conference, is “Bringing Order to Chaos: Barrier-Enabled I/O Stack for Flash Storage” by Youjip Won, Joontaek Oh, Jaemin Jung, Gyeongyeol Choi, Seongbae Son, Jo","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"22 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133162850","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
ROS
Pub Date : 2018-11-09 DOI: 10.1145/3231599
Wenrui Yan, Jie Yao, Q. Cao, C. Xie, Hong Jiang
The combination of the explosive growth in digital data and the demand to preserve much of these data in the long term has made it imperative to find a more cost-effective way than HDD arrays and a more easily accessible way than tape libraries to store massive amounts of data. While modern optical discs are capable of guaranteeing more than 50-year data preservation without media replacement, individual optical discs’ lack of the performance and capacity relative to HDDs or tapes has significantly limited their use in datacenters. This article presents a Rack-scale Optical disc library System, or ROS in short, which provides a PB-level total capacity and inline accessibility on thousands of optical discs built within a 42U Rack. A rotatable roller and robotic arm separating and fetching discs are designed to improve disc placement density and simplify the mechanical structure. A hierarchical storage system based on SSDs, hard disks, and optical discs is proposed to effectively hide the delay of mechanical operation. However, an optical library file system (OLFS) based on FUSE is proposed to schedule mechanical operation and organize data on the tiered storage with a POSIX user interface to provide an illusion of inline data accessibility. We further optimize OLFS by reducing unnecessary user/kernel context switches inheriting from legacy FUSE framework. We evaluate ROS on a few key performance metrics, including operation delays of the mechanical structure and software overhead in a prototype PB-level ROS system. The results show that ROS stacked on Samba and FUSE as network-attached storage (NAS) mode almost saturates the throughput provided by underlying samba via 10GbE network for external users, as well as in this scenario provides about 53ms file write and 15ms read latency, exhibiting its inline accessibility. Besides, ROS is able to effectively hide and virtualize internal complex operational behaviors and be easily deployable in datacenters.
数字数据的爆炸性增长和长期保存这些数据的需求相结合,使得必须找到一种比HDD阵列更具成本效益的方式,一种比磁带库更容易访问的方式来存储大量数据。虽然现代光盘能够保证在不更换介质的情况下保存50年以上的数据,但相对于硬盘驱动器或磁带,单个光盘缺乏性能和容量,这大大限制了它们在数据中心的使用。本文介绍了一个机架级光盘库系统(简称ROS),它提供了pb级的总容量和内置在42U机架内的数千张光盘的联机访问。设计了旋转滚轮和机械臂分离取盘,提高了盘的放置密度,简化了机械结构。提出了一种基于ssd、硬盘和光盘的分层存储系统,有效地隐藏了机械操作的延迟。然而,提出了一种基于FUSE的光学库文件系统(OLFS),通过POSIX用户界面在分层存储上调度机械操作和组织数据,以提供内联数据可访问性的假象。通过减少从遗留FUSE框架继承的不必要的用户/内核上下文切换,我们进一步优化了OLFS。我们通过几个关键性能指标来评估ROS,包括机械结构的操作延迟和原型pb级ROS系统的软件开销。结果表明,ROS作为网络附加存储(NAS)模式堆叠在Samba和FUSE上,几乎饱和了底层Samba通过10GbE网络为外部用户提供的吞吐量,并且在此场景中提供了大约53ms的文件写入和15ms的读取延迟,显示了其内联可访问性。此外,ROS能够有效地隐藏和虚拟化内部复杂的操作行为,并且易于在数据中心中部署。
{"title":"ROS","authors":"Wenrui Yan, Jie Yao, Q. Cao, C. Xie, Hong Jiang","doi":"10.1145/3231599","DOIUrl":"https://doi.org/10.1145/3231599","url":null,"abstract":"The combination of the explosive growth in digital data and the demand to preserve much of these data in the long term has made it imperative to find a more cost-effective way than HDD arrays and a more easily accessible way than tape libraries to store massive amounts of data. While modern optical discs are capable of guaranteeing more than 50-year data preservation without media replacement, individual optical discs’ lack of the performance and capacity relative to HDDs or tapes has significantly limited their use in datacenters. This article presents a Rack-scale Optical disc library System, or ROS in short, which provides a PB-level total capacity and inline accessibility on thousands of optical discs built within a 42U Rack. A rotatable roller and robotic arm separating and fetching discs are designed to improve disc placement density and simplify the mechanical structure. A hierarchical storage system based on SSDs, hard disks, and optical discs is proposed to effectively hide the delay of mechanical operation. However, an optical library file system (OLFS) based on FUSE is proposed to schedule mechanical operation and organize data on the tiered storage with a POSIX user interface to provide an illusion of inline data accessibility. We further optimize OLFS by reducing unnecessary user/kernel context switches inheriting from legacy FUSE framework. We evaluate ROS on a few key performance metrics, including operation delays of the mechanical structure and software overhead in a prototype PB-level ROS system. The results show that ROS stacked on Samba and FUSE as network-attached storage (NAS) mode almost saturates the throughput provided by underlying samba via 10GbE network for external users, as well as in this scenario provides about 53ms file write and 15ms read latency, exhibiting its inline accessibility. Besides, ROS is able to effectively hide and virtualize internal complex operational behaviors and be easily deployable in datacenters.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"125 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-11-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122895605","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
DIDACache
Pub Date : 2018-10-31 DOI: 10.1145/3203410
Zhaoyan Shen, Feng Chen, Yichen Jia, Z. Shao
Key-value caching is crucial to today’s low-latency Internet services. Conventional key-value cache systems, such as Memcached, heavily rely on expensive DRAM memory. To lower Total Cost of Ownership, the industry recently is moving toward more cost-efficient flash-based solutions, such as Facebook’s McDipper [14] and Twitter’s Fatcache [56]. These cache systems typically take commercial SSDs and adopt a Memcached-like scheme to store and manage key-value cache data in flash. Such a practice, though simple, is inefficient due to the huge semantic gap between the key-value cache manager and the underlying flash devices. In this article, we advocate to reconsider the cache system design and directly open device-level details of the underlying flash storage for key-value caching. We propose an enhanced flash-aware key-value cache manager, which consists of a novel unified address mapping module, an integrated garbage collection policy, a dynamic over-provisioning space management, and a customized wear-leveling policy, to directly drive the flash management. A thin intermediate library layer provides a slab-based abstraction of low-level flash memory space and an API interface for directly and easily operating flash devices. A special flash memory SSD hardware that exposes flash physical details is adopted to store key-value items. This co-design approach bridges the semantic gap and well connects the two layers together, which allows us to leverage both the domain knowledge of key-value caches and the unique device properties. In this way, we can maximize the efficiency of key-value caching on flash devices while minimizing its weakness. We implemented a prototype, called DIDACache, based on the Open-Channel SSD platform. Our experiments on real hardware show that we can significantly increase the throughput by 35.5%, reduce the latency by 23.6%, and remove unnecessary erase operations by 28%.
键值缓存对于当今的低延迟Internet服务至关重要。传统的键值缓存系统(如Memcached)严重依赖昂贵的DRAM内存。为了降低总拥有成本,该行业最近正在转向更具成本效益的基于闪存的解决方案,如Facebook的McDipper[14]和Twitter的Fatcache[56]。这些缓存系统通常使用商用ssd,并采用类似memcached的方案在闪存中存储和管理键值缓存数据。这种做法虽然简单,但效率很低,因为键值缓存管理器和底层闪存设备之间存在巨大的语义差距。在本文中,我们建议重新考虑缓存系统设计,并直接打开用于键值缓存的底层闪存的设备级详细信息。我们提出了一种增强的闪存感知键值缓存管理器,它由一个新颖的统一地址映射模块、一个集成的垃圾收集策略、一个动态的过度供应空间管理和一个定制的损耗均衡策略组成,直接驱动闪存管理。一个瘦的中间库层提供了一个基于slab的低级闪存空间抽象和一个API接口,用于直接和轻松地操作闪存设备。采用一种特殊的闪存SSD硬件,可以暴露闪存的物理细节,用于存储键值项。这种协同设计方法弥合了语义上的差距,并将两层很好地连接在一起,这使我们能够利用键值缓存的领域知识和独特的设备属性。通过这种方式,我们可以最大化闪存设备上键值缓存的效率,同时最小化其弱点。我们基于Open-Channel SSD平台实现了一个名为DIDACache的原型。我们在真实硬件上的实验表明,我们可以显着提高吞吐量35.5%,减少延迟23.6%,并消除不必要的擦除操作28%。
{"title":"DIDACache","authors":"Zhaoyan Shen, Feng Chen, Yichen Jia, Z. Shao","doi":"10.1145/3203410","DOIUrl":"https://doi.org/10.1145/3203410","url":null,"abstract":"Key-value caching is crucial to today’s low-latency Internet services. Conventional key-value cache systems, such as Memcached, heavily rely on expensive DRAM memory. To lower Total Cost of Ownership, the industry recently is moving toward more cost-efficient flash-based solutions, such as Facebook’s McDipper [14] and Twitter’s Fatcache [56]. These cache systems typically take commercial SSDs and adopt a Memcached-like scheme to store and manage key-value cache data in flash. Such a practice, though simple, is inefficient due to the huge semantic gap between the key-value cache manager and the underlying flash devices. In this article, we advocate to reconsider the cache system design and directly open device-level details of the underlying flash storage for key-value caching. We propose an enhanced flash-aware key-value cache manager, which consists of a novel unified address mapping module, an integrated garbage collection policy, a dynamic over-provisioning space management, and a customized wear-leveling policy, to directly drive the flash management. A thin intermediate library layer provides a slab-based abstraction of low-level flash memory space and an API interface for directly and easily operating flash devices. A special flash memory SSD hardware that exposes flash physical details is adopted to store key-value items. This co-design approach bridges the semantic gap and well connects the two layers together, which allows us to leverage both the domain knowledge of key-value caches and the unique device properties. In this way, we can maximize the efficiency of key-value caching on flash devices while minimizing its weakness. We implemented a prototype, called DIDACache, based on the Open-Channel SSD platform. Our experiments on real hardware show that we can significantly increase the throughput by 35.5%, reduce the latency by 23.6%, and remove unnecessary erase operations by 28%.","PeriodicalId":273014,"journal":{"name":"ACM Transactions on Storage (TOS)","volume":"173 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-10-31","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124218494","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
期刊
ACM Transactions on Storage (TOS)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1