首页 > 最新文献

2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)最新文献

英文 中文
Red: An efficient replacement algorithm based on REsident Distance for exclusive storage caches 红色:基于REsident Distance的高效的独占缓存替换算法
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496988
Yingjie Zhao, Nong Xiao, Fang Liu
This paper presents our replacement algorithm named RED for storage caches. RED is exclusive. It can eliminate the duplications between a storage cache and its client cache. RED is high performance. A new criterion Resident Distance is proposed for making an efficient replacement decision instead of Recency and Frequency. Moreover, RED is non-intrusive to a storage client. It does not need to change client software and could be used in a real-life system. Previous work on the management of a storage cache can attain one or two of above benefits, but not all of them. We have evaluated the performance of RED by using simulations with both synthetic and real-life traces. The simulation results show that RED significantly outperforms LRU, ARC, MQ, and is better than DEMOTE, PROMOTE for a wide range of cache sizes.
本文提出了一种名为RED的存储缓存替换算法。RED是独家的。它可以消除存储缓存与其客户端缓存之间的重复。RED是高性能。提出了一种新的“驻留距离”准则来代替“近因”和“频率”来进行有效的替换决策。此外,RED对存储客户端是非侵入性的。它不需要更改客户端软件,可以在实际系统中使用。以前关于存储缓存管理的工作可以获得上述好处中的一两个,但不是全部。我们通过模拟合成轨迹和真实轨迹来评估RED的性能。仿真结果表明,RED算法明显优于LRU、ARC、MQ,并且在大范围的缓存大小下优于DEMOTE、PROMOTE。
{"title":"Red: An efficient replacement algorithm based on REsident Distance for exclusive storage caches","authors":"Yingjie Zhao, Nong Xiao, Fang Liu","doi":"10.1109/MSST.2010.5496988","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496988","url":null,"abstract":"This paper presents our replacement algorithm named RED for storage caches. RED is exclusive. It can eliminate the duplications between a storage cache and its client cache. RED is high performance. A new criterion Resident Distance is proposed for making an efficient replacement decision instead of Recency and Frequency. Moreover, RED is non-intrusive to a storage client. It does not need to change client software and could be used in a real-life system. Previous work on the management of a storage cache can attain one or two of above benefits, but not all of them. We have evaluated the performance of RED by using simulations with both synthetic and real-life traces. The simulation results show that RED significantly outperforms LRU, ARC, MQ, and is better than DEMOTE, PROMOTE for a wide range of cache sizes.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124922828","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
BPAC: An adaptive write buffer management scheme for flash-based Solid State Drives BPAC:一种基于闪存的固态硬盘的自适应写缓冲区管理方案
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496998
Guanying Wu, B. Eckart, Xubin He
Solid State Drives (SSD's) have shown promise to be a candidate to replace traditional hard disk drives, but due to certain physical characteristics of NAND flash, there are some challenging areas of improvement and further research. We focus on the layout and management of the small amount of RAM that serves as a cache between the SSD and the system that uses it. Of the techniques that have previously been proposed to manage this cache, we identify several sources of inefficient cache space management due to the way pages are clustered in blocks and the limited replacement policy. We develop a hybrid page/block architecture along with an advanced replacement policy, called BPAC, or Block-Page Adaptive Cache, to exploit both temporal and spatial locality. Our technique involves adaptively partitioning the SSD on-disk cache to separately hold pages with high temporal locality in a page list and clusters of pages with low temporal but high spatial locality in a block list. We run trace-driven simulations to verify our design and find that it outperforms other popular flash-aware cache schemes under different workloads.
固态硬盘(SSD)已经显示出取代传统硬盘驱动器的潜力,但由于NAND闪存的某些物理特性,还有一些具有挑战性的改进和进一步研究领域。我们关注的是作为SSD和使用它的系统之间的缓存的少量RAM的布局和管理。在先前提出的用于管理此缓存的技术中,我们确定了由于页面聚集在块中的方式和有限的替换策略而导致缓存空间管理效率低下的几个来源。我们开发了一种混合页面/块架构以及一种高级替换策略,称为BPAC,或块页面自适应缓存,以利用时间和空间局域性。我们的技术包括自适应地对磁盘上的SSD缓存进行分区,在页面列表中分别保存具有高时间局部性的页面,在块列表中分别保存具有低时间局部性但高空间局部性的页面集群。我们运行跟踪驱动模拟来验证我们的设计,并发现它在不同工作负载下优于其他流行的闪存感知缓存方案。
{"title":"BPAC: An adaptive write buffer management scheme for flash-based Solid State Drives","authors":"Guanying Wu, B. Eckart, Xubin He","doi":"10.1109/MSST.2010.5496998","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496998","url":null,"abstract":"Solid State Drives (SSD's) have shown promise to be a candidate to replace traditional hard disk drives, but due to certain physical characteristics of NAND flash, there are some challenging areas of improvement and further research. We focus on the layout and management of the small amount of RAM that serves as a cache between the SSD and the system that uses it. Of the techniques that have previously been proposed to manage this cache, we identify several sources of inefficient cache space management due to the way pages are clustered in blocks and the limited replacement policy. We develop a hybrid page/block architecture along with an advanced replacement policy, called BPAC, or Block-Page Adaptive Cache, to exploit both temporal and spatial locality. Our technique involves adaptively partitioning the SSD on-disk cache to separately hold pages with high temporal locality in a page list and clusters of pages with low temporal but high spatial locality in a block list. We run trace-driven simulations to verify our design and find that it outperforms other popular flash-aware cache schemes under different workloads.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126132931","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 51
MAD2: A scalable high-throughput exact deduplication approach for network backup services MAD2:用于网络备份服务的可扩展高吞吐量精确重复数据删除方法
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496987
Jiansheng Wei, Hong Jiang, Ke Zhou, D. Feng
Deduplication has been widely used in disk-based secondary storage systems to improve space efficiency. However, there are two challenges facing scalable high-throughput deduplication storage. The first is the duplicate-lookup disk bottleneck due to the large size of data index that usually exceeds the available RAM space, which limits the deduplication throughput. The second is the storage node island effect resulting from duplicate data among multiple storage nodes that are difficult to eliminate. Existing approaches fail to completely eliminate the duplicates while simultaneously addressing the challenges. This paper proposes MAD2, a scalable high-throughput exact deduplication approach for network backup services. MAD2 eliminates duplicate data both at the file level and at the chunk level by employing four techniques to accelerate the deduplication process and evenly distribute data. First, MAD2 organizes fingerprints into a Hash Bucket Matrix (HBM), whose rows can be used to preserve the data locality in backups. Second, MAD2 uses Bloom Filter Array (BFA) as a quick index to quickly identify non-duplicate incoming data objects or indicate where to find a possible duplicate. Third, Dual Cache is integrated in MAD2 to effectively capture and exploit data locality. Finally, MAD2 employs a DHT-based Load-Balance technique to evenly distribute data objects among multiple storage nodes in their backup sequences to further enhance performance with a well-balanced load. We evaluate our MAD2 approach on the backend storage of B-Cloud, a research-oriented distributed system that provides network backup services. Experimental results show that MAD2 significantly outperforms the state-of-the-art approximate deduplication approaches in terms of deduplication efficiency, supporting a deduplication throughput of at least 100MB/s for each storage component.
重复数据删除技术被广泛应用于基于磁盘的二级存储系统中,以提高存储空间的使用效率。但是,可扩展的高吞吐量重复数据删除存储面临两个挑战。首先是重复查找磁盘瓶颈,这是由于数据索引的大小通常超过可用的RAM空间,从而限制了重复数据删除吞吐量。二是由于多个存储节点之间的重复数据难以消除而产生的存储节点孤岛效应。现有的方法无法在解决挑战的同时完全消除重复。本文提出了一种可扩展的、高吞吐量的精确重复数据删除方法MAD2,用于网络备份服务。MAD2通过采用四种技术来加速重复数据删除过程并均匀分布数据,从而消除了文件级和块级的重复数据。首先,MAD2将指纹组织到哈希桶矩阵(HBM)中,该矩阵的行可用于在备份中保留数据局部性。其次,MAD2使用Bloom Filter Array (BFA)作为快速索引来快速识别非重复的传入数据对象或指示在何处查找可能的重复。第三,MAD2集成了双缓存,有效地捕获和利用数据局部性。最后,MAD2采用基于dht的负载平衡技术,在多个存储节点的备份序列中均匀分布数据对象,从而通过负载均衡进一步提高性能。我们在B-Cloud的后端存储上评估了我们的MAD2方法,B-Cloud是一个提供网络备份服务的研究型分布式系统。实验结果表明,MAD2在重复数据删除效率方面明显优于最先进的近似重复数据删除方法,支持每个存储组件至少100MB/s的重复数据删除吞吐量。
{"title":"MAD2: A scalable high-throughput exact deduplication approach for network backup services","authors":"Jiansheng Wei, Hong Jiang, Ke Zhou, D. Feng","doi":"10.1109/MSST.2010.5496987","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496987","url":null,"abstract":"Deduplication has been widely used in disk-based secondary storage systems to improve space efficiency. However, there are two challenges facing scalable high-throughput deduplication storage. The first is the duplicate-lookup disk bottleneck due to the large size of data index that usually exceeds the available RAM space, which limits the deduplication throughput. The second is the storage node island effect resulting from duplicate data among multiple storage nodes that are difficult to eliminate. Existing approaches fail to completely eliminate the duplicates while simultaneously addressing the challenges. This paper proposes MAD2, a scalable high-throughput exact deduplication approach for network backup services. MAD2 eliminates duplicate data both at the file level and at the chunk level by employing four techniques to accelerate the deduplication process and evenly distribute data. First, MAD2 organizes fingerprints into a Hash Bucket Matrix (HBM), whose rows can be used to preserve the data locality in backups. Second, MAD2 uses Bloom Filter Array (BFA) as a quick index to quickly identify non-duplicate incoming data objects or indicate where to find a possible duplicate. Third, Dual Cache is integrated in MAD2 to effectively capture and exploit data locality. Finally, MAD2 employs a DHT-based Load-Balance technique to evenly distribute data objects among multiple storage nodes in their backup sequences to further enhance performance with a well-balanced load. We evaluate our MAD2 approach on the backend storage of B-Cloud, a research-oriented distributed system that provides network backup services. Experimental results show that MAD2 significantly outperforms the state-of-the-art approximate deduplication approaches in terms of deduplication efficiency, supporting a deduplication throughput of at least 100MB/s for each storage component.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122485808","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 98
Delayed partial parity scheme for reliable and high-performance flash memory SSD 延迟部分奇偶校验方案为可靠和高性能闪存SSD
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496997
Soojun Im, Dongkun Shin
The I/O performances of flash memory solidstate disks (SSDs) are increasing by exploiting parallel I/O architectures. However, the reliability problem is a critical issue in building a large-scale flash storage. We propose a novel Redundant Arrays of Inexpensive Disks (RAID) architecture which uses the delayed parity update and partial parity caching techniques for reliable and high-performance flash memory SSDs. The proposed techniques improve the performance of the RAID-5 SSD by 38% and 30% on average in comparison to the original RAID-5 technique and the previous delayed parity update technique, respectively.
通过利用并行I/O架构,闪存固态硬盘(ssd)的I/O性能不断提高。然而,可靠性问题是构建大规模闪存的关键问题。我们提出了一种新的廉价磁盘冗余阵列(RAID)架构,该架构使用延迟奇偶更新和部分奇偶缓存技术来实现可靠和高性能的闪存ssd。与原始的RAID-5技术和先前的延迟奇偶更新技术相比,所提出的技术将RAID-5 SSD的性能平均提高38%和30%。
{"title":"Delayed partial parity scheme for reliable and high-performance flash memory SSD","authors":"Soojun Im, Dongkun Shin","doi":"10.1109/MSST.2010.5496997","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496997","url":null,"abstract":"The I/O performances of flash memory solidstate disks (SSDs) are increasing by exploiting parallel I/O architectures. However, the reliability problem is a critical issue in building a large-scale flash storage. We propose a novel Redundant Arrays of Inexpensive Disks (RAID) architecture which uses the delayed parity update and partial parity caching techniques for reliable and high-performance flash memory SSDs. The proposed techniques improve the performance of the RAID-5 SSD by 38% and 30% on average in comparison to the original RAID-5 technique and the previous delayed parity update technique, respectively.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"54 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127961390","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
Design issues for a shingled write disk system 瓦片式写磁盘系统的设计问题
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496991
A. Amer, D. Long, E. L. Miller, Jehan-Francois Pâris, T. Schwarz
If the data density of magnetic disks is to continue its current 30–50% annual growth, new recording techniques are required. Among the actively considered options, shingled writing is currently the most attractive one because it is the easiest to implement at the device level. Shingled write recording trades the inconvenience of the inability to update in-place for a much higher data density by a using a different write technique that overlaps the currently written track with the previous track. Random reads are still possible on such devices, but writes must be done largely sequentially. In this paper, we discuss possible changes to disk-based data structures that the adoption of shingled writing will require. We first explore disk structures that are optimized for large sequential writes with little or no sequential writing, even of metadata structures, while providing acceptable read performance. We also examine the usefulness of non-volatile RAM and the benefits of object-based interfaces in the context of shingled disks. Finally, through the analysis of recent device traces, we demonstrate the surprising stability of written device blocks, with general purpose workloads showing that more than 93% of device blocks remain unchanged over a day, and that for more specialized workloads less than 0.5% of a shingled-write disk's capacity would be needed to hold randomly updated blocks.
如果磁盘的数据密度要保持目前每年30-50%的增长,就需要新的记录技术。在积极考虑的选项中,带状写入是目前最具吸引力的选项,因为它在设备级别上最容易实现。通过使用一种不同的写入技术,将当前写入的磁道与先前的磁道重叠,Shingled写入记录避免了无法就地更新的不便,从而获得了更高的数据密度。在这样的设备上仍然可以进行随机读取,但是写操作基本上必须按顺序进行。在本文中,我们讨论了对基于磁盘的数据结构可能进行的更改,这些更改是采用打包写入所必需的。我们首先探索为大规模顺序写而很少或没有顺序写(甚至是元数据结构)优化的磁盘结构,同时提供可接受的读性能。我们还研究了非易失性RAM的有用性和基于对象的接口在拼接磁盘环境中的好处。最后,通过对最近设备跟踪的分析,我们展示了写入设备块的惊人稳定性,通用工作负载显示,超过93%的设备块在一天内保持不变,对于更专业的工作负载,只需不到0.5%的带状写入磁盘容量就可以保存随机更新的块。
{"title":"Design issues for a shingled write disk system","authors":"A. Amer, D. Long, E. L. Miller, Jehan-Francois Pâris, T. Schwarz","doi":"10.1109/MSST.2010.5496991","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496991","url":null,"abstract":"If the data density of magnetic disks is to continue its current 30–50% annual growth, new recording techniques are required. Among the actively considered options, shingled writing is currently the most attractive one because it is the easiest to implement at the device level. Shingled write recording trades the inconvenience of the inability to update in-place for a much higher data density by a using a different write technique that overlaps the currently written track with the previous track. Random reads are still possible on such devices, but writes must be done largely sequentially. In this paper, we discuss possible changes to disk-based data structures that the adoption of shingled writing will require. We first explore disk structures that are optimized for large sequential writes with little or no sequential writing, even of metadata structures, while providing acceptable read performance. We also examine the usefulness of non-volatile RAM and the benefits of object-based interfaces in the context of shingled disks. Finally, through the analysis of recent device traces, we demonstrate the surprising stability of written device blocks, with general purpose workloads showing that more than 93% of device blocks remain unchanged over a day, and that for more specialized workloads less than 0.5% of a shingled-write disk's capacity would be needed to hold randomly updated blocks.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"67 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122812542","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 75
Energy and thermal aware buffer cache replacement algorithm 能量和热感知缓冲区缓存替换算法
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496982
Jianhui Yue, Yifeng Zhu, Zhao Cai, Lin Lin
Power consumption is an increasingly impressing concern for data servers as it directly affects running costs and system reliability. Prior studies have shown most memory space on data servers are used for buffer caching and thus cache replacement becomes critical. Temporally concentrating memory accesses to a smaller set of memory chips increases the chances of free riding through DMA overlapping and also enlarges the opportunities for other ranks to power down. This paper proposes a power and thermal-aware buffer cache replacement algorithm. It conjectures that the memory rank that holds the most amount of cold blocks are very likely to be accessed in the near future. Choosing the victim block from this rank can help reduce the number of memory ranks that are active simultaneously. We use three real-world I/O server traces, including TPC-C, LM-TBF and MSN-BEFS to evaluate our algorithm. Experimental results show that our algorithm can save up to 27% energy than LRU and reduce the temperature of memory up to 5.45°C with little or no performance degradation.
功耗是数据服务器日益关注的问题,因为它直接影响运行成本和系统可靠性。先前的研究表明,数据服务器上的大部分内存空间用于缓冲缓存,因此缓存替换变得至关重要。暂时地将内存访问集中到一组较小的内存芯片上,增加了通过DMA重叠搭便车的机会,也增加了其他队列断电的机会。提出了一种功率和热感知的缓存替换算法。它推测持有最多冷块的内存等级很可能在不久的将来被访问。从这个等级中选择受害块可以帮助减少同时活动的内存等级的数量。我们使用三个真实的I/O服务器跟踪,包括TPC-C, LM-TBF和MSN-BEFS来评估我们的算法。实验结果表明,该算法可以比LRU节省27%的能量,并在性能几乎没有下降的情况下将内存温度降低到5.45°C。
{"title":"Energy and thermal aware buffer cache replacement algorithm","authors":"Jianhui Yue, Yifeng Zhu, Zhao Cai, Lin Lin","doi":"10.1109/MSST.2010.5496982","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496982","url":null,"abstract":"Power consumption is an increasingly impressing concern for data servers as it directly affects running costs and system reliability. Prior studies have shown most memory space on data servers are used for buffer caching and thus cache replacement becomes critical. Temporally concentrating memory accesses to a smaller set of memory chips increases the chances of free riding through DMA overlapping and also enlarges the opportunities for other ranks to power down. This paper proposes a power and thermal-aware buffer cache replacement algorithm. It conjectures that the memory rank that holds the most amount of cold blocks are very likely to be accessed in the near future. Choosing the victim block from this rank can help reduce the number of memory ranks that are active simultaneously. We use three real-world I/O server traces, including TPC-C, LM-TBF and MSN-BEFS to evaluate our algorithm. Experimental results show that our algorithm can save up to 27% energy than LRU and reduce the temperature of memory up to 5.45°C with little or no performance degradation.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115440079","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
A performance model and file system space allocation scheme for SSDs 介绍了ssd硬盘的性能模型和文件系统空间分配方案
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496986
Choulseung Hyun, Jongmoo Choi, Y. Oh, Donghee Lee, Eunsam Kim, S. Noh
Solid State Drives (SSDs) are now becoming a part of main stream computers. Even though disk scheduling algorithms and file systems of today have been optimized to exploit the characteristics of hard drives, relatively little attention has been paid to model and exploit the characteristics of SSDs. In this paper, we consider the use of SSDs from the file system standpoint. To do so, we derive a performance model for the SSDs. Based on this model, we devise a file system space allocation scheme, which we call Greedy-Space, for block or hybrid mapping SSDs. From the Postmark benchmark results, we observe substantial performance improvements when employing the Greedy-Space scheme in ext3 and Reiser file systems running on three SSDs available in the market.
固态硬盘(ssd)现在正在成为主流计算机的一部分。尽管今天的磁盘调度算法和文件系统已经优化到可以利用硬盘驱动器的特性,但对ssd的特性进行建模和利用的关注相对较少。在本文中,我们从文件系统的角度考虑ssd的使用。为此,我们推导了ssd的性能模型。基于这个模型,我们设计了一个文件系统空间分配方案,我们称之为贪心空间,用于块或混合映射ssd。从邮戳基准测试结果来看,我们观察到在ext3和Reiser文件系统中使用Greedy-Space方案时,性能有了很大的提高,这些文件系统运行在市场上的三块ssd上。
{"title":"A performance model and file system space allocation scheme for SSDs","authors":"Choulseung Hyun, Jongmoo Choi, Y. Oh, Donghee Lee, Eunsam Kim, S. Noh","doi":"10.1109/MSST.2010.5496986","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496986","url":null,"abstract":"Solid State Drives (SSDs) are now becoming a part of main stream computers. Even though disk scheduling algorithms and file systems of today have been optimized to exploit the characteristics of hard drives, relatively little attention has been paid to model and exploit the characteristics of SSDs. In this paper, we consider the use of SSDs from the file system standpoint. To do so, we derive a performance model for the SSDs. Based on this model, we devise a file system space allocation scheme, which we call Greedy-Space, for block or hybrid mapping SSDs. From the Postmark benchmark results, we observe substantial performance improvements when employing the Greedy-Space scheme in ext3 and Reiser file systems running on three SSDs available in the market.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125061261","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Observations made while running a multi-petabyte storage system 在运行多pb存储系统时进行的观察
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496984
M. Santos, Dennis Waldron
We take an overview of the CERN Advanced Storage (CASTOR) version 2 system and its usage at CERN while serving the High Energy Physics community. We further explore some of the observations made between 2005 and 2010 while managing this multi-petabyte distributed storage system.
我们概述了CERN高级存储(CASTOR)版本2系统及其在CERN的使用情况,同时为高能物理社区服务。我们进一步探讨了2005年至2010年间在管理这个多pb的分布式存储系统时所做的一些观察。
{"title":"Observations made while running a multi-petabyte storage system","authors":"M. Santos, Dennis Waldron","doi":"10.1109/MSST.2010.5496984","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496984","url":null,"abstract":"We take an overview of the CERN Advanced Storage (CASTOR) version 2 system and its usage at CERN while serving the High Energy Physics community. We further explore some of the observations made between 2005 and 2010 while managing this multi-petabyte distributed storage system.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126815942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Block storage listener for detecting file-level intrusions 用于检测文件级入侵的块存储监听器
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496974
M. Allalouf, Muli Ben-Yehuda, J. Satran, Itai Segall
An intrusion detection system (IDS) is usually located and operated at the host, where it captures local suspicious events, or at an appliance that listens to the network activity. Providing an online IDS to the storage controller is essential for dealing with compromised hosts or coordinated attacks by multiple hosts. SAN block storage controllers are connected to the world via block-level protocols, such as iSCSI and Fibre Channel. Usually, block-level storage systems do not maintain information specific to the file-system using them. The range of threats that can be handled at the block level is limited. A file system view at the controller, together with the knowledge of which arriving block belongs to which file or inode, will enable the detection of file-level threats. In this paper, we present IDStor, an IDS for block-based storage. IDStor acts as a listener to storage traffic, out of the controller's I/O path, and is therefore attractive for integration into existing SAN-based storage solutions. IDStor maintains a block-to-file mapping that is updated online. Using this mapping, IDStor infers the semantics of file-level commands from the intercepted block-level operations, thereby detecting file-level intrusions by merely observing the block read and write commands passing between the hosts and the controller.
入侵检测系统(IDS)通常位于主机上并在主机上操作,它在主机上捕获本地可疑事件,或者位于侦听网络活动的设备上。为存储控制器提供在线IDS对于处理受损主机或多主机协同攻击至关重要。SAN块存储控制器通过块级协议(如iSCSI和光纤通道)连接到世界。通常,块级存储系统不维护特定于使用它们的文件系统的信息。可以在块级别处理的威胁范围是有限的。控制器上的文件系统视图,以及哪个到达的块属于哪个文件或索引节点的知识,将使检测文件级威胁成为可能。在本文中,我们提出了IDStor,一个基于块的存储IDS。IDStor在控制器的I/O路径之外充当存储流量的侦听器,因此对于集成到现有的基于san的存储解决方案非常有吸引力。IDStor维护一个在线更新的块到文件的映射。使用这种映射,IDStor从截获的块级操作推断文件级命令的语义,从而仅通过观察在主机和控制器之间传递的块读写命令来检测文件级入侵。
{"title":"Block storage listener for detecting file-level intrusions","authors":"M. Allalouf, Muli Ben-Yehuda, J. Satran, Itai Segall","doi":"10.1109/MSST.2010.5496974","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496974","url":null,"abstract":"An intrusion detection system (IDS) is usually located and operated at the host, where it captures local suspicious events, or at an appliance that listens to the network activity. Providing an online IDS to the storage controller is essential for dealing with compromised hosts or coordinated attacks by multiple hosts. SAN block storage controllers are connected to the world via block-level protocols, such as iSCSI and Fibre Channel. Usually, block-level storage systems do not maintain information specific to the file-system using them. The range of threats that can be handled at the block level is limited. A file system view at the controller, together with the knowledge of which arriving block belongs to which file or inode, will enable the detection of file-level threats. In this paper, we present IDStor, an IDS for block-based storage. IDStor acts as a listener to storage traffic, out of the controller's I/O path, and is therefore attractive for integration into existing SAN-based storage solutions. IDStor maintains a block-to-file mapping that is updated online. Using this mapping, IDStor infers the semantics of file-level commands from the intercepted block-level operations, thereby detecting file-level intrusions by merely observing the block read and write commands passing between the hosts and the controller.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124005766","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Deferred updates for flash-based storage 基于闪存的存储延迟更新
Pub Date : 2010-05-03 DOI: 10.1109/MSST.2010.5496994
Biplob K. Debnath, M. Mokbel, D. Lilja, D. Du
The NAND flash memory based storage has faster read, higher power savings, and lower cooling cost compared to the conventional rotating magnetic disk drive. However, in case of flash memory, read and write operations are not symmetric. Write operations are much slower than read operations. Moreover, frequent update operations reduce the lifetime of the flash memory. Due to the faster read performance, flash-based storage is particularly attractive for the read-intensive database workloads, while it can produce poor performance when used for the update-intensive database workloads. This paper aims to improve write performance and lifetime of flash-based storage for the update-intensive workloads. In particular, we propose a new hierarchical approach named as deferred update methodology. Instead of directly updating the data records, first we buffer the changes due to update operations as logs in two intermediate in-flash layers. Next, we apply multiple update logs in bulk to the data records. Experimental results show that our proposed methodology significantly improves update processing overhead and longevity of the flash-based storages.
与传统的旋转磁盘驱动器相比,基于NAND闪存的存储具有更快的读取速度、更高的功耗和更低的冷却成本。然而,在闪存的情况下,读和写操作是不对称的。写操作比读操作慢得多。此外,频繁的更新操作降低了闪存的使用寿命。由于更快的读取性能,基于闪存的存储对于读取密集型数据库工作负载特别有吸引力,而当用于更新密集型数据库工作负载时,它可能会产生较差的性能。本文旨在为更新密集型工作负载提高基于闪存的存储的写性能和寿命。特别地,我们提出了一种新的分层方法,称为延迟更新方法。我们没有直接更新数据记录,而是首先将由于更新操作而产生的更改作为日志缓冲在两个中间的闪存层中。接下来,我们将多个更新日志批量应用于数据记录。实验结果表明,我们提出的方法显著改善了基于闪存的存储的更新处理开销和寿命。
{"title":"Deferred updates for flash-based storage","authors":"Biplob K. Debnath, M. Mokbel, D. Lilja, D. Du","doi":"10.1109/MSST.2010.5496994","DOIUrl":"https://doi.org/10.1109/MSST.2010.5496994","url":null,"abstract":"The NAND flash memory based storage has faster read, higher power savings, and lower cooling cost compared to the conventional rotating magnetic disk drive. However, in case of flash memory, read and write operations are not symmetric. Write operations are much slower than read operations. Moreover, frequent update operations reduce the lifetime of the flash memory. Due to the faster read performance, flash-based storage is particularly attractive for the read-intensive database workloads, while it can produce poor performance when used for the update-intensive database workloads. This paper aims to improve write performance and lifetime of flash-based storage for the update-intensive workloads. In particular, we propose a new hierarchical approach named as deferred update methodology. Instead of directly updating the data records, first we buffer the changes due to update operations as logs in two intermediate in-flash layers. Next, we apply multiple update logs in bulk to the data records. Experimental results show that our proposed methodology significantly improves update processing overhead and longevity of the flash-based storages.","PeriodicalId":350968,"journal":{"name":"2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2010-05-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116723587","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
期刊
2010 IEEE 26th Symposium on Mass Storage Systems and Technologies (MSST)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1