首页 > 最新文献

2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)最新文献

英文 中文
TrustFS: An SGX-Enabled Stackable File System Framework TrustFS:一个启用sgx的可堆叠文件系统框架
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00012
Tânia Esteves, Ricardo Macedo, Alberto Faria, Bernardo Portela, J. Paulo, J. Pereira, Danny Harnik
Data confidentiality in cloud services is commonly ensured by encrypting information before uploading it. However, this approach limits the use of content-aware functionalities, such as deduplication and compression. Although this issue has been addressed individually for some of these functionalities, no unified framework for building secure storage systems exists that can leverage such operations over encrypted data. We present TrustFS, a programmable and modular stackable file system framework for implementing secure content-aware storage functionalities over hardware-assisted trusted execution environments. This framework extends the original SafeFS architecture to provide the isolated execution guarantees of Intel SGX. We demonstrate its usability by implementing an SGX-enabled stackable file system prototype while a preliminary evaluation shows that it incurs reasonable performance overhead when compared to conventional storage systems. Finally, we highlight open research challenges that must be further pursued in order for TrustFS to be fully adequate for building production-ready secure storage solutions.
云服务中的数据保密性通常通过在上传信息之前对信息进行加密来保证。但是,这种方法限制了内容感知功能的使用,例如重复数据删除和压缩。尽管已经针对其中一些功能单独解决了这个问题,但目前还没有构建安全存储系统的统一框架,可以在加密数据上利用这些操作。我们提出了TrustFS,一个可编程的模块化可堆叠文件系统框架,用于在硬件辅助的可信执行环境中实现安全的内容感知存储功能。该框架扩展了原来的saff体系结构,以提供英特尔SGX的隔离执行保证。我们通过实现一个支持sgx的可堆叠文件系统原型来证明它的可用性,而初步评估表明,与传统存储系统相比,它会产生合理的性能开销。最后,我们强调了开放的研究挑战,为了使TrustFS完全适合构建生产就绪的安全存储解决方案,必须进一步追求。
{"title":"TrustFS: An SGX-Enabled Stackable File System Framework","authors":"Tânia Esteves, Ricardo Macedo, Alberto Faria, Bernardo Portela, J. Paulo, J. Pereira, Danny Harnik","doi":"10.1109/SRDSW49218.2019.00012","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00012","url":null,"abstract":"Data confidentiality in cloud services is commonly ensured by encrypting information before uploading it. However, this approach limits the use of content-aware functionalities, such as deduplication and compression. Although this issue has been addressed individually for some of these functionalities, no unified framework for building secure storage systems exists that can leverage such operations over encrypted data. We present TrustFS, a programmable and modular stackable file system framework for implementing secure content-aware storage functionalities over hardware-assisted trusted execution environments. This framework extends the original SafeFS architecture to provide the isolated execution guarantees of Intel SGX. We demonstrate its usability by implementing an SGX-enabled stackable file system prototype while a preliminary evaluation shows that it incurs reasonable performance overhead when compared to conventional storage systems. Finally, we highlight open research challenges that must be further pursued in order for TrustFS to be fully adequate for building production-ready secure storage solutions.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129403811","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Towards A Unified Programming Model for Blockchain Smart Contract dApp Systems 区块链智能合约dApp系统的统一编程模型
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00017
J. Ellul, Gordon J. Pace
Developing smart contract decentralised application based systems typically involves writing code for various platforms, from the smart contract code residing on the underlying distributed ledger technology implementation to back end oracles and front end websites or mobile apps. In addition to the different technologies used for the different parts, the programmer is also burdened with implementing communication channels between the various parts. In this paper we propose a unified programming model allowing for developers to build such systems through a single code artifact, using a macroprogramming approach.
开发基于智能合约去中心化应用程序的系统通常涉及为各种平台编写代码,从驻留在底层分布式账本技术实现上的智能合约代码到后端预言机和前端网站或移动应用程序。除了不同部分使用的不同技术之外,程序员还负担着实现各个部分之间的通信通道。在本文中,我们提出了一个统一的编程模型,允许开发人员使用宏编程方法通过单个代码工件构建这样的系统。
{"title":"Towards A Unified Programming Model for Blockchain Smart Contract dApp Systems","authors":"J. Ellul, Gordon J. Pace","doi":"10.1109/SRDSW49218.2019.00017","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00017","url":null,"abstract":"Developing smart contract decentralised application based systems typically involves writing code for various platforms, from the smart contract code residing on the underlying distributed ledger technology implementation to back end oracles and front end websites or mobile apps. In addition to the different technologies used for the different parts, the programmer is also burdened with implementing communication channels between the various parts. In this paper we propose a unified programming model allowing for developers to build such systems through a single code artifact, using a macroprogramming approach.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131646115","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Forkmon: Monitoring the Networks Supporting Bitcoin Hard Forks Forkmon:监控支持比特币硬分叉的网络
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00016
Thiago Luiz Gontijo de Almeida, Pierre Francois, S. Frénot
In this article, we shed light on the peer-to-peer networks which support public blockchains stemming from Bitcoin forks. While the Bitcoin network has undergone a lot of attention, little has been discovered on the size, geographical spread, and general dependability of the networks supporting such forks. In this paper, we first discuss the various types of Bitcoin forks. We identify the case of hard forks which essentially consist in independent crypto-currencies that become completely separated from the reference Bitcoin network. We present a set of tools that are used to gather information on the Bitcoin forks networks. Finally, we provide preliminary analysis results regarding the size, IP layer localization, concentration, and overlap of ten Bitcoin forks.
在本文中,我们阐明了支持源自比特币分叉的公共区块链的点对点网络。虽然比特币网络受到了很多关注,但很少有人发现支持此类分叉的网络的规模、地理分布和总体可靠性。在本文中,我们首先讨论了各种类型的比特币分叉。我们确定了硬分叉的情况,硬分叉本质上由独立的加密货币组成,这些货币与参考比特币网络完全分离。我们提供了一套用于收集比特币分叉网络信息的工具。最后,我们提供了关于十个比特币分叉的大小、IP层定位、集中度和重叠的初步分析结果。
{"title":"Forkmon: Monitoring the Networks Supporting Bitcoin Hard Forks","authors":"Thiago Luiz Gontijo de Almeida, Pierre Francois, S. Frénot","doi":"10.1109/SRDSW49218.2019.00016","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00016","url":null,"abstract":"In this article, we shed light on the peer-to-peer networks which support public blockchains stemming from Bitcoin forks. While the Bitcoin network has undergone a lot of attention, little has been discovered on the size, geographical spread, and general dependability of the networks supporting such forks. In this paper, we first discuss the various types of Bitcoin forks. We identify the case of hard forks which essentially consist in independent crypto-currencies that become completely separated from the reference Bitcoin network. We present a set of tools that are used to gather information on the Bitcoin forks networks. Finally, we provide preliminary analysis results regarding the size, IP layer localization, concentration, and overlap of ten Bitcoin forks.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132668157","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Pooled Mining is Driving Blockchains Toward Centralized Systems 池挖矿正在推动区块链走向中心化系统
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00015
Liuyang Ren, Paul A. S. Ward
The decentralization property of blockchains stems from the fact that each miner accepts or refuses transactions and blocks based on its own verification results. However, pooled mining causes blockchains to evolve into centralized systems because pool participants delegate their decision-making rights to pool managers. In this paper, we established and validated a model for Proof-of-Work mining, introduced the concept of equivalent blocks, and quantitatively derived that pooling effectively lowers the income variance of miners. We also analyzed Bitcoin and Ethereum data to prove that pooled mining has become prevalent in the real world. The percentage of pool-mined blocks increased from 49.91% to 91.12% within four months in Bitcoin and from 76.9% to 92.2% within five months in Ethereum. In July 2018, Bitcoin and Ethereum mining were dominated by only six and five pools respectively.
区块链的去中心化属性源于每个矿工根据自己的验证结果接受或拒绝交易和区块。然而,池挖矿导致区块链演变成中心化系统,因为池参与者将他们的决策权委托给了池管理者。本文建立并验证了工作量证明挖矿模型,引入等效区块的概念,定量推导出池化有效降低了矿工的收入差异。我们还分析了比特币和以太坊的数据,以证明池化采矿在现实世界中已经变得普遍。在比特币中,池开采区块的比例在四个月内从49.91%增加到91.12%,在以太坊中,在五个月内从76.9%增加到92.2%。2018年7月,比特币和以太坊挖矿分别仅由6个和5个矿池主导。
{"title":"Pooled Mining is Driving Blockchains Toward Centralized Systems","authors":"Liuyang Ren, Paul A. S. Ward","doi":"10.1109/SRDSW49218.2019.00015","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00015","url":null,"abstract":"The decentralization property of blockchains stems from the fact that each miner accepts or refuses transactions and blocks based on its own verification results. However, pooled mining causes blockchains to evolve into centralized systems because pool participants delegate their decision-making rights to pool managers. In this paper, we established and validated a model for Proof-of-Work mining, introduced the concept of equivalent blocks, and quantitatively derived that pooling effectively lowers the income variance of miners. We also analyzed Bitcoin and Ethereum data to prove that pooled mining has become prevalent in the real world. The percentage of pool-mined blocks increased from 49.91% to 91.12% within four months in Bitcoin and from 76.9% to 92.2% within five months in Ethereum. In July 2018, Bitcoin and Ethereum mining were dominated by only six and five pools respectively.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123483398","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 8
A Comprehensive Rearranging Priority Based Method To Accelerate the Reconstruction of RAID Arrays 一种基于综合重排优先级的加速RAID重构方法
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00010
Xin Xie, Chentao Wu, Chao Li, Jie Li, M. Guo, Fang Xu
With the development of cloud computing, the reliability of disk arrays are increasingly concerned. Data centers usually use erasure codes to provide high reliability. However, most of reconstruction methods on disk arrays focus on single/ multiple disk(s) recovery, which ignores how to efficiently reconstruct the lost data such as Latent Sector Errors (LSEs), etc. In real situations, local stripe errors are much more common than disk failures. It has become an urgent problem that how to improve reconstruction efficiently for stripes. This paper proposes a comprehensive rearranging priority reconstruction(CRPR), which combines temporal locality, spatial locality and coding characteristics together. CRPR divides different blocks into various priorities and recovers them sequentially. To demonstrate the effectiveness of CRPR, we conduct several simulations via disksim. The simulations results show that, the comprehensive rearranging priority reconstruction method keeps up with previous methods and can save up to 63.9% in terms of waiting time.
随着云计算的发展,磁盘阵列的可靠性越来越受到人们的关注。数据中心通常使用擦除码来提供高可靠性。然而,大多数磁盘阵列的重构方法都集中在单盘/多盘恢复上,而忽略了如何有效地重构丢失的数据,如潜在扇区错误(Latent Sector error, lse)等。在实际情况下,本地分条错误比磁盘故障更常见。如何提高条纹图像的重构效率已成为一个亟待解决的问题。提出了一种综合时间局部性、空间局部性和编码特征的重排优先级重构方法。CRPR将不同的块划分为不同的优先级,并依次恢复。为了证明CRPR的有效性,我们通过disksim进行了多次仿真。仿真结果表明,综合重排优先级重建方法与以往方法基本一致,可节省63.9%的等待时间。
{"title":"A Comprehensive Rearranging Priority Based Method To Accelerate the Reconstruction of RAID Arrays","authors":"Xin Xie, Chentao Wu, Chao Li, Jie Li, M. Guo, Fang Xu","doi":"10.1109/SRDSW49218.2019.00010","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00010","url":null,"abstract":"With the development of cloud computing, the reliability of disk arrays are increasingly concerned. Data centers usually use erasure codes to provide high reliability. However, most of reconstruction methods on disk arrays focus on single/ multiple disk(s) recovery, which ignores how to efficiently reconstruct the lost data such as Latent Sector Errors (LSEs), etc. In real situations, local stripe errors are much more common than disk failures. It has become an urgent problem that how to improve reconstruction efficiently for stripes. This paper proposes a comprehensive rearranging priority reconstruction(CRPR), which combines temporal locality, spatial locality and coding characteristics together. CRPR divides different blocks into various priorities and recovers them sequentially. To demonstrate the effectiveness of CRPR, we conduct several simulations via disksim. The simulations results show that, the comprehensive rearranging priority reconstruction method keeps up with previous methods and can save up to 63.9% in terms of waiting time.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127681304","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Enabling the Efficient, Dependable Cloud-Based Storage of Human Genomes 实现高效、可靠的人类基因组云存储
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00011
V. Cogo, A. Bessani
Efficiently storing large data sets of human genomes is a long-term ambition from both the research and clinical life sciences communities. For instance, biobanks stock thousands to millions of biological physical samples and have been under pressure to store also their resulting digitized genomes. However, these and other life sciences institutions lack the infrastructure and expertise to efficiently store this data. Cloud computing is a natural economic alternative to private infrastructures, but it is not as good an alternative in terms of security and privacy. In this work, we present an end-to-end composite pipeline intended to enable the efficient, dependable cloud-based storage of human genomes by integrating three mechanisms we have recently proposed. These mechanisms encompass (1) a privacy-sensitivity detector for human genomes, (2) a similarity-based deduplication and delta-encoding algorithm for sequencing data, and (3) an auditability scheme to verify who has effectively read data in storage systems that use secure information dispersal. By integrating them with appropriate storage configurations, one can obtain reasonable privacy protection, security, and dependability guarantees at modest costs (e.g., less than $1/Genome/Year). Our preliminary analysis indicates that this pipeline costs only 3% more than non-replicated systems, 48% less than fully-replicating all data, and 31% less than secure information dispersal schemes.
有效地存储大量人类基因组数据集是研究和临床生命科学界的长期目标。例如,生物银行储存了成千上万的生物物理样本,并且一直面临着存储这些样本的数字化基因组的压力。然而,这些机构和其他生命科学机构缺乏有效存储这些数据的基础设施和专业知识。云计算是私有基础设施的一种自然的经济替代方案,但就安全性和隐私性而言,它不是一种好的替代方案。在这项工作中,我们提出了一个端到端复合管道,旨在通过整合我们最近提出的三种机制,实现高效、可靠的基于云的人类基因组存储。这些机制包括(1)针对人类基因组的隐私敏感检测器,(2)针对测序数据的基于相似性的重复数据删除和增量编码算法,以及(3)可审计方案,以验证谁有效地读取了使用安全信息分散的存储系统中的数据。通过将它们与适当的存储配置集成,人们可以以适度的成本(例如,低于1美元/基因组/年)获得合理的隐私保护、安全性和可靠性保证。我们的初步分析表明,这种管道的成本仅比非复制系统高3%,比完全复制所有数据的系统低48%,比安全信息分散方案低31%。
{"title":"Enabling the Efficient, Dependable Cloud-Based Storage of Human Genomes","authors":"V. Cogo, A. Bessani","doi":"10.1109/SRDSW49218.2019.00011","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00011","url":null,"abstract":"Efficiently storing large data sets of human genomes is a long-term ambition from both the research and clinical life sciences communities. For instance, biobanks stock thousands to millions of biological physical samples and have been under pressure to store also their resulting digitized genomes. However, these and other life sciences institutions lack the infrastructure and expertise to efficiently store this data. Cloud computing is a natural economic alternative to private infrastructures, but it is not as good an alternative in terms of security and privacy. In this work, we present an end-to-end composite pipeline intended to enable the efficient, dependable cloud-based storage of human genomes by integrating three mechanisms we have recently proposed. These mechanisms encompass (1) a privacy-sensitivity detector for human genomes, (2) a similarity-based deduplication and delta-encoding algorithm for sequencing data, and (3) an auditability scheme to verify who has effectively read data in storage systems that use secure information dispersal. By integrating them with appropriate storage configurations, one can obtain reasonable privacy protection, security, and dependability guarantees at modest costs (e.g., less than $1/Genome/Year). Our preliminary analysis indicates that this pipeline costs only 3% more than non-replicated systems, 48% less than fully-replicating all data, and 31% less than secure information dispersal schemes.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127183983","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Evaluation and Ranking of Replica Deployments in Geographic State Machine Replication 地理状态机复制中副本部署的评估和排序
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00014
Shota Numakura, Junya Nakamura, Ren Ohmura
Geographic state machine replication (SMR) is a replication method in which replicas of a service are located on multiple continents to improve the fault tolerance of a general service. Nowadays, geographic SMR is easily realized using public cloud services; SMR provides extraordinary resilience against catastrophic disasters. Previous studies have revealed that the geographic distribution of the replicas has a significant influence on the performance of the geographic SMR; however, the optimal way for a system integrator to deploy replicas remains unknown. In this paper, we propose a method to evaluate and rank replica deployments to assist a system integrator in deciding a final replica deployment. In the method, we also propose a novel evaluation function that estimates a latency of SMR protocols with round-trip time (RTT). To demonstrate the effectiveness of the proposed method, we build thousands of geographic SMRs on Amazon Web Services and present experimental results. The results show that the proposed method that estimates a latency based on RTTs can generate consistent rankings with reasonable calculation time.
地理状态机复制(SMR)是一种复制方法,其中服务的副本位于多个大洲,以提高一般服务的容错性。如今,地理SMR很容易通过公共云服务实现;SMR提供了非凡的抗灾能力。已有研究表明,副本的地理分布对地理SMR的绩效有显著影响;然而,系统集成商部署副本的最佳方式仍然未知。在本文中,我们提出了一种评估和排序副本部署的方法,以帮助系统集成商决定最终的副本部署。在该方法中,我们还提出了一个新的评估函数,该函数可以用往返时间(RTT)来估计SMR协议的延迟。为了证明该方法的有效性,我们在Amazon Web Services上构建了数千个地理smr,并给出了实验结果。结果表明,基于rtt估计时延的方法可以在合理的计算时间内生成一致的排名。
{"title":"Evaluation and Ranking of Replica Deployments in Geographic State Machine Replication","authors":"Shota Numakura, Junya Nakamura, Ren Ohmura","doi":"10.1109/SRDSW49218.2019.00014","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00014","url":null,"abstract":"Geographic state machine replication (SMR) is a replication method in which replicas of a service are located on multiple continents to improve the fault tolerance of a general service. Nowadays, geographic SMR is easily realized using public cloud services; SMR provides extraordinary resilience against catastrophic disasters. Previous studies have revealed that the geographic distribution of the replicas has a significant influence on the performance of the geographic SMR; however, the optimal way for a system integrator to deploy replicas remains unknown. In this paper, we propose a method to evaluate and rank replica deployments to assist a system integrator in deciding a final replica deployment. In the method, we also propose a novel evaluation function that estimates a latency of SMR protocols with round-trip time (RTT). To demonstrate the effectiveness of the proposed method, we build thousands of geographic SMRs on Amazon Web Services and present experimental results. The results show that the proposed method that estimates a latency based on RTTs can generate consistent rankings with reasonable calculation time.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"6 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125165048","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
On the Impact of log Compaction on Incrementally Checkpointing Stateful Stream-Processing Operators 日志压缩对增量检查点状态流处理操作符的影响
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00008
Aris Chronarakis, Antonis Papaioannou, K. Magoutis
Incremental checkpointing (IC) is a fault-tolerance technique used in several stateful distributed stream processing systems. It relies on continuously logging state updates to a remote storage service and periodically compacting the update-log via a background process. We highlight a tradeoff between the intensity of compaction of the IC update-log (and the associated resource overhead) and its impact on recovery time in such systems. We also highlight the control parameters that can be used to adjust this tradeoff in the Apache Samza stream processing system, and demonstrate this tradeoff experimentally.
增量检查点(IC)是几种有状态分布式流处理系统中使用的容错技术。它依赖于连续地将状态更新记录到远程存储服务,并通过后台进程定期压缩更新日志。我们强调在IC更新日志的压缩强度(以及相关的资源开销)及其对此类系统中恢复时间的影响之间进行权衡。我们还强调了在Apache Samza流处理系统中可用于调整这种权衡的控制参数,并通过实验演示了这种权衡。
{"title":"On the Impact of log Compaction on Incrementally Checkpointing Stateful Stream-Processing Operators","authors":"Aris Chronarakis, Antonis Papaioannou, K. Magoutis","doi":"10.1109/SRDSW49218.2019.00008","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00008","url":null,"abstract":"Incremental checkpointing (IC) is a fault-tolerance technique used in several stateful distributed stream processing systems. It relies on continuously logging state updates to a remote storage service and periodically compacting the update-log via a background process. We highlight a tradeoff between the intensity of compaction of the IC update-log (and the associated resource overhead) and its impact on recovery time in such systems. We also highlight the control parameters that can be used to adjust this tradeoff in the Apache Samza stream processing system, and demonstrate this tradeoff experimentally.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"63 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127240107","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
A Cost-Effective Cloud Event Archival for SIEMs 面向SIEMs的经济高效的云事件归档
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00013
Adriano Serckumecka, Ibéria Medeiros, Bernardo Ferreira
Security Information and Event Management (SIEM) systems have been adopted by organizations to enable holistic monitoring of malicious activities in their IT infrastructures. SIEMs receive events from diverse devices of the organization's IT infrastructure (e.g., servers, firewalls, IDS), correlate these events, and present reports for security analysts. Given the large number of events collected by SIEMs, it is costly to store such data for long periods. Since organizations store a relatively limited time-frame of events, the forensic analysis capabilities severely become reduced. This concern limits the organizations' ability to store important information about the past cybersecurity-related activity, limiting forensic analysis. A possible solution for this issue is to leverage public cloud storage services, exploiting their low cost and "infinite" scalability. We present SLiCER an archival system for long-term storage that makes use of a multi-cloud-based storage system to guarantee data security and ensures cost-effectiveness by grouping events in blocks and using indexing techniques to recover them. The system was evaluated using a real dataset and the results show that it is significantly more cost-efficient than competing alternatives.
组织已经采用了安全信息和事件管理(SIEM)系统来全面监控其IT基础设施中的恶意活动。siem从组织的IT基础设施的不同设备(例如,服务器、防火墙、IDS)接收事件,关联这些事件,并为安全分析师提供报告。考虑到siem收集的大量事件,长时间存储这些数据的成本很高。由于组织存储相对有限的事件时间框架,因此取证分析能力严重降低。这种担忧限制了组织存储有关过去网络安全相关活动的重要信息的能力,限制了取证分析。这个问题的一个可能的解决方案是利用公共云存储服务,利用它们的低成本和“无限”的可扩展性。我们为SLiCER提供了一个用于长期存储的归档系统,该系统利用基于多云的存储系统来保证数据安全,并通过将事件分组在块中并使用索引技术来恢复它们来确保成本效益。使用真实数据集对该系统进行了评估,结果表明该系统比竞争方案具有更高的成本效益。
{"title":"A Cost-Effective Cloud Event Archival for SIEMs","authors":"Adriano Serckumecka, Ibéria Medeiros, Bernardo Ferreira","doi":"10.1109/SRDSW49218.2019.00013","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00013","url":null,"abstract":"Security Information and Event Management (SIEM) systems have been adopted by organizations to enable holistic monitoring of malicious activities in their IT infrastructures. SIEMs receive events from diverse devices of the organization's IT infrastructure (e.g., servers, firewalls, IDS), correlate these events, and present reports for security analysts. Given the large number of events collected by SIEMs, it is costly to store such data for long periods. Since organizations store a relatively limited time-frame of events, the forensic analysis capabilities severely become reduced. This concern limits the organizations' ability to store important information about the past cybersecurity-related activity, limiting forensic analysis. A possible solution for this issue is to leverage public cloud storage services, exploiting their low cost and \"infinite\" scalability. We present SLiCER an archival system for long-term storage that makes use of a multi-cloud-based storage system to guarantee data security and ensures cost-effectiveness by grouping events in blocks and using indexing techniques to recover them. The system was evaluated using a real dataset and the results show that it is significantly more cost-efficient than competing alternatives.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129242754","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
A Case for Dynamically Programmable Storage Background Tasks 动态可编程存储后台任务的案例
Pub Date : 2019-10-01 DOI: 10.1109/SRDSW49218.2019.00009
Ricardo Macedo, Alberto Faria, J. Paulo, J. Pereira
Modern storage infrastructures feature long and complicated I/O paths composed of several layers, each employing their own optimizations to serve varied applications with fluctuating requirements. However, as these layers do not have global infrastructure visibility, they are unable to optimally tune their behavior to achieve maximum performance. Background storage tasks, in particular, can rapidly overload shared resources, but are executed either periodically or whenever a certain threshold is hit regardless of the overall load on the system. In this paper, we argue that to achieve optimal holistic performance, these tasks should be dynamically programmable and handled by a controller with global visibility. To support this argument, we evaluate the impact on performance of compaction and checkpointing in the context of HBase and PostgreSQL. We find that these tasks can respectively increase 99th percentile latencies by 955.2% and 61.9%. We also identify future research directions to achieve programmable background tasks.
现代存储基础设施的特点是由多层组成的长而复杂的I/O路径,每层都采用自己的优化来满足需求波动的各种应用程序。然而,由于这些层不具有全局基础设施可见性,因此它们无法优化其行为以实现最大性能。特别是后台存储任务,它可以快速地使共享资源过载,但无论系统的总体负载如何,都会定期或在达到某个阈值时执行。在本文中,我们认为为了达到最佳的整体性能,这些任务应该是动态可编程的,并由具有全局可见性的控制器处理。为了支持这个论点,我们评估了HBase和PostgreSQL上下文中压缩和检查点对性能的影响。我们发现这些任务可以分别使第99百分位延迟增加955.2%和61.9%。我们还确定了实现可编程后台任务的未来研究方向。
{"title":"A Case for Dynamically Programmable Storage Background Tasks","authors":"Ricardo Macedo, Alberto Faria, J. Paulo, J. Pereira","doi":"10.1109/SRDSW49218.2019.00009","DOIUrl":"https://doi.org/10.1109/SRDSW49218.2019.00009","url":null,"abstract":"Modern storage infrastructures feature long and complicated I/O paths composed of several layers, each employing their own optimizations to serve varied applications with fluctuating requirements. However, as these layers do not have global infrastructure visibility, they are unable to optimally tune their behavior to achieve maximum performance. Background storage tasks, in particular, can rapidly overload shared resources, but are executed either periodically or whenever a certain threshold is hit regardless of the overall load on the system. In this paper, we argue that to achieve optimal holistic performance, these tasks should be dynamically programmable and handled by a controller with global visibility. To support this argument, we evaluate the impact on performance of compaction and checkpointing in the context of HBase and PostgreSQL. We find that these tasks can respectively increase 99th percentile latencies by 955.2% and 61.9%. We also identify future research directions to achieve programmable background tasks.","PeriodicalId":297328,"journal":{"name":"2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-10-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123135527","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
期刊
2019 38th International Symposium on Reliable Distributed Systems Workshops (SRDSW)
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1