首页 > 最新文献

Proceedings of the Sixth ACM Symposium on Cloud Computing最新文献

英文 中文
Harnessing data loss with forgetful data structures 利用遗忘型数据结构控制数据丢失
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806936
A. Abouzeid, Jay Chen
Forgetting, losing, or corrupting data is almost universally considered harmful in computer science and blasphemous in database and file systems. Typically, loss of data is a consequence of unmanageable or unexpected lower layer deficiencies that the user process must be protected from through multiple layers of storage abstractions and redundancies. We argue that forgetfulness can be a resource for system design and that, like durability, security or integrity, can be used to achieve uncommon, but potentially important goals such as privacy, plausible deniability, and the right to be forgotten. We define the key properties of forgetfulness and draw inspiration from human memory. We develop a data structure, the forgit, that can be used to store images, audio files, videos or numerical data and eventually forget. Forgits are a natural data store for a multitude of today's cloud-based applications and we discuss their use, effectiveness, and limitations in this paper.
在计算机科学中,忘记、丢失或损坏数据几乎被普遍认为是有害的,在数据库和文件系统中也是亵渎神明的。通常,数据丢失是由于无法管理或意想不到的较低层缺陷造成的,必须通过多层存储抽象和冗余来保护用户进程免受这些缺陷的影响。我们认为,遗忘可以成为系统设计的资源,就像持久性、安全性或完整性一样,可以用来实现不常见但潜在重要的目标,如隐私、合理的可否认性和被遗忘权。我们定义了健忘的关键属性,并从人类记忆中汲取灵感。我们开发了一种数据结构,即遗忘器,它可以用来存储图像、音频文件、视频或数字数据,并最终忘记。遗忘是当今众多基于云的应用程序的天然数据存储,我们将在本文中讨论它们的使用、有效性和局限性。
{"title":"Harnessing data loss with forgetful data structures","authors":"A. Abouzeid, Jay Chen","doi":"10.1145/2806777.2806936","DOIUrl":"https://doi.org/10.1145/2806777.2806936","url":null,"abstract":"Forgetting, losing, or corrupting data is almost universally considered harmful in computer science and blasphemous in database and file systems. Typically, loss of data is a consequence of unmanageable or unexpected lower layer deficiencies that the user process must be protected from through multiple layers of storage abstractions and redundancies. We argue that forgetfulness can be a resource for system design and that, like durability, security or integrity, can be used to achieve uncommon, but potentially important goals such as privacy, plausible deniability, and the right to be forgotten. We define the key properties of forgetfulness and draw inspiration from human memory. We develop a data structure, the forgit, that can be used to store images, audio files, videos or numerical data and eventually forget. Forgits are a natural data store for a multitude of today's cloud-based applications and we discuss their use, effectiveness, and limitations in this paper.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127259383","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Database high availability using SHADOW systems 使用SHADOW系统的数据库高可用性
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806841
Jaemyung Kim, K. Salem, Khuzaima S. Daudjee, Ashraf Aboulnaga, Xin Pan
Hot standby techniques are widely used to implement highly available database systems. These techniques make use of two separate copies of the database, an active copy and a backup that is managed by the standby. The two database copies are stored independently and synchronized by the database systems that manage them. However, database systems deployed in computing clouds often have access to reliable persistent storage that can be shared by multiple servers. In this paper we consider how hot standby techniques can be improved in such settings. We present SHADOW systems, a novel approach to hot standby high availability. Like other database systems that use shared storage, SHADOW systems push the task of managing database replication out of the database system and into the underlying storage service, simplifying the database system. Unlike other systems, SHADOW systems also provide write offloading, which frees the active database system from the need to update the persistent database. Instead, that responsibility is placed on the standby system. We present the results of a performance evaluation using a SHADOW prototype on Amazon's cloud, showing that write offloading enables SHADOW to outperform traditional hot standby replication and even a standalone DBMS that does not provide high availability.
热备技术被广泛用于实现高可用性数据库系统。这些技术利用数据库的两个独立副本,一个是活动副本,另一个是由备用副本管理的备份。这两个数据库副本是独立存储的,并由管理它们的数据库系统同步。但是,部署在计算云中的数据库系统通常可以访问可由多个服务器共享的可靠持久存储。在本文中,我们考虑如何在这种情况下改进热备用技术。我们提出了SHADOW系统,这是一种实现热备用高可用性的新方法。与其他使用共享存储的数据库系统一样,SHADOW系统将管理数据库复制的任务从数据库系统中推到底层存储服务中,从而简化了数据库系统。与其他系统不同,SHADOW系统还提供写卸载,从而使活动数据库系统不必更新持久数据库。相反,这个责任被放在备用系统上。我们给出了在Amazon云上使用SHADOW原型进行性能评估的结果,结果显示写卸载使SHADOW优于传统的热备复制,甚至优于不提供高可用性的独立DBMS。
{"title":"Database high availability using SHADOW systems","authors":"Jaemyung Kim, K. Salem, Khuzaima S. Daudjee, Ashraf Aboulnaga, Xin Pan","doi":"10.1145/2806777.2806841","DOIUrl":"https://doi.org/10.1145/2806777.2806841","url":null,"abstract":"Hot standby techniques are widely used to implement highly available database systems. These techniques make use of two separate copies of the database, an active copy and a backup that is managed by the standby. The two database copies are stored independently and synchronized by the database systems that manage them. However, database systems deployed in computing clouds often have access to reliable persistent storage that can be shared by multiple servers. In this paper we consider how hot standby techniques can be improved in such settings. We present SHADOW systems, a novel approach to hot standby high availability. Like other database systems that use shared storage, SHADOW systems push the task of managing database replication out of the database system and into the underlying storage service, simplifying the database system. Unlike other systems, SHADOW systems also provide write offloading, which frees the active database system from the need to update the persistent database. Instead, that responsibility is placed on the standby system. We present the results of a performance evaluation using a SHADOW prototype on Amazon's cloud, showing that write offloading enables SHADOW to outperform traditional hot standby replication and even a standalone DBMS that does not provide high availability.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115196992","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 12
The nearest replica can be farther than you think 最近的复制品可能比你想象的要远
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806939
Kirill Bogdanov, Miguel Peón Quirós, Gerald Q. Maguire, Dejan Kostic
Modern distributed systems are geo-distributed for reasons of increased performance, reliability, and survivability. At the heart of many such systems, e.g., the widely used Cassandra and MongoDB data stores, is an algorithm for choosing a closest set of replicas to service a client request. Suboptimal replica choices due to dynamically changing network conditions result in reduced performance as a result of increased response latency. We present GeoPerf, a tool that tries to automate the process of systematically testing the performance of replica selection algorithms for geo-distributed storage systems. Our key idea is to combine symbolic execution and lightweight modeling to generate a set of inputs that can expose weaknesses in replica selection. As part of our evaluation, we analyzed network round trip times between geographically distributed Amazon EC2 regions, and showed a significant number of daily changes in nearest-K replica orders. We tested Cassandra and MongoDB using our tool, and found bugs in each of these systems. Finally, we use our collected Amazon EC2 latency traces to quantify the time lost due to these bugs. For example due to the bug in Cassandra, the median wasted time for 10% of all requests is above 50 ms.
现代分布式系统是地理分布式的,这是为了提高性能、可靠性和生存能力。在许多这样的系统的核心,例如,广泛使用的Cassandra和MongoDB数据存储,是一种算法,用于选择最接近的一组副本来服务客户端请求。由于动态变化的网络条件而导致的次优副本选择导致响应延迟增加,从而降低了性能。我们介绍了GeoPerf,这是一个工具,它试图自动化系统测试地理分布式存储系统的副本选择算法的性能。我们的关键思想是结合符号执行和轻量级建模来生成一组可以暴露副本选择中的弱点的输入。作为评估的一部分,我们分析了地理上分布的Amazon EC2区域之间的网络往返时间,并显示了最近k副本订单的大量每日变化。我们使用我们的工具测试了Cassandra和MongoDB,并在每个系统中发现了错误。最后,我们使用收集到的Amazon EC2延迟跟踪来量化由于这些错误而损失的时间。例如,由于Cassandra的bug, 10%的请求浪费的时间中值超过50毫秒。
{"title":"The nearest replica can be farther than you think","authors":"Kirill Bogdanov, Miguel Peón Quirós, Gerald Q. Maguire, Dejan Kostic","doi":"10.1145/2806777.2806939","DOIUrl":"https://doi.org/10.1145/2806777.2806939","url":null,"abstract":"Modern distributed systems are geo-distributed for reasons of increased performance, reliability, and survivability. At the heart of many such systems, e.g., the widely used Cassandra and MongoDB data stores, is an algorithm for choosing a closest set of replicas to service a client request. Suboptimal replica choices due to dynamically changing network conditions result in reduced performance as a result of increased response latency. We present GeoPerf, a tool that tries to automate the process of systematically testing the performance of replica selection algorithms for geo-distributed storage systems. Our key idea is to combine symbolic execution and lightweight modeling to generate a set of inputs that can expose weaknesses in replica selection. As part of our evaluation, we analyzed network round trip times between geographically distributed Amazon EC2 regions, and showed a significant number of daily changes in nearest-K replica orders. We tested Cassandra and MongoDB using our tool, and found bugs in each of these systems. Finally, we use our collected Amazon EC2 latency traces to quantify the time lost due to these bugs. For example due to the bug in Cassandra, the median wasted time for 10% of all requests is above 50 ms.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122907789","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
MemcachedGPU: scaling-up scale-out key-value stores MemcachedGPU: scale- up -out键值存储
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806836
Tayler H. Hetherington, Mike O'Connor, Tor M. Aamodt
This paper tackles the challenges of obtaining more efficient data center computing while maintaining low latency, low cost, programmability, and the potential for workload consolidation. We introduce GNoM, a software framework enabling energy-efficient, latency bandwidth optimized UDP network and application processing on GPUs. GNoM handles the data movement and task management to facilitate the development of high-throughput UDP network services on GPUs. We use GNoM to develop MemcachedGPU, an accelerated key-value store, and evaluate the full system on contemporary hardware. MemcachedGPU achieves ~10 GbE line-rate processing of ~13 million requests per second (MRPS) while delivering an efficiency of 62 thousand RPS per Watt (KRPS/W) on a high-performance GPU and 84.8 KRPS/W on a low-power GPU. This closely matches the throughput of an optimized FPGA implementation while providing up to 79% of the energy-efficiency on the low-power GPU. Additionally, the low-power GPU can potentially improve cost-efficiency (KRPS/$) up to 17% over a state-of-the-art CPU implementation. At 8 MRPS, MemcachedGPU achieves a 95-percentile RTT latency under 300μs on both GPUs. An offline limit study on the low-power GPU suggests that MemcachedGPU may continue scaling throughput and energy-efficiency up to 28.5 MRPS and 127 KRPS/W respectively.
本文解决了在保持低延迟、低成本、可编程性和工作负载整合潜力的同时获得更高效的数据中心计算的挑战。我们介绍GNoM,一个软件框架,实现节能,延迟带宽优化的UDP网络和gpu上的应用程序处理。GNoM处理数据移动和任务管理,以便在gpu上开发高吞吐量的UDP网络服务。我们使用GNoM开发MemcachedGPU(一个加速键值存储),并在现代硬件上评估整个系统。MemcachedGPU实现了~ 10gbe的线率处理,每秒~ 1300万请求(MRPS),同时在高性能GPU上提供62000 RPS/W (KRPS/W)的效率,在低功耗GPU上提供84.8 KRPS/W。这与优化FPGA实现的吞吐量非常接近,同时在低功耗GPU上提供高达79%的能效。此外,与最先进的CPU实现相比,低功耗GPU可以潜在地提高成本效率(KRPS/$)高达17%。在8 MRPS时,MemcachedGPU在两个gpu上实现了低于300μs的95百分位RTT延迟。对低功耗GPU的离线限制研究表明,MemcachedGPU可以继续扩展吞吐量和能效,分别达到28.5 MRPS和127 KRPS/W。
{"title":"MemcachedGPU: scaling-up scale-out key-value stores","authors":"Tayler H. Hetherington, Mike O'Connor, Tor M. Aamodt","doi":"10.1145/2806777.2806836","DOIUrl":"https://doi.org/10.1145/2806777.2806836","url":null,"abstract":"This paper tackles the challenges of obtaining more efficient data center computing while maintaining low latency, low cost, programmability, and the potential for workload consolidation. We introduce GNoM, a software framework enabling energy-efficient, latency bandwidth optimized UDP network and application processing on GPUs. GNoM handles the data movement and task management to facilitate the development of high-throughput UDP network services on GPUs. We use GNoM to develop MemcachedGPU, an accelerated key-value store, and evaluate the full system on contemporary hardware. MemcachedGPU achieves ~10 GbE line-rate processing of ~13 million requests per second (MRPS) while delivering an efficiency of 62 thousand RPS per Watt (KRPS/W) on a high-performance GPU and 84.8 KRPS/W on a low-power GPU. This closely matches the throughput of an optimized FPGA implementation while providing up to 79% of the energy-efficiency on the low-power GPU. Additionally, the low-power GPU can potentially improve cost-efficiency (KRPS/$) up to 17% over a state-of-the-art CPU implementation. At 8 MRPS, MemcachedGPU achieves a 95-percentile RTT latency under 300μs on both GPUs. An offline limit study on the low-power GPU suggests that MemcachedGPU may continue scaling throughput and energy-efficiency up to 28.5 MRPS and 127 KRPS/W respectively.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"95 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114817480","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 62
Automating model search for large scale machine learning 大规模机器学习的自动化模型搜索
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806945
Evan R. Sparks, Ameet Talwalkar, D. Haas, M. Franklin, Michael I. Jordan, Tim Kraska
The proliferation of massive datasets combined with the development of sophisticated analytical techniques has enabled a wide variety of novel applications such as improved product recommendations, automatic image tagging, and improved speech-driven interfaces. A major obstacle to supporting these predictive applications is the challenging and expensive process of identifying and training an appropriate predictive model. Recent efforts aiming to automate this process have focused on single node implementations and have assumed that model training itself is a black box, limiting their usefulness for applications driven by large-scale datasets. In this work, we build upon these recent efforts and propose an architecture for automatic machine learning at scale comprised of a cost-based cluster resource allocation estimator, advanced hyper-parameter tuning techniques, bandit resource allocation via runtime algorithm introspection, and physical optimization via batching and optimal resource allocation. The result is TuPAQ, a component of the MLbase system that automatically finds and trains models for a user's predictive application with comparable quality to those found using exhaustive strategies, but an order of magnitude more efficiently than the standard baseline approach. TuPAQ scales to models trained on Terabytes of data across hundreds of machines.
大量数据集的激增与复杂分析技术的发展相结合,使各种各样的新应用成为可能,例如改进的产品推荐、自动图像标记和改进的语音驱动界面。支持这些预测应用程序的主要障碍是识别和训练适当的预测模型的过程具有挑战性和昂贵。最近致力于自动化这一过程的努力主要集中在单节点实现上,并假设模型训练本身是一个黑箱,限制了它们对大规模数据集驱动的应用程序的有用性。在这项工作中,我们以这些最近的努力为基础,提出了一种大规模自动机器学习的体系结构,包括基于成本的集群资源分配估计器、先进的超参数调优技术、通过运行时算法自省进行的资源分配,以及通过批处理和最优资源分配进行的物理优化。结果是TuPAQ, MLbase系统的一个组件,可以自动为用户的预测应用程序找到和训练模型,其质量与使用穷举策略的模型相当,但比标准基线方法效率高一个数量级。TuPAQ可以扩展到数百台机器上的tb级数据训练模型。
{"title":"Automating model search for large scale machine learning","authors":"Evan R. Sparks, Ameet Talwalkar, D. Haas, M. Franklin, Michael I. Jordan, Tim Kraska","doi":"10.1145/2806777.2806945","DOIUrl":"https://doi.org/10.1145/2806777.2806945","url":null,"abstract":"The proliferation of massive datasets combined with the development of sophisticated analytical techniques has enabled a wide variety of novel applications such as improved product recommendations, automatic image tagging, and improved speech-driven interfaces. A major obstacle to supporting these predictive applications is the challenging and expensive process of identifying and training an appropriate predictive model. Recent efforts aiming to automate this process have focused on single node implementations and have assumed that model training itself is a black box, limiting their usefulness for applications driven by large-scale datasets. In this work, we build upon these recent efforts and propose an architecture for automatic machine learning at scale comprised of a cost-based cluster resource allocation estimator, advanced hyper-parameter tuning techniques, bandit resource allocation via runtime algorithm introspection, and physical optimization via batching and optimal resource allocation. The result is TuPAQ, a component of the MLbase system that automatically finds and trains models for a user's predictive application with comparable quality to those found using exhaustive strategies, but an order of magnitude more efficiently than the standard baseline approach. TuPAQ scales to models trained on Terabytes of data across hundreds of machines.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"45 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131117206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 144
Managed communication and consistency for fast data-parallel iterative analytics 管理通信和一致性,用于快速数据并行迭代分析
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806778
Jinliang Wei, Wei Dai, Aurick Qiao, Qirong Ho, Henggang Cui, G. Ganger, Phillip B. Gibbons, Garth A. Gibson, E. Xing
At the core of Machine Learning (ML) analytics is often an expert-suggested model, whose parameters are refined by iteratively processing a training dataset until convergence. The completion time (i.e. convergence time) and quality of the learned model not only depends on the rate at which the refinements are generated but also the quality of each refinement. While data-parallel ML applications often employ a loose consistency model when updating shared model parameters to maximize parallelism, the accumulated error may seriously impact the quality of refinements and thus delay completion time, a problem that usually gets worse with scale. Although more immediate propagation of updates reduces the accumulated error, this strategy is limited by physical network bandwidth. Additionally, the performance of the widely used stochastic gradient descent (SGD) algorithm is sensitive to step size. Simply increasing communication often fails to bring improvement without tuning step size accordingly and tedious hand tuning is usually needed to achieve optimal performance. This paper presents Bösen, a system that maximizes the network communication efficiency under a given inter-machine network bandwidth budget to minimize parallel error, while ensuring theoretical convergence guarantees for large-scale data-parallel ML applications. Furthermore, Bösen prioritizes messages most significant to algorithm convergence, further enhancing algorithm convergence. Finally, Bösen is the first distributed implementation of the recently presented adaptive revision algorithm, which provides orders of magnitude improvement over a carefully tuned fixed schedule of step size refinements for some SGD algorithms. Experiments on two clusters with up to 1024 cores show that our mechanism significantly improves upon static communication schedules.
机器学习(ML)分析的核心通常是专家建议的模型,其参数通过迭代处理训练数据集直到收敛来改进。学习模型的完成时间(即收敛时间)和质量不仅取决于生成细化的速率,还取决于每个细化的质量。虽然数据并行ML应用程序在更新共享模型参数以最大化并行性时通常采用松散的一致性模型,但累积的误差可能会严重影响改进的质量,从而延迟完成时间,这个问题通常会随着规模的扩大而变得更糟。虽然更直接的更新传播减少了累积的错误,但这种策略受到物理网络带宽的限制。此外,广泛使用的随机梯度下降(SGD)算法的性能对步长很敏感。如果不相应地调优步长,简单地增加通信通常无法带来改进,而且通常需要繁琐的手动调优才能实现最佳性能。本文提出了Bösen,一个在给定的机器间网络带宽预算下最大化网络通信效率以最小化并行误差的系统,同时保证了大规模数据并行ML应用的理论收敛保证。此外,Bösen对算法收敛最重要的消息进行优先级排序,进一步增强算法收敛性。最后,Bösen是最近提出的自适应修正算法的第一个分布式实现,它比一些SGD算法精心调整的固定步长优化时间表提供了数量级的改进。在两个最多1024个内核的集群上进行的实验表明,我们的机制显著改善了静态通信调度。
{"title":"Managed communication and consistency for fast data-parallel iterative analytics","authors":"Jinliang Wei, Wei Dai, Aurick Qiao, Qirong Ho, Henggang Cui, G. Ganger, Phillip B. Gibbons, Garth A. Gibson, E. Xing","doi":"10.1145/2806777.2806778","DOIUrl":"https://doi.org/10.1145/2806777.2806778","url":null,"abstract":"At the core of Machine Learning (ML) analytics is often an expert-suggested model, whose parameters are refined by iteratively processing a training dataset until convergence. The completion time (i.e. convergence time) and quality of the learned model not only depends on the rate at which the refinements are generated but also the quality of each refinement. While data-parallel ML applications often employ a loose consistency model when updating shared model parameters to maximize parallelism, the accumulated error may seriously impact the quality of refinements and thus delay completion time, a problem that usually gets worse with scale. Although more immediate propagation of updates reduces the accumulated error, this strategy is limited by physical network bandwidth. Additionally, the performance of the widely used stochastic gradient descent (SGD) algorithm is sensitive to step size. Simply increasing communication often fails to bring improvement without tuning step size accordingly and tedious hand tuning is usually needed to achieve optimal performance. This paper presents Bösen, a system that maximizes the network communication efficiency under a given inter-machine network bandwidth budget to minimize parallel error, while ensuring theoretical convergence guarantees for large-scale data-parallel ML applications. Furthermore, Bösen prioritizes messages most significant to algorithm convergence, further enhancing algorithm convergence. Finally, Bösen is the first distributed implementation of the recently presented adaptive revision algorithm, which provides orders of magnitude improvement over a carefully tuned fixed schedule of step size refinements for some SGD algorithms. Experiments on two clusters with up to 1024 cores show that our mechanism significantly improves upon static communication schedules.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"65 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114694779","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 127
SpotOn: a batch computing service for the spot market spot:现货市场的批量计算服务
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806851
S. Subramanya, Tian Guo, Prateek Sharma, David E. Irwin, P. Shenoy
Cloud spot markets enable users to bid for compute resources, such that the cloud platform may revoke them if the market price rises too high. Due to their increased risk, revocable resources in the spot market are often significantly cheaper (by as much as 10×) than the equivalent non-revocable on-demand resources. One way to mitigate spot market risk is to use various fault-tolerance mechanisms, such as checkpointing or replication, to limit the work lost on revocation. However, the additional performance overhead and cost for a particular fault-tolerance mechanism is a complex function of both an application's resource usage and the magnitude and volatility of spot market prices. We present the design of a batch computing service for the spot market, called SpotOn, that automatically selects a spot market and fault-tolerance mechanism to mitigate the impact of spot revocations without requiring application modification. SpotOn's goal is to execute jobs with the performance of on-demand resources, but at a cost near that of the spot market. We implement and evaluate SpotOn in simulation and using a prototype on Amazon's EC2 that packages jobs in Linux Containers. Our simulation results using a job trace from a Google cluster indicate that SpotOn lowers costs by 91.9% compared to using on-demand resources with little impact on performance.
云现货市场允许用户竞标计算资源,如果市场价格上涨过高,云平台可能会撤销这些资源。由于风险增加,现货市场上的可撤销资源往往比同等的不可撤销按需资源便宜得多(低10倍)。减轻现货市场风险的一种方法是使用各种容错机制,例如检查点或复制,以限制撤销时损失的工作。然而,特定容错机制的额外性能开销和成本是应用程序资源使用和现货市场价格的大小和波动性的复杂函数。我们提出了一种用于现货市场的批量计算服务的设计,称为SpotOn,它可以自动选择现货市场和容错机制,以减轻现货撤销的影响,而无需修改应用程序。SpotOn的目标是按照按需资源的性能执行作业,但成本接近现货市场。我们在模拟中实现并评估了SpotOn,并在Amazon的EC2上使用了一个原型,该原型在Linux容器中封装了作业。我们使用来自Google集群的作业跟踪的模拟结果表明,与使用按需资源相比,SpotOn降低了91.9%的成本,对性能的影响很小。
{"title":"SpotOn: a batch computing service for the spot market","authors":"S. Subramanya, Tian Guo, Prateek Sharma, David E. Irwin, P. Shenoy","doi":"10.1145/2806777.2806851","DOIUrl":"https://doi.org/10.1145/2806777.2806851","url":null,"abstract":"Cloud spot markets enable users to bid for compute resources, such that the cloud platform may revoke them if the market price rises too high. Due to their increased risk, revocable resources in the spot market are often significantly cheaper (by as much as 10×) than the equivalent non-revocable on-demand resources. One way to mitigate spot market risk is to use various fault-tolerance mechanisms, such as checkpointing or replication, to limit the work lost on revocation. However, the additional performance overhead and cost for a particular fault-tolerance mechanism is a complex function of both an application's resource usage and the magnitude and volatility of spot market prices. We present the design of a batch computing service for the spot market, called SpotOn, that automatically selects a spot market and fault-tolerance mechanism to mitigate the impact of spot revocations without requiring application modification. SpotOn's goal is to execute jobs with the performance of on-demand resources, but at a cost near that of the spot market. We implement and evaluate SpotOn in simulation and using a prototype on Amazon's EC2 that packages jobs in Linux Containers. Our simulation results using a job trace from a Google cluster indicate that SpotOn lowers costs by 91.9% compared to using on-demand resources with little impact on performance.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"11 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126403819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 111
Online parameter optimization for elastic data stream processing 弹性数据流处理的在线参数优化
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806847
Thomas S. Heinze, Lars Roediger, A. Meister, Yuanzhen Ji, Zbigniew Jerzak, C. Fetzer
Elastic scaling allows data stream processing systems to dynamically scale in and out to react to workload changes. As a consequence, unexpected load peaks can be handled and the extent of the overprovisioning can be reduced. However, the strategies used for elastic scaling of such systems need to be tuned manually by the user. This is an error prone and cumbersome task, because it requires a detailed knowledge of the underlying system and workload characteristics. In addition, the resulting quality of service for a specific scaling strategy is unknown a priori and can be measured only during runtime. In this paper we present an elastic scaling data stream processing prototype, which allows to trade off monetary cost against the offered quality of service. To that end, we use an online parameter optimization, which minimizes the monetary cost for the user. Using our prototype a user is able to specify the expected quality of service as an input to the optimization, which automatically detects significant changes of the workload pattern and adjusts the elastic scaling strategy based on the current workload characteristics. Our prototype is able to reduce the costs for three real-world use cases by 19% compared to a naive parameter setting and by 10% compared to a manually tuned system. In contrast to state of the art solutions, our system provides a stable and good trade-off between monetary cost and quality of service.
弹性伸缩允许数据流处理系统动态伸缩以响应工作负载变化。因此,可以处理意外的负载峰值,并且可以减少过度供应的程度。然而,用于此类系统弹性扩展的策略需要由用户手动调整。这是一项容易出错且繁琐的任务,因为它需要详细了解底层系统和工作负载特征。此外,特定扩展策略的结果服务质量是先验未知的,只能在运行时进行测量。在本文中,我们提出了一个弹性扩展的数据流处理原型,它允许在货币成本与提供的服务质量之间进行权衡。为此,我们使用在线参数优化,这将使用户的货币成本最小化。使用我们的原型,用户能够指定预期的服务质量作为优化的输入,优化会自动检测工作负载模式的重大变化,并根据当前工作负载特征调整弹性扩展策略。与简单的参数设置相比,我们的原型能够将三个实际用例的成本降低19%,与手动调整系统相比降低10%。与最先进的解决方案相比,我们的系统在货币成本和服务质量之间提供了稳定而良好的权衡。
{"title":"Online parameter optimization for elastic data stream processing","authors":"Thomas S. Heinze, Lars Roediger, A. Meister, Yuanzhen Ji, Zbigniew Jerzak, C. Fetzer","doi":"10.1145/2806777.2806847","DOIUrl":"https://doi.org/10.1145/2806777.2806847","url":null,"abstract":"Elastic scaling allows data stream processing systems to dynamically scale in and out to react to workload changes. As a consequence, unexpected load peaks can be handled and the extent of the overprovisioning can be reduced. However, the strategies used for elastic scaling of such systems need to be tuned manually by the user. This is an error prone and cumbersome task, because it requires a detailed knowledge of the underlying system and workload characteristics. In addition, the resulting quality of service for a specific scaling strategy is unknown a priori and can be measured only during runtime. In this paper we present an elastic scaling data stream processing prototype, which allows to trade off monetary cost against the offered quality of service. To that end, we use an online parameter optimization, which minimizes the monetary cost for the user. Using our prototype a user is able to specify the expected quality of service as an input to the optimization, which automatically detects significant changes of the workload pattern and adjusts the elastic scaling strategy based on the current workload characteristics. Our prototype is able to reduce the costs for three real-world use cases by 19% compared to a naive parameter setting and by 10% compared to a manually tuned system. In contrast to state of the art solutions, our system provides a stable and good trade-off between monetary cost and quality of service.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"55 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116126211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 73
Algebricks: a data model-agnostic compiler backend for big data languages Algebricks:用于大数据语言的数据模型无关的编译器后端
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806941
V. Borkar, Yingyi Bu, E. Carman, Nicola Onose, T. Westmann, Pouria Pirzadeh, M. Carey, V. Tsotras
A number of high-level query languages, such as Hive, Pig, Flume, and Jaql, have been developed in recent years to increase analyst productivity when processing and analyzing very large datasets. The implementation of each of these languages includes a complete, data model-dependent query compiler, yet each involves a number of similar optimizations. In this work, we describe a new query compiler architecture that separates language-specific and data model-dependent aspects from a more general query compiler backend that can generate executable data-parallel programs for shared-nothing clusters and can be used to develop multiple languages with different data models. We have built such a data model-agnostic query compiler substrate, called Algebricks, and have used it to implement three different query languages --- HiveQL, AQL, and XQuery --- to validate the efficacy of this approach. Experiments show that all three query languages benefit from the parallelization and optimization that Algebricks provides and thus have good parallel speedup and scaleup characteristics for large datasets.
近年来开发了许多高级查询语言,如Hive、Pig、Flume和Jaql,以提高分析人员在处理和分析非常大的数据集时的工作效率。每种语言的实现都包括一个完整的、依赖于数据模型的查询编译器,但每种语言都涉及许多类似的优化。在这项工作中,我们描述了一种新的查询编译器架构,它将特定于语言和数据模型相关的方面与更通用的查询编译器后端分离开来,后者可以为无共享集群生成可执行的数据并行程序,并可用于开发具有不同数据模型的多种语言。我们已经构建了这样一个与数据模型无关的查询编译器底层,称为Algebricks,并使用它来实现三种不同的查询语言——HiveQL、AQL和XQuery——以验证这种方法的有效性。实验表明,这三种查询语言都受益于Algebricks提供的并行化和优化,因此对于大型数据集具有良好的并行加速和缩放特性。
{"title":"Algebricks: a data model-agnostic compiler backend for big data languages","authors":"V. Borkar, Yingyi Bu, E. Carman, Nicola Onose, T. Westmann, Pouria Pirzadeh, M. Carey, V. Tsotras","doi":"10.1145/2806777.2806941","DOIUrl":"https://doi.org/10.1145/2806777.2806941","url":null,"abstract":"A number of high-level query languages, such as Hive, Pig, Flume, and Jaql, have been developed in recent years to increase analyst productivity when processing and analyzing very large datasets. The implementation of each of these languages includes a complete, data model-dependent query compiler, yet each involves a number of similar optimizations. In this work, we describe a new query compiler architecture that separates language-specific and data model-dependent aspects from a more general query compiler backend that can generate executable data-parallel programs for shared-nothing clusters and can be used to develop multiple languages with different data models. We have built such a data model-agnostic query compiler substrate, called Algebricks, and have used it to implement three different query languages --- HiveQL, AQL, and XQuery --- to validate the efficacy of this approach. Experiments show that all three query languages benefit from the parallelization and optimization that Algebricks provides and thus have good parallel speedup and scaleup characteristics for large datasets.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"7 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133346838","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
Software-defined caching: managing caches in multi-tenant data centers 软件定义缓存:管理多租户数据中心中的缓存
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806933
Ioan A. Stefanovici, Eno Thereska, G. O'Shea, Bianca Schroeder, Hitesh Ballani, T. Karagiannis, A. Rowstron, T. Talpey
In data centers, caches work both to provide low IO latencies and to reduce the load on the back-end network and storage. But they are not designed for multi-tenancy; system-level caches today cannot be configured to match tenant or provider objectives. Exacerbating the problem is the increasing number of un-coordinated caches on the IO data plane. The lack of global visibility on the control plane to coordinate this distributed set of caches leads to inefficiencies, increasing cloud provider cost. We present Moirai, a tenant- and workload-aware system that allows data center providers to control their distributed caching infrastructure. Moirai can help ease the management of the cache infrastructure and achieve various objectives, such as improving overall resource utilization or providing tenant isolation and QoS guarantees, as we show through several use cases. A key benefit of Moirai is that it is transparent to applications or VMs deployed in data centers. Our prototype runs unmodified OSes and databases, providing immediate benefit to existing applications.
在数据中心中,缓存既可以提供低IO延迟,又可以减少后端网络和存储的负载。但它们不是为多租户设计的;目前无法将系统级缓存配置为匹配租户或提供者的目标。使问题恶化的是IO数据平面上越来越多的不协调缓存。在控制平面上缺乏全局可见性来协调这些分布式缓存集,导致效率低下,增加了云提供商的成本。我们介绍Moirai,一个租户和工作负载感知系统,它允许数据中心提供商控制他们的分布式缓存基础设施。Moirai可以帮助简化缓存基础设施的管理,并实现各种目标,例如提高整体资源利用率或提供租户隔离和QoS保证,正如我们通过几个用例所展示的那样。Moirai的一个关键优点是,它对部署在数据中心的应用程序或虚拟机是透明的。我们的原型运行未经修改的操作系统和数据库,为现有的应用程序提供直接的好处。
{"title":"Software-defined caching: managing caches in multi-tenant data centers","authors":"Ioan A. Stefanovici, Eno Thereska, G. O'Shea, Bianca Schroeder, Hitesh Ballani, T. Karagiannis, A. Rowstron, T. Talpey","doi":"10.1145/2806777.2806933","DOIUrl":"https://doi.org/10.1145/2806777.2806933","url":null,"abstract":"In data centers, caches work both to provide low IO latencies and to reduce the load on the back-end network and storage. But they are not designed for multi-tenancy; system-level caches today cannot be configured to match tenant or provider objectives. Exacerbating the problem is the increasing number of un-coordinated caches on the IO data plane. The lack of global visibility on the control plane to coordinate this distributed set of caches leads to inefficiencies, increasing cloud provider cost. We present Moirai, a tenant- and workload-aware system that allows data center providers to control their distributed caching infrastructure. Moirai can help ease the management of the cache infrastructure and achieve various objectives, such as improving overall resource utilization or providing tenant isolation and QoS guarantees, as we show through several use cases. A key benefit of Moirai is that it is transparent to applications or VMs deployed in data centers. Our prototype runs unmodified OSes and databases, providing immediate benefit to existing applications.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114235507","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 53
期刊
Proceedings of the Sixth ACM Symposium on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1