首页 > 最新文献

Proceedings of the 2017 Symposium on Cloud Computing最新文献

英文 中文
BestConfig: tapping the performance potential of systems via automatic configuration tuning BestConfig:通过自动配置调优挖掘系统的性能潜力
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128605
Yuqing Zhu, Jianxun Liu, Mengying Guo, Yungang Bao, Wenlong Ma, Zhuoyue Liu, Kunpeng Song, Y. Yang
An ever increasing number of configuration parameters are provided to system users. But many users have used one configuration setting across different workloads, leaving untapped the performance potential of systems. A good configuration setting can greatly improve the performance of a deployed system under certain workloads. But with tens or hundreds of parameters, it becomes a highly costly task to decide which configuration setting leads to the best performance. While such task requires the strong expertise in both the system and the application, users commonly lack such expertise. To help users tap the performance potential of systems, we present Best Config, a system for automatically finding a best configuration setting within a resource limit for a deployed system under a given application workload. BestConfig is designed with an extensible architecture to automate the configuration tuning for general systems. To tune system configurations within a resource limit, we propose the divide-and-diverge sampling method and the recursive bound-and-search algorithm. BestConfig can improve the throughput of Tomcat by 75%, that of Cassandra by 63%, that of MySQL by 430%, and reduce the running time of Hive join job by about 50% and that of Spark join job by about 80%, solely by configuration adjustment.
越来越多的配置参数被提供给系统用户。但是,许多用户在不同的工作负载中使用一个配置设置,从而没有充分利用系统的性能潜力。良好的配置设置可以极大地提高部署系统在某些工作负载下的性能。但是,由于有数十或数百个参数,决定哪种配置设置可以带来最佳性能成为一项代价高昂的任务。虽然这样的任务需要系统和应用程序方面的专业知识,但用户通常缺乏这样的专业知识。为了帮助用户挖掘系统的性能潜力,我们提供了Best Config,这是一个在给定应用程序工作负载下的已部署系统的资源限制内自动查找最佳配置设置的系统。BestConfig设计了一个可扩展的体系结构,可以自动对一般系统进行配置调优。为了在有限的资源范围内优化系统配置,我们提出了分散采样方法和递归定界搜索算法。仅通过配置调整,BestConfig就可以使Tomcat的吞吐量提高75%,Cassandra的吞吐量提高63%,MySQL的吞吐量提高430%,Hive join job的运行时间减少约50%,Spark join job的运行时间减少约80%。
{"title":"BestConfig: tapping the performance potential of systems via automatic configuration tuning","authors":"Yuqing Zhu, Jianxun Liu, Mengying Guo, Yungang Bao, Wenlong Ma, Zhuoyue Liu, Kunpeng Song, Y. Yang","doi":"10.1145/3127479.3128605","DOIUrl":"https://doi.org/10.1145/3127479.3128605","url":null,"abstract":"An ever increasing number of configuration parameters are provided to system users. But many users have used one configuration setting across different workloads, leaving untapped the performance potential of systems. A good configuration setting can greatly improve the performance of a deployed system under certain workloads. But with tens or hundreds of parameters, it becomes a highly costly task to decide which configuration setting leads to the best performance. While such task requires the strong expertise in both the system and the application, users commonly lack such expertise. To help users tap the performance potential of systems, we present Best Config, a system for automatically finding a best configuration setting within a resource limit for a deployed system under a given application workload. BestConfig is designed with an extensible architecture to automate the configuration tuning for general systems. To tune system configurations within a resource limit, we propose the divide-and-diverge sampling method and the recursive bound-and-search algorithm. BestConfig can improve the throughput of Tomcat by 75%, that of Cassandra by 63%, that of MySQL by 430%, and reduce the running time of Hive join job by about 50% and that of Spark join job by about 80%, solely by configuration adjustment.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"34 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75006110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 165
SQML: large-scale in-database machine learning with pure SQL SQL:使用纯SQL的大规模数据库内机器学习
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132746
Umar Syed, Sergei Vassilvitskii
Many enterprises have migrated their data from an on-site database to a cloud-based database-as-a-service that handles all database-related administrative tasks while providing a simple SQL interface to the end user. Businesses are also increasingly relying on machine learning to understand their customers and develop new products. Given these converging trends, there is a pressing need for database-as-a-service providers to add support for sophisticated machine learning algorithms to the core functionality of their products.
许多企业已将其数据从现场数据库迁移到基于云的数据库即服务,该服务处理所有与数据库相关的管理任务,同时向最终用户提供简单的SQL接口。企业也越来越依赖机器学习来了解客户和开发新产品。鉴于这些趋同的趋势,数据库即服务提供商迫切需要在其产品的核心功能中添加对复杂机器学习算法的支持。
{"title":"SQML: large-scale in-database machine learning with pure SQL","authors":"Umar Syed, Sergei Vassilvitskii","doi":"10.1145/3127479.3132746","DOIUrl":"https://doi.org/10.1145/3127479.3132746","url":null,"abstract":"Many enterprises have migrated their data from an on-site database to a cloud-based database-as-a-service that handles all database-related administrative tasks while providing a simple SQL interface to the end user. Businesses are also increasingly relying on machine learning to understand their customers and develop new products. Given these converging trends, there is a pressing need for database-as-a-service providers to add support for sophisticated machine learning algorithms to the core functionality of their products.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"95 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73625262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Sketches of space: ownership accounting for shared storage 空间草图:共享存储的所有权
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132021
Jake Wires, P. Ganesan, A. Warfield
Efficient snapshots are an important feature of modern storage systems. However, the implicit sharing underlying most snapshot implementations makes it difficult to answer basic questions about the storage costs of individual snapshots. Traditional techniques for answering these questions incur significant performance penalties due to expensive metadata overheads. We present a novel probabilistic data structure, compatible with existing storage systems, that can provide approximate answers about snapshot costs with very low computational and storage overheads while achieving better than 95% accuracy for real-world data sets.
高效快照是现代存储系统的一个重要特性。然而,大多数快照实现的隐式共享使得很难回答有关单个快照的存储成本的基本问题。由于昂贵的元数据开销,回答这些问题的传统技术会导致显著的性能损失。我们提出了一种新的概率数据结构,与现有的存储系统兼容,可以以非常低的计算和存储开销提供关于快照成本的近似答案,同时对现实世界的数据集实现95%以上的准确率。
{"title":"Sketches of space: ownership accounting for shared storage","authors":"Jake Wires, P. Ganesan, A. Warfield","doi":"10.1145/3127479.3132021","DOIUrl":"https://doi.org/10.1145/3127479.3132021","url":null,"abstract":"Efficient snapshots are an important feature of modern storage systems. However, the implicit sharing underlying most snapshot implementations makes it difficult to answer basic questions about the storage costs of individual snapshots. Traditional techniques for answering these questions incur significant performance penalties due to expensive metadata overheads. We present a novel probabilistic data structure, compatible with existing storage systems, that can provide approximate answers about snapshot costs with very low computational and storage overheads while achieving better than 95% accuracy for real-world data sets.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75382740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
UNO: uniflying host and smart NIC offload for flexible packet processing UNO:统一主机和智能网卡卸载,灵活处理报文
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132252
Yanfang Le, Hyunseok Chang, S. Mukherjee, Limin Wang, Aditya Akella, M. Swift, T. V. Lakshman
Increasingly, smart Network Interface Cards (sNICs) are being used in data centers to offload networking functions (NFs) from host processors thereby making these processors available for tenant applications. Modern sNICs have fully programmable, energy-efficient multi-core processors on which many packet processing functions, including a full-blown programmable switch, can run. However, having multiple switch instances deployed across the host hypervisor and the attached sNICs makes controlling them difficult and data plane operations more complex. This paper proposes a generalized SDN-controlled NF offload architecture called UNO. It can transparently offload dynamically selected host processors' packet processing functions to sNICs by using multiple switches in the host while keeping the data centerwide network control and management planes unmodified. UNO exposes a single virtual control plane to the SDN controller and hides dynamic NF offload behind a unified virtual management plane. This enables UNO to make optimal use of host's and sNIC's combined packet processing capabilities with local optimization based on locally observed traffic patterns and resource consumption, and without central controller involvement. Experimental results based on a real UNO prototype in realistic scenarios show promising results: it can save processing worth up to 8 CPU cores, reduce power usage by up to 2x, and reduce the control plane overhead by more than 50%.
数据中心中越来越多地使用智能网络接口卡(snic)从主机处理器卸载网络功能(NFs),从而使这些处理器可用于租户应用程序。现代snic具有完全可编程的、节能的多核处理器,可以在其上运行许多包处理功能,包括一个成熟的可编程交换机。但是,跨主机管理程序和附加的snic部署多个交换机实例使得控制它们变得困难,数据平面操作变得更加复杂。本文提出了一种通用的sdn控制的NF卸载体系结构UNO。它可以在保持数据中心范围的网络控制和管理平面不变的情况下,通过使用主机中的多个交换机,透明地将所选主机处理器的数据包处理功能动态卸载到snic上。UNO将单个虚拟控制平面暴露给SDN控制器,将NF动态卸载隐藏在统一的虚拟管理平面之后。这使UNO能够在没有中央控制器参与的情况下,根据本地观察到的流量模式和资源消耗进行本地优化,最优地利用主机和sNIC的组合数据包处理能力。基于真实UNO原型在现实场景中的实验结果显示出令人鼓舞的结果:它可以节省多达8个CPU核心的处理,减少高达2倍的功耗,并将控制平面开销降低50%以上。
{"title":"UNO: uniflying host and smart NIC offload for flexible packet processing","authors":"Yanfang Le, Hyunseok Chang, S. Mukherjee, Limin Wang, Aditya Akella, M. Swift, T. V. Lakshman","doi":"10.1145/3127479.3132252","DOIUrl":"https://doi.org/10.1145/3127479.3132252","url":null,"abstract":"Increasingly, smart Network Interface Cards (sNICs) are being used in data centers to offload networking functions (NFs) from host processors thereby making these processors available for tenant applications. Modern sNICs have fully programmable, energy-efficient multi-core processors on which many packet processing functions, including a full-blown programmable switch, can run. However, having multiple switch instances deployed across the host hypervisor and the attached sNICs makes controlling them difficult and data plane operations more complex. This paper proposes a generalized SDN-controlled NF offload architecture called UNO. It can transparently offload dynamically selected host processors' packet processing functions to sNICs by using multiple switches in the host while keeping the data centerwide network control and management planes unmodified. UNO exposes a single virtual control plane to the SDN controller and hides dynamic NF offload behind a unified virtual management plane. This enables UNO to make optimal use of host's and sNIC's combined packet processing capabilities with local optimization based on locally observed traffic patterns and resource consumption, and without central controller involvement. Experimental results based on a real UNO prototype in realistic scenarios show promising results: it can save processing worth up to 8 CPU cores, reduce power usage by up to 2x, and reduce the control plane overhead by more than 50%.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"42 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80942035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
FSP: towards flexible synchronous parallel framework for expectation-maximization based algorithms on cloud FSP:面向云上基于期望最大化算法的灵活同步并行框架
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128612
Zhigang Wang, Lixin Gao, Yu Gu, Y. Bao, Ge Yu
Myriad of parameter estimation algorithms can be performed by an Expectation-Maximization (EM) approach. Traditional synchronous frameworks can parallelize these EM algorithms on the cloud to accelerate computation while guaranteeing the convergence. However, expensive synchronization costs pose great challenges for efficiency. Asynchronous solutions have been recently designed to bypass high-cost synchronous barriers but at expense of potentially losing convergence guarantee. This paper first proposes a flexible synchronous parallel framework (FSP) that provides the capability of synchronous EM algorithms implementations, as well as significantly reduces the barrier cost. Under FSP, every distributed worker can immediately suspend local computation when necessary, to quickly synchronize with each other. That maximizes the time fast workers spend doing useful work, instead of waiting for slow, straggling workers. We then formally prove the algorithm convergence. Further, we analyze how to automatically identify a proper barrier interval to strike a nice balance between reduced synchronization costs and the convergence speed. Empirical results demonstrate that on a broad spectrum of real-world and synthetic datasets, FSP achieves as much as 3x speedup over the up-to-date synchronous solution.
期望最大化(EM)方法可以实现无数的参数估计算法。传统的同步框架可以在云上并行处理这些EM算法,在保证收敛性的同时加快计算速度。然而,昂贵的同步成本给效率带来了巨大的挑战。异步解决方案最近被设计为绕过高成本的同步障碍,但代价是可能失去收敛保证。本文首先提出了一种灵活的同步并行框架(FSP),该框架提供了同步EM算法实现的能力,并显著降低了屏障成本。在FSP下,每个分布式worker可以在必要时立即暂停本地计算,以快速相互同步。这将使速度快的工人花在做有用工作上的时间最大化,而不是等待速度慢、行动迟缓的工人。然后正式证明了算法的收敛性。此外,我们还分析了如何自动识别适当的屏障间隔,以在降低同步成本和收敛速度之间取得良好的平衡。经验结果表明,在广泛的现实世界和合成数据集上,FSP比最新的同步解决方案实现了多达3倍的加速。
{"title":"FSP: towards flexible synchronous parallel framework for expectation-maximization based algorithms on cloud","authors":"Zhigang Wang, Lixin Gao, Yu Gu, Y. Bao, Ge Yu","doi":"10.1145/3127479.3128612","DOIUrl":"https://doi.org/10.1145/3127479.3128612","url":null,"abstract":"Myriad of parameter estimation algorithms can be performed by an Expectation-Maximization (EM) approach. Traditional synchronous frameworks can parallelize these EM algorithms on the cloud to accelerate computation while guaranteeing the convergence. However, expensive synchronization costs pose great challenges for efficiency. Asynchronous solutions have been recently designed to bypass high-cost synchronous barriers but at expense of potentially losing convergence guarantee. This paper first proposes a flexible synchronous parallel framework (FSP) that provides the capability of synchronous EM algorithms implementations, as well as significantly reduces the barrier cost. Under FSP, every distributed worker can immediately suspend local computation when necessary, to quickly synchronize with each other. That maximizes the time fast workers spend doing useful work, instead of waiting for slow, straggling workers. We then formally prove the algorithm convergence. Further, we analyze how to automatically identify a proper barrier interval to strike a nice balance between reduced synchronization costs and the convergence speed. Empirical results demonstrate that on a broad spectrum of real-world and synthetic datasets, FSP achieves as much as 3x speedup over the up-to-date synchronous solution.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"18 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78047355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
An implementation of fast memset() using hardware accelerators: extended abstract 使用硬件加速器的快速memset()实现:扩展抽象
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132573
K. Pusukuri, R. Gardner, Jared C. Smolens
Multicore systems with large caches and huge main memories have become ubiquitous. They provide an attractive opportunity to maximize performance of big-memory applications such as in-memory databases, key-value stores, and graph analytics. However, these big-memory applications require many virtual-to-physical address translations, which increase TLB miss rate and hurt performance. To address this problem, modern hardware and OSes introduced support for huge pages. For example, on SPARC M7, Linux supports 8MB, 2GB, and 16GB huge pages (in addition to the default 8KB). Likewise, Linux supports 2MB and 1GB huge pages on Intel Xeon (E5-2630) platforms.
具有大缓存和大内存的多核系统已经变得无处不在。它们为最大化大内存应用程序(如内存数据库、键值存储和图形分析)的性能提供了一个有吸引力的机会。然而,这些大内存应用程序需要许多虚拟到物理地址的转换,这会增加TLB失误率并损害性能。为了解决这个问题,现代硬件和操作系统引入了对大页面的支持。例如,在SPARC M7上,Linux支持8MB、2GB和16GB的大页面(除了默认的8KB之外)。同样,Linux在Intel Xeon (E5-2630)平台上支持2MB和1GB的大页面。
{"title":"An implementation of fast memset() using hardware accelerators: extended abstract","authors":"K. Pusukuri, R. Gardner, Jared C. Smolens","doi":"10.1145/3127479.3132573","DOIUrl":"https://doi.org/10.1145/3127479.3132573","url":null,"abstract":"Multicore systems with large caches and huge main memories have become ubiquitous. They provide an attractive opportunity to maximize performance of big-memory applications such as in-memory databases, key-value stores, and graph analytics. However, these big-memory applications require many virtual-to-physical address translations, which increase TLB miss rate and hurt performance. To address this problem, modern hardware and OSes introduced support for huge pages. For example, on SPARC M7, Linux supports 8MB, 2GB, and 16GB huge pages (in addition to the default 8KB). Likewise, Linux supports 2MB and 1GB huge pages on Intel Xeon (E5-2630) platforms.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"24 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76943924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Remote memory in the age of fast networks 高速网络时代的远程存储器
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131612
M. Aguilera, Nadav Amit, I. Calciu, Xavier Deguillard, Jayneel Gandhi, Pratap Subrahmanyam, L. Suresh, K. Tati, Rajesh Venkatasubramanian, M. Wei
As the latency of the network approaches that of memory, it becomes increasingly attractive for applications to use remote memory---random-access memory at another computer that is accessed using the virtual memory subsystem. This is an old idea whose time has come, in the age of fast networks. To work effectively, remote memory must address many technical challenges. In this paper, we enumerate these challenges, discuss their feasibility, explain how some of them are addressed by recent work, and indicate other promising ways to tackle them. Some challenges remain as open problems, while others deserve more study. In this paper, we hope to provide a broad research agenda around this topic, by proposing more problems than solutions.
随着网络延迟接近内存延迟,应用程序越来越倾向于使用远程内存——使用虚拟内存子系统访问另一台计算机上的随机访问内存。这是一个古老的想法,在快速网络时代,它的时代已经到来。为了有效地工作,远程内存必须解决许多技术挑战。在本文中,我们列举了这些挑战,讨论了它们的可行性,解释了其中一些是如何通过最近的工作来解决的,并指出了其他有希望的解决方法。有些挑战仍然是悬而未决的问题,而另一些则值得进一步研究。在本文中,我们希望通过提出更多的问题而不是解决方案,围绕这一主题提供一个广泛的研究议程。
{"title":"Remote memory in the age of fast networks","authors":"M. Aguilera, Nadav Amit, I. Calciu, Xavier Deguillard, Jayneel Gandhi, Pratap Subrahmanyam, L. Suresh, K. Tati, Rajesh Venkatasubramanian, M. Wei","doi":"10.1145/3127479.3131612","DOIUrl":"https://doi.org/10.1145/3127479.3131612","url":null,"abstract":"As the latency of the network approaches that of memory, it becomes increasingly attractive for applications to use remote memory---random-access memory at another computer that is accessed using the virtual memory subsystem. This is an old idea whose time has come, in the age of fast networks. To work effectively, remote memory must address many technical challenges. In this paper, we enumerate these challenges, discuss their feasibility, explain how some of them are addressed by recent work, and indicate other promising ways to tackle them. Some challenges remain as open problems, while others deserve more study. In this paper, we hope to provide a broad research agenda around this topic, by proposing more problems than solutions.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82436920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Rethinking reinforcement learning for cloud elasticity 重新思考云弹性的强化学习
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131211
K. Lolos, I. Konstantinou, Verena Kantere, N. Koziris
Cloud elasticity, i.e., the dynamic allocation of resources to applications to meet fluctuating workload demands, has been one of the greatest challenges in cloud computing. Approaches based on reinforcement learning have been proposed but they require a large number of states in order to model complex application behavior. In this work we propose a novel reinforcement learning approach that employs adaptive state space partitioning. The idea is to start from one state that represents the entire environment and partition this into finer-grained states adaptively to the observed workload and system behavior following a decision-tree approach. We explore novel statistical criteria and strategies that decide both the correct parameters and the appropriate time to perform the partitioning.
云弹性,即向应用程序动态分配资源以满足波动的工作负载需求,一直是云计算中的最大挑战之一。基于强化学习的方法已经被提出,但它们需要大量的状态来建模复杂的应用程序行为。在这项工作中,我们提出了一种采用自适应状态空间划分的新型强化学习方法。其思想是从代表整个环境的一个状态开始,并按照决策树方法,根据观察到的工作负载和系统行为自适应地将其划分为更细粒度的状态。我们探索新的统计标准和策略,决定正确的参数和适当的时间来执行分区。
{"title":"Rethinking reinforcement learning for cloud elasticity","authors":"K. Lolos, I. Konstantinou, Verena Kantere, N. Koziris","doi":"10.1145/3127479.3131211","DOIUrl":"https://doi.org/10.1145/3127479.3131211","url":null,"abstract":"Cloud elasticity, i.e., the dynamic allocation of resources to applications to meet fluctuating workload demands, has been one of the greatest challenges in cloud computing. Approaches based on reinforcement learning have been proposed but they require a large number of states in order to model complex application behavior. In this work we propose a novel reinforcement learning approach that employs adaptive state space partitioning. The idea is to start from one state that represents the entire environment and partition this into finer-grained states adaptively to the observed workload and system behavior following a decision-tree approach. We explore novel statistical criteria and strategies that decide both the correct parameters and the appropriate time to perform the partitioning.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82508476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Workload analysis and caching strategies for search advertising systems 搜索广告系统的工作负载分析和缓存策略
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3129255
Conglong Li, D. Andersen, Qiang Fu, S. Elnikety, Yuxiong He
Search advertising depends on accurate predictions of user behavior and interest, accomplished today using complex and computationally expensive machine learning algorithms that estimate the potential revenue gain of thousands of candidate advertisements per search query. The accuracy of this estimation is important for revenue, but the cost of these computations represents a substantial expense, e.g., 10% to 30% of the total gross revenue. Caching the results of previous computations is a potential path to reducing this expense, but traditional domain-agnostic and revenue-agnostic approaches to do so result in substantial revenue loss. This paper presents three domain-specific caching mechanisms that successfully optimize for both factors. Simulations on a trace from the Bing advertising system show that a traditional cache can reduce cost by up to 27.7% but has negative revenue impact as bad as -14.1%. On the other hand, the proposed mechanisms can reduce cost by up to 20.6% while capping revenue impact between -1.3% and 0%. Based on Microsoft's earnings release for FY16 Q4, the traditional cache would reduce the net profit of Bing Ads by $84.9 to $166.1 million in the quarter, while our proposed cache could increase the net profit by $11.1 to $71.5 million.
搜索广告依赖于对用户行为和兴趣的准确预测,目前使用复杂且计算成本高昂的机器学习算法来完成,这些算法可以估计每个搜索查询中数千个候选广告的潜在收入。这种估算的准确性对收入很重要,但这些计算的成本代表了一笔可观的费用,例如,占总收入的10%到30%。缓存以前的计算结果是减少这种开销的潜在途径,但是传统的领域不可知和收入不可知的方法会导致大量的收入损失。本文提出了三个领域特定的缓存机制,成功地针对这两个因素进行了优化。对必应广告系统的跟踪模拟显示,传统的缓存可以降低27.7%的成本,但对收入的负面影响高达-14.1%。另一方面,拟议的机制可以将成本降低20.6%,同时将收入影响限制在-1.3%至0%之间。根据微软2016财年第四季度的财报,传统缓存将使必应广告的净利润减少8490美元至1.661亿美元,而我们提议的缓存将使净利润增加111美元至7150万美元。
{"title":"Workload analysis and caching strategies for search advertising systems","authors":"Conglong Li, D. Andersen, Qiang Fu, S. Elnikety, Yuxiong He","doi":"10.1145/3127479.3129255","DOIUrl":"https://doi.org/10.1145/3127479.3129255","url":null,"abstract":"Search advertising depends on accurate predictions of user behavior and interest, accomplished today using complex and computationally expensive machine learning algorithms that estimate the potential revenue gain of thousands of candidate advertisements per search query. The accuracy of this estimation is important for revenue, but the cost of these computations represents a substantial expense, e.g., 10% to 30% of the total gross revenue. Caching the results of previous computations is a potential path to reducing this expense, but traditional domain-agnostic and revenue-agnostic approaches to do so result in substantial revenue loss. This paper presents three domain-specific caching mechanisms that successfully optimize for both factors. Simulations on a trace from the Bing advertising system show that a traditional cache can reduce cost by up to 27.7% but has negative revenue impact as bad as -14.1%. On the other hand, the proposed mechanisms can reduce cost by up to 20.6% while capping revenue impact between -1.3% and 0%. Based on Microsoft's earnings release for FY16 Q4, the traditional cache would reduce the net profit of Bing Ads by $84.9 to $166.1 million in the quarter, while our proposed cache could increase the net profit by $11.1 to $71.5 million.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"1992 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88997614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Mitigating multi-tenant interference on mobile offloading servers: poster abstract 减轻移动卸载服务器上的多租户干扰:海报摘要
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132563
Zhou Fang, Mulong Luo, Tong Yu, O. Mengshoel, M. Srivastava, Rajesh K. Gupta
This work considers that multiple mobile clients offload various continuous sensing applications with end-to-end delay constraints, to a cluster of machines as the server. Contention for shared computing resources on a server can result in delay degradation and application malfunction. We present ATOMS (Accurate Timing prediction and Offloading for Mobile Systems), a framework to mitigate multi-tenant resource contention and to improve delay using a two-phase Plan-Schedule approach. The planning phase includes methods to predict future workloads from all clients, to estimate contention, and to devise offloading schedule to reduce contention. The scheduling phase dispatches arriving offloaded workload to the server machine that minimizes contention, based on the running workloads on each machine.
这项工作考虑了多个移动客户端卸载各种具有端到端延迟约束的连续传感应用程序,作为服务器的机器集群。对服务器上共享计算资源的争用可能导致延迟退化和应用程序故障。我们提出了atom(移动系统的精确定时预测和卸载),这是一个框架,用于缓解多租户资源争用并使用两阶段计划调度方法改善延迟。计划阶段包括预测来自所有客户机的未来工作负载、估计争用和设计卸载计划以减少争用的方法。调度阶段根据每台机器上运行的工作负载,将到达的卸载工作负载分派到最大限度减少争用的服务器机器。
{"title":"Mitigating multi-tenant interference on mobile offloading servers: poster abstract","authors":"Zhou Fang, Mulong Luo, Tong Yu, O. Mengshoel, M. Srivastava, Rajesh K. Gupta","doi":"10.1145/3127479.3132563","DOIUrl":"https://doi.org/10.1145/3127479.3132563","url":null,"abstract":"This work considers that multiple mobile clients offload various continuous sensing applications with end-to-end delay constraints, to a cluster of machines as the server. Contention for shared computing resources on a server can result in delay degradation and application malfunction. We present ATOMS (Accurate Timing prediction and Offloading for Mobile Systems), a framework to mitigate multi-tenant resource contention and to improve delay using a two-phase Plan-Schedule approach. The planning phase includes methods to predict future workloads from all clients, to estimate contention, and to devise offloading schedule to reduce contention. The scheduling phase dispatches arriving offloaded workload to the server machine that minimizes contention, based on the running workloads on each machine.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":"14 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74995439","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the 2017 Symposium on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1