首页 > 最新文献

Proceedings of the 2017 Symposium on Cloud Computing最新文献

英文 中文
Sketches of space: ownership accounting for shared storage 空间草图:共享存储的所有权
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132021
Jake Wires, P. Ganesan, A. Warfield
Efficient snapshots are an important feature of modern storage systems. However, the implicit sharing underlying most snapshot implementations makes it difficult to answer basic questions about the storage costs of individual snapshots. Traditional techniques for answering these questions incur significant performance penalties due to expensive metadata overheads. We present a novel probabilistic data structure, compatible with existing storage systems, that can provide approximate answers about snapshot costs with very low computational and storage overheads while achieving better than 95% accuracy for real-world data sets.
高效快照是现代存储系统的一个重要特性。然而,大多数快照实现的隐式共享使得很难回答有关单个快照的存储成本的基本问题。由于昂贵的元数据开销,回答这些问题的传统技术会导致显著的性能损失。我们提出了一种新的概率数据结构,与现有的存储系统兼容,可以以非常低的计算和存储开销提供关于快照成本的近似答案,同时对现实世界的数据集实现95%以上的准确率。
{"title":"Sketches of space: ownership accounting for shared storage","authors":"Jake Wires, P. Ganesan, A. Warfield","doi":"10.1145/3127479.3132021","DOIUrl":"https://doi.org/10.1145/3127479.3132021","url":null,"abstract":"Efficient snapshots are an important feature of modern storage systems. However, the implicit sharing underlying most snapshot implementations makes it difficult to answer basic questions about the storage costs of individual snapshots. Traditional techniques for answering these questions incur significant performance penalties due to expensive metadata overheads. We present a novel probabilistic data structure, compatible with existing storage systems, that can provide approximate answers about snapshot costs with very low computational and storage overheads while achieving better than 95% accuracy for real-world data sets.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75382740","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
SQML: large-scale in-database machine learning with pure SQL SQL:使用纯SQL的大规模数据库内机器学习
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132746
Umar Syed, Sergei Vassilvitskii
Many enterprises have migrated their data from an on-site database to a cloud-based database-as-a-service that handles all database-related administrative tasks while providing a simple SQL interface to the end user. Businesses are also increasingly relying on machine learning to understand their customers and develop new products. Given these converging trends, there is a pressing need for database-as-a-service providers to add support for sophisticated machine learning algorithms to the core functionality of their products.
许多企业已将其数据从现场数据库迁移到基于云的数据库即服务,该服务处理所有与数据库相关的管理任务,同时向最终用户提供简单的SQL接口。企业也越来越依赖机器学习来了解客户和开发新产品。鉴于这些趋同的趋势,数据库即服务提供商迫切需要在其产品的核心功能中添加对复杂机器学习算法的支持。
{"title":"SQML: large-scale in-database machine learning with pure SQL","authors":"Umar Syed, Sergei Vassilvitskii","doi":"10.1145/3127479.3132746","DOIUrl":"https://doi.org/10.1145/3127479.3132746","url":null,"abstract":"Many enterprises have migrated their data from an on-site database to a cloud-based database-as-a-service that handles all database-related administrative tasks while providing a simple SQL interface to the end user. Businesses are also increasingly relying on machine learning to understand their customers and develop new products. Given these converging trends, there is a pressing need for database-as-a-service providers to add support for sophisticated machine learning algorithms to the core functionality of their products.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73625262","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
An implementation of fast memset() using hardware accelerators: extended abstract 使用硬件加速器的快速memset()实现:扩展抽象
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132573
K. Pusukuri, R. Gardner, Jared C. Smolens
Multicore systems with large caches and huge main memories have become ubiquitous. They provide an attractive opportunity to maximize performance of big-memory applications such as in-memory databases, key-value stores, and graph analytics. However, these big-memory applications require many virtual-to-physical address translations, which increase TLB miss rate and hurt performance. To address this problem, modern hardware and OSes introduced support for huge pages. For example, on SPARC M7, Linux supports 8MB, 2GB, and 16GB huge pages (in addition to the default 8KB). Likewise, Linux supports 2MB and 1GB huge pages on Intel Xeon (E5-2630) platforms.
具有大缓存和大内存的多核系统已经变得无处不在。它们为最大化大内存应用程序(如内存数据库、键值存储和图形分析)的性能提供了一个有吸引力的机会。然而,这些大内存应用程序需要许多虚拟到物理地址的转换,这会增加TLB失误率并损害性能。为了解决这个问题,现代硬件和操作系统引入了对大页面的支持。例如,在SPARC M7上,Linux支持8MB、2GB和16GB的大页面(除了默认的8KB之外)。同样,Linux在Intel Xeon (E5-2630)平台上支持2MB和1GB的大页面。
{"title":"An implementation of fast memset() using hardware accelerators: extended abstract","authors":"K. Pusukuri, R. Gardner, Jared C. Smolens","doi":"10.1145/3127479.3132573","DOIUrl":"https://doi.org/10.1145/3127479.3132573","url":null,"abstract":"Multicore systems with large caches and huge main memories have become ubiquitous. They provide an attractive opportunity to maximize performance of big-memory applications such as in-memory databases, key-value stores, and graph analytics. However, these big-memory applications require many virtual-to-physical address translations, which increase TLB miss rate and hurt performance. To address this problem, modern hardware and OSes introduced support for huge pages. For example, on SPARC M7, Linux supports 8MB, 2GB, and 16GB huge pages (in addition to the default 8KB). Likewise, Linux supports 2MB and 1GB huge pages on Intel Xeon (E5-2630) platforms.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76943924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BestConfig: tapping the performance potential of systems via automatic configuration tuning BestConfig:通过自动配置调优挖掘系统的性能潜力
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128605
Yuqing Zhu, Jianxun Liu, Mengying Guo, Yungang Bao, Wenlong Ma, Zhuoyue Liu, Kunpeng Song, Y. Yang
An ever increasing number of configuration parameters are provided to system users. But many users have used one configuration setting across different workloads, leaving untapped the performance potential of systems. A good configuration setting can greatly improve the performance of a deployed system under certain workloads. But with tens or hundreds of parameters, it becomes a highly costly task to decide which configuration setting leads to the best performance. While such task requires the strong expertise in both the system and the application, users commonly lack such expertise. To help users tap the performance potential of systems, we present Best Config, a system for automatically finding a best configuration setting within a resource limit for a deployed system under a given application workload. BestConfig is designed with an extensible architecture to automate the configuration tuning for general systems. To tune system configurations within a resource limit, we propose the divide-and-diverge sampling method and the recursive bound-and-search algorithm. BestConfig can improve the throughput of Tomcat by 75%, that of Cassandra by 63%, that of MySQL by 430%, and reduce the running time of Hive join job by about 50% and that of Spark join job by about 80%, solely by configuration adjustment.
越来越多的配置参数被提供给系统用户。但是,许多用户在不同的工作负载中使用一个配置设置,从而没有充分利用系统的性能潜力。良好的配置设置可以极大地提高部署系统在某些工作负载下的性能。但是,由于有数十或数百个参数,决定哪种配置设置可以带来最佳性能成为一项代价高昂的任务。虽然这样的任务需要系统和应用程序方面的专业知识,但用户通常缺乏这样的专业知识。为了帮助用户挖掘系统的性能潜力,我们提供了Best Config,这是一个在给定应用程序工作负载下的已部署系统的资源限制内自动查找最佳配置设置的系统。BestConfig设计了一个可扩展的体系结构,可以自动对一般系统进行配置调优。为了在有限的资源范围内优化系统配置,我们提出了分散采样方法和递归定界搜索算法。仅通过配置调整,BestConfig就可以使Tomcat的吞吐量提高75%,Cassandra的吞吐量提高63%,MySQL的吞吐量提高430%,Hive join job的运行时间减少约50%,Spark join job的运行时间减少约80%。
{"title":"BestConfig: tapping the performance potential of systems via automatic configuration tuning","authors":"Yuqing Zhu, Jianxun Liu, Mengying Guo, Yungang Bao, Wenlong Ma, Zhuoyue Liu, Kunpeng Song, Y. Yang","doi":"10.1145/3127479.3128605","DOIUrl":"https://doi.org/10.1145/3127479.3128605","url":null,"abstract":"An ever increasing number of configuration parameters are provided to system users. But many users have used one configuration setting across different workloads, leaving untapped the performance potential of systems. A good configuration setting can greatly improve the performance of a deployed system under certain workloads. But with tens or hundreds of parameters, it becomes a highly costly task to decide which configuration setting leads to the best performance. While such task requires the strong expertise in both the system and the application, users commonly lack such expertise. To help users tap the performance potential of systems, we present Best Config, a system for automatically finding a best configuration setting within a resource limit for a deployed system under a given application workload. BestConfig is designed with an extensible architecture to automate the configuration tuning for general systems. To tune system configurations within a resource limit, we propose the divide-and-diverge sampling method and the recursive bound-and-search algorithm. BestConfig can improve the throughput of Tomcat by 75%, that of Cassandra by 63%, that of MySQL by 430%, and reduce the running time of Hive join job by about 50% and that of Spark join job by about 80%, solely by configuration adjustment.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75006110","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 165
FSP: towards flexible synchronous parallel framework for expectation-maximization based algorithms on cloud FSP:面向云上基于期望最大化算法的灵活同步并行框架
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128612
Zhigang Wang, Lixin Gao, Yu Gu, Y. Bao, Ge Yu
Myriad of parameter estimation algorithms can be performed by an Expectation-Maximization (EM) approach. Traditional synchronous frameworks can parallelize these EM algorithms on the cloud to accelerate computation while guaranteeing the convergence. However, expensive synchronization costs pose great challenges for efficiency. Asynchronous solutions have been recently designed to bypass high-cost synchronous barriers but at expense of potentially losing convergence guarantee. This paper first proposes a flexible synchronous parallel framework (FSP) that provides the capability of synchronous EM algorithms implementations, as well as significantly reduces the barrier cost. Under FSP, every distributed worker can immediately suspend local computation when necessary, to quickly synchronize with each other. That maximizes the time fast workers spend doing useful work, instead of waiting for slow, straggling workers. We then formally prove the algorithm convergence. Further, we analyze how to automatically identify a proper barrier interval to strike a nice balance between reduced synchronization costs and the convergence speed. Empirical results demonstrate that on a broad spectrum of real-world and synthetic datasets, FSP achieves as much as 3x speedup over the up-to-date synchronous solution.
期望最大化(EM)方法可以实现无数的参数估计算法。传统的同步框架可以在云上并行处理这些EM算法,在保证收敛性的同时加快计算速度。然而,昂贵的同步成本给效率带来了巨大的挑战。异步解决方案最近被设计为绕过高成本的同步障碍,但代价是可能失去收敛保证。本文首先提出了一种灵活的同步并行框架(FSP),该框架提供了同步EM算法实现的能力,并显著降低了屏障成本。在FSP下,每个分布式worker可以在必要时立即暂停本地计算,以快速相互同步。这将使速度快的工人花在做有用工作上的时间最大化,而不是等待速度慢、行动迟缓的工人。然后正式证明了算法的收敛性。此外,我们还分析了如何自动识别适当的屏障间隔,以在降低同步成本和收敛速度之间取得良好的平衡。经验结果表明,在广泛的现实世界和合成数据集上,FSP比最新的同步解决方案实现了多达3倍的加速。
{"title":"FSP: towards flexible synchronous parallel framework for expectation-maximization based algorithms on cloud","authors":"Zhigang Wang, Lixin Gao, Yu Gu, Y. Bao, Ge Yu","doi":"10.1145/3127479.3128612","DOIUrl":"https://doi.org/10.1145/3127479.3128612","url":null,"abstract":"Myriad of parameter estimation algorithms can be performed by an Expectation-Maximization (EM) approach. Traditional synchronous frameworks can parallelize these EM algorithms on the cloud to accelerate computation while guaranteeing the convergence. However, expensive synchronization costs pose great challenges for efficiency. Asynchronous solutions have been recently designed to bypass high-cost synchronous barriers but at expense of potentially losing convergence guarantee. This paper first proposes a flexible synchronous parallel framework (FSP) that provides the capability of synchronous EM algorithms implementations, as well as significantly reduces the barrier cost. Under FSP, every distributed worker can immediately suspend local computation when necessary, to quickly synchronize with each other. That maximizes the time fast workers spend doing useful work, instead of waiting for slow, straggling workers. We then formally prove the algorithm convergence. Further, we analyze how to automatically identify a proper barrier interval to strike a nice balance between reduced synchronization costs and the convergence speed. Empirical results demonstrate that on a broad spectrum of real-world and synthetic datasets, FSP achieves as much as 3x speedup over the up-to-date synchronous solution.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78047355","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
UNO: uniflying host and smart NIC offload for flexible packet processing UNO:统一主机和智能网卡卸载,灵活处理报文
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132252
Yanfang Le, Hyunseok Chang, S. Mukherjee, Limin Wang, Aditya Akella, M. Swift, T. V. Lakshman
Increasingly, smart Network Interface Cards (sNICs) are being used in data centers to offload networking functions (NFs) from host processors thereby making these processors available for tenant applications. Modern sNICs have fully programmable, energy-efficient multi-core processors on which many packet processing functions, including a full-blown programmable switch, can run. However, having multiple switch instances deployed across the host hypervisor and the attached sNICs makes controlling them difficult and data plane operations more complex. This paper proposes a generalized SDN-controlled NF offload architecture called UNO. It can transparently offload dynamically selected host processors' packet processing functions to sNICs by using multiple switches in the host while keeping the data centerwide network control and management planes unmodified. UNO exposes a single virtual control plane to the SDN controller and hides dynamic NF offload behind a unified virtual management plane. This enables UNO to make optimal use of host's and sNIC's combined packet processing capabilities with local optimization based on locally observed traffic patterns and resource consumption, and without central controller involvement. Experimental results based on a real UNO prototype in realistic scenarios show promising results: it can save processing worth up to 8 CPU cores, reduce power usage by up to 2x, and reduce the control plane overhead by more than 50%.
数据中心中越来越多地使用智能网络接口卡(snic)从主机处理器卸载网络功能(NFs),从而使这些处理器可用于租户应用程序。现代snic具有完全可编程的、节能的多核处理器,可以在其上运行许多包处理功能,包括一个成熟的可编程交换机。但是,跨主机管理程序和附加的snic部署多个交换机实例使得控制它们变得困难,数据平面操作变得更加复杂。本文提出了一种通用的sdn控制的NF卸载体系结构UNO。它可以在保持数据中心范围的网络控制和管理平面不变的情况下,通过使用主机中的多个交换机,透明地将所选主机处理器的数据包处理功能动态卸载到snic上。UNO将单个虚拟控制平面暴露给SDN控制器,将NF动态卸载隐藏在统一的虚拟管理平面之后。这使UNO能够在没有中央控制器参与的情况下,根据本地观察到的流量模式和资源消耗进行本地优化,最优地利用主机和sNIC的组合数据包处理能力。基于真实UNO原型在现实场景中的实验结果显示出令人鼓舞的结果:它可以节省多达8个CPU核心的处理,减少高达2倍的功耗,并将控制平面开销降低50%以上。
{"title":"UNO: uniflying host and smart NIC offload for flexible packet processing","authors":"Yanfang Le, Hyunseok Chang, S. Mukherjee, Limin Wang, Aditya Akella, M. Swift, T. V. Lakshman","doi":"10.1145/3127479.3132252","DOIUrl":"https://doi.org/10.1145/3127479.3132252","url":null,"abstract":"Increasingly, smart Network Interface Cards (sNICs) are being used in data centers to offload networking functions (NFs) from host processors thereby making these processors available for tenant applications. Modern sNICs have fully programmable, energy-efficient multi-core processors on which many packet processing functions, including a full-blown programmable switch, can run. However, having multiple switch instances deployed across the host hypervisor and the attached sNICs makes controlling them difficult and data plane operations more complex. This paper proposes a generalized SDN-controlled NF offload architecture called UNO. It can transparently offload dynamically selected host processors' packet processing functions to sNICs by using multiple switches in the host while keeping the data centerwide network control and management planes unmodified. UNO exposes a single virtual control plane to the SDN controller and hides dynamic NF offload behind a unified virtual management plane. This enables UNO to make optimal use of host's and sNIC's combined packet processing capabilities with local optimization based on locally observed traffic patterns and resource consumption, and without central controller involvement. Experimental results based on a real UNO prototype in realistic scenarios show promising results: it can save processing worth up to 8 CPU cores, reduce power usage by up to 2x, and reduce the control plane overhead by more than 50%.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80942035","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 79
Workload analysis and caching strategies for search advertising systems 搜索广告系统的工作负载分析和缓存策略
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3129255
Conglong Li, D. Andersen, Qiang Fu, S. Elnikety, Yuxiong He
Search advertising depends on accurate predictions of user behavior and interest, accomplished today using complex and computationally expensive machine learning algorithms that estimate the potential revenue gain of thousands of candidate advertisements per search query. The accuracy of this estimation is important for revenue, but the cost of these computations represents a substantial expense, e.g., 10% to 30% of the total gross revenue. Caching the results of previous computations is a potential path to reducing this expense, but traditional domain-agnostic and revenue-agnostic approaches to do so result in substantial revenue loss. This paper presents three domain-specific caching mechanisms that successfully optimize for both factors. Simulations on a trace from the Bing advertising system show that a traditional cache can reduce cost by up to 27.7% but has negative revenue impact as bad as -14.1%. On the other hand, the proposed mechanisms can reduce cost by up to 20.6% while capping revenue impact between -1.3% and 0%. Based on Microsoft's earnings release for FY16 Q4, the traditional cache would reduce the net profit of Bing Ads by $84.9 to $166.1 million in the quarter, while our proposed cache could increase the net profit by $11.1 to $71.5 million.
搜索广告依赖于对用户行为和兴趣的准确预测,目前使用复杂且计算成本高昂的机器学习算法来完成,这些算法可以估计每个搜索查询中数千个候选广告的潜在收入。这种估算的准确性对收入很重要,但这些计算的成本代表了一笔可观的费用,例如,占总收入的10%到30%。缓存以前的计算结果是减少这种开销的潜在途径,但是传统的领域不可知和收入不可知的方法会导致大量的收入损失。本文提出了三个领域特定的缓存机制,成功地针对这两个因素进行了优化。对必应广告系统的跟踪模拟显示,传统的缓存可以降低27.7%的成本,但对收入的负面影响高达-14.1%。另一方面,拟议的机制可以将成本降低20.6%,同时将收入影响限制在-1.3%至0%之间。根据微软2016财年第四季度的财报,传统缓存将使必应广告的净利润减少8490美元至1.661亿美元,而我们提议的缓存将使净利润增加111美元至7150万美元。
{"title":"Workload analysis and caching strategies for search advertising systems","authors":"Conglong Li, D. Andersen, Qiang Fu, S. Elnikety, Yuxiong He","doi":"10.1145/3127479.3129255","DOIUrl":"https://doi.org/10.1145/3127479.3129255","url":null,"abstract":"Search advertising depends on accurate predictions of user behavior and interest, accomplished today using complex and computationally expensive machine learning algorithms that estimate the potential revenue gain of thousands of candidate advertisements per search query. The accuracy of this estimation is important for revenue, but the cost of these computations represents a substantial expense, e.g., 10% to 30% of the total gross revenue. Caching the results of previous computations is a potential path to reducing this expense, but traditional domain-agnostic and revenue-agnostic approaches to do so result in substantial revenue loss. This paper presents three domain-specific caching mechanisms that successfully optimize for both factors. Simulations on a trace from the Bing advertising system show that a traditional cache can reduce cost by up to 27.7% but has negative revenue impact as bad as -14.1%. On the other hand, the proposed mechanisms can reduce cost by up to 20.6% while capping revenue impact between -1.3% and 0%. Based on Microsoft's earnings release for FY16 Q4, the traditional cache would reduce the net profit of Bing Ads by $84.9 to $166.1 million in the quarter, while our proposed cache could increase the net profit by $11.1 to $71.5 million.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88997614","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Remote memory in the age of fast networks 高速网络时代的远程存储器
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131612
M. Aguilera, Nadav Amit, I. Calciu, Xavier Deguillard, Jayneel Gandhi, Pratap Subrahmanyam, L. Suresh, K. Tati, Rajesh Venkatasubramanian, M. Wei
As the latency of the network approaches that of memory, it becomes increasingly attractive for applications to use remote memory---random-access memory at another computer that is accessed using the virtual memory subsystem. This is an old idea whose time has come, in the age of fast networks. To work effectively, remote memory must address many technical challenges. In this paper, we enumerate these challenges, discuss their feasibility, explain how some of them are addressed by recent work, and indicate other promising ways to tackle them. Some challenges remain as open problems, while others deserve more study. In this paper, we hope to provide a broad research agenda around this topic, by proposing more problems than solutions.
随着网络延迟接近内存延迟,应用程序越来越倾向于使用远程内存——使用虚拟内存子系统访问另一台计算机上的随机访问内存。这是一个古老的想法,在快速网络时代,它的时代已经到来。为了有效地工作,远程内存必须解决许多技术挑战。在本文中,我们列举了这些挑战,讨论了它们的可行性,解释了其中一些是如何通过最近的工作来解决的,并指出了其他有希望的解决方法。有些挑战仍然是悬而未决的问题,而另一些则值得进一步研究。在本文中,我们希望通过提出更多的问题而不是解决方案,围绕这一主题提供一个广泛的研究议程。
{"title":"Remote memory in the age of fast networks","authors":"M. Aguilera, Nadav Amit, I. Calciu, Xavier Deguillard, Jayneel Gandhi, Pratap Subrahmanyam, L. Suresh, K. Tati, Rajesh Venkatasubramanian, M. Wei","doi":"10.1145/3127479.3131612","DOIUrl":"https://doi.org/10.1145/3127479.3131612","url":null,"abstract":"As the latency of the network approaches that of memory, it becomes increasingly attractive for applications to use remote memory---random-access memory at another computer that is accessed using the virtual memory subsystem. This is an old idea whose time has come, in the age of fast networks. To work effectively, remote memory must address many technical challenges. In this paper, we enumerate these challenges, discuss their feasibility, explain how some of them are addressed by recent work, and indicate other promising ways to tackle them. Some challenges remain as open problems, while others deserve more study. In this paper, we hope to provide a broad research agenda around this topic, by proposing more problems than solutions.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82436920","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Rethinking reinforcement learning for cloud elasticity 重新思考云弹性的强化学习
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131211
K. Lolos, I. Konstantinou, Verena Kantere, N. Koziris
Cloud elasticity, i.e., the dynamic allocation of resources to applications to meet fluctuating workload demands, has been one of the greatest challenges in cloud computing. Approaches based on reinforcement learning have been proposed but they require a large number of states in order to model complex application behavior. In this work we propose a novel reinforcement learning approach that employs adaptive state space partitioning. The idea is to start from one state that represents the entire environment and partition this into finer-grained states adaptively to the observed workload and system behavior following a decision-tree approach. We explore novel statistical criteria and strategies that decide both the correct parameters and the appropriate time to perform the partitioning.
云弹性,即向应用程序动态分配资源以满足波动的工作负载需求,一直是云计算中的最大挑战之一。基于强化学习的方法已经被提出,但它们需要大量的状态来建模复杂的应用程序行为。在这项工作中,我们提出了一种采用自适应状态空间划分的新型强化学习方法。其思想是从代表整个环境的一个状态开始,并按照决策树方法,根据观察到的工作负载和系统行为自适应地将其划分为更细粒度的状态。我们探索新的统计标准和策略,决定正确的参数和适当的时间来执行分区。
{"title":"Rethinking reinforcement learning for cloud elasticity","authors":"K. Lolos, I. Konstantinou, Verena Kantere, N. Koziris","doi":"10.1145/3127479.3131211","DOIUrl":"https://doi.org/10.1145/3127479.3131211","url":null,"abstract":"Cloud elasticity, i.e., the dynamic allocation of resources to applications to meet fluctuating workload demands, has been one of the greatest challenges in cloud computing. Approaches based on reinforcement learning have been proposed but they require a large number of states in order to model complex application behavior. In this work we propose a novel reinforcement learning approach that employs adaptive state space partitioning. The idea is to start from one state that represents the entire environment and partition this into finer-grained states adaptively to the observed workload and system behavior following a decision-tree approach. We explore novel statistical criteria and strategies that decide both the correct parameters and the appropriate time to perform the partitioning.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82508476","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Incentivizing self-capping to increase cloud utilization 激励自我封顶以提高云利用率
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128611
Mohammad Shahrad, C. Klein, Liang Zheng, M. Chiang, E. Elmroth, D. Wentzlaff
Cloud Infrastructure as a Service (IaaS) providers continually seek higher resource utilization to better amortize capital costs. Higher utilization not only can enable higher profit for IaaS providers but also provides a mechanism to raise energy efficiency; therefore creating greener cloud services. Unfortunately, achieving high utilization is difficult mainly due to infrastructure providers needing to maintain spare capacity to service demand fluctuations. Graceful degradation is a self-adaptation technique originally designed for constructing robust services that survive resource shortages. Previous work has shown that graceful degradation can also be used to improve resource utilization in the cloud by absorbing demand fluctuations and reducing spare capacity. In this work, we build a system and pricing model that enables infrastructure providers to incentivize their tenants to use graceful degradation. By using graceful degradation with an appropriate pricing model, the infrastructure provider can realize higher resource utilization while simultaneously, its tenants can increase their profit. Our proposed solution is based on a hybrid model which guarantees both reserved and peak on-demand capacities over flexible periods. It also includes a global dynamic price pair for capacity which remains uniform during each tenant's Service Level Agreement (SLA) term. We evaluate our scheme using simulations based on real-world traces and also implement a prototype using RUBiS on the Xen hypervisor as an end-to-end demonstration. Our analysis shows that the proposed scheme never hurts a tenant's net profit, but can improve it by as much as 93%. Simultaneously, it can also improve the effective utilization of contracts from 42% to as high as 99%.
云基础设施即服务(IaaS)提供商不断寻求更高的资源利用率,以更好地摊销资本成本。更高的利用率不仅可以使IaaS提供商获得更高的利润,而且还提供了提高能源效率的机制;因此,创建更绿色的云服务。不幸的是,实现高利用率是困难的,主要原因是基础设施提供商需要保持闲置产能以满足需求波动。优雅退化是一种自适应技术,最初设计用于构建在资源短缺情况下存活的健壮服务。以前的工作表明,优雅退化也可用于通过吸收需求波动和减少备用容量来提高云中的资源利用率。在这项工作中,我们建立了一个系统和定价模型,使基础设施提供商能够激励他们的租户使用优雅的降级。通过采用适当的定价模型和优雅的退化,基础设施提供者可以实现更高的资源利用率,同时租户也可以增加他们的利润。我们提出的解决方案是基于一种混合模型,该模型可以在灵活的时间段内保证保留容量和峰值按需容量。它还包括容量的全局动态价格对,该价格对在每个租户的服务水平协议(SLA)期限内保持一致。我们使用基于真实跟踪的模拟来评估我们的方案,并在Xen管理程序上使用rubi实现原型,作为端到端演示。我们的分析表明,拟议的方案不会损害租户的净利润,但可以提高高达93%。同时,它还可以将合同的有效利用率从42%提高到99%。
{"title":"Incentivizing self-capping to increase cloud utilization","authors":"Mohammad Shahrad, C. Klein, Liang Zheng, M. Chiang, E. Elmroth, D. Wentzlaff","doi":"10.1145/3127479.3128611","DOIUrl":"https://doi.org/10.1145/3127479.3128611","url":null,"abstract":"Cloud Infrastructure as a Service (IaaS) providers continually seek higher resource utilization to better amortize capital costs. Higher utilization not only can enable higher profit for IaaS providers but also provides a mechanism to raise energy efficiency; therefore creating greener cloud services. Unfortunately, achieving high utilization is difficult mainly due to infrastructure providers needing to maintain spare capacity to service demand fluctuations. Graceful degradation is a self-adaptation technique originally designed for constructing robust services that survive resource shortages. Previous work has shown that graceful degradation can also be used to improve resource utilization in the cloud by absorbing demand fluctuations and reducing spare capacity. In this work, we build a system and pricing model that enables infrastructure providers to incentivize their tenants to use graceful degradation. By using graceful degradation with an appropriate pricing model, the infrastructure provider can realize higher resource utilization while simultaneously, its tenants can increase their profit. Our proposed solution is based on a hybrid model which guarantees both reserved and peak on-demand capacities over flexible periods. It also includes a global dynamic price pair for capacity which remains uniform during each tenant's Service Level Agreement (SLA) term. We evaluate our scheme using simulations based on real-world traces and also implement a prototype using RUBiS on the Xen hypervisor as an end-to-end demonstration. Our analysis shows that the proposed scheme never hurts a tenant's net profit, but can improve it by as much as 93%. Simultaneously, it can also improve the effective utilization of contracts from 42% to as high as 99%.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78035924","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
期刊
Proceedings of the 2017 Symposium on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1