首页 > 最新文献

Proceedings of the Sixth ACM Symposium on Cloud Computing最新文献

英文 中文
Microsoft azure SQL database telemetry 微软azure SQL数据库遥测
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806845
Willis Lang, Frank Bertsch, D. DeWitt, Nigel Ellis
Microsoft operates the Azure SQL Database (ASD) cloud service, one of the dominant relational cloud database services in the market today. To aid the academic community in their research on designing and efficiently operating cloud database services, Microsoft is introducing the release of production-level telemetry traces from the ASD service. This telemetry data set provides, over a wide set of important hardware resources and counters, the consumption level of each customer database replica. The first release will be a multi-month time-series data set that includes the full cluster traces from two different ASD global regions.
微软运营Azure SQL数据库(ASD)云服务,这是当今市场上占主导地位的关系云数据库服务之一。为了帮助学术界研究设计和高效运行云数据库服务,微软推出了ASD服务的生产级遥测跟踪版本。这个遥测数据集通过一系列重要的硬件资源和计数器提供了每个客户数据库副本的消耗级别。第一个版本将是一个多个月的时间序列数据集,其中包括来自两个不同的ASD全球区域的完整集群轨迹。
{"title":"Microsoft azure SQL database telemetry","authors":"Willis Lang, Frank Bertsch, D. DeWitt, Nigel Ellis","doi":"10.1145/2806777.2806845","DOIUrl":"https://doi.org/10.1145/2806777.2806845","url":null,"abstract":"Microsoft operates the Azure SQL Database (ASD) cloud service, one of the dominant relational cloud database services in the market today. To aid the academic community in their research on designing and efficiently operating cloud database services, Microsoft is introducing the release of production-level telemetry traces from the ASD service. This telemetry data set provides, over a wide set of important hardware resources and counters, the consumption level of each customer database replica. The first release will be a multi-month time-series data set that includes the full cluster traces from two different ASD global regions.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"398 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116691015","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 19
Domino: understanding wide-area, asynchronous event causality in web applications Domino:理解web应用程序中的广域异步事件因果关系
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806940
Ding Li, James W. Mickens, Suman Nath, Lenin Ravindranath
In a modern web application, a single high-level action like a mouse click triggers a flurry of asynchronous events on the client browser and remote web servers. We introduce Domino, a new tool which automatically captures and analyzes end-to-end, asynchronous causal relationship of events that span clients and servers. Using Domino, we found uncharacteristically long event chains in Bing Maps, discovered data races in the WinJS implementation of promises, and developed a new server-side scheduling algorithm for reducing the tail latency of server responses.
在现代web应用程序中,像鼠标点击这样的单个高级操作会在客户端浏览器和远程web服务器上触发一系列异步事件。我们将介绍Domino,这是一种新工具,可以自动捕获和分析跨客户机和服务器的端到端异步事件因果关系。使用Domino,我们在Bing Maps中发现了异常长的事件链,在承诺的WinJS实现中发现了数据竞争,并开发了一种新的服务器端调度算法,用于减少服务器响应的尾部延迟。
{"title":"Domino: understanding wide-area, asynchronous event causality in web applications","authors":"Ding Li, James W. Mickens, Suman Nath, Lenin Ravindranath","doi":"10.1145/2806777.2806940","DOIUrl":"https://doi.org/10.1145/2806777.2806940","url":null,"abstract":"In a modern web application, a single high-level action like a mouse click triggers a flurry of asynchronous events on the client browser and remote web servers. We introduce Domino, a new tool which automatically captures and analyzes end-to-end, asynchronous causal relationship of events that span clients and servers. Using Domino, we found uncharacteristically long event chains in Bing Maps, discovered data races in the WinJS implementation of promises, and developed a new server-side scheduling algorithm for reducing the tail latency of server responses.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"38 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132015403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
dJay: enabling high-density multi-tenancy for cloud gaming servers with dynamic cost-benefit GPU load balancing dJay:为云游戏服务器提供高密度多租户,具有动态成本效益的GPU负载平衡
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806942
Sergey Grizan, David Chu, A. Wolman, Roger Wattenhofer
In cloud gaming, servers perform remote rendering on behalf of thin clients. Such a server must deliver sufficient frame rate (at least 30fps) to each of its clients. At the same time, each client desires an immersive experience, and therefore the server should also provide the best graphics quality possible to each client. Statically provisioning time slices of the server GPU for each client suffers from severe underutilization because clients can come and go, and scenes that the clients need rendered can vary greatly in terms of GPU resource usage over time. In this work, we present dJay, a utility-maximizing cloud gaming server that dynamically tunes client GPU rendering workloads in order to 1) ensure all clients get satisfactory frame rate, and 2) provide the best possible graphics quality across clients. To accomplish this, we develop three main components. First, we build an online profiler that collects key cost and benefit data, and distills the data into a reusable regression model. Second, we build an online utility optimizer that uses the regression model to tune GPU workloads for better graphics quality. The optimizer solves the Multiple Choice Knapsack problem. We demonstrate dJay on two high quality commercial games, Doom 3 and Fable 3. Our results show that when compared to a static configuration, we can respond much better to peaks and troughs, achieving up to four times the multi-tenant density on a single server while offering clients the best possible graphics quality.
在云游戏中,服务器代表瘦客户机执行远程呈现。这样的服务器必须向每个客户端提供足够的帧速率(至少30fps)。同时,每个客户端都希望获得身临其境的体验,因此服务器也应该为每个客户端提供最好的图形质量。静态地为每个客户端配置服务器GPU的时间片会遭受严重的利用率不足,因为客户端可以来来去去,并且客户端需要渲染的场景在GPU资源使用方面会随着时间的推移而变化很大。在这项工作中,我们提出了dJay,一个实用最大化的云游戏服务器,它可以动态调整客户端GPU渲染工作负载,以便1)确保所有客户端获得满意的帧速率,2)在客户端之间提供最佳的图形质量。为了实现这一点,我们开发了三个主要组件。首先,我们构建一个在线分析器,它收集关键的成本和收益数据,并将数据提取到一个可重用的回归模型中。其次,我们构建了一个在线实用程序优化器,它使用回归模型来调整GPU工作负载以获得更好的图形质量。优化器解决了多项选择背包问题。我们在两款高质量的商业游戏《毁灭战士3》和《神鬼寓言3》中展示了dJay。我们的结果表明,与静态配置相比,我们可以更好地响应高峰和低谷,在单个服务器上实现多达四倍的多租户密度,同时为客户机提供最佳的图形质量。
{"title":"dJay: enabling high-density multi-tenancy for cloud gaming servers with dynamic cost-benefit GPU load balancing","authors":"Sergey Grizan, David Chu, A. Wolman, Roger Wattenhofer","doi":"10.1145/2806777.2806942","DOIUrl":"https://doi.org/10.1145/2806777.2806942","url":null,"abstract":"In cloud gaming, servers perform remote rendering on behalf of thin clients. Such a server must deliver sufficient frame rate (at least 30fps) to each of its clients. At the same time, each client desires an immersive experience, and therefore the server should also provide the best graphics quality possible to each client. Statically provisioning time slices of the server GPU for each client suffers from severe underutilization because clients can come and go, and scenes that the clients need rendered can vary greatly in terms of GPU resource usage over time. In this work, we present dJay, a utility-maximizing cloud gaming server that dynamically tunes client GPU rendering workloads in order to 1) ensure all clients get satisfactory frame rate, and 2) provide the best possible graphics quality across clients. To accomplish this, we develop three main components. First, we build an online profiler that collects key cost and benefit data, and distills the data into a reusable regression model. Second, we build an online utility optimizer that uses the regression model to tune GPU workloads for better graphics quality. The optimizer solves the Multiple Choice Knapsack problem. We demonstrate dJay on two high quality commercial games, Doom 3 and Fable 3. Our results show that when compared to a static configuration, we can respond much better to peaks and troughs, achieving up to four times the multi-tenant density on a single server while offering clients the best possible graphics quality.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117192255","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Zorro: zero-cost reactive failure recovery in distributed graph processing 佐罗:分布式图处理中的零成本无功故障恢复
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806934
Mayank Pundir, Luke M. Leslie, Indranil Gupta, R. Campbell
Distributed graph processing systems largely rely on proactive techniques for failure recovery. Unfortunately, these approaches (such as checkpointing) entail a significant overhead. In this paper, we argue that distributed graph processing systems should instead use a reactive approach to failure recovery. The reactive approach trades off completeness of the result (generating a slightly inaccurate result) while reducing the overhead during failure-free execution to zero. We build a system called Zorro that imbues this reactive approach, and integrate Zorro into two graph processing systems -- PowerGraph and LFGraph. When a failure occurs, Zorro opportunistically exploits vertex replication inherent in today's graph processing systems to quickly rebuild the state of failed servers. Experiments using real-world graphs demonstrate that Zorro is able to recover over 99% of the graph state when 6--12% of the servers fail, and between 87--95% when half the cluster fails. Furthermore, using various graph processing algorithms, Zorro incurs little to no accuracy loss in all experimental failure scenarios, and achieves a worst-case accuracy of 97%.
分布式图形处理系统在很大程度上依赖于故障恢复的主动技术。不幸的是,这些方法(比如检查点)会带来很大的开销。在本文中,我们认为分布式图形处理系统应该使用响应式方法来恢复故障。响应式方法折衷了结果的完整性(生成稍微不准确的结果),同时将无故障执行期间的开销减少到零。我们构建了一个名为佐罗的系统,它融入了这种反应式方法,并将佐罗集成到两个图形处理系统中——PowerGraph和LFGraph。当发生故障时,佐罗会利用当前图形处理系统中固有的顶点复制来快速重建故障服务器的状态。使用真实图的实验表明,当6- 12%的服务器故障时,Zorro能够恢复99%以上的图状态,当一半的集群故障时,Zorro能够恢复87- 95%。此外,使用各种图形处理算法,佐罗在所有实验故障场景中几乎没有精度损失,最坏情况下精度达到97%。
{"title":"Zorro: zero-cost reactive failure recovery in distributed graph processing","authors":"Mayank Pundir, Luke M. Leslie, Indranil Gupta, R. Campbell","doi":"10.1145/2806777.2806934","DOIUrl":"https://doi.org/10.1145/2806777.2806934","url":null,"abstract":"Distributed graph processing systems largely rely on proactive techniques for failure recovery. Unfortunately, these approaches (such as checkpointing) entail a significant overhead. In this paper, we argue that distributed graph processing systems should instead use a reactive approach to failure recovery. The reactive approach trades off completeness of the result (generating a slightly inaccurate result) while reducing the overhead during failure-free execution to zero. We build a system called Zorro that imbues this reactive approach, and integrate Zorro into two graph processing systems -- PowerGraph and LFGraph. When a failure occurs, Zorro opportunistically exploits vertex replication inherent in today's graph processing systems to quickly rebuild the state of failed servers. Experiments using real-world graphs demonstrate that Zorro is able to recover over 99% of the graph state when 6--12% of the servers fail, and between 87--95% when half the cluster fails. Furthermore, using various graph processing algorithms, Zorro incurs little to no accuracy loss in all experimental failure scenarios, and achieves a worst-case accuracy of 97%.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117237552","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 30
Response time service level agreements for cloud-hosted web applications 云托管web应用程序的响应时间服务水平协议
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806842
Hiranya Jayathilaka, C. Krintz, R. Wolski
Cloud computing is a successful model for hosting web-facing applications that are accessed by their users as services. While clouds currently offer Service Level Agreements (SLAs) containing guarantees of availability, they do not make performance guarantees for deployed applications. In this work we present Cerebro -- a system for establishing statistical guarantees of application response time in cloud settings. Cerebro combines off-line static analysis of application control structure with on-line cloud performance monitoring and statistical forecasting to predict bounds on the response time of web-facing application programming interfaces (APIs). Because Cerebro does not require application instrumentation or per-application cloud benchmarking, it does not impose any runtime overhead, and is suitable for use at cloud scales. Also, because the bounds are statistical, they are appropriate for use as the basis for SLAs between cloud-hosted applications and their users. We investigate the correctness of Cerebro predictions, the tightness of their bounds, and the duration over which the bounds persist in both Google App Engine and AppScale (public and private cloud platforms respectively). We also detail the effectiveness of our SLA prediction methodology compared to other performance bound estimation methods based on simple statistical analysis.
云计算是托管面向web的应用程序的一种成功模式,用户可以将这些应用程序作为服务访问。虽然云目前提供了包含可用性保证的服务水平协议(sla),但它们并没有为已部署的应用程序提供性能保证。在这项工作中,我们提出了Cerebro——一个用于在云设置中建立应用程序响应时间统计保证的系统。Cerebro将应用程序控制结构的离线静态分析与在线云性能监测和统计预测相结合,以预测面向web的应用程序编程接口(api)的响应时间界限。因为Cerebro不需要应用程序检测或每个应用程序的云基准测试,所以它不会强加任何运行时开销,并且适合在云规模上使用。此外,由于边界是统计性的,因此它们适合用作云托管应用程序与其用户之间sla的基础。我们调查了Cerebro预测的正确性、边界的紧密性,以及边界在Google App Engine和AppScale(分别为公共和私有云平台)中持续的时间。我们还详细介绍了与其他基于简单统计分析的性能界限估计方法相比,我们的SLA预测方法的有效性。
{"title":"Response time service level agreements for cloud-hosted web applications","authors":"Hiranya Jayathilaka, C. Krintz, R. Wolski","doi":"10.1145/2806777.2806842","DOIUrl":"https://doi.org/10.1145/2806777.2806842","url":null,"abstract":"Cloud computing is a successful model for hosting web-facing applications that are accessed by their users as services. While clouds currently offer Service Level Agreements (SLAs) containing guarantees of availability, they do not make performance guarantees for deployed applications. In this work we present Cerebro -- a system for establishing statistical guarantees of application response time in cloud settings. Cerebro combines off-line static analysis of application control structure with on-line cloud performance monitoring and statistical forecasting to predict bounds on the response time of web-facing application programming interfaces (APIs). Because Cerebro does not require application instrumentation or per-application cloud benchmarking, it does not impose any runtime overhead, and is suitable for use at cloud scales. Also, because the bounds are statistical, they are appropriate for use as the basis for SLAs between cloud-hosted applications and their users. We investigate the correctness of Cerebro predictions, the tightness of their bounds, and the duration over which the bounds persist in both Google App Engine and AppScale (public and private cloud platforms respectively). We also detail the effectiveness of our SLA prediction methodology compared to other performance bound estimation methods based on simple statistical analysis.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"51 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131672403","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Potassium: penetration testing as a service 钾:渗透测试即服务
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806935
Richard Li, Dallin Abendroth, Xing Lin, Yuankai Guo, H. Baek, E. Eide, R. Ricci, J. Merwe
Penetration testing---the process of probing a deployed system for security vulnerabilities---involves a fundamental tension. If one tests a production system, there is a real danger of collateral damage; this is particularly true for systems hosted in the cloud due to the presence of other tenants. If one tests against a separate system brought up to model the live one, the dynamic state of the production system is not captured, and the value of the test is reduced. This paper presents Potassium, which provides penetration testing as a service (PTaaS) and resolves this tension for system owners, penetration testers, and cloud providers. Potassium uses techniques originally developed for live migration of virtual machines to clone them instead, capturing their full disk, memory, and network state. Potassium isolates the cloned system from the rest of the cloud, providing confidence that side effects of the penetration test will not harm other tenants. The penetration tester effectively owns the cloned system, allowing testing to be more thorough, efficient, and automatable. Experiments with our Potassium prototype show that PTaaS can detect real-world vulnerabilities while having minimal impact on cloud-based production systems.
渗透测试——探测已部署系统的安全漏洞的过程——涉及一种根本性的紧张关系。如果测试一个生产系统,就会有附带损害的真正危险;对于由于其他租户的存在而托管在云中的系统尤其如此。如果对一个单独的系统进行测试,以对活动系统进行建模,则不会捕获生产系统的动态状态,并且测试的价值会减少。本文介绍了钾,它提供了作为服务的渗透测试(PTaaS),并解决了系统所有者、渗透测试人员和云提供商之间的紧张关系。钾使用最初为虚拟机的实时迁移而开发的技术来克隆它们,从而捕获它们的完整磁盘、内存和网络状态。钾将克隆系统与云的其余部分隔离开来,提供了渗透测试的副作用不会伤害其他租户的信心。渗透测试人员有效地拥有克隆系统,允许测试更加彻底、有效和自动化。对我们的钾原型的实验表明,PTaaS可以检测现实世界的漏洞,同时对基于云的生产系统的影响最小。
{"title":"Potassium: penetration testing as a service","authors":"Richard Li, Dallin Abendroth, Xing Lin, Yuankai Guo, H. Baek, E. Eide, R. Ricci, J. Merwe","doi":"10.1145/2806777.2806935","DOIUrl":"https://doi.org/10.1145/2806777.2806935","url":null,"abstract":"Penetration testing---the process of probing a deployed system for security vulnerabilities---involves a fundamental tension. If one tests a production system, there is a real danger of collateral damage; this is particularly true for systems hosted in the cloud due to the presence of other tenants. If one tests against a separate system brought up to model the live one, the dynamic state of the production system is not captured, and the value of the test is reduced. This paper presents Potassium, which provides penetration testing as a service (PTaaS) and resolves this tension for system owners, penetration testers, and cloud providers. Potassium uses techniques originally developed for live migration of virtual machines to clone them instead, capturing their full disk, memory, and network state. Potassium isolates the cloned system from the rest of the cloud, providing confidence that side effects of the penetration test will not harm other tenants. The penetration tester effectively owns the cloned system, allowing testing to be more thorough, efficient, and automatable. Experiments with our Potassium prototype show that PTaaS can detect real-world vulnerabilities while having minimal impact on cloud-based production systems.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"8 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115736524","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
DSwitch: a dual mode direct and network attached disk DSwitch:双模式直连和网络挂载磁盘
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806850
Quanlu Zhang, Yafei Dai, Lintao Zhang
Putting computers into low power mode (e.g., suspend-to-RAM) could potentially save significant amount of power when the computers are not in use. Unfortunately, this is often infeasible in practice because data stored on the computers (i.e., directly attached disks, DAS) might need to be accessed by others. Separating storage from computation by attaching storage on the network (e.g., NAS and SAN) could potentially solve this problem, at the cost of lower performance, more network congestion, increased peak power consumption, and higher equipment cost. Though DAS does not suffer these problems, it is not flexible for power saving. In this paper, we present DSwitch, an architecture that, depending on the workload, allows a disk to be attached either directly or through network. We design flexible workload migration based on DSwitch, and show that a wide variety of applications in both data center and home/office settings can be well supported. The experiments demonstrate that our prototype DSwitch achieves a power savings of 91.9% to 97.5% when a disk is in low power network attached mode, while incurring no performance degradation and minimal power overhead when it is in high performance directly attached mode.
当计算机不使用时,将计算机置于低功耗模式(例如,挂起到内存)可能会节省大量的电力。不幸的是,这在实践中通常是不可行的,因为存储在计算机上的数据(即直接连接的磁盘,DAS)可能需要其他人访问。通过在网络上附加存储(例如NAS和SAN)来分离存储和计算可能潜在地解决这个问题,但代价是性能降低、网络拥塞加剧、峰值功耗增加和设备成本增加。虽然DAS没有这些问题,但它在节能方面并不灵活。在本文中,我们介绍了DSwitch,这是一种架构,根据工作负载的不同,它允许直接或通过网络连接磁盘。我们基于DSwitch设计了灵活的工作负载迁移,并表明可以很好地支持数据中心和家庭/办公室设置中的各种应用程序。实验表明,当磁盘处于低功耗网络连接模式时,我们的原型DSwitch可以节省91.9%至97.5%的功耗,而当磁盘处于高性能直接连接模式时,不会产生性能下降和最小的功耗开销。
{"title":"DSwitch: a dual mode direct and network attached disk","authors":"Quanlu Zhang, Yafei Dai, Lintao Zhang","doi":"10.1145/2806777.2806850","DOIUrl":"https://doi.org/10.1145/2806777.2806850","url":null,"abstract":"Putting computers into low power mode (e.g., suspend-to-RAM) could potentially save significant amount of power when the computers are not in use. Unfortunately, this is often infeasible in practice because data stored on the computers (i.e., directly attached disks, DAS) might need to be accessed by others. Separating storage from computation by attaching storage on the network (e.g., NAS and SAN) could potentially solve this problem, at the cost of lower performance, more network congestion, increased peak power consumption, and higher equipment cost. Though DAS does not suffer these problems, it is not flexible for power saving. In this paper, we present DSwitch, an architecture that, depending on the workload, allows a disk to be attached either directly or through network. We design flexible workload migration based on DSwitch, and show that a wide variety of applications in both data center and home/office settings can be well supported. The experiments demonstrate that our prototype DSwitch achieves a power savings of 91.9% to 97.5% when a disk is in low power network attached mode, while incurring no performance degradation and minimal power overhead when it is in high performance directly attached mode.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116888432","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards a comprehensive performance model of virtual machine live migration 迈向一个全面的虚拟机动态迁移性能模型
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806838
Senthil Nathan, U. Bellur, Purushottam Kulkarni
Although many models exist to predict the time taken to migrate a virtual machine from one physical machine to another, our empirical validation of these models has shown the 90th percentile error to be 46% (43 secs) and 159% (112 secs) for KVM and Xen live migration, respectively. Our analysis reveals that these models are fundamentally flawed as they all fail to take into account the following three critical parameters: (i) the writable working set size, (ii) the number of pages eligible for the skip technique, (iii) the relation of the number of skipped pages with the page dirty rate and the page transfer rate, and incorrectly model the key parameter---the number of new pages dirtied per unit time. In this paper, we propose a novel model that takes all these parameters into account. We present a thorough validation with 53 workloads and show that the 90th percentile error in the estimated migration times is only 12% (8 secs) and 19% (14 secs) for KVM and Xen live migration, respectively.
尽管存在许多模型来预测将虚拟机从一台物理机迁移到另一台物理机所花费的时间,但我们对这些模型的经验验证表明,KVM和Xen实时迁移的第90个百分点误差分别为46%(43秒)和159%(112秒)。我们的分析表明,这些模型从根本上是有缺陷的,因为它们都没有考虑到以下三个关键参数:(i)可写工作集大小,(ii)符合跳过技术的页面数量,(iii)跳过页面数量与页面脏率和页面传输速率的关系,并且错误地建模关键参数-每单位时间脏的新页面数量。在本文中,我们提出了一个考虑所有这些参数的新模型。我们对53个工作负载进行了彻底的验证,并显示KVM和Xen实时迁移的估计迁移时间的第90个百分位数误差分别为12%(8秒)和19%(14秒)。
{"title":"Towards a comprehensive performance model of virtual machine live migration","authors":"Senthil Nathan, U. Bellur, Purushottam Kulkarni","doi":"10.1145/2806777.2806838","DOIUrl":"https://doi.org/10.1145/2806777.2806838","url":null,"abstract":"Although many models exist to predict the time taken to migrate a virtual machine from one physical machine to another, our empirical validation of these models has shown the 90th percentile error to be 46% (43 secs) and 159% (112 secs) for KVM and Xen live migration, respectively. Our analysis reveals that these models are fundamentally flawed as they all fail to take into account the following three critical parameters: (i) the writable working set size, (ii) the number of pages eligible for the skip technique, (iii) the relation of the number of skipped pages with the page dirty rate and the page transfer rate, and incorrectly model the key parameter---the number of new pages dirtied per unit time. In this paper, we propose a novel model that takes all these parameters into account. We present a thorough validation with 53 workloads and show that the 90th percentile error in the estimated migration times is only 12% (8 secs) and 19% (14 secs) for KVM and Xen live migration, respectively.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126907786","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 54
Scheduling jobs across geo-distributed datacenters 跨地理分布数据中心调度作业
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806780
Chien-Chun Hung, L. Golubchik, Minlan Yu
With growing data volumes generated and stored across geo-distributed datacenters, it is becoming increasingly inefficient to aggregate all data required for computation at a single datacenter. Instead, a recent trend is to distribute computation to take advantage of data locality, thus reducing the resource (e.g., bandwidth) costs while improving performance. In this trend, new challenges are emerging in job scheduling, which requires coordination among the datacenters as each job runs across geo-distributed sites. In this paper, we propose novel job scheduling algorithms that coordinate job scheduling across datacenters with low overhead, while achieving near-optimal performance. Our extensive simulation study with realistic job traces shows that the proposed scheduling algorithms result in up to 50% improvement in average job completion time over the Shortest Remaining Processing Time (SRPT) based approaches.
随着跨地理分布式数据中心生成和存储的数据量不断增长,在单个数据中心聚合计算所需的所有数据变得越来越低效。相反,最近的趋势是分配计算以利用数据局部性,从而在提高性能的同时减少资源(例如带宽)成本。在这种趋势中,作业调度出现了新的挑战,因为每个作业在地理分布的站点上运行,需要在数据中心之间进行协调。在本文中,我们提出了新的作业调度算法,以低开销协调跨数据中心的作业调度,同时实现接近最优的性能。我们对实际作业轨迹进行了广泛的仿真研究,结果表明,与基于最短剩余处理时间(SRPT)的方法相比,所提出的调度算法的平均作业完成时间提高了50%。
{"title":"Scheduling jobs across geo-distributed datacenters","authors":"Chien-Chun Hung, L. Golubchik, Minlan Yu","doi":"10.1145/2806777.2806780","DOIUrl":"https://doi.org/10.1145/2806777.2806780","url":null,"abstract":"With growing data volumes generated and stored across geo-distributed datacenters, it is becoming increasingly inefficient to aggregate all data required for computation at a single datacenter. Instead, a recent trend is to distribute computation to take advantage of data locality, thus reducing the resource (e.g., bandwidth) costs while improving performance. In this trend, new challenges are emerging in job scheduling, which requires coordination among the datacenters as each job runs across geo-distributed sites. In this paper, we propose novel job scheduling algorithms that coordinate job scheduling across datacenters with low overhead, while achieving near-optimal performance. Our extensive simulation study with realistic job traces shows that the proposed scheduling algorithms result in up to 50% improvement in average job completion time over the Shortest Remaining Processing Time (SRPT) based approaches.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128394769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 132
Energy proportionality and workload consolidation for latency-critical applications 延迟关键型应用程序的能量比例和工作负载整合
Pub Date : 2015-08-27 DOI: 10.1145/2806777.2806848
G. Prekas, Mia Primorac, A. Belay, C. Kozyrakis, Edouard Bugnion
Energy proportionality and workload consolidation are important objectives towards increasing efficiency in large-scale datacenters. Our work focuses on achieving these goals in the presence of applications with μs-scale tail latency requirements. Such applications represent a growing subset of datacenter workloads and are typically deployed on dedicated servers, which is the simplest way to ensure low tail latency across all loads. Unfortunately, it also leads to low energy efficiency and low resource utilization during the frequent periods of medium or low load. We present the OS mechanisms and dynamic control needed to adjust core allocation and voltage/frequency settings based on the measured delays for latency-critical workloads. This allows for energy proportionality and frees the maximum amount of resources per server for other background applications, while respecting service-level objectives. Monitoring hardware queue depths allows us to detect increases in queuing latencies. Carefully coordinated adjustments to the NIC's packet redirection table enable us to reassign flow groups between the threads of a latency-critical application in milliseconds without dropping or reordering packets. We compare the efficiency of our solution to the Pareto-optimal frontier of 224 distinct static configurations. Dynamic resource control saves 44%--54% of processor energy, which corresponds to 85%--93% of the Pareto-optimal upper bound. Dynamic resource control also allows background jobs to run at 32%--46% of their standalone throughput, which corresponds to 82%--92% of the Pareto bound.
能源比例和工作负载整合是提高大型数据中心效率的重要目标。我们的工作重点是在具有μs级尾延迟需求的应用程序中实现这些目标。这类应用程序代表了数据中心工作负载的一个不断增长的子集,通常部署在专用服务器上,这是确保跨所有负载的低尾部延迟的最简单方法。然而,在频繁的中负荷或低负荷期间,这也导致能源效率低,资源利用率低。我们介绍了基于延迟关键工作负载的测量延迟来调整核心分配和电压/频率设置所需的操作系统机制和动态控制。这允许能量比例,并为其他后台应用程序释放每台服务器的最大资源量,同时尊重服务级目标。监视硬件队列深度使我们能够检测队列延迟的增加。仔细协调调整NIC的数据包重定向表,使我们能够在毫秒内重新分配延迟关键应用程序线程之间的流组,而不会丢弃或重新排序数据包。我们将我们的解决方案的效率与224种不同静态配置的帕累托最优边界进行了比较。动态资源控制可以节省44% ~ 54%的处理器能量,相当于pareto最优上界的85% ~ 93%。动态资源控制还允许后台作业以独立吞吐量的32%- 46%运行,这相当于帕累托界限的82%- 92%。
{"title":"Energy proportionality and workload consolidation for latency-critical applications","authors":"G. Prekas, Mia Primorac, A. Belay, C. Kozyrakis, Edouard Bugnion","doi":"10.1145/2806777.2806848","DOIUrl":"https://doi.org/10.1145/2806777.2806848","url":null,"abstract":"Energy proportionality and workload consolidation are important objectives towards increasing efficiency in large-scale datacenters. Our work focuses on achieving these goals in the presence of applications with μs-scale tail latency requirements. Such applications represent a growing subset of datacenter workloads and are typically deployed on dedicated servers, which is the simplest way to ensure low tail latency across all loads. Unfortunately, it also leads to low energy efficiency and low resource utilization during the frequent periods of medium or low load. We present the OS mechanisms and dynamic control needed to adjust core allocation and voltage/frequency settings based on the measured delays for latency-critical workloads. This allows for energy proportionality and frees the maximum amount of resources per server for other background applications, while respecting service-level objectives. Monitoring hardware queue depths allows us to detect increases in queuing latencies. Carefully coordinated adjustments to the NIC's packet redirection table enable us to reassign flow groups between the threads of a latency-critical application in milliseconds without dropping or reordering packets. We compare the efficiency of our solution to the Pareto-optimal frontier of 224 distinct static configurations. Dynamic resource control saves 44%--54% of processor energy, which corresponds to 85%--93% of the Pareto-optimal upper bound. Dynamic resource control also allows background jobs to run at 32%--46% of their standalone throughput, which corresponds to 82%--92% of the Pareto bound.","PeriodicalId":275158,"journal":{"name":"Proceedings of the Sixth ACM Symposium on Cloud Computing","volume":"214 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2015-08-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115739050","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 71
期刊
Proceedings of the Sixth ACM Symposium on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1