首页 > 最新文献

Proceedings of the 2017 Symposium on Cloud Computing最新文献

英文 中文
Reducing tail latencies in micro-batch streaming workloads 减少微批处理流工作负载的尾部延迟
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3134433
Faria Kalim, A. Tantawi, S. Costache, A. Youssef
Spark Streaming discretizes streams of data into micro-batches, each of which is further sub-divided into tasks and processed in parallel to improve job throughput. Previous work [2, 3] has lowered end-to-end latency in Spark Streaming. However, two causes of high tail latencies remain unaddressed: 1) data is not load-balanced across tasks, and 2) straggler tasks can increase end-to-end latency by 8 times more than the median task on a production cluster [1]. We propose a feedback-control mechanism that allows frameworks to adaptively load-balance workloads across tasks according to their processing speeds. The task runtimes are thus equalized, lowering end-to-end tail latency. Further, this reduces load on machines that have transient resource bottlenecks, thus resolving the bottlenecks and preventing them from having an enduring impact on task runtimes.
Spark Streaming将数据流离散为微批,每个微批进一步细分为任务并并行处理,以提高作业吞吐量。之前的工作[2,3]已经降低了Spark Streaming的端到端延迟。然而,高尾延迟的两个原因仍然没有得到解决:1)数据在任务之间没有负载均衡,2)离散任务可能会增加端到端延迟,比生产集群上的中位数任务多8倍[1]。我们提出了一种反馈控制机制,允许框架根据任务的处理速度自适应地平衡负载。因此,任务运行时是均衡的,降低了端到端的尾部延迟。此外,这减少了具有瞬时资源瓶颈的机器上的负载,从而解决了瓶颈并防止它们对任务运行时产生持久的影响。
{"title":"Reducing tail latencies in micro-batch streaming workloads","authors":"Faria Kalim, A. Tantawi, S. Costache, A. Youssef","doi":"10.1145/3127479.3134433","DOIUrl":"https://doi.org/10.1145/3127479.3134433","url":null,"abstract":"Spark Streaming discretizes streams of data into micro-batches, each of which is further sub-divided into tasks and processed in parallel to improve job throughput. Previous work [2, 3] has lowered end-to-end latency in Spark Streaming. However, two causes of high tail latencies remain unaddressed: 1) data is not load-balanced across tasks, and 2) straggler tasks can increase end-to-end latency by 8 times more than the median task on a production cluster [1]. We propose a feedback-control mechanism that allows frameworks to adaptively load-balance workloads across tasks according to their processing speeds. The task runtimes are thus equalized, lowering end-to-end tail latency. Further, this reduces load on machines that have transient resource bottlenecks, thus resolving the bottlenecks and preventing them from having an enduring impact on task runtimes.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74164057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
AKC: advanced KSM for cloud computing AKC:用于云计算的高级KSM
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131616
Sioh Lee, Bongkyu Kim, Youngpil Kim, C. Yoo
Kernel samepage merging (KSM) in Linux kernel archive is a memory deduplication scheme that finds duplicate pages and shares the page in order to alleviate memory bottleneck in cloud. However, because the KSM has to scan all pages in memory to find duplicate pages, KSM consumes high CPU cycles and so causes virtual machines (VMs) performance degradation [1]. This degradation of VMs performance is an obstacle in cloud to service real-time applications (i.e. Netflix) [3]. A previous work, CMD [1] proposed page grouping scheme to reduce page comparisons, but it requires special monitoring hardware, XLH [2] enhanced page sharing with the information of guest VM I/O operation. However, the CPU overhead of XLH is still very high - similar to the default KSM. to make KSM more useful, we need an optimization scheme that consume less CPU cycles. Therefore, we first profile the CPU cycle consumption of KSM and the results show that page comparison (28.77%) and page checksum (26.14%) take most of cycles. Based on the results, we propose advanced KSM for cloud computing (AKC) that consumes less CPU cycles than the default KSM. to reduce the number of page comparisons, we apply checksum based RB-tree structure. In addition, AKC decreases page checksum overhead with hardware-accelerated crc32 hash function.
Linux内核存档中的内核同页合并(Kernel samepage merge, KSM)是一种查找重复页面并共享页面的内存重复数据删除方案,以缓解云环境中的内存瓶颈。然而,由于KSM必须扫描内存中的所有页面才能找到重复的页面,因此KSM消耗很高的CPU周期,从而导致虚拟机(vm)性能下降[1]。这种虚拟机性能的下降是云服务实时应用程序(即Netflix)的障碍[3]。先前的工作CMD[1]提出了页面分组方案来减少页面比较,但它需要特殊的监控硬件,XLH[2]增强了与guest VM I/O操作信息的页面共享。然而,XLH的CPU开销仍然非常高——与默认的KSM类似。为了使KSM更有用,我们需要一个消耗更少CPU周期的优化方案。因此,我们首先分析了KSM的CPU周期消耗,结果表明页面比较(28.77%)和页面校验和(26.14%)占用了大部分周期。基于结果,我们提出了用于云计算(AKC)的高级KSM,它比默认KSM消耗更少的CPU周期。为了减少页面比较的次数,我们采用了基于校验和的rb树结构。此外,AKC通过硬件加速的crc32哈希函数减少了页面校验和开销。
{"title":"AKC: advanced KSM for cloud computing","authors":"Sioh Lee, Bongkyu Kim, Youngpil Kim, C. Yoo","doi":"10.1145/3127479.3131616","DOIUrl":"https://doi.org/10.1145/3127479.3131616","url":null,"abstract":"Kernel samepage merging (KSM) in Linux kernel archive is a memory deduplication scheme that finds duplicate pages and shares the page in order to alleviate memory bottleneck in cloud. However, because the KSM has to scan all pages in memory to find duplicate pages, KSM consumes high CPU cycles and so causes virtual machines (VMs) performance degradation [1]. This degradation of VMs performance is an obstacle in cloud to service real-time applications (i.e. Netflix) [3]. A previous work, CMD [1] proposed page grouping scheme to reduce page comparisons, but it requires special monitoring hardware, XLH [2] enhanced page sharing with the information of guest VM I/O operation. However, the CPU overhead of XLH is still very high - similar to the default KSM. to make KSM more useful, we need an optimization scheme that consume less CPU cycles. Therefore, we first profile the CPU cycle consumption of KSM and the results show that page comparison (28.77%) and page checksum (26.14%) take most of cycles. Based on the results, we propose advanced KSM for cloud computing (AKC) that consumes less CPU cycles than the default KSM. to reduce the number of page comparisons, we apply checksum based RB-tree structure. In addition, AKC decreases page checksum overhead with hardware-accelerated crc32 hash function.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"75893415","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Analysis of TPC-DS: the first standard benchmark for SQL-based big data systems TPC-DS分析:基于sql的大数据系统的第一个标准基准
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128603
Meikel Pöss, T. Rabl, H. Jacobsen
The advent of Web 2.0 companies, such as Facebook, Google, and Amazon with their insatiable appetite for vast amounts of structured, semi-structured, and unstructured data, triggered the development of Hadoop and related tools, e.g., YARN, MapReduce, and Pig, as well as NoSQL databases. These tools form an open source software stack to support the processing of large and diverse data sets on clustered systems to perform decision support tasks. Recently, SQL is resurrecting in many of these solutions, e.g., Hive, Stinger, Impala, Shark, and Presto. At the same time, RDBMS vendors are adding Hadoop support into their SQL engines, e.g., IBM's Big SQL, Actian's Vortex, Oracle's Big Data SQL, and SAP's HANA. Because there was no industry standard benchmark that could measure the performance of SQL-based big data solutions, marketing claims were mostly based on "cherry picked" subsets of the TPC-DS benchmark to suit individual companies strengths, while blending out their weaknesses. In this paper, we present and analyze our work on modifying TPC-DS to fill the void for an industry standard benchmark that is able to measure the performance of SQL-based big data solutions. The new benchmark was ratified by the TPC in early 2016. To show the significance of the new benchmark, we analyze performance data obtained on four different systems running big data, traditional RDBMS, and columnar in-memory architectures.
Web 2.0公司的出现,如Facebook、谷歌和Amazon,他们对大量结构化、半结构化和非结构化数据的贪欲无法满足,引发了Hadoop和相关工具的发展,如YARN、MapReduce和Pig,以及NoSQL数据库。这些工具形成了一个开源软件堆栈,以支持处理集群系统上的大型和不同的数据集,从而执行决策支持任务。最近,SQL在许多这些解决方案中复活,例如Hive、Stinger、Impala、Shark和Presto。与此同时,RDBMS供应商正在将Hadoop支持添加到他们的SQL引擎中,例如IBM的Big SQL、Actian的Vortex、Oracle的Big Data SQL和SAP的HANA。由于没有行业标准基准可以衡量基于sql的大数据解决方案的性能,营销主张主要是基于“精心挑选”的TPC-DS基准子集,以适应个别公司的优势,同时融合他们的弱点。在本文中,我们介绍并分析了我们在修改TPC-DS方面的工作,以填补能够衡量基于sql的大数据解决方案性能的行业标准基准的空白。2016年初,TPC批准了新的基准。为了展示新基准的重要性,我们分析了在运行大数据、传统RDBMS和列式内存架构的四个不同系统上获得的性能数据。
{"title":"Analysis of TPC-DS: the first standard benchmark for SQL-based big data systems","authors":"Meikel Pöss, T. Rabl, H. Jacobsen","doi":"10.1145/3127479.3128603","DOIUrl":"https://doi.org/10.1145/3127479.3128603","url":null,"abstract":"The advent of Web 2.0 companies, such as Facebook, Google, and Amazon with their insatiable appetite for vast amounts of structured, semi-structured, and unstructured data, triggered the development of Hadoop and related tools, e.g., YARN, MapReduce, and Pig, as well as NoSQL databases. These tools form an open source software stack to support the processing of large and diverse data sets on clustered systems to perform decision support tasks. Recently, SQL is resurrecting in many of these solutions, e.g., Hive, Stinger, Impala, Shark, and Presto. At the same time, RDBMS vendors are adding Hadoop support into their SQL engines, e.g., IBM's Big SQL, Actian's Vortex, Oracle's Big Data SQL, and SAP's HANA. Because there was no industry standard benchmark that could measure the performance of SQL-based big data solutions, marketing claims were mostly based on \"cherry picked\" subsets of the TPC-DS benchmark to suit individual companies strengths, while blending out their weaknesses. In this paper, we present and analyze our work on modifying TPC-DS to fill the void for an industry standard benchmark that is able to measure the performance of SQL-based big data solutions. The new benchmark was ratified by the TPC in early 2016. To show the significance of the new benchmark, we analyze performance data obtained on four different systems running big data, traditional RDBMS, and columnar in-memory architectures.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"76669164","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 36
STYX: a trusted and accelerated hierarchical SSL key management and distribution system for cloud based CDN application STYX:一个可信和加速的分层SSL密钥管理和分发系统,用于基于云的CDN应用
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3127482
Changzheng Wei, Jian Li, Weigang Li, Ping Yu, Haibing Guan
Protecting the customer's SSL private key is the paramount issue to persuade the website owners to migrate their contents onto the cloud infrastructure, besides the advantages of cloud infrastructure in terms of flexibility, efficiency, scalability and elasticity. The emerging Keyless SSL solution retains on-premise custody of customers' SSL private keys on their own servers. However, it suffers from significant performance degradation and limited scalability, caused by the long distance connection to Key Server for each new coming end-user request. The performance improvements using persistent session and key caching onto cloud will degrade the key invulnerability and discourage the website owners because of the cloud's security bugs. In this paper, the challenges of secured key protection and distribution are addressed in philosophy of "Storing the trusted DATA on untrusted platform and transmitting through untrusted channel". To this end, a three-phase hierarchical key management scheme, called STYX1 is proposed to provide the secured key protection together with hardware assisted service acceleration for cloud-based content delivery network (CCDN) applications. The STYX is implemented based on Intel Software Guard Extensions (SGX), Intel QuickAssist Technology (QAT) and SIGMA (SIGn-and-MAc) protocol. STYX can provide the tight key security guarantee by SGX based key distribution with a light overhead, and it can further significantly enhance the system performance with QAT based acceleration. The comprehensive evaluations show that the STYX not only guarantees the absolute security but also outperforms the direct HTTPS server deployed CDN without QAT by up to 5x throughput with significant latency reduction at the same time.
保护客户的SSL私钥是说服网站所有者将其内容迁移到云基础设施上的首要问题,除了云基础设施在灵活性、效率、可扩展性和弹性方面的优势之外。新兴的无密钥SSL解决方案在客户自己的服务器上保留客户SSL私钥的本地托管。但是,由于每个新到来的最终用户请求都需要与Key Server进行长距离连接,因此它的性能明显下降,可伸缩性有限。使用持久会话和密钥缓存到云上的性能改进将降低密钥的不受攻击性,并且由于云的安全漏洞而使网站所有者感到沮丧。本文以“可信的数据存储在不可信的平台上,通过不可信的通道传输”的理念,解决了安全密钥保护和分发面临的挑战。为此,提出了一种称为STYX1的三阶段分层密钥管理方案,为基于云的内容分发网络(CCDN)应用提供安全的密钥保护和硬件辅助的业务加速。STYX基于Intel Software Guard Extensions (SGX)、Intel QuickAssist Technology (QAT)和SIGMA (SIGn-and-MAc)协议实现。STYX可以通过基于SGX的密钥分发以较低的开销提供严密的密钥安全保证,并且可以通过基于QAT的加速进一步显著提高系统性能。综合评估表明,STYX不仅保证了绝对的安全性,而且比没有QAT的直接HTTPS服务器部署的CDN的吞吐量提高了5倍,同时显著降低了延迟。
{"title":"STYX: a trusted and accelerated hierarchical SSL key management and distribution system for cloud based CDN application","authors":"Changzheng Wei, Jian Li, Weigang Li, Ping Yu, Haibing Guan","doi":"10.1145/3127479.3127482","DOIUrl":"https://doi.org/10.1145/3127479.3127482","url":null,"abstract":"Protecting the customer's SSL private key is the paramount issue to persuade the website owners to migrate their contents onto the cloud infrastructure, besides the advantages of cloud infrastructure in terms of flexibility, efficiency, scalability and elasticity. The emerging Keyless SSL solution retains on-premise custody of customers' SSL private keys on their own servers. However, it suffers from significant performance degradation and limited scalability, caused by the long distance connection to Key Server for each new coming end-user request. The performance improvements using persistent session and key caching onto cloud will degrade the key invulnerability and discourage the website owners because of the cloud's security bugs. In this paper, the challenges of secured key protection and distribution are addressed in philosophy of \"Storing the trusted DATA on untrusted platform and transmitting through untrusted channel\". To this end, a three-phase hierarchical key management scheme, called STYX1 is proposed to provide the secured key protection together with hardware assisted service acceleration for cloud-based content delivery network (CCDN) applications. The STYX is implemented based on Intel Software Guard Extensions (SGX), Intel QuickAssist Technology (QAT) and SIGMA (SIGn-and-MAc) protocol. STYX can provide the tight key security guarantee by SGX based key distribution with a light overhead, and it can further significantly enhance the system performance with QAT based acceleration. The comprehensive evaluations show that the STYX not only guarantees the absolute security but also outperforms the direct HTTPS server deployed CDN without QAT by up to 5x throughput with significant latency reduction at the same time.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87431214","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Polygravity: traffic usage accountability via coarse-grained measurements in multi-tenant data centers Polygravity:在多租户数据中心中通过粗粒度测量实现流量使用责任
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3129258
H. Baek, Cheng Jin, Guofei Jiang, C. Lumezanu, J. Merwe, Ning Xia, Qiang Xu
Network usage accountability is critical in helping operators and customers of multi-tenant data centers deal with concerns such as capacity planning, resource allocation, hotspot detection, link failure detection, and troubleshooting. However, the cost of measurements and instrumentation to achieve flow-level accountability is non-trivial. We propose Polygravity to determine tenant traffic usage via lightweight measurements in multi-tenant data centers. We adopt a tomogravity model widely used in ISP networks, and adapt it to a multi-tenant data center environment. By integrating datacenter-specific domain knowledge, sampling-based partial estimation and gravity-based internal sinks/sources estimation, Polygravity addresses two key challenges for adapting tomogravity to a data center environment: sparse traffic matrices and internal traffic sinks/sources. We conducted extensive evaluation of our approach using realistic data center workloads. Our results show that Polygravity can determine tenant IP flow usage with less than 1% average relative error for tenants with fine-grained domain knowledge. In addition, for tenants with coarse-grained domain knowledge and with partial host-based sampling, Polygravity reduces the relative error of sampling-based estimation by 1/3.
网络使用责任对于帮助多租户数据中心的运营商和客户处理诸如容量规划、资源分配、热点检测、链路故障检测和故障排除等问题至关重要。然而,实现流级责任的测量和仪器的成本不是微不足道的。我们建议Polygravity通过多租户数据中心中的轻量级测量来确定租户流量使用情况。我们采用了在ISP网络中广泛使用的一种自重力模型,并将其适应于多租户数据中心环境。通过集成数据中心特定的领域知识、基于采样的部分估计和基于重力的内部汇/源估计,Polygravity解决了使tomogravity适应数据中心环境的两个关键挑战:稀疏流量矩阵和内部流量汇/源。我们使用实际的数据中心工作负载对我们的方法进行了广泛的评估。我们的结果表明,对于具有细粒度领域知识的租户,Polygravity可以以小于1%的平均相对误差确定租户IP流使用情况。此外,对于具有粗粒度领域知识和部分基于主机采样的租户,Polygravity将基于采样的估计的相对误差降低了1/3。
{"title":"Polygravity: traffic usage accountability via coarse-grained measurements in multi-tenant data centers","authors":"H. Baek, Cheng Jin, Guofei Jiang, C. Lumezanu, J. Merwe, Ning Xia, Qiang Xu","doi":"10.1145/3127479.3129258","DOIUrl":"https://doi.org/10.1145/3127479.3129258","url":null,"abstract":"Network usage accountability is critical in helping operators and customers of multi-tenant data centers deal with concerns such as capacity planning, resource allocation, hotspot detection, link failure detection, and troubleshooting. However, the cost of measurements and instrumentation to achieve flow-level accountability is non-trivial. We propose Polygravity to determine tenant traffic usage via lightweight measurements in multi-tenant data centers. We adopt a tomogravity model widely used in ISP networks, and adapt it to a multi-tenant data center environment. By integrating datacenter-specific domain knowledge, sampling-based partial estimation and gravity-based internal sinks/sources estimation, Polygravity addresses two key challenges for adapting tomogravity to a data center environment: sparse traffic matrices and internal traffic sinks/sources. We conducted extensive evaluation of our approach using realistic data center workloads. Our results show that Polygravity can determine tenant IP flow usage with less than 1% average relative error for tenants with fine-grained domain knowledge. In addition, for tenants with coarse-grained domain knowledge and with partial host-based sampling, Polygravity reduces the relative error of sampling-based estimation by 1/3.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"91199938","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
SEaMLESS: a SErvice migration cLoud architecture for energy saving and memory releaSing capabilities 无缝:一种服务迁移云架构,具有节能和内存释放功能
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3128604
Dino Lopez Pacheco, Quentin Jacquemart, A. Segalini, M. Rifai, M. Dione, G. Urvoy-Keller
Idle virtual machines (VMs) are a waste of resources in data centers. We introduce SEaMLESS, which transforms a fully-Hedged idle VM into a lightweight and resourceless Virtual Network Function (VNF). Idle VMs can then be saved to disk and release their memory. Simultaneously, the VNF provides service availability. Upon user activity, the appropriate VM is restored, without introducing any interruption for service users. Tens of VNFs can be contained within the same memory space required for one single VM, thereby facilitating ample resources savings when scaled up to a data center.
空闲的虚拟机是数据中心资源的一种浪费。我们介绍SEaMLESS,它将完全对冲的空闲VM转换为轻量级和无资源的虚拟网络功能(VNF)。空闲的虚拟机可以保存到磁盘并释放其内存。同时,VNF提供业务可用性。在用户活动时,恢复相应的虚拟机,而不会对业务用户造成任何中断。在单个VM所需的相同内存空间中可以包含数十个VNFs,从而在扩展到数据中心时可以节省大量资源。
{"title":"SEaMLESS: a SErvice migration cLoud architecture for energy saving and memory releaSing capabilities","authors":"Dino Lopez Pacheco, Quentin Jacquemart, A. Segalini, M. Rifai, M. Dione, G. Urvoy-Keller","doi":"10.1145/3127479.3128604","DOIUrl":"https://doi.org/10.1145/3127479.3128604","url":null,"abstract":"Idle virtual machines (VMs) are a waste of resources in data centers. We introduce SEaMLESS, which transforms a fully-Hedged idle VM into a lightweight and resourceless Virtual Network Function (VNF). Idle VMs can then be saved to disk and release their memory. Simultaneously, the VNF provides service availability. Upon user activity, the appropriate VM is restored, without introducing any interruption for service users. Tens of VNFs can be contained within the same memory space required for one single VM, thereby facilitating ample resources savings when scaled up to a data center.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89249351","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
CapNet: security and least authority in a capability-enabled cloud CapNet:启用功能的云中的安全性和最低权限
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3131209
A. Burtsev, David Johnson, Josh Kunz, E. Eide, J. Merwe
We present CapNet, a capability-based network architecture designed to enable least authority and secure collaboration in the cloud. CapNet allows fine-grained management of rights, recursive delegation, hierarchical policies, and least privilege. To enable secure collaboration, CapNet extends a classical capability model with support for decentralized authority. We implement CapNet in the substrate of a software-defined network, integrate it with the OpenStack cloud, and develop protocols enabling secure multi-party collaboration.
我们提出CapNet,这是一种基于能力的网络架构,旨在实现云中的最小权限和安全协作。CapNet允许对权限、递归委托、分层策略和最小权限进行细粒度管理。为了实现安全协作,CapNet扩展了经典的能力模型,支持分散的权限。我们在软件定义网络的基础上实现CapNet,将其与OpenStack云集成,并开发支持安全多方协作的协议。
{"title":"CapNet: security and least authority in a capability-enabled cloud","authors":"A. Burtsev, David Johnson, Josh Kunz, E. Eide, J. Merwe","doi":"10.1145/3127479.3131209","DOIUrl":"https://doi.org/10.1145/3127479.3131209","url":null,"abstract":"We present CapNet, a capability-based network architecture designed to enable least authority and secure collaboration in the cloud. CapNet allows fine-grained management of rights, recursive delegation, hierarchical policies, and least privilege. To enable secure collaboration, CapNet extends a classical capability model with support for decentralized authority. We implement CapNet in the substrate of a software-defined network, integrate it with the OpenStack cloud, and develop protocols enabling secure multi-party collaboration.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78304622","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
To edge or not to edge? 边缘还是不边缘?
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3132572
Faria Kalim, S. Noghabi, Shiv Verma
Edge computing caters to a wide range of use cases from latency sensitive to bandwidth constrained applications. However, the exact specifications of the edge that give the most benefit for each type of application are still unclear. We investigate the concrete conditions when the edge is feasible, i.e., when users observe performance gains from the edge while costs remain low for the providers, for an application that requires both low latency and high bandwidth: video analytics.
边缘计算满足从延迟敏感到带宽受限应用的广泛用例。然而,为每种类型的应用程序提供最大好处的边缘的确切规格仍然不清楚。我们研究了当边缘可行时的具体情况,即当用户观察到边缘的性能提升,同时提供商的成本保持较低时,对于需要低延迟和高带宽的应用程序:视频分析。
{"title":"To edge or not to edge?","authors":"Faria Kalim, S. Noghabi, Shiv Verma","doi":"10.1145/3127479.3132572","DOIUrl":"https://doi.org/10.1145/3127479.3132572","url":null,"abstract":"Edge computing caters to a wide range of use cases from latency sensitive to bandwidth constrained applications. However, the exact specifications of the edge that give the most benefit for each type of application are still unclear. We investigate the concrete conditions when the edge is feasible, i.e., when users observe performance gains from the edge while costs remain low for the providers, for an application that requires both low latency and high bandwidth: video analytics.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"73199845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
On-demand virtualization for live migration in bare metal cloud 裸机云中实时迁移的按需虚拟化
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3129254
Jae-Hwa Im, Jongyul Kim, Jonguk Kim, Seongwook Jin, S. Maeng
The level of demand for bare-metal cloud services has increased rapidly because such services are cost-effective for several types of workloads, and some cloud clients prefer a single-tenant environment due to the lower security vulnerability of such enviornments. However, as the bare-metal cloud does not utilize a virtualization layer, it cannot use live migration. Thus, there is a lack of manageability with the bare-metal cloud. Live migration support can improve the manageability of bare-metal cloud services significantly. This paper suggests an on-demand virtualization technique to improve the manageability of bare-metal cloud services. A thin virtualization layer is inserted into the bare-metal cloud when live migration is requested. After the completion of the live migration process, the thin virtualization layer is removed from the host. We modified BitVisor [19] to implement on-demand virtualization and live migration on the x86 architecture. The elapsed time of on-demand virtualization was negligible. It takes about 20 ms to insert the virtualization layer and 30 ms to remove the one. After removing the virtualization layer, the host machine works with bare-metal performance.
裸机云服务的需求水平迅速增加,因为此类服务对于几种类型的工作负载具有成本效益,并且由于此类环境的安全性漏洞较低,一些云客户更喜欢单租户环境。但是,由于裸机云不利用虚拟化层,因此它不能使用实时迁移。因此,裸机云缺乏可管理性。实时迁移支持可以显著提高裸机云服务的可管理性。本文提出了一种按需虚拟化技术,以提高裸机云服务的可管理性。当需要实时迁移时,瘦虚拟化层被插入裸机云中。热迁移过程完成后,将从主机上移除精简虚拟化层。我们修改了BitVisor[19],在x86架构上实现按需虚拟化和实时迁移。按需虚拟化的运行时间可以忽略不计。插入虚拟化层耗时约20ms,移除虚拟化层耗时约30ms。移除虚拟化层后,主机可以使用裸机性能。
{"title":"On-demand virtualization for live migration in bare metal cloud","authors":"Jae-Hwa Im, Jongyul Kim, Jonguk Kim, Seongwook Jin, S. Maeng","doi":"10.1145/3127479.3129254","DOIUrl":"https://doi.org/10.1145/3127479.3129254","url":null,"abstract":"The level of demand for bare-metal cloud services has increased rapidly because such services are cost-effective for several types of workloads, and some cloud clients prefer a single-tenant environment due to the lower security vulnerability of such enviornments. However, as the bare-metal cloud does not utilize a virtualization layer, it cannot use live migration. Thus, there is a lack of manageability with the bare-metal cloud. Live migration support can improve the manageability of bare-metal cloud services significantly. This paper suggests an on-demand virtualization technique to improve the manageability of bare-metal cloud services. A thin virtualization layer is inserted into the bare-metal cloud when live migration is requested. After the completion of the live migration process, the thin virtualization layer is removed from the host. We modified BitVisor [19] to implement on-demand virtualization and live migration on the x86 architecture. The elapsed time of on-demand virtualization was negligible. It takes about 20 ms to insert the virtualization layer and 30 ms to remove the one. After removing the virtualization layer, the host machine works with bare-metal performance.","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"72852558","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Stocator: an object store aware connector for apache spark Stocator:一个用于apache spark的对象存储感知连接器
Pub Date : 2017-09-24 DOI: 10.1145/3127479.3134761
G. Vernik, M. Factor, E. K. Kolodner, Effi Ofer, P. Michiardi, Francesco Pace
Data is the natural resource of the 21st century. It is being produced at dizzying rates, e.g., for genomics, for media and entertainment, and for Internet of Things. Object storage systems such as Amazon S3, Azure Blob storage, and IBM Cloud Object Storage, are highly scalable distributed storage systems that offer high capacity, cost effective storage. But it is not enough just to store data; we also need to derive value from it. Apache Spark is the leading big data analytics processing engine combining MapReduce, SQL, streaming, and complex analytics. We present Stocator, a high performance storage connector, enabling Spark to work directly on data stored in object storage systems, while providing the same correctness guarantees as Hadoop's original storage system, HDFS. Current object storage connectors from the Hadoop community, e.g., for the S3 and Swift APIs, do not deal well with eventual consistency, which can lead to failure. These connectors assume file system semantics, which is natural given that their model of operation is based on interaction with HDFS. In particular, Spark and Hadoop achieve fault tolerance and enable speculative execution by creating temporary files, listing directories to identify these files, and then renaming them. This paradigm avoids interference between tasks doing the same work and thus writing output with the same name. However, with eventually consistent object storage, a container listing may not yet include a recently created object, and thus an object may not be renamed, leading to incomplete or incorrect results. Solutions such as EMRFS [1] from Amazon, S3mper [4] from Netflix, and S3Guard [2], attempt to overcome eventual consistency by requiring additional strongly consistent data storage. These solutions require multiple storage systems, are costly, and can introduce issues of consistency between the stores. Current object storage connectors from the Hadoop community are also notorious for their poor performance for write workloads. This, too, stems from their use of the rename operation, which is not a native object storage operation; not only is it not atomic, but it must be implemented using a costly copy operation, followed by delete. Others have tried to improve the performance of object storage connectors by eliminating rename, e.g., the Direct-ParquetOutputCommitter [5] for S3a introduced by Databricks, but have failed to preserve fault tolerance and speculation. Stocator takes advantage of object storage semantics to achieve both high performance and fault tolerance. It eliminates the rename paradigm by writing each output object to its final name. The name includes both the part number and the attempt number, so that multiple attempts to write the same part use different objects. Stocator proposes to extend an already existing success indicator object written at the end of a Spark job, to include a manifest with the names of all the objects that compose the final output; this ensures that
数据是21世纪的自然资源。它正在以令人眼花缭乱的速度生产,例如基因组学,媒体和娱乐以及物联网。对象存储系统(如Amazon S3、Azure Blob存储和IBM Cloud Object storage)是高度可扩展的分布式存储系统,提供高容量、高成本效益的存储。但仅仅存储数据是不够的;我们还需要从中获得价值。Apache Spark是领先的大数据分析处理引擎,结合了MapReduce、SQL、流和复杂分析。我们介绍了Stocator,一个高性能的存储连接器,使Spark能够直接处理存储在对象存储系统中的数据,同时提供与Hadoop原始存储系统HDFS相同的正确性保证。目前来自Hadoop社区的对象存储连接器,例如S3和Swift api,不能很好地处理最终的一致性,这可能导致失败。这些连接器假定文件系统语义,这是很自然的,因为它们的操作模型是基于与HDFS的交互。特别是,Spark和Hadoop通过创建临时文件,列出目录来识别这些文件,然后重命名它们来实现容错和推测执行。此范例避免了执行相同工作的任务之间的干扰,从而避免了使用相同名称编写输出。然而,对于最终一致的对象存储,容器清单可能还没有包含最近创建的对象,因此对象可能没有被重命名,从而导致不完整或不正确的结果。Amazon的EMRFS[1]、Netflix的S3mper[4]和S3Guard[2]等解决方案试图通过需要额外的强一致性数据存储来克服最终的一致性。这些解决方案需要多个存储系统,成本很高,并且可能导致存储之间的一致性问题。目前来自Hadoop社区的对象存储连接器也因其糟糕的写工作负载性能而臭名昭著。这也源于它们使用的重命名操作,该操作不是本机对象存储操作;它不仅不是原子的,而且必须使用代价高昂的复制操作来实现,然后再执行删除操作。其他人试图通过消除重命名来提高对象存储连接器的性能,例如,Databricks为S3a引入的Direct-ParquetOutputCommitter[5],但未能保持容错和推测性。Stocator利用对象存储语义来实现高性能和容错性。它通过将每个输出对象写入其最终名称来消除重命名范例。名称包括部件号和尝试号,以便多次尝试写入相同的部件时使用不同的对象。Stocator建议扩展已经存在的在Spark作业结束时编写的成功指示器对象,以包含包含组成最终输出的所有对象名称的清单;这确保后续作业将正确读取输出,而无需诉诸结果可能不一致的列表操作。通过利用对象创建的固有原子性和使用清单,我们获得容错性并启用推测执行;通过避免重命名范例,我们大大降低了连接器的复杂性和对象存储上的操作数量。我们已经实现了我们的连接器,并在开源中共享了它[3]。我们将其性能与S3a和Hadoop Swift连接器在一系列工作负载下的性能进行了比较,发现它在对象存储上执行的操作要少得多,在某些情况下只有三十分之一。由于对象存储服务的价格通常包括基于执行的操作数量的收费,因此这种操作的减少除了减少客户端软件的负载外,还降低了客户端的成本。它还降低了对象存储提供商的成本和负载,因为它可以用相同的处理能力为更多的客户端提供服务。Stocator还大大提高了运行在对象存储上的Spark工作负载的性能,特别是对于写密集型工作负载,它的速度可以提高18倍。
{"title":"Stocator: an object store aware connector for apache spark","authors":"G. Vernik, M. Factor, E. K. Kolodner, Effi Ofer, P. Michiardi, Francesco Pace","doi":"10.1145/3127479.3134761","DOIUrl":"https://doi.org/10.1145/3127479.3134761","url":null,"abstract":"Data is the natural resource of the 21st century. It is being produced at dizzying rates, e.g., for genomics, for media and entertainment, and for Internet of Things. Object storage systems such as Amazon S3, Azure Blob storage, and IBM Cloud Object Storage, are highly scalable distributed storage systems that offer high capacity, cost effective storage. But it is not enough just to store data; we also need to derive value from it. Apache Spark is the leading big data analytics processing engine combining MapReduce, SQL, streaming, and complex analytics. We present Stocator, a high performance storage connector, enabling Spark to work directly on data stored in object storage systems, while providing the same correctness guarantees as Hadoop's original storage system, HDFS. Current object storage connectors from the Hadoop community, e.g., for the S3 and Swift APIs, do not deal well with eventual consistency, which can lead to failure. These connectors assume file system semantics, which is natural given that their model of operation is based on interaction with HDFS. In particular, Spark and Hadoop achieve fault tolerance and enable speculative execution by creating temporary files, listing directories to identify these files, and then renaming them. This paradigm avoids interference between tasks doing the same work and thus writing output with the same name. However, with eventually consistent object storage, a container listing may not yet include a recently created object, and thus an object may not be renamed, leading to incomplete or incorrect results. Solutions such as EMRFS [1] from Amazon, S3mper [4] from Netflix, and S3Guard [2], attempt to overcome eventual consistency by requiring additional strongly consistent data storage. These solutions require multiple storage systems, are costly, and can introduce issues of consistency between the stores. Current object storage connectors from the Hadoop community are also notorious for their poor performance for write workloads. This, too, stems from their use of the rename operation, which is not a native object storage operation; not only is it not atomic, but it must be implemented using a costly copy operation, followed by delete. Others have tried to improve the performance of object storage connectors by eliminating rename, e.g., the Direct-ParquetOutputCommitter [5] for S3a introduced by Databricks, but have failed to preserve fault tolerance and speculation. Stocator takes advantage of object storage semantics to achieve both high performance and fault tolerance. It eliminates the rename paradigm by writing each output object to its final name. The name includes both the part number and the attempt number, so that multiple attempts to write the same part use different objects. Stocator proposes to extend an already existing success indicator object written at the end of a Spark job, to include a manifest with the names of all the objects that compose the final output; this ensures that","PeriodicalId":20679,"journal":{"name":"Proceedings of the 2017 Symposium on Cloud Computing","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2017-09-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83221819","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
期刊
Proceedings of the 2017 Symposium on Cloud Computing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1