首页 > 最新文献

Companion of the 2018 ACM/SPEC International Conference on Performance Engineering最新文献

英文 中文
SPECnet: Predicting SPEC Scores using Deep Learning SPECnet:使用深度学习预测SPEC分数
Dibyendu Das, Prakash S. Raghavendra, Arun Ramachandran
In this work we show how to build a deep neural network (DNN) to predict SPEC® scores - called the SPECnet. More than ten years have passed since the introduction of the SPEC CPU2006 suite (retired in January 2018) and thousands of submissions are available for CPU2006 integer and floating point benchmarks. We build a DNN which inputs hardware and software features from these submissions and is subsequently trained on the corresponding reported SPEC scores. We then use the trained DNN to predict scores for upcoming machine configurations. We achieve 5%-7% training and dev/test errors pointing to pretty high accuracy rates (93%-95%) for prediction. Such a prediction rate is very comparable to expected human-level accuracy of 97%-98% achieved via careful performance modelling of the core and un-core system components. In addition to the CPU2006 suite, we also apply SPECnet to SPEComp2012 and SPECjbb2015. Though the reported submissions for these benchmark suites number in hundreds only, we show that such a DNN is able to predict for these benchmarks reasonably well (~85% accuracy) too. Our SPECnet implementation uses state-of-the-art Tensorflow infrastructure and is extremely flexible and extensible.
在这项工作中,我们展示了如何构建一个深度神经网络(DNN)来预测SPEC®分数-称为SPECnet。自引入SPEC CPU2006套件(2018年1月退役)以来,已经过去了十多年,CPU2006整数和浮点基准测试已经提交了数千份。我们构建了一个DNN,它从这些提交中输入硬件和软件特征,并随后根据相应的报告SPEC分数进行训练。然后,我们使用训练好的DNN来预测即将到来的机器配置的分数。我们实现了5%-7%的训练和开发/测试错误,这表明预测的准确率相当高(93%-95%)。这样的预测率与通过对核心和非核心系统组件进行仔细的性能建模而达到的预期人类水平的97%-98%的准确率非常相似。除了CPU2006套件,我们还将SPECnet应用于speccomp2012和SPECjbb2015。虽然这些基准测试套件的报告提交数量只有数百个,但我们表明这样的DNN也能够相当好地预测这些基准测试(准确率约为85%)。我们的SPECnet实现使用最先进的Tensorflow基础设施,非常灵活和可扩展。
{"title":"SPECnet: Predicting SPEC Scores using Deep Learning","authors":"Dibyendu Das, Prakash S. Raghavendra, Arun Ramachandran","doi":"10.1145/3185768.3186301","DOIUrl":"https://doi.org/10.1145/3185768.3186301","url":null,"abstract":"In this work we show how to build a deep neural network (DNN) to predict SPEC® scores - called the SPECnet. More than ten years have passed since the introduction of the SPEC CPU2006 suite (retired in January 2018) and thousands of submissions are available for CPU2006 integer and floating point benchmarks. We build a DNN which inputs hardware and software features from these submissions and is subsequently trained on the corresponding reported SPEC scores. We then use the trained DNN to predict scores for upcoming machine configurations. We achieve 5%-7% training and dev/test errors pointing to pretty high accuracy rates (93%-95%) for prediction. Such a prediction rate is very comparable to expected human-level accuracy of 97%-98% achieved via careful performance modelling of the core and un-core system components. In addition to the CPU2006 suite, we also apply SPECnet to SPEComp2012 and SPECjbb2015. Though the reported submissions for these benchmark suites number in hundreds only, we show that such a DNN is able to predict for these benchmarks reasonably well (~85% accuracy) too. Our SPECnet implementation uses state-of-the-art Tensorflow infrastructure and is extremely flexible and extensible.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"21 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89203271","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards the Performance Analysis of Apache Tez Applications Apache Tez应用程序的性能分析
J. I. Requeno, Iñigo Gascón, J. Merseguer
Apache Tez is an application framework for large data processing using interactive queries. When a Tez developer faces the fulfillment of performance requirements s/he needs to configure and optimize the Tez application to specific execution contexts. However, these are not easy tasks, though the Apache Tez configuration will impact in the performance of the application significantly. Therefore, we propose some steps, towards the modeling and simulation of Apache Tez applications, that can help in the performance assessment of Tez designs. For the modeling, we propose a UML profile for Apache Tez. For the simulation, we propose to transform the stereotypes of the profile into stochastic Petri nets, which can be eventually used for computing performance metrics.
Apache Tez是一个使用交互式查询进行大数据处理的应用程序框架。当Tez开发人员面临性能需求的实现时,他/她需要配置和优化Tez应用程序以适应特定的执行上下文。然而,这些都不是简单的任务,尽管Apache Tez配置将显著影响应用程序的性能。因此,我们提出了一些步骤,朝着Apache Tez应用程序的建模和仿真,可以帮助Tez设计的性能评估。为了建模,我们为Apache Tez提出了一个UML概要文件。对于模拟,我们建议将轮廓的原型转换为随机Petri网,最终可用于计算性能指标。
{"title":"Towards the Performance Analysis of Apache Tez Applications","authors":"J. I. Requeno, Iñigo Gascón, J. Merseguer","doi":"10.1145/3185768.3186284","DOIUrl":"https://doi.org/10.1145/3185768.3186284","url":null,"abstract":"Apache Tez is an application framework for large data processing using interactive queries. When a Tez developer faces the fulfillment of performance requirements s/he needs to configure and optimize the Tez application to specific execution contexts. However, these are not easy tasks, though the Apache Tez configuration will impact in the performance of the application significantly. Therefore, we propose some steps, towards the modeling and simulation of Apache Tez applications, that can help in the performance assessment of Tez designs. For the modeling, we propose a UML profile for Apache Tez. For the simulation, we propose to transform the stereotypes of the profile into stochastic Petri nets, which can be eventually used for computing performance metrics.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"22 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"80439248","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
CAUS: An Elasticity Controller for a Containerized Microservice 原因:容器化微服务的弹性控制器
Floriment Klinaku, Markus Frank, Steffen Becker
Recent trends towards microservice architectures and containers as a deployment unit raise the need for novel adaptation processes to enable elasticity for containerized microservices. Microservices facing unpredictable workloads need to react fast and match the supply as closely as possible to the demand in order to guarantee quality objectives and to keep costs at a minimum. Current state-of-the-art approaches, that react on conditions which reflect the need to scale, are either slow or lack precision in supplying the demand with the adequate capacity. Therefore, we propose a novel heuristic adaptation process which enables elasticity for a particular containerized microservice. The proposed method consists of two mechanisms that complement each other. One part reacts to changes in load intensity by scaling container instances depending on their processing capability. The other mechanism manages additional containers as a buffer to handle unpredictable workload changes. We evaluate the proposed adaptation process and discuss its effectiveness and feasibility in controlling autonomously the number of replicated containers.
微服务架构和容器作为部署单元的最新趋势提出了对新的适应流程的需求,以实现容器化微服务的弹性。面对不可预测工作负载的微服务需要快速响应,并尽可能地将供应与需求匹配起来,以保证质量目标并将成本保持在最低水平。目前最先进的办法是根据反映规模需要的条件作出反应,在向需求提供足够的能力方面要么缓慢,要么缺乏准确性。因此,我们提出了一种新的启发式适应过程,该过程可以为特定的容器化微服务提供弹性。所提出的方法由两个相互补充的机制组成。一部分通过根据容器实例的处理能力对负载强度的变化做出反应。另一种机制管理额外的容器作为缓冲区,以处理不可预测的工作负载变化。我们评估了提出的适应过程,并讨论了其在自主控制复制容器数量方面的有效性和可行性。
{"title":"CAUS: An Elasticity Controller for a Containerized Microservice","authors":"Floriment Klinaku, Markus Frank, Steffen Becker","doi":"10.1145/3185768.3186296","DOIUrl":"https://doi.org/10.1145/3185768.3186296","url":null,"abstract":"Recent trends towards microservice architectures and containers as a deployment unit raise the need for novel adaptation processes to enable elasticity for containerized microservices. Microservices facing unpredictable workloads need to react fast and match the supply as closely as possible to the demand in order to guarantee quality objectives and to keep costs at a minimum. Current state-of-the-art approaches, that react on conditions which reflect the need to scale, are either slow or lack precision in supplying the demand with the adequate capacity. Therefore, we propose a novel heuristic adaptation process which enables elasticity for a particular containerized microservice. The proposed method consists of two mechanisms that complement each other. One part reacts to changes in load intensity by scaling container instances depending on their processing capability. The other mechanism manages additional containers as a buffer to handle unpredictable workload changes. We evaluate the proposed adaptation process and discuss its effectiveness and feasibility in controlling autonomously the number of replicated containers.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"20 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74801687","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
PROWL: Towards Predicting the Runtime of Batch Workloads PROWL:预测批处理工作负载的运行时间
Dheeraj Chahal, Benny Mathew
Many applications in the enterprise domain require batch processing to perform business critical operations. Batch jobs perform automated, complex processing of large volumes of data without human intervention. Parallel processing allows multiple batch jobs to run concurrently to minimize the total completion time. However, this may result in one or more jobs exceeding their individual completion deadline due to resource sharing. The objective of this work is to predict the completion time of a batch job when it is running in conjunction with other batch jobs. Batch jobs may be multi-threaded and threads can have distinct CPU requirements. Our predictions are based on a simulation model using the service demand (total CPU time required) of each thread in the job. Moreover, for multi-threaded jobs, we simulate the server with instantaneous CPU utilization of each job in the small intervals instead of aggregate value while predicting the completion time. In this paper, a simulation based method is presented to predict the completion time of each batch job in a concurrent run of multiple jobs. A validation study with synthetic benchmark FIO shows that the job completion time prediction error is less than 15% in the worst case.
企业域中的许多应用程序都需要批处理来执行业务关键操作。批处理作业在没有人工干预的情况下对大量数据执行自动化、复杂的处理。并行处理允许多个批处理作业并发运行,以最小化总完成时间。然而,由于资源共享,这可能导致一个或多个作业超过其各自的完成期限。这项工作的目标是预测批处理作业与其他批处理作业一起运行时的完成时间。批处理作业可能是多线程的,并且线程可能有不同的CPU需求。我们的预测基于使用作业中每个线程的服务需求(所需的总CPU时间)的模拟模型。此外,对于多线程作业,在预测完成时间时,我们使用每个作业在小间隔内的瞬时CPU利用率而不是汇总值来模拟服务器。本文提出了一种基于仿真的方法来预测多批作业并发运行中每个批作业的完成时间。综合基准FIO验证研究表明,在最坏情况下,作业完成时间预测误差小于15%。
{"title":"PROWL: Towards Predicting the Runtime of Batch Workloads","authors":"Dheeraj Chahal, Benny Mathew","doi":"10.1145/3185768.3186407","DOIUrl":"https://doi.org/10.1145/3185768.3186407","url":null,"abstract":"Many applications in the enterprise domain require batch processing to perform business critical operations. Batch jobs perform automated, complex processing of large volumes of data without human intervention. Parallel processing allows multiple batch jobs to run concurrently to minimize the total completion time. However, this may result in one or more jobs exceeding their individual completion deadline due to resource sharing. The objective of this work is to predict the completion time of a batch job when it is running in conjunction with other batch jobs. Batch jobs may be multi-threaded and threads can have distinct CPU requirements. Our predictions are based on a simulation model using the service demand (total CPU time required) of each thread in the job. Moreover, for multi-threaded jobs, we simulate the server with instantaneous CPU utilization of each job in the small intervals instead of aggregate value while predicting the completion time. In this paper, a simulation based method is presented to predict the completion time of each batch job in a concurrent run of multiple jobs. A validation study with synthetic benchmark FIO shows that the job completion time prediction error is less than 15% in the worst case.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"14 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88948128","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Package-Aware Scheduling of FaaS Functions FaaS功能的包感知调度
Cristina L. Abad, Edwin F. Boza, Erwin Van Eyk
We consider the problem of scheduling small cloud functions on serverless computing platforms. Fast deployment and execution of these functions is critical, for example, for microservices architectures. However, functions that require large packages or libraries are bloated and start slowly. A solution is to cache packages at the worker nodes instead of bundling them with the functions. However, existing FaaS schedulers are vanilla load balancers, agnostic of any packages that may have been cached in response to prior function executions, and cannot reap the benefits of package caching (other than by chance). To address this problem, we propose a package-aware scheduling algorithm that tries to assign functions that require the same package to the same worker node. Our algorithm increases the hit rate of the package cache and, as a result, reduces the latency of the cloud functions. At the same time, we consider the load sustained by the workers and actively seek to avoid imbalance beyond a configurable threshold. Our preliminary evaluation shows that, even with our limited exploration of the configuration space so-far, we can achieve 66% performance improvement at the cost of a (manageable) higher node imbalance.
我们考虑在无服务器计算平台上调度小型云功能的问题。这些功能的快速部署和执行至关重要,例如,对于微服务架构而言。但是,需要大型包或库的函数会变得臃肿,并且启动缓慢。一种解决方案是在工作节点上缓存包,而不是将它们与函数捆绑在一起。但是,现有的FaaS调度器是普通的负载平衡器,不知道可能为响应先前的函数执行而缓存的任何包,并且无法获得包缓存的好处(除非偶然)。为了解决这个问题,我们提出了一种包感知调度算法,该算法试图将需要相同包的功能分配给相同的工作节点。我们的算法提高了包缓存的命中率,从而减少了云功能的延迟。同时,我们考虑工人承受的负载,并积极寻求避免超过可配置阈值的不平衡。我们的初步评估表明,即使到目前为止我们对配置空间的探索有限,我们也可以以(可管理的)更高的节点不平衡为代价实现66%的性能改进。
{"title":"Package-Aware Scheduling of FaaS Functions","authors":"Cristina L. Abad, Edwin F. Boza, Erwin Van Eyk","doi":"10.1145/3185768.3186294","DOIUrl":"https://doi.org/10.1145/3185768.3186294","url":null,"abstract":"We consider the problem of scheduling small cloud functions on serverless computing platforms. Fast deployment and execution of these functions is critical, for example, for microservices architectures. However, functions that require large packages or libraries are bloated and start slowly. A solution is to cache packages at the worker nodes instead of bundling them with the functions. However, existing FaaS schedulers are vanilla load balancers, agnostic of any packages that may have been cached in response to prior function executions, and cannot reap the benefits of package caching (other than by chance). To address this problem, we propose a package-aware scheduling algorithm that tries to assign functions that require the same package to the same worker node. Our algorithm increases the hit rate of the package cache and, as a result, reduces the latency of the cloud functions. At the same time, we consider the load sustained by the workers and actively seek to avoid imbalance beyond a configurable threshold. Our preliminary evaluation shows that, even with our limited exploration of the configuration space so-far, we can achieve 66% performance improvement at the cost of a (manageable) higher node imbalance.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"s3-49 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"90842895","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 27
On the Simulation of Processors Enhanced for Security in Virtualization 虚拟化中增强安全性的处理器仿真研究
Swapneel C. Mhatre, P. Chandran, J. R
Computer system simulators model the hardware and reduce the time required for the design of the hardware, by exploring the design space and thereby eliminating the time-consuming process of testing each possibility by actually building the hardware. Simulators used in Computer Architecture typically model processors, memories and disks. With the regaining of the importance of virtualization, there is a need for a simulator for virtualization-enhanced processors, especially, with security related enhancements. But adapting existing simulators to model virtualization-enabled processors is a deceptively difficult task, wrought with complications like multiple access levels. In our research, we aim to identify methods for effectively simulating virtualization-enabled processors. This paper reports the results of a preliminary simulation of Architectural Support for Memory Isolation (ASMI), a memory architecture model that provides memory isolation, using ModelSim and highlights the need for a hardware-based simulator for processors enhanced for security in virtualization.
计算机系统模拟器对硬件进行建模,并通过探索设计空间来减少硬件设计所需的时间,从而消除了通过实际构建硬件来测试每种可能性的耗时过程。计算机体系结构中使用的模拟器通常对处理器、存储器和磁盘进行建模。随着虚拟化重要性的恢复,有必要为虚拟化增强的处理器提供模拟器,特别是与安全性相关的增强。但是,使现有的模拟器适应支持虚拟化的处理器是一项看似困难的任务,其中存在多访问级别等复杂问题。在我们的研究中,我们的目标是确定有效地模拟支持虚拟化的处理器的方法。本文报告了使用ModelSim对内存隔离体系结构支持(ASMI)(一种提供内存隔离的内存体系结构模型)进行初步模拟的结果,并强调了对基于硬件的处理器模拟器的需求,以增强虚拟化中的安全性。
{"title":"On the Simulation of Processors Enhanced for Security in Virtualization","authors":"Swapneel C. Mhatre, P. Chandran, J. R","doi":"10.1145/3185768.3185774","DOIUrl":"https://doi.org/10.1145/3185768.3185774","url":null,"abstract":"Computer system simulators model the hardware and reduce the time required for the design of the hardware, by exploring the design space and thereby eliminating the time-consuming process of testing each possibility by actually building the hardware. Simulators used in Computer Architecture typically model processors, memories and disks. With the regaining of the importance of virtualization, there is a need for a simulator for virtualization-enhanced processors, especially, with security related enhancements. But adapting existing simulators to model virtualization-enabled processors is a deceptively difficult task, wrought with complications like multiple access levels. In our research, we aim to identify methods for effectively simulating virtualization-enabled processors. This paper reports the results of a preliminary simulation of Architectural Support for Memory Isolation (ASMI), a memory architecture model that provides memory isolation, using ModelSim and highlights the need for a hardware-based simulator for processors enhanced for security in virtualization.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"16 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84414154","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
An Auto-Tuning Framework for a NUMA-Aware Hessenberg Reduction Algorithm 基于numa感知的Hessenberg约简算法的自动调优框架
Mahmoud Eljammaly, L. Karlsson, B. Kågström
The performance of a recently developed Hessenberg reduction algorithm greatly depends on the values chosen for its tunable parameters. The problem is hard to solve effectively with generic methods and tools. We describe a modular auto-tuning framework in which the underlying optimization algorithm is easy to substitute. The framework exposes sub-problems of standard auto-tuning type for which existing generic methods can be reused. The outputs of concurrently executing sub-tuners are assembled by the framework into a solution to the original problem. This paper presents work-in-progress.
最近开发的海森伯格约简算法的性能在很大程度上取决于其可调参数的选择值。用通用的方法和工具很难有效地解决这个问题。我们描述了一个模块化的自动调优框架,其中底层的优化算法易于替换。该框架暴露了标准自动调优类型的子问题,现有的泛型方法可以被重用。并发执行子调谐器的输出由框架组装成原始问题的解决方案。本文介绍了正在进行的工作。
{"title":"An Auto-Tuning Framework for a NUMA-Aware Hessenberg Reduction Algorithm","authors":"Mahmoud Eljammaly, L. Karlsson, B. Kågström","doi":"10.1145/3185768.3186304","DOIUrl":"https://doi.org/10.1145/3185768.3186304","url":null,"abstract":"The performance of a recently developed Hessenberg reduction algorithm greatly depends on the values chosen for its tunable parameters. The problem is hard to solve effectively with generic methods and tools. We describe a modular auto-tuning framework in which the underlying optimization algorithm is easy to substitute. The framework exposes sub-problems of standard auto-tuning type for which existing generic methods can be reused. The outputs of concurrently executing sub-tuners are assembled by the framework into a solution to the original problem. This paper presents work-in-progress.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89562654","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Autoscaling Performance Measurement Tool 自动缩放性能测量工具
Anshul Jindal, Vladimir Podolskiy, M. Gerndt
More companies are shifting focus to adding more layers of virtualization for their cloud applications thus increasing the flexibility in development, deployment and management of applications. Increase in the number of layers can result in additional overhead during autoscaling and also in coordination issues while layers may use the same resources while managed by different software. In order to capture these multilayered autoscaling performance issues, an Autoscaling Performance Measurement Tool (APMT) was developed. This tool evaluates the performance of cloud autoscaling solutions and combinations thereof for varying types of load patterns. In the paper, we highlight the architecture of the tool and its configuration. An autoscaling behavior for major IaaS providers with Kubernetes pods as the second layer of virtualization is illustrated using the data collected by APMT.
越来越多的公司将注意力转移到为他们的云应用程序添加更多的虚拟化层,从而增加应用程序开发、部署和管理的灵活性。层数的增加可能会导致自动伸缩期间的额外开销,也会导致协调问题,因为层可能使用相同的资源,但由不同的软件管理。为了捕获这些多层自动缩放性能问题,开发了一个自动缩放性能测量工具(APMT)。此工具评估云自动缩放解决方案的性能及其组合,以适应不同类型的负载模式。在本文中,我们重点介绍了该工具的体系结构及其配置。使用APMT收集的数据说明了使用Kubernetes pod作为第二层虚拟化的主要IaaS提供商的自动伸缩行为。
{"title":"Autoscaling Performance Measurement Tool","authors":"Anshul Jindal, Vladimir Podolskiy, M. Gerndt","doi":"10.1145/3185768.3186293","DOIUrl":"https://doi.org/10.1145/3185768.3186293","url":null,"abstract":"More companies are shifting focus to adding more layers of virtualization for their cloud applications thus increasing the flexibility in development, deployment and management of applications. Increase in the number of layers can result in additional overhead during autoscaling and also in coordination issues while layers may use the same resources while managed by different software. In order to capture these multilayered autoscaling performance issues, an Autoscaling Performance Measurement Tool (APMT) was developed. This tool evaluates the performance of cloud autoscaling solutions and combinations thereof for varying types of load patterns. In the paper, we highlight the architecture of the tool and its configuration. An autoscaling behavior for major IaaS providers with Kubernetes pods as the second layer of virtualization is illustrated using the data collected by APMT.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"88129184","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Evaluation of Energy Consumption of Replicated Tasks in a Volunteer Computing Environment 志愿计算环境下重复任务的能量消耗评价
A. McGough, M. Forshaw
High Throughput Computing allows workloads of many thousands of tasks to be performed efficiently over many distributed resources and frees the user from the laborious process of managing task deployment, execution and result collection. However, in many cases the High Throughput Computing system is comprised from volunteer computational resources where tasks may be evicted by the owner of the resource. This has two main disadvantages. First, tasks may take longer to run as they may require multiple deployments before finally obtaining enough time on a resource to complete. Second, the wasted computation time will lead to wasted energy. We may be able to reduce the effect of the first disadvantage here by submitting multiple replicas of the task and take the results from the first one to complete. This, though, could lead to a significant increase in energy consumption. Thus we desire to only ever submit the minimum number of replicas required to run the task in the allocated time whilst simultaneously minimising energy. In this work we evaluate the use of fixed replica counts and Reinforcement Learning on the proportion of task which fail to finish in a given time-frame and the energy consumed by the system.
高吞吐量计算允许在许多分布式资源上高效地执行数千个任务的工作负载,并将用户从管理任务部署、执行和结果收集的繁重过程中解放出来。然而,在许多情况下,高吞吐量计算系统由志愿计算资源组成,其中任务可能被资源所有者驱逐。这有两个主要缺点。首先,任务可能需要更长的时间来运行,因为它们可能需要多次部署才能最终在资源上获得足够的时间来完成。其次,浪费的计算时间会导致能源的浪费。我们可以通过提交任务的多个副本并从第一个副本获取结果来减少第一个缺点的影响。然而,这可能会导致能源消耗的显著增加。因此,我们希望只提交在分配的时间内运行任务所需的最小数量的副本,同时最大限度地减少能量。在这项工作中,我们评估了固定副本计数和强化学习在给定时间框架内无法完成的任务比例和系统消耗的能量的使用。
{"title":"Evaluation of Energy Consumption of Replicated Tasks in a Volunteer Computing Environment","authors":"A. McGough, M. Forshaw","doi":"10.1145/3185768.3186313","DOIUrl":"https://doi.org/10.1145/3185768.3186313","url":null,"abstract":"High Throughput Computing allows workloads of many thousands of tasks to be performed efficiently over many distributed resources and frees the user from the laborious process of managing task deployment, execution and result collection. However, in many cases the High Throughput Computing system is comprised from volunteer computational resources where tasks may be evicted by the owner of the resource. This has two main disadvantages. First, tasks may take longer to run as they may require multiple deployments before finally obtaining enough time on a resource to complete. Second, the wasted computation time will lead to wasted energy. We may be able to reduce the effect of the first disadvantage here by submitting multiple replicas of the task and take the results from the first one to complete. This, though, could lead to a significant increase in energy consumption. Thus we desire to only ever submit the minimum number of replicas required to run the task in the allocated time whilst simultaneously minimising energy. In this work we evaluate the use of fixed replica counts and Reinforcement Learning on the proportion of task which fail to finish in a given time-frame and the energy consumed by the system.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"1 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"84152071","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
DIBS: A Data Integration Benchmark Suite DIBS:数据集成基准套件
A. Cabrera, Clayton J. Faber, Kyle Cepeda, Robert Derber, Cooper Epstein, Jason Zheng, R. Cytron, R. Chamberlain
As the generation of data becomes more prolific, the amount of time and resources necessary to perform analyses on these data increases. What is less understood, however, is the data preprocessing steps that must be applied before any meaningful analysis can begin. This problem of taking data in some initial form and transforming it into a desired one is known as data integration. Here, we introduce the Data Integration Benchmarking Suite (DIBS), a suite of applications that are representative of data integration workloads across many disciplines. We apply a comprehensive characterization to these applications to better understand the general behavior of data integration tasks. As a result of our benchmark suite and characterization methods, we offer insight regarding data integration tasks that will guide other researchers designing solutions in this area.
随着数据的生成越来越多,对这些数据执行分析所需的时间和资源也在增加。然而,人们不太了解的是,在任何有意义的分析开始之前必须应用的数据预处理步骤。这种以某种初始形式获取数据并将其转换为所需形式的问题称为数据集成。在这里,我们介绍数据集成基准套件(Data Integration Benchmarking Suite, DIBS),这是一组应用程序,代表了跨许多学科的数据集成工作负载。我们对这些应用程序进行了全面的描述,以便更好地理解数据集成任务的一般行为。由于我们的基准套件和表征方法,我们提供了有关数据集成任务的见解,这将指导其他研究人员在该领域设计解决方案。
{"title":"DIBS: A Data Integration Benchmark Suite","authors":"A. Cabrera, Clayton J. Faber, Kyle Cepeda, Robert Derber, Cooper Epstein, Jason Zheng, R. Cytron, R. Chamberlain","doi":"10.1145/3185768.3186307","DOIUrl":"https://doi.org/10.1145/3185768.3186307","url":null,"abstract":"As the generation of data becomes more prolific, the amount of time and resources necessary to perform analyses on these data increases. What is less understood, however, is the data preprocessing steps that must be applied before any meaningful analysis can begin. This problem of taking data in some initial form and transforming it into a desired one is known as data integration. Here, we introduce the Data Integration Benchmarking Suite (DIBS), a suite of applications that are representative of data integration workloads across many disciplines. We apply a comprehensive characterization to these applications to better understand the general behavior of data integration tasks. As a result of our benchmark suite and characterization methods, we offer insight regarding data integration tasks that will guide other researchers designing solutions in this area.","PeriodicalId":10596,"journal":{"name":"Companion of the 2018 ACM/SPEC International Conference on Performance Engineering","volume":"2010 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2018-04-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"82595037","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
期刊
Companion of the 2018 ACM/SPEC International Conference on Performance Engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1