首页 > 最新文献

Proceedings of the 5th ACM/SPEC international conference on Performance engineering最新文献

英文 中文
Extreme big data processing in large-scale graph analytics and billion-scale social simulation 大规模图形分析和十亿规模社会模拟中的极端大数据处理
T. Suzumura
This paper introduces some of the example applications handling extremely big data with supercomputers such as large-scale network analysis, X10-based large-scale graph analytics library, Graph500 benchmark, and billion-scale social simulation.
本文介绍了超大规模网络分析、基于x10的大规模图分析库、Graph500基准测试、十亿规模社会模拟等在超级计算机上处理超大数据的应用实例。
{"title":"Extreme big data processing in large-scale graph analytics and billion-scale social simulation","authors":"T. Suzumura","doi":"10.1145/2568088.2576096","DOIUrl":"https://doi.org/10.1145/2568088.2576096","url":null,"abstract":"This paper introduces some of the example applications handling extremely big data with supercomputers such as large-scale network analysis, X10-based large-scale graph analytics library, Graph500 benchmark, and billion-scale social simulation.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"252 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131555936","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Model-driven engineering in practice: integrated performance decision support for process-centric business impact analysis 实践中的模型驱动工程:为以流程为中心的业务影响分析提供集成的性能决策支持
D. Redlich, Ulrich Winkler, T. Molka, Wasif Gilani
Modern businesses and business processes depend on an increasingly interconnected set of resources, which can be affected by external and internal factors at any time. Threats like natural disasters, terrorism, or even power blackouts potentially cause disruptions in an organisation's resource infrastructure which in turn negatively impacts the performance of dependent business processes. In order to assist business analysts dealing with this ever increasing complexity of interdependent business structures a model-driven workbench named Model-Driven Business Impact Analysis (MDBIA) has been developed with the purpose of predicting consequences on the business process level for an organisation in case of disruptions. An already existing Model-Driven Performance Engineering (MDPE) workbench, which originally provided process-centric performance decision support, has been adapted and extended to meet the additional requirements of business impact analysis. The fundamental concepts of the resulting MDBIA workbench, which include the introduction of the applied key models and transformation chain, are presented and evaluated in this paper.
现代企业和业务流程依赖于一组日益相互关联的资源,这些资源随时可能受到外部和内部因素的影响。像自然灾害、恐怖主义、甚至停电这样的威胁可能会导致组织的资源基础设施中断,进而对相关业务流程的性能产生负面影响。为了帮助业务分析人员处理这种日益复杂的相互依赖的业务结构,已经开发了一个名为模型驱动业务影响分析(MDBIA)的模型驱动工作台,其目的是在中断的情况下预测组织在业务流程级别上的后果。已经存在的模型驱动性能工程(Model-Driven Performance Engineering, MDPE)工作台最初提供以流程为中心的性能决策支持,现在已经进行了调整和扩展,以满足业务影响分析的额外需求。本文介绍并评估了由此产生的MDBIA工作台的基本概念,其中包括所应用的关键模型和转换链的介绍。
{"title":"Model-driven engineering in practice: integrated performance decision support for process-centric business impact analysis","authors":"D. Redlich, Ulrich Winkler, T. Molka, Wasif Gilani","doi":"10.1145/2568088.2576797","DOIUrl":"https://doi.org/10.1145/2568088.2576797","url":null,"abstract":"Modern businesses and business processes depend on an increasingly interconnected set of resources, which can be affected by external and internal factors at any time. Threats like natural disasters, terrorism, or even power blackouts potentially cause disruptions in an organisation's resource infrastructure which in turn negatively impacts the performance of dependent business processes. In order to assist business analysts dealing with this ever increasing complexity of interdependent business structures a model-driven workbench named Model-Driven Business Impact Analysis (MDBIA) has been developed with the purpose of predicting consequences on the business process level for an organisation in case of disruptions. An already existing Model-Driven Performance Engineering (MDPE) workbench, which originally provided process-centric performance decision support, has been adapted and extended to meet the additional requirements of business impact analysis. The fundamental concepts of the resulting MDBIA workbench, which include the introduction of the applied key models and transformation chain, are presented and evaluated in this paper.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"187 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127136254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Run-time performance optimization of a BigData query language 大数据查询语言的运行时性能优化
Yanbin Liu, Parijat Dube, Scott Gray
JAQL is a query language for large-scale data that connects BigData analytics and MapReduce framework together. Also an IBM product, JAQL's performance is critical for IBM InfoSphere BigInsights, a BigData analytics platform. In this paper, we report our work on improving JAQL performance from multiple perspectives. We explore the parallelism of JAQL, profile JAQL for performance analysis, identify I/O as the dominant performance bottleneck, and improve JAQL performance with an emphasis on reducing I/O data size and increasing (de)serialization efficiency. With TPCH benchmark on a simple Hadoop cluster, we report up to 2x performance improvements in JAQL with our optimization fixes.
JAQL是一种连接BigData分析和MapReduce框架的大规模数据查询语言。JAQL也是一款IBM产品,它的性能对IBM InfoSphere BigInsights(一个大数据分析平台)至关重要。在本文中,我们从多个角度报告了我们在提高JAQL性能方面的工作。我们将探讨JAQL的并行性,对JAQL进行性能分析,确定I/O是主要的性能瓶颈,并通过减少I/O数据大小和提高(反)序列化效率来提高JAQL性能。在一个简单的Hadoop集群上使用TPCH基准测试,我们报告通过我们的优化修复,JAQL的性能提高了2倍。
{"title":"Run-time performance optimization of a BigData query language","authors":"Yanbin Liu, Parijat Dube, Scott Gray","doi":"10.1145/2568088.2576800","DOIUrl":"https://doi.org/10.1145/2568088.2576800","url":null,"abstract":"JAQL is a query language for large-scale data that connects BigData analytics and MapReduce framework together. Also an IBM product, JAQL's performance is critical for IBM InfoSphere BigInsights, a BigData analytics platform. In this paper, we report our work on improving JAQL performance from multiple perspectives. We explore the parallelism of JAQL, profile JAQL for performance analysis, identify I/O as the dominant performance bottleneck, and improve JAQL performance with an emphasis on reducing I/O data size and increasing (de)serialization efficiency. With TPCH benchmark on a simple Hadoop cluster, we report up to 2x performance improvements in JAQL with our optimization fixes.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128388677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Application performance management using learning, optimization, and control 使用学习、优化和控制的应用程序性能管理
Xiaoyun Zhu
In the past decade, the IT industry has experienced a paradigm shift as computing resources became available as a utility through cloud based services. In spite of the wider adoption of cloud computing platforms, some businesses and organizations hesitate to move all their applications to the cloud due to performance concerns. Existing practices in application performance management rely heavily on white-box modeling and diagnosis approaches or on performance troubleshooting "cookbooks" to find potential bottlenecks and remediation steps. However, the scalability and adaptivity of such approaches remain severely constrained, especially in a highly-dynamic, consolidated cloud environment. For performance isolation and differentiation, most modern hypervisors offer powerful resource control primitives such as reservations, limits, and shares for individual virtual machines (VMs). Even so, with the exploding growth of virtual machine sprawl, setting these controls properly such that co-located virtualized applications get enough resources to meet their respective service level objectives (SLOs) becomes a nearly insoluble task. These challenges present unique opportunities in leveraging the rich telemetry collected from applications and systems in the cloud, and in applying statistical learning, optimization, and control based techniques to developing model-based, automated application performance management frameworks. There has been a large body of research in this area in the last several years, but many problems remain. In this talk, I'll highlight some of the automated and data-driven performance management techniques we have developed, along with related technical challenges. I'll then discuss open research problems, in hope to attract more innovative ideas and solutions from a larger community of researchers and practitioners.
在过去的十年中,随着计算资源通过基于云的服务作为实用程序可用,IT行业经历了范式转变。尽管云计算平台得到了更广泛的采用,但由于性能问题,一些企业和组织对将所有应用程序迁移到云中犹豫不决。应用程序性能管理中的现有实践严重依赖于白盒建模和诊断方法,或者依赖于性能故障排除“菜谱”来发现潜在的瓶颈和补救步骤。但是,这些方法的可伸缩性和适应性仍然受到严重限制,特别是在高度动态的合并云环境中。为了实现性能隔离和区分,大多数现代管理程序都提供了强大的资源控制原语,例如针对单个虚拟机(vm)的保留、限制和共享。即便如此,随着虚拟机扩展的爆炸式增长,适当地设置这些控制以使位于同一位置的虚拟化应用程序获得足够的资源来满足各自的服务水平目标(slo)几乎成为一项无法解决的任务。这些挑战为利用从云中的应用程序和系统收集的丰富遥测数据,以及应用统计学习、优化和基于控制的技术来开发基于模型的自动化应用程序性能管理框架提供了独特的机会。在过去的几年里,这一领域已经有了大量的研究,但仍然存在许多问题。在这次演讲中,我将重点介绍我们开发的一些自动化和数据驱动的性能管理技术,以及相关的技术挑战。然后,我将讨论开放的研究问题,希望从更大的研究人员和实践者群体中吸引更多的创新想法和解决方案。
{"title":"Application performance management using learning, optimization, and control","authors":"Xiaoyun Zhu","doi":"10.1145/2568088.2576098","DOIUrl":"https://doi.org/10.1145/2568088.2576098","url":null,"abstract":"In the past decade, the IT industry has experienced a paradigm shift as computing resources became available as a utility through cloud based services. In spite of the wider adoption of cloud computing platforms, some businesses and organizations hesitate to move all their applications to the cloud due to performance concerns. Existing practices in application performance management rely heavily on white-box modeling and diagnosis approaches or on performance troubleshooting \"cookbooks\" to find potential bottlenecks and remediation steps. However, the scalability and adaptivity of such approaches remain severely constrained, especially in a highly-dynamic, consolidated cloud environment. For performance isolation and differentiation, most modern hypervisors offer powerful resource control primitives such as reservations, limits, and shares for individual virtual machines (VMs). Even so, with the exploding growth of virtual machine sprawl, setting these controls properly such that co-located virtualized applications get enough resources to meet their respective service level objectives (SLOs) becomes a nearly insoluble task. These challenges present unique opportunities in leveraging the rich telemetry collected from applications and systems in the cloud, and in applying statistical learning, optimization, and control based techniques to developing model-based, automated application performance management frameworks. There has been a large body of research in this area in the last several years, but many problems remain. In this talk, I'll highlight some of the automated and data-driven performance management techniques we have developed, along with related technical challenges. I'll then discuss open research problems, in hope to attract more innovative ideas and solutions from a larger community of researchers and practitioners.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"16 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115599254","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Continuous validation of load test suites 负载测试套件的持续验证
Mark D. Syer, Z. Jiang, M. Nagappan, A. Hassan, Mohamed N. Nasser, P. Flora
Ultra-Large-Scale (ULS) systems face continuously evolving field workloads in terms of activated/disabled feature sets, varying usage patterns and changing deployment configurations. These evolving workloads often have a large impact, on the performance of a ULS system. Hence, continuous load testing is critical to ensuring the error-free operation of such systems. A common challenge facing performance analysts is to validate if a load test closely resembles the current field workloads. Such validation may be performed by comparing execution logs from the load test and the field. However, the size and unstructured nature of execution logs makes such a comparison unfeasible without automated support. In this paper, we propose an automated approach to validate whether a load test resembles the field workload and, if not, determines how they differ by compare execution logs from a load test and the field. Performance analysts can then update their load test cases to eliminate such differences, hence creating more realistic load test cases. We perform three case studies on two large systems: one open-source system and one enterprise system. Our approach identifies differences between load tests and the field with a precision of >75% compared to only >16% for the state-of-the-practice.
超大规模(ULS)系统在激活/禁用功能集、不同的使用模式和不断变化的部署配置方面面临着不断变化的现场工作负载。这些不断变化的工作负载通常对ULS系统的性能有很大的影响。因此,持续的负载测试对于确保此类系统的无错误运行至关重要。性能分析人员面临的一个常见挑战是验证负载测试是否与当前字段工作负载非常相似。这种验证可以通过比较负载测试和现场的执行日志来执行。然而,执行日志的大小和非结构化性质使得没有自动化支持就无法进行这种比较。在本文中,我们提出了一种自动化的方法来验证负载测试是否与字段工作负载相似,如果不相似,则通过比较负载测试和字段的执行日志来确定它们之间的差异。性能分析人员可以更新他们的负载测试用例来消除这些差异,从而创建更真实的负载测试用例。我们在两个大型系统上进行了三个案例研究:一个是开源系统,一个是企业系统。我们的方法以>75%的精度识别负载测试和现场之间的差异,而实践状态的精度仅为>16%。
{"title":"Continuous validation of load test suites","authors":"Mark D. Syer, Z. Jiang, M. Nagappan, A. Hassan, Mohamed N. Nasser, P. Flora","doi":"10.1145/2568088.2568101","DOIUrl":"https://doi.org/10.1145/2568088.2568101","url":null,"abstract":"Ultra-Large-Scale (ULS) systems face continuously evolving field workloads in terms of activated/disabled feature sets, varying usage patterns and changing deployment configurations. These evolving workloads often have a large impact, on the performance of a ULS system. Hence, continuous load testing is critical to ensuring the error-free operation of such systems. A common challenge facing performance analysts is to validate if a load test closely resembles the current field workloads. Such validation may be performed by comparing execution logs from the load test and the field. However, the size and unstructured nature of execution logs makes such a comparison unfeasible without automated support. In this paper, we propose an automated approach to validate whether a load test resembles the field workload and, if not, determines how they differ by compare execution logs from a load test and the field. Performance analysts can then update their load test cases to eliminate such differences, hence creating more realistic load test cases. We perform three case studies on two large systems: one open-source system and one enterprise system. Our approach identifies differences between load tests and the field with a precision of >75% compared to only >16% for the state-of-the-practice.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"48 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127587218","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 26
PowerPerfCenter: a power and performance prediction tool for multi-tier applications PowerPerfCenter:用于多层应用程序的功率和性能预测工具
V. Apte, Bhavin J. Doshi
The performance analysis of a server application and the sizing of the hardware required to host it in a data center continue to be pressing issues today. With most server-grade computers now built with "frequency-scaled CPUs" and other such devices, it has become important to answer performance and sizing questions in the presence of such hardware. PowerPerfCenter is an application performance modeling tool that allows specification of devices whose operating speeds can change dynamically. It also estimates power usage by the machines in presence of such devices. Furthermore, it allows specification of a dynamic workload which is required to understand the impact of power management. We validated the performance metrics predicted by PowerPerfCenter against measured ones of an application deployed on a test-bed consisting of frequency-scaled CPUs, and found the match to be good. We also used PowerPerfCenter to show that power savings may not be significant if a device does not have different idle power consumption when configured with different operating speeds.
服务器应用程序的性能分析和在数据中心中托管它所需的硬件的大小仍然是当今的紧迫问题。由于现在大多数服务器级计算机都使用“频率缩放的cpu”和其他此类设备构建,因此在存在此类硬件的情况下回答性能和大小问题变得非常重要。PowerPerfCenter是一个应用程序性能建模工具,它允许对运行速度可以动态变化的设备进行规范。它还估计了在这些设备存在的情况下机器的用电量。此外,它允许指定动态工作负载,这是理解电源管理的影响所必需的。我们将PowerPerfCenter预测的性能指标与部署在由按频率缩放的cpu组成的测试平台上的应用程序的实际性能指标进行了验证,发现两者的匹配非常好。我们还使用PowerPerfCenter显示,如果设备在配置不同的运行速度时没有不同的空闲功耗,那么省电可能并不显著。
{"title":"PowerPerfCenter: a power and performance prediction tool for multi-tier applications","authors":"V. Apte, Bhavin J. Doshi","doi":"10.1145/2568088.2576758","DOIUrl":"https://doi.org/10.1145/2568088.2576758","url":null,"abstract":"The performance analysis of a server application and the sizing of the hardware required to host it in a data center continue to be pressing issues today. With most server-grade computers now built with \"frequency-scaled CPUs\" and other such devices, it has become important to answer performance and sizing questions in the presence of such hardware. PowerPerfCenter is an application performance modeling tool that allows specification of devices whose operating speeds can change dynamically. It also estimates power usage by the machines in presence of such devices. Furthermore, it allows specification of a dynamic workload which is required to understand the impact of power management. We validated the performance metrics predicted by PowerPerfCenter against measured ones of an application deployed on a test-bed consisting of frequency-scaled CPUs, and found the match to be good. We also used PowerPerfCenter to show that power savings may not be significant if a device does not have different idle power consumption when configured with different operating speeds.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"141 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129844234","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Constructing performance model of JMS middleware platform 构建JMS中间件平台的性能模型
Tomás Martinec, L. Marek, A. Steinhauser, P. Tůma, Qais Noorshams, A. Rentschler, Ralf H. Reussner
Middleware performance models are useful building blocks in the performance models of distributed software applications. We focus on performance models of messaging middleware implementing the Java Message Service standard, showing how certain system design properties -- including pipelined processing and message coalescing -- interact to create performance behavior that the existing models do not capture accurately. We construct a performance model of the ActiveMQ messaging middleware that addresses the outlined issues and discuss how the approach extends to other middleware implementations.
中间件性能模型是分布式软件应用程序性能模型中有用的构建块。我们将重点关注实现Java Message Service标准的消息传递中间件的性能模型,展示某些系统设计属性(包括流水线处理和消息合并)如何交互以创建现有模型无法准确捕获的性能行为。我们构建了ActiveMQ消息传递中间件的性能模型,该模型解决了上述问题,并讨论了该方法如何扩展到其他中间件实现。
{"title":"Constructing performance model of JMS middleware platform","authors":"Tomás Martinec, L. Marek, A. Steinhauser, P. Tůma, Qais Noorshams, A. Rentschler, Ralf H. Reussner","doi":"10.1145/2568088.2568096","DOIUrl":"https://doi.org/10.1145/2568088.2568096","url":null,"abstract":"Middleware performance models are useful building blocks in the performance models of distributed software applications. We focus on performance models of messaging middleware implementing the Java Message Service standard, showing how certain system design properties -- including pipelined processing and message coalescing -- interact to create performance behavior that the existing models do not capture accurately. We construct a performance model of the ActiveMQ messaging middleware that addresses the outlined issues and discuss how the approach extends to other middleware implementations.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"198 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123429726","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
LibReDE: a library for resource demand estimation 一个用于资源需求估计的库
Simon Spinner, G. Casale, Xiaoyun Zhu, Samuel Kounev
When creating a performance model, it is necessary to quantify the amount of resources consumed by an application serving individual requests. In distributed enterprise systems, these resource demands usually cannot be observed directly, their estimation is a major challenge. Different statistical approaches to resource demand estimation based on monitoring data have been proposed, e.g., using linear regression or Kalman filtering techniques. In this paper, we present LibReDE, a library of ready-to-use implementations of approaches to resource demand estimation that can be used for online and offline analysis. It is the first publicly available tool for this task and aims at supporting performance engineers during performance model construction. The library enables the quick comparison of the estimation accuracy of different approaches in a given context and thus helps to select an optimal one.
在创建性能模型时,有必要量化服务于单个请求的应用程序所消耗的资源量。在分布式企业系统中,这些资源需求通常不能直接观察到,它们的估计是一个主要的挑战。已经提出了根据监测数据估计资源需求的不同统计方法,例如使用线性回归或卡尔曼滤波技术。在本文中,我们提出了LibReDE,这是一个可用于在线和离线分析的资源需求估计方法的现成实现库。它是用于此任务的第一个公开可用的工具,旨在在性能模型构建期间为性能工程师提供支持。该库可以快速比较给定环境下不同方法的估计精度,从而帮助选择最优方法。
{"title":"LibReDE: a library for resource demand estimation","authors":"Simon Spinner, G. Casale, Xiaoyun Zhu, Samuel Kounev","doi":"10.1145/2568088.2576093","DOIUrl":"https://doi.org/10.1145/2568088.2576093","url":null,"abstract":"When creating a performance model, it is necessary to quantify the amount of resources consumed by an application serving individual requests. In distributed enterprise systems, these resource demands usually cannot be observed directly, their estimation is a major challenge. Different statistical approaches to resource demand estimation based on monitoring data have been proposed, e.g., using linear regression or Kalman filtering techniques. In this paper, we present LibReDE, a library of ready-to-use implementations of approaches to resource demand estimation that can be used for online and offline analysis. It is the first publicly available tool for this task and aims at supporting performance engineers during performance model construction. The library enables the quick comparison of the estimation accuracy of different approaches in a given context and thus helps to select an optimal one.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130324742","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 41
Agile middleware for scheduling: meeting competing performance requirements of diverse tasks 用于调度的敏捷中间件:满足不同任务的竞争性性能需求
Feng Yan, S. Hughes, Alma Riska, E. Smirni
As the need for scaled-out systems increases, it is paramount to architect them as large distributed systems consisting of off-the-shelf basic computing components known as compute or data nodes. These nodes are expected to handle their work independently, and often utilize off-the-shelf management tools, like those offered by Linux, to differentiate priorities of tasks. While prioritization of background tasks in server nodes takes center stage in scaled-out systems, with many tasks associated with salient features such as eventual consistency, data analytics, and garbage collection, the standard Linux tools such as nice and ionice fail to adapt to the dynamic behavior of high priority tasks in order to achieve the best trade-off between protecting the performance of high priority workload and completing as much low priority work as possible. In this paper, we provide a solution by proposing a priority scheduling middleware that employs different policies to schedule background tasks based on the instantaneous resource requirements of the high priority applications running on the server node. The selection of policies is based on off-line and on-line learning of the high priority workload characteristics and the imposed performance impact due to low priority work. In effect, this middleware uses a {em hybrid} approach to scheduling rather than a monolithic policy. We prototype and evaluate it via measurements on a test-bed and show that this scheduling middleware is robust as it effectively and autonomically changes the relative priorities between high and low priority tasks, consistently meeting their competing performance targets.
随着向外扩展系统需求的增加,将它们架构为大型分布式系统至关重要,这些系统由现成的基本计算组件(称为计算或数据节点)组成。这些节点被期望独立地处理它们的工作,并且经常使用现成的管理工具(如Linux提供的工具)来区分任务的优先级。虽然服务器节点中后台任务的优先级在向外扩展的系统中占据中心位置,并且许多任务与最终一致性、数据分析和垃圾收集等显著特性相关联,但是标准的Linux工具(如nice和ionice)无法适应高优先级任务的动态行为,以便在保护高优先级工作负载的性能和完成尽可能多的低优先级工作之间实现最佳权衡。在本文中,我们通过提出一个优先级调度中间件提供了一个解决方案,该中间件采用不同的策略来调度后台任务,该策略基于运行在服务器节点上的高优先级应用程序的瞬时资源需求。策略的选择基于对高优先级工作负载特征的离线和在线学习,以及由于低优先级工作而造成的性能影响。实际上,该中间件使用{em hybrid}方法来调度,而不是单一策略。我们通过测试平台上的测量对其进行了原型化和评估,并表明该调度中间件是健壮的,因为它有效且自主地更改高优先级和低优先级任务之间的相对优先级,始终满足它们相互竞争的性能目标。
{"title":"Agile middleware for scheduling: meeting competing performance requirements of diverse tasks","authors":"Feng Yan, S. Hughes, Alma Riska, E. Smirni","doi":"10.1145/2568088.2568104","DOIUrl":"https://doi.org/10.1145/2568088.2568104","url":null,"abstract":"As the need for scaled-out systems increases, it is paramount to architect them as large distributed systems consisting of off-the-shelf basic computing components known as compute or data nodes. These nodes are expected to handle their work independently, and often utilize off-the-shelf management tools, like those offered by Linux, to differentiate priorities of tasks. While prioritization of background tasks in server nodes takes center stage in scaled-out systems, with many tasks associated with salient features such as eventual consistency, data analytics, and garbage collection, the standard Linux tools such as nice and ionice fail to adapt to the dynamic behavior of high priority tasks in order to achieve the best trade-off between protecting the performance of high priority workload and completing as much low priority work as possible. In this paper, we provide a solution by proposing a priority scheduling middleware that employs different policies to schedule background tasks based on the instantaneous resource requirements of the high priority applications running on the server node. The selection of policies is based on off-line and on-line learning of the high priority workload characteristics and the imposed performance impact due to low priority work. In effect, this middleware uses a {em hybrid} approach to scheduling rather than a monolithic policy. We prototype and evaluate it via measurements on a test-bed and show that this scheduling middleware is robust as it effectively and autonomically changes the relative priorities between high and low priority tasks, consistently meeting their competing performance targets.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128897771","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Engineering resource management middleware for optimizing the performance of clouds processing mapreduce jobs with deadlines 工程资源管理中间件,用于优化具有最后期限的云处理mapreduce作业的性能
Norman Lim, S. Majumdar, P. Ashwood-Smith
This paper focuses on devising efficient resource management techniques used by the resource management middleware in clouds that handle MapReduce jobs with end-to-end service level agreements (SLAs) comprising an earliest start time, execution time, and a deadline. This research and development work, performed in collaboration with our industrial partner, presents the formulation of the matchmaking and scheduling problem for MapReduce jobs as an optimization problem using: Mixed Integer Linear Programming (MILP) and Constraint Programming (CP) techniques. In addition to the formulations devised, our experience in implementing the MILP and CP models using various open source as well as commercial software packages is described. Furthermore, a performance evaluation of the different approaches used to implement the formulations is conducted using a variety of different workloads.
本文的重点是设计有效的资源管理技术,用于云中的资源管理中间件,这些中间件处理具有端到端服务级别协议(sla)的MapReduce作业,包括最早的开始时间、执行时间和截止日期。这项研究和开发工作是与我们的工业合作伙伴合作完成的,使用混合整数线性规划(MILP)和约束规划(CP)技术,将MapReduce作业的配对和调度问题作为优化问题提出。除了所设计的公式之外,我们还描述了使用各种开放源代码和商业软件包实现MILP和CP模型的经验。此外,还使用各种不同的工作负载对用于实现配方的不同方法进行了性能评估。
{"title":"Engineering resource management middleware for optimizing the performance of clouds processing mapreduce jobs with deadlines","authors":"Norman Lim, S. Majumdar, P. Ashwood-Smith","doi":"10.1145/2568088.2576796","DOIUrl":"https://doi.org/10.1145/2568088.2576796","url":null,"abstract":"This paper focuses on devising efficient resource management techniques used by the resource management middleware in clouds that handle MapReduce jobs with end-to-end service level agreements (SLAs) comprising an earliest start time, execution time, and a deadline. This research and development work, performed in collaboration with our industrial partner, presents the formulation of the matchmaking and scheduling problem for MapReduce jobs as an optimization problem using: Mixed Integer Linear Programming (MILP) and Constraint Programming (CP) techniques. In addition to the formulations devised, our experience in implementing the MILP and CP models using various open source as well as commercial software packages is described. Furthermore, a performance evaluation of the different approaches used to implement the formulations is conducted using a variety of different workloads.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129003014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 21
期刊
Proceedings of the 5th ACM/SPEC international conference on Performance engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1