首页 > 最新文献

Proceedings of the 5th ACM/SPEC international conference on Performance engineering最新文献

英文 中文
Benchmarking graph-processing platforms: a vision 对图形处理平台进行基准测试:远景
Yong Guo, A. Varbanescu, A. Iosup, Claudio Martella, Theodore L. Willke
Processing graphs, especially at large scale, is an increasingly useful activity in a variety of business, engineering, and scientific domains. Already, there are tens of graph-processing platforms, such as Hadoop, Giraph, GraphLab, etc., each with a different design and functionality. For graph-processing to continue to evolve, users have to find it easy to select a graph-processing platform, and developers and system integrators have to find it easy to quantify the performance and other non-functional aspects of interest. However, the state of performance analysis of graph-processing platforms is still immature: there are few studies and, for the few that exist, there are few similarities, and relatively little understanding of the impact of dataset and algorithm diversity on performance. Our vision is to develop, with the help of the performance-savvy community, a comprehensive benchmarking suite for graph-processing platforms. In this work, we take a step in this direction, by proposing a set of seven challenges, summarizing our previous work on performance evaluation of distributed graph-processing platforms, and introducing our on-going work within the SPEC Research Group's Cloud Working Group.
处理图,尤其是大规模的图,在各种商业、工程和科学领域都是越来越有用的活动。现在已经有几十个图形处理平台,如Hadoop、Giraph、GraphLab等,每一个都有不同的设计和功能。为了使图形处理继续发展,用户必须能够轻松地选择图形处理平台,开发人员和系统集成商必须能够轻松地量化性能和其他感兴趣的非功能方面。然而,对于图处理平台的性能分析,目前的研究还不成熟:研究很少,而且存在的研究很少有相似之处,对数据集和算法多样性对性能的影响的理解也相对较少。我们的愿景是在性能专家社区的帮助下,为图形处理平台开发一个全面的基准测试套件。在这项工作中,我们向这个方向迈出了一步,提出了一组七个挑战,总结了我们以前在分布式图形处理平台的性能评估方面的工作,并介绍了我们在SPEC研究小组的云工作组中正在进行的工作。
{"title":"Benchmarking graph-processing platforms: a vision","authors":"Yong Guo, A. Varbanescu, A. Iosup, Claudio Martella, Theodore L. Willke","doi":"10.1145/2568088.2576761","DOIUrl":"https://doi.org/10.1145/2568088.2576761","url":null,"abstract":"Processing graphs, especially at large scale, is an increasingly useful activity in a variety of business, engineering, and scientific domains. Already, there are tens of graph-processing platforms, such as Hadoop, Giraph, GraphLab, etc., each with a different design and functionality. For graph-processing to continue to evolve, users have to find it easy to select a graph-processing platform, and developers and system integrators have to find it easy to quantify the performance and other non-functional aspects of interest. However, the state of performance analysis of graph-processing platforms is still immature: there are few studies and, for the few that exist, there are few similarities, and relatively little understanding of the impact of dataset and algorithm diversity on performance. Our vision is to develop, with the help of the performance-savvy community, a comprehensive benchmarking suite for graph-processing platforms. In this work, we take a step in this direction, by proposing a set of seven challenges, summarizing our previous work on performance evaluation of distributed graph-processing platforms, and introducing our on-going work within the SPEC Research Group's Cloud Working Group.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133615644","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 40
An evaluation of zookeeper for high availability in system S 系统S中zookeeper的高可用性评价
C. Pham, Victor Dogaru, R. Wagle, C. Venkatramani, Z. Kalbarczyk, R. Iyer
ZooKeeper provides scalable, highly available coordination services for distributed applications. In this paper, we evaluate the use of ZooKeeper in a distributed stream computing system called System S to provide a resilient name service, dynamic configuration management, and system state management. The evaluation shed light on the advantages of using ZooKeeper in these contexts as well as its limitations. We also describe design changes we made to handle named objects in System S to overcome the limitations. We present detailed experimental results, which we believe will be beneficial to the community.
ZooKeeper为分布式应用提供可扩展、高可用性的协调服务。在本文中,我们评估了ZooKeeper在分布式流计算系统system S中的使用,以提供弹性名称服务、动态配置管理和系统状态管理。评估揭示了在这些环境中使用ZooKeeper的优点以及它的局限性。我们还描述了在System S中处理命名对象所做的设计更改,以克服这些限制。我们提出了详细的实验结果,我们相信这将对社区有益。
{"title":"An evaluation of zookeeper for high availability in system S","authors":"C. Pham, Victor Dogaru, R. Wagle, C. Venkatramani, Z. Kalbarczyk, R. Iyer","doi":"10.1145/2568088.2576801","DOIUrl":"https://doi.org/10.1145/2568088.2576801","url":null,"abstract":"ZooKeeper provides scalable, highly available coordination services for distributed applications. In this paper, we evaluate the use of ZooKeeper in a distributed stream computing system called System S to provide a resilient name service, dynamic configuration management, and system state management. The evaluation shed light on the advantages of using ZooKeeper in these contexts as well as its limitations. We also describe design changes we made to handle named objects in System S to overcome the limitations. We present detailed experimental results, which we believe will be beneficial to the community.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133744571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Software contention aware queueing network model of three-tier web systems 三层web系统的软件竞争感知排队网络模型
Shadi Ghaith, Miao Wang, Philip Perry, L. Murphy
Using modelling to predict the performance characteristics of software applications typically uses Queueing Network Models representing the various system hardware resources. Leaving out the software resources, such as the limited number of threads, in such models leads to a reduced prediction accuracy. Accounting for Software Contention is a challenging task as existing techniques to model software components are complex and require deep knowledge of the software architecture. Furthermore, they also require complex measurement processes to obtain the model's service demands. In addition, solving the resultant model usually require simulation solvers which are often time consuming. In this work, we aim to provide a simpler model for three-tier web software systems which accounts for Software Contention that can be solved by time efficient analytical solvers. We achieve this by expanding the existing "Two-Level Iterative Queuing Modelling of Software Contention" method to handle the number of threads at the Application Server tier and the number of Data Sources at the Database Server tier. This is done in a generic manner to allow for extending the solution to other software components like memory and critical sections. Initial results show that our technique clearly outperforms existing techniques.
使用建模来预测软件应用程序的性能特征通常使用表示各种系统硬件资源的排队网络模型。在这样的模型中忽略软件资源,比如线程数量有限,会导致预测精度降低。考虑软件争用是一项具有挑战性的任务,因为现有的软件组件建模技术非常复杂,并且需要对软件体系结构有深入的了解。此外,它们还需要复杂的测量过程来获得模型的服务需求。此外,求解得到的模型通常需要仿真求解器,这往往很耗时。在这项工作中,我们的目标是为三层web软件系统提供一个更简单的模型,该模型解释了可以通过时间效率分析求解器解决的软件争用。我们通过扩展现有的“软件争用的两级迭代排队建模”方法来处理应用程序服务器层的线程数量和数据库服务器层的数据源数量来实现这一目标。这是以一种通用的方式完成的,以便将解决方案扩展到其他软件组件,如内存和关键区。初步结果表明,我们的技术明显优于现有的技术。
{"title":"Software contention aware queueing network model of three-tier web systems","authors":"Shadi Ghaith, Miao Wang, Philip Perry, L. Murphy","doi":"10.1145/2568088.2576760","DOIUrl":"https://doi.org/10.1145/2568088.2576760","url":null,"abstract":"Using modelling to predict the performance characteristics of software applications typically uses Queueing Network Models representing the various system hardware resources. Leaving out the software resources, such as the limited number of threads, in such models leads to a reduced prediction accuracy. Accounting for Software Contention is a challenging task as existing techniques to model software components are complex and require deep knowledge of the software architecture. Furthermore, they also require complex measurement processes to obtain the model's service demands. In addition, solving the resultant model usually require simulation solvers which are often time consuming. In this work, we aim to provide a simpler model for three-tier web software systems which accounts for Software Contention that can be solved by time efficient analytical solvers. We achieve this by expanding the existing \"Two-Level Iterative Queuing Modelling of Software Contention\" method to handle the number of threads at the Application Server tier and the number of Data Sources at the Database Server tier. This is done in a generic manner to allow for extending the solution to other software components like memory and critical sections. Initial results show that our technique clearly outperforms existing techniques.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"81 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115376487","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
LIMBO: a tool for modeling variable load intensities LIMBO:一个模拟可变载荷强度的工具
J. V. Kistowski, N. Herbst, Samuel Kounev
Modern software systems are expected to deliver reliable performance under highly variable load intensities while at the same time making efficient use of dynamically allocated resources. Conventional benchmarking frameworks provide limited support for emulating such highly variable and dynamic load profiles and workload scenarios. Industrial benchmarks typically use workloads with constant or stepwise increasing load intensity, or they simply replay recorded workload traces. In this paper, we present LIMBO - an Eclipse-based tool for modeling variable load intensity profiles based on the Descartes Load Intensity Model as an underlying modeling formalism.
现代软件系统期望在高度可变的负载强度下提供可靠的性能,同时有效地利用动态分配的资源。传统的基准测试框架对模拟这种高度可变的动态负载配置文件和工作负载场景提供了有限的支持。工业基准测试通常使用恒定或逐步增加负载强度的工作负载,或者它们只是重放记录的工作负载跟踪。在本文中,我们提出了LIMBO——一个基于eclipse的工具,用于基于笛卡尔负载强度模型作为底层建模形式来建模可变负载强度轮廓。
{"title":"LIMBO: a tool for modeling variable load intensities","authors":"J. V. Kistowski, N. Herbst, Samuel Kounev","doi":"10.1145/2568088.2576092","DOIUrl":"https://doi.org/10.1145/2568088.2576092","url":null,"abstract":"Modern software systems are expected to deliver reliable performance under highly variable load intensities while at the same time making efficient use of dynamically allocated resources. Conventional benchmarking frameworks provide limited support for emulating such highly variable and dynamic load profiles and workload scenarios. Industrial benchmarks typically use workloads with constant or stepwise increasing load intensity, or they simply replay recorded workload traces. In this paper, we present LIMBO - an Eclipse-based tool for modeling variable load intensity profiles based on the Descartes Load Intensity Model as an underlying modeling formalism.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"513 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123071845","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
System performance analyses through object-oriented fault and coupling prisms 基于面向对象故障和耦合棱镜的系统性能分析
Alessandro Murgia, R. Tonelli, M. Marchesi, G. Concas, S. Counsell, S. Swift
A fundamental aspect of a system's performance over time is the number of faults it generates. The relationship between the software engineering concept of "coupling" (i.e., the degree of inter-connectedness of a system's components) and faults is still a research question attracting attention and a relationship with strong implications for performance; excessive coupling is generally acknowledged to contribute to fault-proneness. In this paper, we explore the relationship between faults and coupling. Two releases from each of three open-source Eclipse projects (six releases in total) were used as an empirical basis and coupling and fault data extracted from those systems. A contrasting coupling profile between fault-free and fault-prone classes was observed and this result was statistically supported. Object-oriented (OO) classes with low values of fan-in (incoming coupling) and fan-out (outgoing coupling) appeared to support fault-free classes, while classes with high fan-out supported relatively fault-prone classes. We also considered size as an influence on fault-proneness. The study thus emphasizes the importance of minimizing coupling where possible (and particularly that of fan-out); failing to control coupling may store up problems for later in a system's life; equally, controlling class size should be a concomitant goal.
随着时间的推移,系统性能的一个基本方面是它产生的故障数量。软件工程概念“耦合”(即系统组件的相互连接程度)与故障之间的关系仍然是一个引人关注的研究问题,并且对性能具有很强的影响;人们普遍认为,过度耦合会导致容易发生故障。本文探讨了故障与耦合之间的关系。三个开源Eclipse项目(总共六个版本)中的每个项目都使用了两个版本作为经验基础,并从这些系统中提取了耦合和故障数据。观察到无故障和易故障类别之间的对比耦合概况,这一结果在统计上得到支持。具有低扇入(传入耦合)和扇出(传出耦合)值的面向对象(OO)类似乎支持无故障类,而具有高扇出的类支持相对容易发生故障的类。我们还考虑了大小对断层倾向的影响。因此,该研究强调了尽可能减少耦合的重要性(特别是扇形输出);未能控制耦合可能会在系统生命周期的后期积累问题;同样,控制班级规模也应该是一个共同的目标。
{"title":"System performance analyses through object-oriented fault and coupling prisms","authors":"Alessandro Murgia, R. Tonelli, M. Marchesi, G. Concas, S. Counsell, S. Swift","doi":"10.1145/2568088.2568089","DOIUrl":"https://doi.org/10.1145/2568088.2568089","url":null,"abstract":"A fundamental aspect of a system's performance over time is the number of faults it generates. The relationship between the software engineering concept of \"coupling\" (i.e., the degree of inter-connectedness of a system's components) and faults is still a research question attracting attention and a relationship with strong implications for performance; excessive coupling is generally acknowledged to contribute to fault-proneness. In this paper, we explore the relationship between faults and coupling. Two releases from each of three open-source Eclipse projects (six releases in total) were used as an empirical basis and coupling and fault data extracted from those systems. A contrasting coupling profile between fault-free and fault-prone classes was observed and this result was statistically supported. Object-oriented (OO) classes with low values of fan-in (incoming coupling) and fan-out (outgoing coupling) appeared to support fault-free classes, while classes with high fan-out supported relatively fault-prone classes. We also considered size as an influence on fault-proneness. The study thus emphasizes the importance of minimizing coupling where possible (and particularly that of fan-out); failing to control coupling may store up problems for later in a system's life; equally, controlling class size should be a concomitant goal.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"109 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134345158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
An experimental methodology to evaluate energy efficiency and performance in an enterprise virtualized environment 一种在企业虚拟化环境中评估能源效率和性能的实验方法
Jesús Omana Iglesias, Philip Perry, L. Murphy, Teodora Sandra Buda, James Thorburn
omputing servers generally have a narrow dynamic power range. For instance, even completely idle servers consume between 50% and 70% of their peak power. Since the usage rate of the server has the main influence on its power consumption, energy-efficiency is achieved whenever the utilization of the servers that are powered on reaches its peak. For this purpose, enterprises generally adopt the following technique: consolidate as many workloads as possible via virtualization in a minimum amount of servers (i.e. maximize utilization) and power down the ones that remain idle (i.e. reduce power consumption). However, such approach can severely impact servers' performance and reliability. In this paper, we propose a methodology to determine the ideal values for power consumption and utilization for a server without performance degradation. We accomplish this through a series of experiments using two typical types of workloads commonly found in enterprises: TPC-H and SPECpower ssj2008 benchmarks. We use the first to measure the amount of queries responded successfully per hour for different numbers of users (i.e. Throughput@Size) in the VM. Moreover, we use the latter to measure the power consumption and number of operations successfully handled by a VM at different target loads. We conducted experiments varying the utilization level and number of users for different VMs and the results show that it is possible to reach the maximum value of power consumption for a server, without experiencing performance degradations when running indi- vidual, or mixing workloads.
计算服务器通常具有较窄的动态功率范围。例如,即使完全空闲的服务器也会消耗其峰值功率的50%到70%。由于服务器的使用率对其功耗有主要影响,因此只要已上电的服务器的利用率达到峰值,就可以实现能源效率。为此,企业通常采用以下技术:通过虚拟化将尽可能多的工作负载整合到最少数量的服务器中(即最大化利用率),并关闭仍然空闲的服务器(即降低功耗)。但是,这种方法会严重影响服务器的性能和可靠性。在本文中,我们提出了一种方法来确定不降低性能的服务器的功耗和利用率的理想值。我们通过使用企业中常见的两种典型工作负载类型进行一系列实验来实现这一点:TPC-H和SPECpower ssj2008基准测试。我们使用第一种方法来衡量虚拟机中不同数量的用户(例如Throughput@Size)每小时成功响应的查询数量。此外,我们使用后者来度量VM在不同目标负载下成功处理的功耗和操作数量。我们对不同vm的利用率和用户数量进行了不同的实验,结果表明,在单独运行或混合工作负载时,可以达到服务器的最大功耗值,而不会出现性能下降。
{"title":"An experimental methodology to evaluate energy efficiency and performance in an enterprise virtualized environment","authors":"Jesús Omana Iglesias, Philip Perry, L. Murphy, Teodora Sandra Buda, James Thorburn","doi":"10.1145/2568088.2568099","DOIUrl":"https://doi.org/10.1145/2568088.2568099","url":null,"abstract":"omputing servers generally have a narrow dynamic power range. For instance, even completely idle servers consume between 50% and 70% of their peak power. Since the usage rate of the server has the main influence on its power consumption, energy-efficiency is achieved whenever the utilization of the servers that are powered on reaches its peak. For this purpose, enterprises generally adopt the following technique: consolidate as many workloads as possible via virtualization in a minimum amount of servers (i.e. maximize utilization) and power down the ones that remain idle (i.e. reduce power consumption). However, such approach can severely impact servers' performance and reliability. In this paper, we propose a methodology to determine the ideal values for power consumption and utilization for a server without performance degradation. We accomplish this through a series of experiments using two typical types of workloads commonly found in enterprises: TPC-H and SPECpower ssj2008 benchmarks. We use the first to measure the amount of queries responded successfully per hour for different numbers of users (i.e. Throughput@Size) in the VM. Moreover, we use the latter to measure the power consumption and number of operations successfully handled by a VM at different target loads. We conducted experiments varying the utilization level and number of users for different VMs and the results show that it is possible to reach the maximum value of power consumption for a server, without experiencing performance degradations when running indi- vidual, or mixing workloads.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128691889","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Speeding up processing data from millions of smart meters 加速处理数百万智能电表的数据
Jiang Zheng, Zhao Li, A. Dagnino
As an important element of the Smart Grid, Advanced Metering Infrastructure (AMI) systems have been implemented and deployed throughout the world in the past several years. An AMI system connects millions of end devices (e.g., smart meters and sensors in the residential level) with utility control centers via an efficient two-way communication infrastructure. AMI systems are able to exchange substantial meter data and control information between utilities and end devices in real-time or near real-time. The major challenge our research was to scale ABB's Meter Data Management System (MDMS) to manage data that originates from millions of smart meters. We designed a lightweight architecture capable of collect ever-increasing large amount of meter data from various metering systems, clean, analyze, and aggregate the meter data to support various smart grid applications. To meet critical high performance requirements, various concurrency processing techniques were implemented and integrated in our prototype. Our experiments showed that on average the implemented data file parser took about 42 minutes to complete parsing, cleaning, and aggregating 5.184 billion meter reads on a single machine with the hardware configuration of 12-core CPU, 32G RAM, and SSD Hard Drives. The throughput is about 7.38 billion meter reads (206.7GB data) per hour (i.e., 1811TB/year). In addition, well-designed publish/subscribe and communication infrastructures ensure the scalability and flexibility of the system.
作为智能电网的重要组成部分,先进计量基础设施(AMI)系统在过去几年中已经在全球范围内实施和部署。AMI系统通过高效的双向通信基础设施将数百万终端设备(例如,住宅级的智能电表和传感器)与公用事业控制中心连接起来。AMI系统能够实时或接近实时地在公用事业和终端设备之间交换大量仪表数据和控制信息。我们研究的主要挑战是扩展ABB的仪表数据管理系统(MDMS),以管理来自数百万智能电表的数据。我们设计了一个轻量级架构,能够从各种计量系统收集不断增加的大量仪表数据,清理、分析和汇总仪表数据,以支持各种智能电网应用。为了满足关键的高性能需求,我们在原型中实现和集成了各种并发处理技术。我们的实验表明,在一台硬件配置为12核CPU、32G RAM和SSD硬盘驱动器的机器上,实现的数据文件解析器平均需要大约42分钟来完成解析、清理和聚合51.84亿个仪表读取。吞吐量约为每小时73.8亿米读取(206.7GB数据)(即1811TB/年)。此外,精心设计的发布/订阅和通信基础设施确保了系统的可伸缩性和灵活性。
{"title":"Speeding up processing data from millions of smart meters","authors":"Jiang Zheng, Zhao Li, A. Dagnino","doi":"10.1145/2568088.2576798","DOIUrl":"https://doi.org/10.1145/2568088.2576798","url":null,"abstract":"As an important element of the Smart Grid, Advanced Metering Infrastructure (AMI) systems have been implemented and deployed throughout the world in the past several years. An AMI system connects millions of end devices (e.g., smart meters and sensors in the residential level) with utility control centers via an efficient two-way communication infrastructure. AMI systems are able to exchange substantial meter data and control information between utilities and end devices in real-time or near real-time. The major challenge our research was to scale ABB's Meter Data Management System (MDMS) to manage data that originates from millions of smart meters. We designed a lightweight architecture capable of collect ever-increasing large amount of meter data from various metering systems, clean, analyze, and aggregate the meter data to support various smart grid applications. To meet critical high performance requirements, various concurrency processing techniques were implemented and integrated in our prototype. Our experiments showed that on average the implemented data file parser took about 42 minutes to complete parsing, cleaning, and aggregating 5.184 billion meter reads on a single machine with the hardware configuration of 12-core CPU, 32G RAM, and SSD Hard Drives. The throughput is about 7.38 billion meter reads (206.7GB data) per hour (i.e., 1811TB/year). In addition, well-designed publish/subscribe and communication infrastructures ensure the scalability and flexibility of the system.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"52 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128067863","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Performance awareness: keynote abstract 表现意识:基调抽象
P. Tůma
The talk will take a broad look at performance awareness, defined as the ability to observe performance and to act on the observations. The implicit question posed in the talk is what can be done to improve various aspects of performance awareness -- be it our awareness of the various performance relevant mechanisms, our awareness of the expected software performance, our ability to attain and exploit performance awareness as software developers, and our options for implementing performance aware applications.
这次演讲将广泛地探讨绩效意识,它被定义为观察绩效并根据观察采取行动的能力。演讲中提出的隐含问题是,我们可以做些什么来提高性能意识的各个方面——无论是我们对各种性能相关机制的意识,我们对预期软件性能的意识,我们作为软件开发人员获得和利用性能意识的能力,以及我们实现性能意识应用程序的选择。
{"title":"Performance awareness: keynote abstract","authors":"P. Tůma","doi":"10.1145/2568088.2576097","DOIUrl":"https://doi.org/10.1145/2568088.2576097","url":null,"abstract":"The talk will take a broad look at performance awareness, defined as the ability to observe performance and to act on the observations. The implicit question posed in the talk is what can be done to improve various aspects of performance awareness -- be it our awareness of the various performance relevant mechanisms, our awareness of the expected software performance, our ability to attain and exploit performance awareness as software developers, and our options for implementing performance aware applications.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"216 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133257727","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
The taming of the shrew: increasing performance by automatic parameter tuning for java garbage collectors 驯悍:通过java垃圾收集器的自动参数调优来提高性能
Philipp Lengauer, H. Mössenböck
Garbage collection, if not tuned properly, can considerably impact application performance. Unfortunately, configuring a garbage collector is a tedious task as only few guidelines exist and tuning is often done by trial and error. We present what is, to our knowledge, the first published work on automatically tuning Java garbage collectors in a black-box manner considering all available parameters. We propose the use of iterated local search methods to automatically compute application-specific garbage collector configurations. Our experiments show that automatic tuning can reduce garbage collection time by up to 77% for a specific application and a specific workload and by 35% on average across all benchmarks (compared to the default configuration). We evaluated our approach for 3 different garbage collectors on the DaCapo and SPECjbb benchmarks, as well as on a real-world industrial application.
垃圾收集如果调优不当,会对应用程序性能产生很大影响。不幸的是,配置垃圾收集器是一项繁琐的任务,因为只有很少的指导方针,而且调优通常是通过反复试验来完成的。据我们所知,这是关于以黑盒方式考虑所有可用参数自动调优Java垃圾收集器的第一篇已发表的文章。我们建议使用迭代本地搜索方法来自动计算特定于应用程序的垃圾收集器配置。我们的实验表明,自动调优可以将特定应用程序和特定工作负载的垃圾收集时间减少77%,在所有基准测试中平均减少35%(与默认配置相比)。我们在DaCapo和SPECjbb基准测试中针对3种不同的垃圾收集器以及实际的工业应用程序评估了我们的方法。
{"title":"The taming of the shrew: increasing performance by automatic parameter tuning for java garbage collectors","authors":"Philipp Lengauer, H. Mössenböck","doi":"10.1145/2568088.2568091","DOIUrl":"https://doi.org/10.1145/2568088.2568091","url":null,"abstract":"Garbage collection, if not tuned properly, can considerably impact application performance. Unfortunately, configuring a garbage collector is a tedious task as only few guidelines exist and tuning is often done by trial and error. We present what is, to our knowledge, the first published work on automatically tuning Java garbage collectors in a black-box manner considering all available parameters. We propose the use of iterated local search methods to automatically compute application-specific garbage collector configurations. Our experiments show that automatic tuning can reduce garbage collection time by up to 77% for a specific application and a specific workload and by 35% on average across all benchmarks (compared to the default configuration). We evaluated our approach for 3 different garbage collectors on the DaCapo and SPECjbb benchmarks, as well as on a real-world industrial application.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"71 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114224560","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 37
Efficient optimization of software performance models via parameter-space pruning 通过参数空间剪枝实现软件性能模型的高效优化
M. Tribastone
When performance characteristics are taken into account in a software design, models can be used to identify optimal configurations of the system's parameters. Unfortunately, for realistic scenarios, the cost of the optimization is typically high, leading to computational difficulties in the exploration of large parameter spaces. This paper proposes an approach to provably exact parameter-space pruning for a class of models of large-scale software systems analyzed with fluid techniques, efficient and scalable deterministic approximations of massively parallel stochastic models. We present a result of monotonicity of fluid solutions with respect to the model parameters, and employ it in the context of optimization programs with evolutionary algorithms by discarding candidate configurations a priori, i.e., without ever solving them, whenever they are proven to give lower fitness than other configurations. An extensive numerical validation shows that this approach yields an average twofold runtime speed-up compared to a baseline optimization algorithm that does not exploit monotonicity. Furthermore, we find that the optimal configuration is within a few percent from the true one obtained by stochastic simulation, whose solution is however orders of magnitude more expensive.
当在软件设计中考虑到性能特征时,可以使用模型来确定系统参数的最佳配置。不幸的是,对于现实场景,优化的成本通常很高,导致在探索大参数空间时出现计算困难。本文提出了用流体技术分析的一类大型软件系统模型的可证明精确参数空间剪枝方法,以及大规模并行随机模型的有效和可扩展的确定性近似。我们提出了流体解相对于模型参数的单调性的结果,并将其应用于具有进化算法的优化程序中,通过先验地丢弃候选配置,即,当它们被证明具有比其他配置更低的适应度时,无需求解它们。广泛的数值验证表明,与不利用单调性的基线优化算法相比,这种方法产生了平均两倍的运行时加速。此外,我们发现最优配置与随机模拟得到的真实配置相差不到几个百分点,但其解要贵几个数量级。
{"title":"Efficient optimization of software performance models via parameter-space pruning","authors":"M. Tribastone","doi":"10.1145/2568088.2568090","DOIUrl":"https://doi.org/10.1145/2568088.2568090","url":null,"abstract":"When performance characteristics are taken into account in a software design, models can be used to identify optimal configurations of the system's parameters. Unfortunately, for realistic scenarios, the cost of the optimization is typically high, leading to computational difficulties in the exploration of large parameter spaces. This paper proposes an approach to provably exact parameter-space pruning for a class of models of large-scale software systems analyzed with fluid techniques, efficient and scalable deterministic approximations of massively parallel stochastic models. We present a result of monotonicity of fluid solutions with respect to the model parameters, and employ it in the context of optimization programs with evolutionary algorithms by discarding candidate configurations a priori, i.e., without ever solving them, whenever they are proven to give lower fitness than other configurations. An extensive numerical validation shows that this approach yields an average twofold runtime speed-up compared to a baseline optimization algorithm that does not exploit monotonicity. Furthermore, we find that the optimal configuration is within a few percent from the true one obtained by stochastic simulation, whose solution is however orders of magnitude more expensive.","PeriodicalId":243233,"journal":{"name":"Proceedings of the 5th ACM/SPEC international conference on Performance engineering","volume":"446 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2014-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115113278","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
期刊
Proceedings of the 5th ACM/SPEC international conference on Performance engineering
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1