首页 > 最新文献

2011 International Green Computing Conference and Workshops最新文献

英文 中文
A fine-grained component-level power measurement method 一种细粒度组件级功率测量方法
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008599
Zehan Cui, Yan Zhu, Yungang Bao, Mingyu Chen
The ever growing energy consumption of computer systems have become a more and more serious problem in the past few years. Power profiling is a fundamental way for us to better understand where, when and how energy is consumed. This paper presents a direct measurement method to measure the power of main computer components with fine time granularity. To achieve this goal, only small amount of extra hardware are employed. An approach to synchronize power dissipation with program phases has also been proposed in this paper. Based on the preliminary version of our tools, we measure the power of CPU, memory and disk when running SPEC CPU2006 benchmarks, and prove that measurement with fine time granularity is essential. The phenomenon we observe from memory power may be served as a guide for memory management or architecture design towards energy efficiency.
在过去的几年里,不断增长的计算机系统能耗已成为一个越来越严重的问题。电力分析是我们更好地了解能源消耗的地点、时间和方式的基本方法。本文提出了一种时间粒度较细的计算机主要部件功率的直接测量方法。为了实现这个目标,只需要使用少量的额外硬件。本文还提出了一种与程序相位同步功耗的方法。基于我们的工具的初步版本,我们在运行SPEC CPU2006基准测试时测量了CPU,内存和磁盘的功率,并证明了精细时间粒度的测量是必不可少的。我们从内存功率中观察到的现象可以作为内存管理或架构设计的能效指南。
{"title":"A fine-grained component-level power measurement method","authors":"Zehan Cui, Yan Zhu, Yungang Bao, Mingyu Chen","doi":"10.1109/IGCC.2011.6008599","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008599","url":null,"abstract":"The ever growing energy consumption of computer systems have become a more and more serious problem in the past few years. Power profiling is a fundamental way for us to better understand where, when and how energy is consumed. This paper presents a direct measurement method to measure the power of main computer components with fine time granularity. To achieve this goal, only small amount of extra hardware are employed. An approach to synchronize power dissipation with program phases has also been proposed in this paper. Based on the preliminary version of our tools, we measure the power of CPU, memory and disk when running SPEC CPU2006 benchmarks, and prove that measurement with fine time granularity is essential. The phenomenon we observe from memory power may be served as a guide for memory management or architecture design towards energy efficiency.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129687921","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 24
Practical performance prediction under Dynamic Voltage Frequency Scaling 动态电压频率标度下的实用性能预测
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008553
B. Rountree, D. Lowenthal, M. Schulz, B. Supinski
Predicting performance under Dynamic Voltage Frequency Scaling (DVFS) remains an open problem. Current best practice explores available performance counters to serve as input to linear regression models that predict performance. However, the inaccuracies of these models require that large-scale DVFS runtime algorithms predict performance conservatively in order to avoid significant consequences of mispredictions. Recent theoretical work based on interval analysis advocates a more accurate and reliable solution based on a single new performance counter, Leading Loads. In this paper, we evaluate a processor-independent analytic framework for existing performance counters based on this interval analysis model. We begin with an analysis of the counters used in many published models. We then briefly describe the Leading Loads architectural model and describe how we can use Leading Loads Cycles to predict performance under DVFS. We validate this approach for the NAS Parallel Benchmarks and SPEC CPU 2006 benchmarks, demonstrating an order of magnitude improvement in both error and standard deviation compared to the best existing approaches.
动态电压频率标度(DVFS)下的性能预测仍然是一个有待解决的问题。当前的最佳实践探索可用的性能计数器,作为预测性能的线性回归模型的输入。然而,这些模型的不准确性要求大规模DVFS运行时算法保守地预测性能,以避免错误预测的严重后果。最近基于区间分析的理论工作提倡一种更准确和可靠的解决方案,该解决方案基于一个新的性能计数器,Leading Loads。在本文中,我们基于这个区间分析模型,评估了一个与处理器无关的现有性能计数器分析框架。我们首先分析许多已发布模型中使用的计数器。然后,我们简要地描述了超前负载架构模型,并描述了如何使用超前负载周期来预测DVFS下的性能。我们在NAS并行基准测试和SPEC CPU 2006基准测试中验证了这种方法,与现有的最佳方法相比,在误差和标准偏差方面都有了数量级的改进。
{"title":"Practical performance prediction under Dynamic Voltage Frequency Scaling","authors":"B. Rountree, D. Lowenthal, M. Schulz, B. Supinski","doi":"10.1109/IGCC.2011.6008553","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008553","url":null,"abstract":"Predicting performance under Dynamic Voltage Frequency Scaling (DVFS) remains an open problem. Current best practice explores available performance counters to serve as input to linear regression models that predict performance. However, the inaccuracies of these models require that large-scale DVFS runtime algorithms predict performance conservatively in order to avoid significant consequences of mispredictions. Recent theoretical work based on interval analysis advocates a more accurate and reliable solution based on a single new performance counter, Leading Loads. In this paper, we evaluate a processor-independent analytic framework for existing performance counters based on this interval analysis model. We begin with an analysis of the counters used in many published models. We then briefly describe the Leading Loads architectural model and describe how we can use Leading Loads Cycles to predict performance under DVFS. We validate this approach for the NAS Parallel Benchmarks and SPEC CPU 2006 benchmarks, demonstrating an order of magnitude improvement in both error and standard deviation compared to the best existing approaches.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"97 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115015768","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 85
Energy efficient virtual machine allocation in the cloud 云中的节能虚拟机分配
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008550
Ryan Jansen, P. Brenner
Reducing energy consumption is a critical step in lowering data center operating costs for various institutions. As such, with the growing popularity of cloud computing, it is necessary to examine various methods by which energy consumption in cloud environments can be reduced. We analyze the effects of virtual machine allocation on energy consumption, using a variety of real-world policies and a realistic testing scenario. We found that by using an allocation policy designed to minimize energy, total energy consumption could be reduced by up to 14%, and total monetary energy costs could be reduced by up to 26%.
降低能源消耗是降低各种机构数据中心运营成本的关键步骤。因此,随着云计算的日益普及,有必要研究降低云环境能耗的各种方法。我们使用各种现实世界的策略和一个现实的测试场景,分析虚拟机分配对能耗的影响。我们发现,通过使用旨在最小化能源的分配政策,总能源消耗可以减少高达14%,总货币能源成本可以减少高达26%。
{"title":"Energy efficient virtual machine allocation in the cloud","authors":"Ryan Jansen, P. Brenner","doi":"10.1109/IGCC.2011.6008550","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008550","url":null,"abstract":"Reducing energy consumption is a critical step in lowering data center operating costs for various institutions. As such, with the growing popularity of cloud computing, it is necessary to examine various methods by which energy consumption in cloud environments can be reduced. We analyze the effects of virtual machine allocation on energy consumption, using a variety of real-world policies and a realistic testing scenario. We found that by using an allocation policy designed to minimize energy, total energy consumption could be reduced by up to 14%, and total monetary energy costs could be reduced by up to 26%.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"26 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123614020","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 42
Green governors: A framework for Continuously Adaptive DVFS 绿色调控器:连续自适应DVFS的框架
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008552
Vasileios Spiliopoulos, S. Kaxiras, G. Keramidas
We present Continuously Adaptive Dynamic Voltage/Frequency scaling in Linux systems running on Intel i7 and AMD Phenom II processors. By exploiting slack, inherent in memory-bound programs, our approach aims to improve power efficiency even when the processor does not sit idle. Our underlying methodology is based on a simple first-order processor performance model in which frequency scaling is expressed as a change (in cycles) of the main memory latency. Utilizing available monitoring hardware we show that our model is powerful enough to i) predict with reasonable accuracy the effect of frequency scaling (in terms of performance loss) and ii) predict the core energy under different V/f combinations. To validate our approach we perform highly accurate, fine-grained power measurements directly on the off-chip voltage regulators. We use our model to implement various DVFS policies as Linux “green” governors to continuously optimize for various power-efficiency metrics such as EDP or ED2P, or achieve energy savings with a user-specified limit on performance loss. Our evaluation shows that, for SPEC2006 workloads, our governors achieve dynamically the same optimal EDP or ED2P (within 2% on avg.) as an exhaustive search of all possible frequencies. Energy savings can reach up to 56% in memory-bound workloads with corresponding improvements of about 55% for EDP or ED2P.
我们介绍了在Intel i7和AMD Phenom II处理器上运行的Linux系统中的连续自适应动态电压/频率缩放。通过利用内存约束程序固有的空闲,我们的方法旨在提高电源效率,即使处理器不处于空闲状态。我们的基本方法是基于一个简单的一阶处理器性能模型,其中频率缩放表示为主存储器延迟的变化(以周期为单位)。利用可用的监测硬件,我们表明我们的模型足够强大,可以i)以合理的精度预测频率缩放的影响(就性能损失而言)和ii)预测不同V/f组合下的核心能量。为了验证我们的方法,我们直接在片外稳压器上执行高精度,细粒度的功率测量。我们使用我们的模型来实现各种DVFS策略作为Linux“绿色”调控器,以不断优化各种能效指标(如EDP或ED2P),或者在用户指定的性能损失限制下实现节能。我们的评估表明,对于SPEC2006工作负载,我们的调控器动态地实现了与穷尽搜索所有可能频率相同的最优EDP或ED2P(平均在2%以内)。在内存受限的工作负载中,节能最多可达56%,而对于EDP或ED2P,节能可相应提高约55%。
{"title":"Green governors: A framework for Continuously Adaptive DVFS","authors":"Vasileios Spiliopoulos, S. Kaxiras, G. Keramidas","doi":"10.1109/IGCC.2011.6008552","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008552","url":null,"abstract":"We present Continuously Adaptive Dynamic Voltage/Frequency scaling in Linux systems running on Intel i7 and AMD Phenom II processors. By exploiting slack, inherent in memory-bound programs, our approach aims to improve power efficiency even when the processor does not sit idle. Our underlying methodology is based on a simple first-order processor performance model in which frequency scaling is expressed as a change (in cycles) of the main memory latency. Utilizing available monitoring hardware we show that our model is powerful enough to i) predict with reasonable accuracy the effect of frequency scaling (in terms of performance loss) and ii) predict the core energy under different V/f combinations. To validate our approach we perform highly accurate, fine-grained power measurements directly on the off-chip voltage regulators. We use our model to implement various DVFS policies as Linux “green” governors to continuously optimize for various power-efficiency metrics such as EDP or ED2P, or achieve energy savings with a user-specified limit on performance loss. Our evaluation shows that, for SPEC2006 workloads, our governors achieve dynamically the same optimal EDP or ED2P (within 2% on avg.) as an exhaustive search of all possible frequencies. Energy savings can reach up to 56% in memory-bound workloads with corresponding improvements of about 55% for EDP or ED2P.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"84 1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128458485","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 140
Improving MapReduce energy efficiency for computation intensive workloads 改进MapReduce能源效率,用于计算密集型工作负载
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008564
Thomas Wirtz, Rong Ge
MapReduce is a programming model for data intensive computing on large-scale distributed systems. With its wide acceptance and deployment, improving the energy efficiency of MapReduce will lead to significant energy savings for data centers and computational grids. In this paper, we study the performance and energy efficiency of the Hadoop implementation of MapReduce under the context of energy-proportional computing. We consider how MapReduce efficiency varies with two runtime configurations: resource allocation that changes the number of available concurrent workers, and DVFS (Dynamic Voltage and Frequency Scaling) that adjusts the processor frequency based on the workloads' computational needs. Our experimental results indicate significant energy savings can be achieved from judicious resource allocation and intelligent DVFS scheduling for computation intensive applications, though the level of improvements depends on both workload characteristic of the MapReduce application and the policy of resource and DVFS scheduling.
MapReduce是一种面向大规模分布式系统的数据密集型计算的编程模型。随着它的广泛接受和部署,提高MapReduce的能源效率将为数据中心和计算网格节省大量的能源。在本文中,我们研究了在能量比例计算背景下Hadoop实现MapReduce的性能和能效。我们考虑了MapReduce效率如何随两种运行时配置而变化:改变可用并发工作数的资源分配,以及根据工作负载的计算需求调整处理器频率的DVFS(动态电压和频率缩放)。我们的实验结果表明,对于计算密集型应用程序,明智的资源分配和智能DVFS调度可以实现显著的节能,尽管改进的程度取决于MapReduce应用程序的工作负载特征以及资源和DVFS调度策略。
{"title":"Improving MapReduce energy efficiency for computation intensive workloads","authors":"Thomas Wirtz, Rong Ge","doi":"10.1109/IGCC.2011.6008564","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008564","url":null,"abstract":"MapReduce is a programming model for data intensive computing on large-scale distributed systems. With its wide acceptance and deployment, improving the energy efficiency of MapReduce will lead to significant energy savings for data centers and computational grids. In this paper, we study the performance and energy efficiency of the Hadoop implementation of MapReduce under the context of energy-proportional computing. We consider how MapReduce efficiency varies with two runtime configurations: resource allocation that changes the number of available concurrent workers, and DVFS (Dynamic Voltage and Frequency Scaling) that adjusts the processor frequency based on the workloads' computational needs. Our experimental results indicate significant energy savings can be achieved from judicious resource allocation and intelligent DVFS scheduling for computation intensive applications, though the level of improvements depends on both workload characteristic of the MapReduce application and the policy of resource and DVFS scheduling.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124067565","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 92
Assessing data deduplication trade-offs from an energy and performance perspective 从能源和性能的角度评估重复数据删除的利弊
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008567
L. Costa, S. Al-Kiswany, R. Lopes, M. Ripeanu
The energy costs of running computer systems are a growing concern: for large data centers, recent estimates put these costs higher than the cost of hardware itself. As a consequence, energy efficiency has become a pervasive theme for designing, deploying, and operating computer systems. This paper evaluates the energy trade-offs brought by data deduplication in distributed storage systems. Depending on the workload, deduplication can enable a lower storage footprint, reduce the I/O pressure on the storage system, and reduce network traffic, at the cost of increased computational overhead. From an energy perspective, data deduplication enables a trade-off between the energy consumed for additional computation and the energy saved by lower storage and network load. The main point our experiments and model bring home is the following: while for non energy-proportional machines performance- and energy-centric optimizations have break-even points that are relatively close, for the newer generation of energy proportional machines the break-even points are significantly different. An important consequence of this difference is that, with newer systems, there are higher energy inefficiencies when the system is optimized for performance.
运行计算机系统的能源成本越来越令人担忧:对于大型数据中心来说,最近的估计表明这些成本高于硬件本身的成本。因此,能源效率已成为设计、部署和操作计算机系统的普遍主题。本文对分布式存储系统中重复数据删除带来的能量权衡进行了评估。根据工作负载的不同,重复数据删除可以降低存储占用,减少存储系统的I/O压力,减少网络流量,但代价是增加计算开销。从能源的角度来看,重复数据删除可以在额外计算所消耗的能量与较低的存储和网络负载所节省的能量之间进行权衡。我们的实验和模型带来的主要观点如下:对于非能量比例机器,性能和以能量为中心的优化具有相对接近的盈亏平衡点,对于新一代的能量比例机器,盈亏平衡点显着不同。这种差异的一个重要后果是,对于较新的系统,当系统进行性能优化时,存在更高的能源低效。
{"title":"Assessing data deduplication trade-offs from an energy and performance perspective","authors":"L. Costa, S. Al-Kiswany, R. Lopes, M. Ripeanu","doi":"10.1109/IGCC.2011.6008567","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008567","url":null,"abstract":"The energy costs of running computer systems are a growing concern: for large data centers, recent estimates put these costs higher than the cost of hardware itself. As a consequence, energy efficiency has become a pervasive theme for designing, deploying, and operating computer systems. This paper evaluates the energy trade-offs brought by data deduplication in distributed storage systems. Depending on the workload, deduplication can enable a lower storage footprint, reduce the I/O pressure on the storage system, and reduce network traffic, at the cost of increased computational overhead. From an energy perspective, data deduplication enables a trade-off between the energy consumed for additional computation and the energy saved by lower storage and network load. The main point our experiments and model bring home is the following: while for non energy-proportional machines performance- and energy-centric optimizations have break-even points that are relatively close, for the newer generation of energy proportional machines the break-even points are significantly different. An important consequence of this difference is that, with newer systems, there are higher energy inefficiencies when the system is optimized for performance.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"4609 3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130794500","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
Fast algorithms for thermal constrained performance optimization in DAG scheduling on multi-core processors 多核DAG调度中热约束性能优化的快速算法
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008554
Hafiz Fahad Sheikh, I. Ahmad
Thermal management is highly crucial for efficient exploitation of the potentially enormous computational power offered by advanced multi-core processors. Higher temperatures can adversely affect these processors. Without any thermal constraint, a task graph may be scheduled to run on the cores at their maximum voltage. Very often, multiple factors lead to imposing constraints on temperature, ensuring that cores remain below a certain temperature range and yet deliver good performance. The challenge is how to schedule the same task graph under the imposed thermal constraints such that the performance degradation is the minimum. In this paper we present two algorithms for minimizing the performance degradation and the corresponding overhead while satisfying the thermal constraints. The proposed algorithms, named PAVD, and TAVD, adjust a given schedule of a task graph by decreasing the voltage level of judiciously selected tasks in each step. The algorithms differ in the way they select a task at each step and the amount of time spent in searching that task. TAVD selects the tasks by prioritizing among the cores and tasks which attained maximum temperature while PAVD selects the tasks with the minimum performance penalty. For comparison, we develop a simpler greedy-based approach to show that the problem is non-trivial. Extensive experiments using both random and application-oriented task graphs demonstrate that all three algorithms satisfy the imposed thermal constraints by trading-off performance, while each showing its own strength.
热管理对于有效利用先进多核处理器提供的潜在巨大计算能力至关重要。较高的温度会对这些处理器产生不利影响。在没有任何热约束的情况下,任务图可以被安排在内核的最大电压下运行。通常,多种因素导致对温度施加限制,以确保核心保持在一定的温度范围内,同时提供良好的性能。挑战在于如何在强加的热约束下调度相同的任务图,从而使性能下降最小。在本文中,我们提出了两种算法来最小化性能下降和相应的开销,同时满足热约束。所提出的算法PAVD和TAVD通过降低每一步中明智选择的任务的电压水平来调整任务图的给定时间表。这些算法的不同之处在于它们在每一步选择任务的方式以及搜索该任务所花费的时间。TAVD通过在内核和达到最高温度的任务之间进行优先级排序来选择任务,而PAVD则选择性能损失最小的任务。为了比较,我们开发了一种更简单的基于贪婪的方法来表明这个问题不是微不足道的。使用随机和面向应用的任务图进行的大量实验表明,所有三种算法都通过权衡性能来满足强加的热约束,同时每种算法都显示出自己的优势。
{"title":"Fast algorithms for thermal constrained performance optimization in DAG scheduling on multi-core processors","authors":"Hafiz Fahad Sheikh, I. Ahmad","doi":"10.1109/IGCC.2011.6008554","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008554","url":null,"abstract":"Thermal management is highly crucial for efficient exploitation of the potentially enormous computational power offered by advanced multi-core processors. Higher temperatures can adversely affect these processors. Without any thermal constraint, a task graph may be scheduled to run on the cores at their maximum voltage. Very often, multiple factors lead to imposing constraints on temperature, ensuring that cores remain below a certain temperature range and yet deliver good performance. The challenge is how to schedule the same task graph under the imposed thermal constraints such that the performance degradation is the minimum. In this paper we present two algorithms for minimizing the performance degradation and the corresponding overhead while satisfying the thermal constraints. The proposed algorithms, named PAVD, and TAVD, adjust a given schedule of a task graph by decreasing the voltage level of judiciously selected tasks in each step. The algorithms differ in the way they select a task at each step and the amount of time spent in searching that task. TAVD selects the tasks by prioritizing among the cores and tasks which attained maximum temperature while PAVD selects the tasks with the minimum performance penalty. For comparison, we develop a simpler greedy-based approach to show that the problem is non-trivial. Extensive experiments using both random and application-oriented task graphs demonstrate that all three algorithms satisfy the imposed thermal constraints by trading-off performance, while each showing its own strength.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"28 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126944574","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Where does the power go in a computer system: Experimental analysis and implications 计算机系统的动力去了哪里:实验分析和启示
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008598
Hui Chen, Shinan Wang, Weisong Shi
In the last few years the power dissipation problem of computer systems has attracted more and more attention. A lot of work has been done to decrease power dissipation and increase energy efficiency. We are still, however, not observing significant decrease of power dissipation. On the contrary, modern computer systems consume ever increasing amounts of energy. Where does the power go in a computer system is a question that many people are concerned with. Through comprehensive experiments and measurements, we observe several phenomenons that are in opposition to our common sense. Many people believe, for instance, that CPU utilization is a good indicator of the power dissipation of CPU. Our experiment results, however, show that CPU utilization is not an accurate reflection of the CPU power. Moreover, we discover that despite the performance improvements it introduces, cache could be a big problem for power reducing. Based on our observations we derive ten implications that are important for energy efficient system design.
近年来,计算机系统的功耗问题越来越受到人们的关注。在降低功耗和提高能效方面已经做了大量的工作。然而,我们仍然没有观察到功耗的显著降低。相反,现代计算机系统消耗越来越多的能源。计算机系统的电源到哪里去了是许多人关心的问题。通过全面的实验和测量,我们观察到一些与我们的常识相反的现象。例如,许多人认为CPU利用率是CPU功耗的一个很好的指标。然而,我们的实验结果表明,CPU利用率并不是CPU功率的准确反映。此外,我们发现,尽管它带来了性能改进,但缓存可能是降低功耗的一个大问题。根据我们的观察,我们得出了十个对节能系统设计很重要的启示。
{"title":"Where does the power go in a computer system: Experimental analysis and implications","authors":"Hui Chen, Shinan Wang, Weisong Shi","doi":"10.1109/IGCC.2011.6008598","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008598","url":null,"abstract":"In the last few years the power dissipation problem of computer systems has attracted more and more attention. A lot of work has been done to decrease power dissipation and increase energy efficiency. We are still, however, not observing significant decrease of power dissipation. On the contrary, modern computer systems consume ever increasing amounts of energy. Where does the power go in a computer system is a question that many people are concerned with. Through comprehensive experiments and measurements, we observe several phenomenons that are in opposition to our common sense. Many people believe, for instance, that CPU utilization is a good indicator of the power dissipation of CPU. Our experiment results, however, show that CPU utilization is not an accurate reflection of the CPU power. Moreover, we discover that despite the performance improvements it introduces, cache could be a big problem for power reducing. Based on our observations we derive ten implications that are important for energy efficient system design.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126746000","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 25
Sequence pair based voltage island floorplanning 基于序列对的电压岛平面规划
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008601
D. Sengupta, A. Veneris, S. Wilton, A. Ivanov, R. Saleh
In the nanometer era of VLSI design, high power consumption is considered to be a “show-stopper” for many applications. Voltage Island design has emerged as a popular method for addressing this issue. This technique requires multiple supply voltages on the same chip with blocks assigned to different supply voltages. Implementation challenges force blocks with similar supply voltages to be placed contiguous to one another, thereby creating “islands”. Classical floorplanners assume a single supply voltage in the entire SoC and thus require additional design steps to realize voltage islands. In this paper we present a new floorplanning algorithm based on the sequence pair representation that can floorplan blocks in the form of islands. Given the possible supply voltage choices for each block, the floorplanner simultaneously attempts to reduce power and area of the chip. Our floorplanner integrates the tasks of assigning blocks to different supply voltages and the placing of the blocks in the chip. Compared to previous work, the proposed floorplanner on average reduces the area overhead of the chip by 13.5% with 34% runtime improvement. Additionally we explore the tradeoff between power and area for different floorplan solutions.
在超大规模集成电路设计的纳米时代,高功耗被认为是许多应用的“阻碍因素”。电压岛设计已经成为解决这个问题的一种流行方法。这种技术要求在同一芯片上有多个电源电压,并将块分配给不同的电源电压。实现挑战迫使具有相似供电电压的模块彼此相邻放置,从而形成“孤岛”。传统的地板规划器在整个SoC中假设单个电源电压,因此需要额外的设计步骤来实现电压岛。本文提出了一种新的基于序列对表示的平面规划算法,该算法能够以岛屿的形式对街区进行平面规划。考虑到每个模块可能的供电电压选择,地板规划师同时试图减少芯片的功率和面积。我们的地板规划器集成了将模块分配到不同的电源电压和将模块放置在芯片中的任务。与之前的工作相比,所提出的地板规划器平均减少了13.5%的芯片面积开销,运行时间提高了34%。此外,我们还探讨了不同平面图解决方案在功率和面积之间的权衡。
{"title":"Sequence pair based voltage island floorplanning","authors":"D. Sengupta, A. Veneris, S. Wilton, A. Ivanov, R. Saleh","doi":"10.1109/IGCC.2011.6008601","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008601","url":null,"abstract":"In the nanometer era of VLSI design, high power consumption is considered to be a “show-stopper” for many applications. Voltage Island design has emerged as a popular method for addressing this issue. This technique requires multiple supply voltages on the same chip with blocks assigned to different supply voltages. Implementation challenges force blocks with similar supply voltages to be placed contiguous to one another, thereby creating “islands”. Classical floorplanners assume a single supply voltage in the entire SoC and thus require additional design steps to realize voltage islands. In this paper we present a new floorplanning algorithm based on the sequence pair representation that can floorplan blocks in the form of islands. Given the possible supply voltage choices for each block, the floorplanner simultaneously attempts to reduce power and area of the chip. Our floorplanner integrates the tasks of assigning blocks to different supply voltages and the placing of the blocks in the chip. Compared to previous work, the proposed floorplanner on average reduces the area overhead of the chip by 13.5% with 34% runtime improvement. Additionally we explore the tradeoff between power and area for different floorplan solutions.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"37 21","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132972942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
Scheduling and mapping of periodic tasks on multi-core embedded systems with energy harvesting 基于能量收集的多核嵌入式系统周期任务调度与映射
Pub Date : 2011-07-25 DOI: 10.1109/IGCC.2011.6008595
Jun Lu, Qinru Qiu
In this paper we propose a low-complexity and effective task mapping, scheduling and power management method for multi-core real-time embedded systems with energy harvesting. The proposed method is based on the concept of task CPU utilization, which is defined as the worst-case task execution time divided by its period. This work mathematically proves that by allocating the new task to the core with the lowest utilization, we can achieve the lowest overall energy dissipation. This method, combined with a new dynamic voltage and frequency selection (DVFS) algorithm with energy harvesting awareness and task slack management (TSM) forms the proposed UTilization Based (UTB) algorithm. With periodical tasks in a multi-core platform, this partitioned scheduling method is optimal for energy dissipation if the proposed utilization-based scheduling and DVFS algorithm is applied on each core. Experimental results show that new algorithm achieves better performance in terms of deadline miss rate in a single-core environment, comparing to the best of existing algorithm. When applied on a multi-core platform, the UTB algorithm achieves better efficiency in utilizing the harvested energy and overflowed energy.
本文提出了一种低复杂度、高效的多核实时嵌入式能量采集系统任务映射、调度和电源管理方法。该方法基于任务CPU利用率的概念,将其定义为最坏情况下任务执行时间除以其周期。本工作从数学上证明了将新任务分配给利用率最低的核心,可以实现最低的总能耗。该方法与具有能量收集意识和任务松弛管理(TSM)的动态电压频率选择(DVFS)算法相结合,形成了基于利用率(UTilization Based, UTB)算法。对于多核平台上的周期性任务,如果在每个核上采用基于利用率的调度和DVFS算法,这种分区调度方法的能量消耗是最优的。实验结果表明,在单核环境下,与现有算法相比,新算法在截止日期缺失率方面取得了更好的性能。当应用于多核平台时,UTB算法在收集能量和溢出能量的利用上取得了更好的效率。
{"title":"Scheduling and mapping of periodic tasks on multi-core embedded systems with energy harvesting","authors":"Jun Lu, Qinru Qiu","doi":"10.1109/IGCC.2011.6008595","DOIUrl":"https://doi.org/10.1109/IGCC.2011.6008595","url":null,"abstract":"In this paper we propose a low-complexity and effective task mapping, scheduling and power management method for multi-core real-time embedded systems with energy harvesting. The proposed method is based on the concept of task CPU utilization, which is defined as the worst-case task execution time divided by its period. This work mathematically proves that by allocating the new task to the core with the lowest utilization, we can achieve the lowest overall energy dissipation. This method, combined with a new dynamic voltage and frequency selection (DVFS) algorithm with energy harvesting awareness and task slack management (TSM) forms the proposed UTilization Based (UTB) algorithm. With periodical tasks in a multi-core platform, this partitioned scheduling method is optimal for energy dissipation if the proposed utilization-based scheduling and DVFS algorithm is applied on each core. Experimental results show that new algorithm achieves better performance in terms of deadline miss rate in a single-core environment, comparing to the best of existing algorithm. When applied on a multi-core platform, the UTB algorithm achieves better efficiency in utilizing the harvested energy and overflowed energy.","PeriodicalId":306876,"journal":{"name":"2011 International Green Computing Conference and Workshops","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2011-07-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134464967","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 38
期刊
2011 International Green Computing Conference and Workshops
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1