HAT: Heterogeneous Adaptive Throttling for On-Chip Networks

K. Chang, Rachata Ausavarungnirun, Chris Fallin, O. Mutlu
{"title":"HAT: Heterogeneous Adaptive Throttling for On-Chip Networks","authors":"K. Chang, Rachata Ausavarungnirun, Chris Fallin, O. Mutlu","doi":"10.1109/SBAC-PAD.2012.44","DOIUrl":null,"url":null,"abstract":"The network-on-chip (NoC) is a primary shared resource in a chip multiprocessor (CMP) system. As core counts continue to increase and applications become increasingly data-intensive, the network load will also increase, leading to more congestion in the network. This network congestion can degrade system performance if the network load is not appropriately controlled. Prior works have proposed source-throttling congestion control, which limits the rate at which new network traffic (packets) enters the NoC in order to reduce congestion and improve performance. These prior congestion control mechanisms have shortcomings that significantly limit their performance: either 1) they are not application-aware, but rather throttle all applications equally regardless of applications' sensitivity to latency, or 2) they are not network-load-aware, throttling according to application characteristics but sometimes under- or over-throttling the cores. In this work, we propose Heterogeneous Adaptive Throttling, or HAT, a new source-throttling congestion control mechanism based on two key principles: application-aware throttling and network-load-aware throttling rate adjustment. First, we observe that only network-bandwidth-intensive applications(those which use the network most heavily) should be throttled, allowing the other latency-sensitive applications to make faster progress without as much interference. Second, we observe that the throttling rate which yields the best performance varies between workloads, a single, static, throttling rate under-throttles some workloads while over-throttling others. Hence, the throttling mechanism should observe network load dynamically and adjust its throttling rate accordingly. While some past works have also used a closed-loop control approach, none have been application-aware. HAT is the first mechanism to combine application-awareness and network-load-aware throttling rate adjustment to address congestion in a NoC. We evaluate HAT using a wide variety of multiprogrammed workloads on several NoC-based CMP systems with 16-, 64-, and 144-cores and compare its performance to two state-of-the-art congestion control mechanisms. Our evaluations show that HAT consistently provides higher system performance and fairness than prior congestion control mechanisms.","PeriodicalId":232444,"journal":{"name":"2012 IEEE 24th International Symposium on Computer Architecture and High Performance Computing","volume":"36 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2012-10-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"78","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2012 IEEE 24th International Symposium on Computer Architecture and High Performance Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/SBAC-PAD.2012.44","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 78

Abstract

The network-on-chip (NoC) is a primary shared resource in a chip multiprocessor (CMP) system. As core counts continue to increase and applications become increasingly data-intensive, the network load will also increase, leading to more congestion in the network. This network congestion can degrade system performance if the network load is not appropriately controlled. Prior works have proposed source-throttling congestion control, which limits the rate at which new network traffic (packets) enters the NoC in order to reduce congestion and improve performance. These prior congestion control mechanisms have shortcomings that significantly limit their performance: either 1) they are not application-aware, but rather throttle all applications equally regardless of applications' sensitivity to latency, or 2) they are not network-load-aware, throttling according to application characteristics but sometimes under- or over-throttling the cores. In this work, we propose Heterogeneous Adaptive Throttling, or HAT, a new source-throttling congestion control mechanism based on two key principles: application-aware throttling and network-load-aware throttling rate adjustment. First, we observe that only network-bandwidth-intensive applications(those which use the network most heavily) should be throttled, allowing the other latency-sensitive applications to make faster progress without as much interference. Second, we observe that the throttling rate which yields the best performance varies between workloads, a single, static, throttling rate under-throttles some workloads while over-throttling others. Hence, the throttling mechanism should observe network load dynamically and adjust its throttling rate accordingly. While some past works have also used a closed-loop control approach, none have been application-aware. HAT is the first mechanism to combine application-awareness and network-load-aware throttling rate adjustment to address congestion in a NoC. We evaluate HAT using a wide variety of multiprogrammed workloads on several NoC-based CMP systems with 16-, 64-, and 144-cores and compare its performance to two state-of-the-art congestion control mechanisms. Our evaluations show that HAT consistently provides higher system performance and fairness than prior congestion control mechanisms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
芯片上网络的异构自适应节流
片上网络(NoC)是片上多处理器(CMP)系统中的主要共享资源。随着核心数量的不断增加和应用程序的数据密集程度越来越高,网络负载也会增加,从而导致网络更加拥塞。如果不适当控制网络负载,这种网络拥塞会降低系统性能。先前的工作已经提出了源节流拥塞控制,它限制了新网络流量(数据包)进入NoC的速率,以减少拥塞并提高性能。这些先前的拥塞控制机制有明显限制其性能的缺点:要么1)它们不是应用程序感知的,而是平等地限制所有应用程序,而不管应用程序对延迟的敏感性如何;要么2)它们不是网络负载感知的,根据应用程序特征进行限制,但有时会对核心进行过少或过少的限制。在这项工作中,我们提出了异构自适应节流(HAT),这是一种新的源节流拥塞控制机制,基于两个关键原则:应用感知节流和网络负载感知节流速率调整。首先,我们观察到只有网络带宽密集型应用程序(那些使用网络最多的应用程序)应该被限制,从而允许其他延迟敏感的应用程序在没有太多干扰的情况下取得更快的进展。其次,我们观察到产生最佳性能的节流率因工作负载而异,单一的静态节流率对某些工作负载的节流不足,而对其他工作负载的节流过高。因此,节流机制应该动态观察网络负载,并相应地调整其节流速率。虽然过去的一些工作也使用了闭环控制方法,但没有一个是应用感知的。HAT是第一个结合应用程序感知和网络负载感知的节流速率调整来解决NoC中的拥塞的机制。我们在几个基于cpu的16核、64核和144核CMP系统上使用各种多程序工作负载来评估HAT,并将其性能与两种最先进的拥塞控制机制进行比较。我们的评估表明,HAT始终比以前的拥塞控制机制提供更高的系统性能和公平性。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Using Heterogeneous Networks to Improve Energy Efficiency in Direct Coherence Protocols for Many-Core CMPs Cloud Workload Analysis with SWAT Energy-Performance Tradeoffs in Software Transactional Memory CSHARP: Coherence and SHaring Aware Cache Replacement Policies for Parallel Applications Exploiting Concurrent GPU Operations for Efficient Work Stealing on Multi-GPUs
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1