Investigating a Dynamic Loop Scheduling with Reinforcement Learning Approach to Load Balancing in Scientific Applications

M. Rashid, I. Banicescu, R. Cariño
{"title":"Investigating a Dynamic Loop Scheduling with Reinforcement Learning Approach to Load Balancing in Scientific Applications","authors":"M. Rashid, I. Banicescu, R. Cariño","doi":"10.1109/ISPDC.2008.25","DOIUrl":null,"url":null,"abstract":"The advantages of integrating reinforcement learning (RL) techniques into scientific parallel time-stepping applications have been revealed in research work over the past few years. The object of the integration was to automatically select the most appropriate dynamic loop scheduling (DLS) algorithm from a set of available algorithms with the purpose of improving the application performance via load balancing during the application execution. This paper investigates the performance of such a dynamic loop scheduling with reinforcement learning (DLS-with-RL) approach to load balancing. The DLS-with-RL is most suitable for use in time-stepping scientific applications with large number of steps. The RL agent's characteristics depend on a learning rate parameter and a discount factor parameter. An application simulating wavepacket dynamics that incorporates a DLS-with-RL approach is allowed to execute on a cluster of workstations to investigate the influences of these parameters. The RL agent implemented two RL algorithms: QLEARN and SARSA learning. Preliminary results indicate that on a fixed number of processors, the simulation completion time is not sensitive to the values of the learning parameters used in the experiments. The results also indicate that for this application, there is no advantage of choosing one RL technique over another, even though the techniques differed significantly in the number of times they selected the various DLS algorithms.","PeriodicalId":125975,"journal":{"name":"2008 International Symposium on Parallel and Distributed Computing","volume":"77 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2008-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"8","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2008 International Symposium on Parallel and Distributed Computing","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISPDC.2008.25","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 8

Abstract

The advantages of integrating reinforcement learning (RL) techniques into scientific parallel time-stepping applications have been revealed in research work over the past few years. The object of the integration was to automatically select the most appropriate dynamic loop scheduling (DLS) algorithm from a set of available algorithms with the purpose of improving the application performance via load balancing during the application execution. This paper investigates the performance of such a dynamic loop scheduling with reinforcement learning (DLS-with-RL) approach to load balancing. The DLS-with-RL is most suitable for use in time-stepping scientific applications with large number of steps. The RL agent's characteristics depend on a learning rate parameter and a discount factor parameter. An application simulating wavepacket dynamics that incorporates a DLS-with-RL approach is allowed to execute on a cluster of workstations to investigate the influences of these parameters. The RL agent implemented two RL algorithms: QLEARN and SARSA learning. Preliminary results indicate that on a fixed number of processors, the simulation completion time is not sensitive to the values of the learning parameters used in the experiments. The results also indicate that for this application, there is no advantage of choosing one RL technique over another, even though the techniques differed significantly in the number of times they selected the various DLS algorithms.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于强化学习的动态循环调度在科学应用中的负载平衡研究
在过去几年的研究工作中,强化学习(RL)技术在科学并行时间步进应用中的优势已经被揭示出来。集成的目标是从一组可用算法中自动选择最合适的动态循环调度(DLS)算法,目的是通过应用程序执行期间的负载平衡来提高应用程序性能。本文研究了采用强化学习(dl -with- rl)方法实现负载平衡的动态循环调度的性能。DLS-with-RL最适合用于有大量步长的时间步进科学应用。RL代理的特征取决于学习率参数和折扣因子参数。一个模拟波包动力学的应用程序,结合了dls和rl方法,允许在一个工作站集群上执行,以研究这些参数的影响。RL代理实现了两种RL算法:QLEARN和SARSA学习。初步结果表明,在一定数量的处理器上,仿真完成时间对实验中使用的学习参数值不敏感。结果还表明,对于这个应用程序,选择一种RL技术比另一种技术没有优势,即使这些技术在选择各种DLS算法的次数上存在显著差异。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Towards Security Hardening of Scientific Demand-Driven and Pipelined Distributed Computing Systems A Runtime System Architecture for Ubiquitous Support of OpenMP Context Modeling for Urban Mobile Applications Derivation Control Environment as a Tool for an Efficient Distributed Graph Transformations Coordination Investigating a Dynamic Loop Scheduling with Reinforcement Learning Approach to Load Balancing in Scientific Applications
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1