Tuning the activation function to optimize the forecast horizon of a reservoir computer

IF 2.6 Q1 MATHEMATICS, INTERDISCIPLINARY APPLICATIONS Journal of Physics Complexity Pub Date : 2024-07-11 DOI:10.1088/2632-072x/ad5e55
L A Hurley, J G Restrepo and S E Shaheen
{"title":"Tuning the activation function to optimize the forecast horizon of a reservoir computer","authors":"L A Hurley, J G Restrepo and S E Shaheen","doi":"10.1088/2632-072x/ad5e55","DOIUrl":null,"url":null,"abstract":"Reservoir computing is a machine learning framework where the readouts from a nonlinear system (the reservoir) are trained so that the output from the reservoir, when forced with an input signal, reproduces a desired output signal. A common implementation of reservoir computers (RCs) is to use a recurrent neural network as the reservoir. The design of this network can have significant effects on the performance of the RC. In this paper we study the effect of the node activation function on the ability of RCs to learn and predict chaotic time series. We find that the Forecast Horizon (FH), the time during which the reservoir’s predictions remain accurate, can vary by an order of magnitude across a set of 16 activation functions used in machine learning. By using different functions from this set, and by modifying their parameters, we explore whether the entropy of node activation levels or the curvature of the activation functions determine the predictive ability of the reservoirs. We find that the FH is low when the activation function is used in a region where it has low curvature, and a positive correlation between curvature and FH. For the activation functions studied we find that the largest FH generally occurs at intermediate levels of the entropy of node activation levels. Our results show that the performance of RCs is very sensitive to the activation function shape. Therefore, modifying this shape in hyperparameter optimization algorithms can lead to improvements in RC performance.","PeriodicalId":53211,"journal":{"name":"Journal of Physics Complexity","volume":"244 1","pages":""},"PeriodicalIF":2.6000,"publicationDate":"2024-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Physics Complexity","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1088/2632-072x/ad5e55","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"MATHEMATICS, INTERDISCIPLINARY APPLICATIONS","Score":null,"Total":0}
引用次数: 0

Abstract

Reservoir computing is a machine learning framework where the readouts from a nonlinear system (the reservoir) are trained so that the output from the reservoir, when forced with an input signal, reproduces a desired output signal. A common implementation of reservoir computers (RCs) is to use a recurrent neural network as the reservoir. The design of this network can have significant effects on the performance of the RC. In this paper we study the effect of the node activation function on the ability of RCs to learn and predict chaotic time series. We find that the Forecast Horizon (FH), the time during which the reservoir’s predictions remain accurate, can vary by an order of magnitude across a set of 16 activation functions used in machine learning. By using different functions from this set, and by modifying their parameters, we explore whether the entropy of node activation levels or the curvature of the activation functions determine the predictive ability of the reservoirs. We find that the FH is low when the activation function is used in a region where it has low curvature, and a positive correlation between curvature and FH. For the activation functions studied we find that the largest FH generally occurs at intermediate levels of the entropy of node activation levels. Our results show that the performance of RCs is very sensitive to the activation function shape. Therefore, modifying this shape in hyperparameter optimization algorithms can lead to improvements in RC performance.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
调整激活函数以优化水库计算机的预测范围
蓄水池计算是一种机器学习框架,通过对非线性系统(蓄水池)的读数进行训练,使蓄水池的输出在输入信号的作用下重现所需的输出信号。水库计算机 (RC) 的常见实现方式是使用递归神经网络作为水库。该网络的设计会对 RC 的性能产生重大影响。本文研究了节点激活函数对 RC 学习和预测混沌时间序列能力的影响。我们发现,在机器学习中使用的一组 16 个激活函数中,预测水平线(FH),即储层预测保持准确的时间,会有数量级的变化。通过使用这组函数中的不同函数并修改其参数,我们探索了节点激活水平的熵或激活函数的曲率是否决定了水库的预测能力。我们发现,当在曲率较低的区域使用激活函数时,FH 值较低,而且曲率与 FH 值之间呈正相关。对于所研究的激活函数,我们发现最大的 FH 通常出现在节点激活水平熵的中间水平。我们的结果表明,RC 的性能对激活函数的形状非常敏感。因此,在超参数优化算法中修改这种形状可以提高 RC 性能。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
Journal of Physics Complexity
Journal of Physics Complexity Computer Science-Information Systems
CiteScore
4.30
自引率
11.10%
发文量
45
审稿时长
14 weeks
期刊最新文献
Persistent Mayer Dirac. Fitness-based growth of directed networks with hierarchy The ultrametric backbone is the union of all minimum spanning forests. Exploring the space of graphs with fixed discrete curvatures Augmentations of Forman’s Ricci curvature and their applications in community detection
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1