Dynamic stock-decision ensemble strategy based on deep reinforcement learning.

Xiaoming Yu, Wenjun Wu, Xingchuang Liao, Yong Han
{"title":"Dynamic stock-decision ensemble strategy based on deep reinforcement learning.","authors":"Xiaoming Yu,&nbsp;Wenjun Wu,&nbsp;Xingchuang Liao,&nbsp;Yong Han","doi":"10.1007/s10489-022-03606-0","DOIUrl":null,"url":null,"abstract":"<p><p>In a complex and changeable stock market, it is very important to design a trading agent that can benefit investors. In this paper, we propose two stock trading decision-making methods. First, we propose a nested reinforcement learning (Nested RL) method based on three deep reinforcement learning models (the Advantage Actor Critic, Deep Deterministic Policy Gradient, and Soft Actor Critic models) that adopts an integration strategy by nesting reinforcement learning on the basic decision-maker. Thus, this strategy can dynamically select agents according to the current situation to generate trading decisions made under different market environments. Second, to inherit the advantages of three basic decision-makers, we consider confidence and propose a weight random selection with confidence (WRSC) strategy. In this way, investors can gain more profits by integrating the advantages of all agents. All the algorithms are validated for the U.S., Japanese and British stocks and evaluated by different performance indicators. The experimental results show that the annualized return, cumulative return, and Sharpe ratio values of our ensemble strategy are higher than those of the baselines, which indicates that our nested RL and WRSC methods can assist investors in their portfolio management with more profits under the same level of investment risk.</p>","PeriodicalId":72260,"journal":{"name":"Applied intelligence (Dordrecht, Netherlands)","volume":"53 2","pages":"2452-2470"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9082989/pdf/","citationCount":"3","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Applied intelligence (Dordrecht, Netherlands)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s10489-022-03606-0","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 3

Abstract

In a complex and changeable stock market, it is very important to design a trading agent that can benefit investors. In this paper, we propose two stock trading decision-making methods. First, we propose a nested reinforcement learning (Nested RL) method based on three deep reinforcement learning models (the Advantage Actor Critic, Deep Deterministic Policy Gradient, and Soft Actor Critic models) that adopts an integration strategy by nesting reinforcement learning on the basic decision-maker. Thus, this strategy can dynamically select agents according to the current situation to generate trading decisions made under different market environments. Second, to inherit the advantages of three basic decision-makers, we consider confidence and propose a weight random selection with confidence (WRSC) strategy. In this way, investors can gain more profits by integrating the advantages of all agents. All the algorithms are validated for the U.S., Japanese and British stocks and evaluated by different performance indicators. The experimental results show that the annualized return, cumulative return, and Sharpe ratio values of our ensemble strategy are higher than those of the baselines, which indicates that our nested RL and WRSC methods can assist investors in their portfolio management with more profits under the same level of investment risk.

Abstract Image

Abstract Image

Abstract Image

查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
基于深度强化学习的动态股票决策集成策略。
在一个复杂多变的股票市场中,设计一个能让投资者受益的交易代理是非常重要的。本文提出了两种股票交易决策方法。首先,我们提出了一种基于三种深度强化学习模型(优势参与者批评模型、深度确定性策略梯度模型和软参与者批评模型)的嵌套强化学习(nested RL)方法,该方法通过在基本决策者上嵌套强化学习来采用集成策略。因此,该策略可以根据当前情况动态选择agent,生成在不同市场环境下的交易决策。其次,为了继承三个基本决策者的优势,考虑置信度,提出了加权随机选择的置信度策略。这样,投资者可以通过整合各代理商的优势来获得更多的利润。所有的算法都在美国、日本和英国股票上进行了验证,并通过不同的业绩指标进行了评估。实验结果表明,集合策略的年化收益率、累积收益率和夏普比率值均高于基线值,表明在相同的投资风险水平下,我们的嵌套RL和WRSC方法可以帮助投资者进行投资组合管理,并获得更高的收益。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
自引率
0.00%
发文量
0
期刊最新文献
Label recovery and label correlation co-learning for multi-view multi-label classification with incomplete labels. DC-CNN: Dual-channel Convolutional Neural Networks with attention-pooling for fake news detection. A dual alignment-based multi-source domain adaptation framework for motor imagery EEG classification. A novel hybrid multi-thread metaheuristic approach for fake news detection in social media. Front-end deep learning web apps development and deployment: a review.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1