Logarithmic Regret Bounds for Continuous-Time Average-Reward Markov Decision Processes

IF 2.2 2区 数学 Q2 AUTOMATION & CONTROL SYSTEMS SIAM Journal on Control and Optimization Pub Date : 2024-09-10 DOI:10.1137/23m1584101
Xuefeng Gao, Xun Yu Zhou
{"title":"Logarithmic Regret Bounds for Continuous-Time Average-Reward Markov Decision Processes","authors":"Xuefeng Gao, Xun Yu Zhou","doi":"10.1137/23m1584101","DOIUrl":null,"url":null,"abstract":"SIAM Journal on Control and Optimization, Volume 62, Issue 5, Page 2529-2556, October 2024. <br/> Abstract. We consider reinforcement learning for continuous-time Markov decision processes (MDPs) in the infinite-horizon, average-reward setting. In contrast to discrete-time MDPs, a continuous-time process moves to a state and stays there for a random holding time after an action is taken. With unknown transition probabilities and rates of exponential holding times, we derive instance-dependent regret lower bounds that are logarithmic in the time horizon. Moreover, we design a learning algorithm and establish a finite-time regret bound that achieves the logarithmic growth rate. Our analysis builds upon upper confidence reinforcement learning, a delicate estimation of the mean holding times, and stochastic comparison of point processes.","PeriodicalId":49531,"journal":{"name":"SIAM Journal on Control and Optimization","volume":"36 1","pages":""},"PeriodicalIF":2.2000,"publicationDate":"2024-09-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"SIAM Journal on Control and Optimization","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1137/23m1584101","RegionNum":2,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"AUTOMATION & CONTROL SYSTEMS","Score":null,"Total":0}
引用次数: 0

Abstract

SIAM Journal on Control and Optimization, Volume 62, Issue 5, Page 2529-2556, October 2024.
Abstract. We consider reinforcement learning for continuous-time Markov decision processes (MDPs) in the infinite-horizon, average-reward setting. In contrast to discrete-time MDPs, a continuous-time process moves to a state and stays there for a random holding time after an action is taken. With unknown transition probabilities and rates of exponential holding times, we derive instance-dependent regret lower bounds that are logarithmic in the time horizon. Moreover, we design a learning algorithm and establish a finite-time regret bound that achieves the logarithmic growth rate. Our analysis builds upon upper confidence reinforcement learning, a delicate estimation of the mean holding times, and stochastic comparison of point processes.
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
连续时间平均回报马尔可夫决策过程的对数回归界线
SIAM 控制与优化期刊》第 62 卷第 5 期第 2529-2556 页,2024 年 10 月。 摘要我们考虑的是无限视距、平均回报环境下连续时间马尔可夫决策过程(MDP)的强化学习。与离散时间马尔可夫决策过程不同,连续时间过程在采取行动后会移动到一个状态,并在该状态停留一段时间。在未知过渡概率和指数保持时间率的情况下,我们推导出了与实例相关的遗憾下限,这些遗憾下限与时间跨度成对数关系。此外,我们还设计了一种学习算法,并建立了能达到对数增长率的有限时间后悔约束。我们的分析建立在上置信度强化学习、平均保持时间的微妙估计和点过程的随机比较之上。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 去求助
来源期刊
CiteScore
4.00
自引率
4.50%
发文量
143
审稿时长
12 months
期刊介绍: SIAM Journal on Control and Optimization (SICON) publishes original research articles on the mathematics and applications of control theory and certain parts of optimization theory. Papers considered for publication must be significant at both the mathematical level and the level of applications or potential applications. Papers containing mostly routine mathematics or those with no discernible connection to control and systems theory or optimization will not be considered for publication. From time to time, the journal will also publish authoritative surveys of important subject areas in control theory and optimization whose level of maturity permits a clear and unified exposition. The broad areas mentioned above are intended to encompass a wide range of mathematical techniques and scientific, engineering, economic, and industrial applications. These include stochastic and deterministic methods in control, estimation, and identification of systems; modeling and realization of complex control systems; the numerical analysis and related computational methodology of control processes and allied issues; and the development of mathematical theories and techniques that give new insights into old problems or provide the basis for further progress in control theory and optimization. Within the field of optimization, the journal focuses on the parts that are relevant to dynamic and control systems. Contributions to numerical methodology are also welcome in accordance with these aims, especially as related to large-scale problems and decomposition as well as to fundamental questions of convergence and approximation.
期刊最新文献
Local Exact Controllability of the One-Dimensional Nonlinear Schrödinger Equation in the Case of Dirichlet Boundary Conditions Backward Stochastic Differential Equations with Conditional Reflection and Related Recursive Optimal Control Problems Logarithmic Regret Bounds for Continuous-Time Average-Reward Markov Decision Processes Optimal Ratcheting of Dividend Payout Under Brownian Motion Surplus An Optimal Spectral Inequality for Degenerate Operators
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1