风险概率标准下的有限视界部分可观测半马尔可夫决策过程

IF 0.8 4区 管理学 Q4 OPERATIONS RESEARCH & MANAGEMENT SCIENCE Operations Research Letters Pub Date : 2024-09-27 DOI:10.1016/j.orl.2024.107187
Xin Wen , Xianping Guo , Li Xia
{"title":"风险概率标准下的有限视界部分可观测半马尔可夫决策过程","authors":"Xin Wen ,&nbsp;Xianping Guo ,&nbsp;Li Xia","doi":"10.1016/j.orl.2024.107187","DOIUrl":null,"url":null,"abstract":"<div><div>This paper deals with a risk probability minimization problem for finite horizon partially observable semi-Markov decision processes, which are the fairly most general models for stochastic dynamic systems. In contrast to the expected discounted and average criteria, the optimality investigated in this paper is to minimize the probability that the accumulated rewards do not reach a prescribed profit level at the finite terminal stage. First, the state space is augmented as the joint conditional distribution of the current unobserved state and the remaining profit goal. We introduce a class of policies depending on observable histories and a class of Markov policies including observable process with the joint conditional distribution. Then under mild assumptions, we prove that the value function is the unique solution to the optimality equation for the probability criterion by using iteration techniques. The existence of (<em>ϵ</em>-)optimal Markov policy for this problem is established. Finally, we use a bandit problem with the probability criterion to demonstrate our main results in which an effective algorithm and the corresponding numerical calculation are given for the semi-Markov model. Moreover, for the case of reduction to the discrete-time Markov model, we derive a concise solution.</div></div>","PeriodicalId":54682,"journal":{"name":"Operations Research Letters","volume":"57 ","pages":"Article 107187"},"PeriodicalIF":0.8000,"publicationDate":"2024-09-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Finite horizon partially observable semi-Markov decision processes under risk probability criteria\",\"authors\":\"Xin Wen ,&nbsp;Xianping Guo ,&nbsp;Li Xia\",\"doi\":\"10.1016/j.orl.2024.107187\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This paper deals with a risk probability minimization problem for finite horizon partially observable semi-Markov decision processes, which are the fairly most general models for stochastic dynamic systems. In contrast to the expected discounted and average criteria, the optimality investigated in this paper is to minimize the probability that the accumulated rewards do not reach a prescribed profit level at the finite terminal stage. First, the state space is augmented as the joint conditional distribution of the current unobserved state and the remaining profit goal. We introduce a class of policies depending on observable histories and a class of Markov policies including observable process with the joint conditional distribution. Then under mild assumptions, we prove that the value function is the unique solution to the optimality equation for the probability criterion by using iteration techniques. The existence of (<em>ϵ</em>-)optimal Markov policy for this problem is established. Finally, we use a bandit problem with the probability criterion to demonstrate our main results in which an effective algorithm and the corresponding numerical calculation are given for the semi-Markov model. Moreover, for the case of reduction to the discrete-time Markov model, we derive a concise solution.</div></div>\",\"PeriodicalId\":54682,\"journal\":{\"name\":\"Operations Research Letters\",\"volume\":\"57 \",\"pages\":\"Article 107187\"},\"PeriodicalIF\":0.8000,\"publicationDate\":\"2024-09-27\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Operations Research Letters\",\"FirstCategoryId\":\"91\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0167637724001238\",\"RegionNum\":4,\"RegionCategory\":\"管理学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q4\",\"JCRName\":\"OPERATIONS RESEARCH & MANAGEMENT SCIENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Operations Research Letters","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0167637724001238","RegionNum":4,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q4","JCRName":"OPERATIONS RESEARCH & MANAGEMENT SCIENCE","Score":null,"Total":0}
引用次数: 0

摘要

本文研究的是有限视距部分可观测半马尔可夫决策过程的风险概率最小化问题,半马尔可夫决策过程是随机动态系统的最一般模型。与预期贴现和平均准则不同,本文研究的最优性是最小化累积奖励在有限终结阶段达不到规定利润水平的概率。首先,状态空间被扩展为当前未观测状态和剩余利润目标的联合条件分布。我们引入了一类取决于可观测历史的策略,以及一类包含联合条件分布的可观测过程的马尔可夫策略。然后,在温和的假设条件下,我们利用迭代技术证明价值函数是概率准则最优方程的唯一解。针对这个问题,我们确定了 (ϵ-)optimal Markov policy 的存在。最后,我们用一个带有概率准则的强盗问题来证明我们的主要成果,其中给出了半马尔可夫模型的有效算法和相应的数值计算。此外,对于还原到离散时间马尔可夫模型的情况,我们还得出了一个简明的解决方案。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
查看原文
分享 分享
微信好友 朋友圈 QQ好友 复制链接
本刊更多论文
Finite horizon partially observable semi-Markov decision processes under risk probability criteria
This paper deals with a risk probability minimization problem for finite horizon partially observable semi-Markov decision processes, which are the fairly most general models for stochastic dynamic systems. In contrast to the expected discounted and average criteria, the optimality investigated in this paper is to minimize the probability that the accumulated rewards do not reach a prescribed profit level at the finite terminal stage. First, the state space is augmented as the joint conditional distribution of the current unobserved state and the remaining profit goal. We introduce a class of policies depending on observable histories and a class of Markov policies including observable process with the joint conditional distribution. Then under mild assumptions, we prove that the value function is the unique solution to the optimality equation for the probability criterion by using iteration techniques. The existence of (ϵ-)optimal Markov policy for this problem is established. Finally, we use a bandit problem with the probability criterion to demonstrate our main results in which an effective algorithm and the corresponding numerical calculation are given for the semi-Markov model. Moreover, for the case of reduction to the discrete-time Markov model, we derive a concise solution.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
Operations Research Letters
Operations Research Letters 管理科学-运筹学与管理科学
CiteScore
2.10
自引率
9.10%
发文量
111
审稿时长
83 days
期刊介绍: Operations Research Letters is committed to the rapid review and fast publication of short articles on all aspects of operations research and analytics. Apart from a limitation to eight journal pages, quality, originality, relevance and clarity are the only criteria for selecting the papers to be published. ORL covers the broad field of optimization, stochastic models and game theory. Specific areas of interest include networks, routing, location, queueing, scheduling, inventory, reliability, and financial engineering. We wish to explore interfaces with other fields such as life sciences and health care, artificial intelligence and machine learning, energy distribution, and computational social sciences and humanities. Our traditional strength is in methodology, including theory, modelling, algorithms and computational studies. We also welcome novel applications and concise literature reviews.
期刊最新文献
Break maximization for round-robin tournaments without consecutive breaks Anchored rescheduling problem with non-availability periods On BASTA for discrete-time queues Assessing the accuracy of externalities prediction in a LCFS-PR M/G/1 queue under partial information Optimal strategies and values for monotone and classical mean-variance preferences coincide when asset prices are continuous
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
已复制链接
已复制链接
快去分享给好友吧!
我知道了
×
扫码分享
扫码分享
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1