Rodrigo e Alvim Alexandre , Marcelo D. Fragoso , Virgílio J.M. Ferreira Filho , Edilson F. Arruda
{"title":"Solving Markov decision processes via state space decomposition and time aggregation","authors":"Rodrigo e Alvim Alexandre , Marcelo D. Fragoso , Virgílio J.M. Ferreira Filho , Edilson F. Arruda","doi":"10.1016/j.ejor.2025.01.037","DOIUrl":null,"url":null,"abstract":"<div><div>Although there are techniques to address large scale Markov decision processes (MDP), a computationally adequate solution of the so-called curse of dimensionality still eludes, in many aspects, a satisfactory treatment. In this paper, we advance in this issue by introducing a novel multi-subset partitioning scheme to allow for a distributed evaluation of the MDP, aiming to accelerate convergence and enable distributed policy improvement across the state space, whereby the value function and the policy improvement step can be performed independently, one subset at a time. The scheme’s innovation hinges on a design that induces communication properties that allow us to evaluate time aggregated trajectories via absorption analysis, thereby limiting the computational effort. The paper introduces and proves the convergence of a class of distributed time aggregation algorithms that combine the partitioning scheme with two-phase time aggregation to distribute the computations and accelerate convergence. In addition, we make use of Foster’s sufficient conditions for stochastic stability to develop a new theoretical result which underpins a partition design that guarantees that large regions of the state space are rarely visited and have a marginal effect on the system’s performance. This enables the design of approximate algorithms to find near-optimal solutions to large scale systems by focusing on the most visited regions of the state space. We validate the approach in a series of experiments featuring production and inventory and queuing applications. The results highlight the potential of the proposed algorithms to rapidly approach the optimal solution under different problem settings.</div></div>","PeriodicalId":55161,"journal":{"name":"European Journal of Operational Research","volume":"324 1","pages":"Pages 155-167"},"PeriodicalIF":6.0000,"publicationDate":"2025-02-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"European Journal of Operational Research","FirstCategoryId":"91","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0377221725000803","RegionNum":2,"RegionCategory":"管理学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"OPERATIONS RESEARCH & MANAGEMENT SCIENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Although there are techniques to address large scale Markov decision processes (MDP), a computationally adequate solution of the so-called curse of dimensionality still eludes, in many aspects, a satisfactory treatment. In this paper, we advance in this issue by introducing a novel multi-subset partitioning scheme to allow for a distributed evaluation of the MDP, aiming to accelerate convergence and enable distributed policy improvement across the state space, whereby the value function and the policy improvement step can be performed independently, one subset at a time. The scheme’s innovation hinges on a design that induces communication properties that allow us to evaluate time aggregated trajectories via absorption analysis, thereby limiting the computational effort. The paper introduces and proves the convergence of a class of distributed time aggregation algorithms that combine the partitioning scheme with two-phase time aggregation to distribute the computations and accelerate convergence. In addition, we make use of Foster’s sufficient conditions for stochastic stability to develop a new theoretical result which underpins a partition design that guarantees that large regions of the state space are rarely visited and have a marginal effect on the system’s performance. This enables the design of approximate algorithms to find near-optimal solutions to large scale systems by focusing on the most visited regions of the state space. We validate the approach in a series of experiments featuring production and inventory and queuing applications. The results highlight the potential of the proposed algorithms to rapidly approach the optimal solution under different problem settings.
尽管有一些技术可以解决大规模马尔可夫决策过程(MDP),但在许多方面,对所谓的“维数诅咒”(curse of dimensionality)的计算上充分的解决方案仍然没有得到令人满意的处理。在本文中,我们通过引入一种新的多子集分区方案来推进这一问题,该方案允许对MDP进行分布式评估,旨在加速收敛并实现跨状态空间的分布式策略改进,从而使值函数和策略改进步骤可以独立执行,每次执行一个子集。该方案的创新在于一种设计,它可以诱导通信特性,使我们能够通过吸收分析来评估时间聚合轨迹,从而限制计算工作量。介绍并证明了一类分布式时间聚合算法的收敛性,该算法将分划方案与两相时间聚合相结合,实现了计算量的分散和收敛速度的加快。此外,我们利用福斯特随机稳定性的充分条件来发展一个新的理论结果,该结果支持分区设计,保证状态空间的大区域很少被访问,并且对系统的性能有边际影响。这使得近似算法的设计能够通过关注状态空间中访问最多的区域来找到大规模系统的近最优解。我们在一系列以生产、库存和排队应用程序为特征的实验中验证了该方法。结果突出了所提出的算法在不同问题设置下快速接近最优解的潜力。
期刊介绍:
The European Journal of Operational Research (EJOR) publishes high quality, original papers that contribute to the methodology of operational research (OR) and to the practice of decision making.