{"title":"Low-complexity algorithm for restless bandits with imperfect observations","authors":"Keqin Liu, Richard Weber, Chengzhong Zhang","doi":"10.1007/s00186-024-00868-x","DOIUrl":null,"url":null,"abstract":"<p>We consider a class of restless bandit problems that finds a broad application area in reinforcement learning and stochastic optimization. We consider <i>N</i> independent discrete-time Markov processes, each of which had two possible states: 1 and 0 (‘good’ and ‘bad’). Only if a process is both in state 1 and observed to be so does reward accrue. The aim is to maximize the expected discounted sum of returns over the infinite horizon subject to a constraint that only <i>M</i> <span>\\((<N)\\)</span> processes may be observed at each step. Observation is error-prone: there are known probabilities that state 1 (0) will be observed as 0 (1). From this one knows, at any time <i>t</i>, a probability that process <i>i</i> is in state 1. The resulting system may be modeled as a restless multi-armed bandit problem with an information state space of uncountable cardinality. Restless bandit problems with even finite state spaces are PSPACE-HARD in general. We propose a novel approach for simplifying the dynamic programming equations of this class of restless bandits and develop a low-complexity algorithm that achieves a strong performance and is readily extensible to the general restless bandit model with observation errors. Under certain conditions, we establish the existence (indexability) of Whittle index and its equivalence to our algorithm. When those conditions do not hold, we show by numerical experiments the near-optimal performance of our algorithm in the general parametric space. Furthermore, we theoretically prove the optimality of our algorithm for homogeneous systems.</p>","PeriodicalId":49862,"journal":{"name":"Mathematical Methods of Operations Research","volume":"15 1","pages":""},"PeriodicalIF":0.9000,"publicationDate":"2024-09-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Mathematical Methods of Operations Research","FirstCategoryId":"100","ListUrlMain":"https://doi.org/10.1007/s00186-024-00868-x","RegionNum":4,"RegionCategory":"数学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"MATHEMATICS, APPLIED","Score":null,"Total":0}
引用次数: 0
Abstract
We consider a class of restless bandit problems that finds a broad application area in reinforcement learning and stochastic optimization. We consider N independent discrete-time Markov processes, each of which had two possible states: 1 and 0 (‘good’ and ‘bad’). Only if a process is both in state 1 and observed to be so does reward accrue. The aim is to maximize the expected discounted sum of returns over the infinite horizon subject to a constraint that only M\((<N)\) processes may be observed at each step. Observation is error-prone: there are known probabilities that state 1 (0) will be observed as 0 (1). From this one knows, at any time t, a probability that process i is in state 1. The resulting system may be modeled as a restless multi-armed bandit problem with an information state space of uncountable cardinality. Restless bandit problems with even finite state spaces are PSPACE-HARD in general. We propose a novel approach for simplifying the dynamic programming equations of this class of restless bandits and develop a low-complexity algorithm that achieves a strong performance and is readily extensible to the general restless bandit model with observation errors. Under certain conditions, we establish the existence (indexability) of Whittle index and its equivalence to our algorithm. When those conditions do not hold, we show by numerical experiments the near-optimal performance of our algorithm in the general parametric space. Furthermore, we theoretically prove the optimality of our algorithm for homogeneous systems.
期刊介绍:
This peer reviewed journal publishes original and high-quality articles on important mathematical and computational aspects of operations research, in particular in the areas of continuous and discrete mathematical optimization, stochastics, and game theory. Theoretically oriented papers are supposed to include explicit motivations of assumptions and results, while application oriented papers need to contain substantial mathematical contributions. Suggestions for algorithms should be accompanied with numerical evidence for their superiority over state-of-the-art methods. Articles must be of interest for a large audience in operations research, written in clear and correct English, and typeset in LaTeX. A special section contains invited tutorial papers on advanced mathematical or computational aspects of operations research, aiming at making such methodologies accessible for a wider audience.
All papers are refereed. The emphasis is on originality, quality, and importance.