Hostname: page-component-89b8bd64d-dvtzq Total loading time: 0 Render date: 2026-05-07T14:00:24.274Z Has data issue: false hasContentIssue false

Asymptotically optimal policies for weakly coupled Markov decision processes

Published online by Cambridge University Press:  25 March 2026

Diego Goldsztajn*
Affiliation:
Universidad ORT Uruguay
Konstantin Avrachenkov*
Affiliation:
Inria
*
*Postal address: 11100 Montevideo, Uruguay. Email: goldsztajn@ort.edu.uy
**Postal address: 06902 Sophia Antipolis Cedex, France. Email: konstantin.avrachenkov@inria.fr

Abstract

We consider the problem of maximizing the expected average reward obtained over an infinite time horizon by n weakly coupled Markov decision processes. Our setup is a substantial generalization of the multi-armed restless bandit problem that allows for multiple actions and constraints. We establish a connection with a deterministic and continuous-variable control problem where the objective is to maximize the average reward derived from an occupancy measure that represents the empirical distribution of the processes when $n \to \infty$. We show that a solution of this fluid problem can be used to construct policies for the weakly coupled processes that achieve the maximum expected average reward as $n \to \infty$, and we give sufficient conditions for the existence of solutions. Under certain assumptions on the constraints, we prove that these conditions are automatically satisfied if the unconstrained single-process problem admits a suitable unichain and aperiodic policy. In particular, the assumptions include multi-armed restless bandits and a broad class of problems with multiple actions and inequality constraints. Also, the policies can be constructed in an explicit way in these cases. Our theoretical results are complemented by several concrete examples and numerical experiments, which include multichain setups that are covered by the theoretical results.

Information

Type
Original Article
Copyright
© The Author(s), 2026. Published by Cambridge University Press on behalf of Applied Probability Trust.

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

Article purchase

Temporarily unavailable