Hostname: page-component-89b8bd64d-9prln Total loading time: 0 Render date: 2026-05-10T23:40:56.314Z Has data issue: false hasContentIssue false

Bayesian dynamic programming

Published online by Cambridge University Press:  01 July 2016

Ulrich Rieder*
Affiliation:
University of Hamburg

Abstract

We consider a non-stationary Bayesian dynamic decision model with general state, action and parameter spaces. It is shown that this model can be reduced to a non-Markovian (resp. Markovian) decision model with completely known transition probabilities. Under rather weak convergence assumptions on the expected total rewards some general results are presented concerning the restriction on deterministic generalized Markov policies, the criteria of optimality and the existence of Bayes policies. These facts are based on the above transformations and on results of Hindererand Schäl.

Information

Type
Research Article
Copyright
Copyright © Applied Probability Trust 1975 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

Article purchase

Temporarily unavailable