Hostname: page-component-cd9895bd7-hc48f Total loading time: 0 Render date: 2024-12-23T05:19:50.910Z Has data issue: false hasContentIssue false

On a class of non-Markov decision processes

Published online by Cambridge University Press:  14 July 2016

K. D. Glazebrook*
Affiliation:
University of Newcastle upon Tyne

Abstract

The optimal strategy for a class of non-Markov decision processes is characterised and has the property that changes of action may occur between successive transitions of the process. Results are given which enable the optimal strategy to be computed iteratively.

Type
Research Papers
Copyright
Copyright © Applied Probability Trust 1978 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

Gittins, J. C. and Glazebrook, K. D. (1977) On Bayesian models in stochastic scheduling. J. Appl. Prob. 14, 556565.Google Scholar
Gittins, J. C. and Jones, D. M. (1972) A dynamic allocation index for the sequential design of experiments. In Progress in Statistics: Proc. 9th European Meeting of Statisticians , ed. Gani, J. et al. North Holland, Amsterdam.Google Scholar
Lippman, S. A. (1971) Maximal average-reward policies for semi-Markov decision processes with arbitrary state and action space. Ann. Math. Statist. 42, 17171726.CrossRefGoogle Scholar
Miller, B. L. (1968) Finite state continuous time Markov decision processes with an infinite planning horizon. J. Math. Anal. Appl. 22, 552569.Google Scholar
Ross, S. M. (1968) Arbitrary state Markovian decision processes. Ann. Math. Statist. 39, 21182122.CrossRefGoogle Scholar
Ross, S. M. (1970) Applied Probability Models with Optimization Applications. Holden-Day, San Francisco.Google Scholar