A Utility Criterion for Markov Decision Processes
Stratton C. Jaquette
Additional contact information
Stratton C. Jaquette: Systems Control, Inc., Palo Alto
Management Science, 1976, vol. 23, issue 1, 43-49
Abstract:
Optimality criteria for Markov decision processes have historically been based on a risk neutral formulation of the decision maker's preferences. An explicit utility formulation, incorporating both risk and time preference and based on some results in the axiomatic theory of choice under uncertainty, is developed. This forms an optimality criterion called utility optimality with constant aversion to risk. The objective is to maximize the expected utility using an exponential utility function. Implicit in the formulation is an interpretation of the decision process which is not sequential. It is shown that optimal policies exist which are not necessarily stationary for an infinite horizon stationary Markov decision process with finite state and action spaces. An example is given.
Date: 1976
References: Add references at CitEc
Citations: View citations in EconPapers (18)
Downloads: (external link)
http://dx.doi.org/10.1287/mnsc.23.1.43 (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:inm:ormnsc:v:23:y:1976:i:1:p:43-49
Access Statistics for this article
More articles in Management Science from INFORMS Contact information at EDIRC.
Bibliographic data for series maintained by Chris Asher ().