EconPapers    
Economics at your fingertips  
 

Computing a Bias-Optimal Policy in a Discrete-Time Markov Decision Problem

Eric V. Denardo
Additional contact information
Eric V. Denardo: Yale University, New Haven, Connecticut

Operations Research, 1970, vol. 18, issue 2, 279-289

Abstract: This paper treats a discrete-time Markov decision model with an infinite planning horizon and no discounting. A “bias-optimal” policy for this decision problem satisfies a criterion that is more selective than maximizing the gain rate. The problem of computing a bias-optimal policy, also treated by Veinott in 1966, is here parsed into a sequence of three simple Markov decision problems, each of which can be solved by linear programming or policy iteration.

Date: 1970
References: Add references at CitEc
Citations:

Downloads: (external link)
http://dx.doi.org/10.1287/opre.18.2.279 (application/pdf)

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:inm:oropre:v:18:y:1970:i:2:p:279-289

Access Statistics for this article

More articles in Operations Research from INFORMS Contact information at EDIRC.
Bibliographic data for series maintained by Chris Asher ().

 
Page updated 2025-03-19
Handle: RePEc:inm:oropre:v:18:y:1970:i:2:p:279-289