On maximizing the average time at a goal
S. Demko and
T. P. Hill
Stochastic Processes and their Applications, 1984, vol. 17, issue 2, 349-357
Abstract:
In a decision process (gambling or dynamic programming problem) with finite state space and arbitrary decision sets (gambles or actions), there is always available a Markov strategy which uniformly (nearly) maximizes the average time spent at a goal. If the decision sets are closed, there is even a stationary strategy with the same property. Examples are given to show that approximations by discounted or finite horizon payoffs are not useful for the general average reward problem.
Keywords: gambling; theory; goal; problems; dynamic; programming; stationary; strategy; Markov; strategy; average; reward; criterion (search for similar items in EconPapers)
Date: 1984
References: Add references at CitEc
Citations:
Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/0304-4149(84)90010-3
Full text for ScienceDirect subscribers only
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:eee:spapps:v:17:y:1984:i:2:p:349-357
Ordering information: This journal article can be ordered from
http://http://www.elsevier.com/wps/find/supportfaq.cws_home/regional
https://shop.elsevie ... _01_ooc_1&version=01
Access Statistics for this article
Stochastic Processes and their Applications is currently edited by T. Mikosch
More articles in Stochastic Processes and their Applications from Elsevier
Bibliographic data for series maintained by Catherine Liu ().