EconPapers    
Economics at your fingertips  
 

On maximizing the average time at a goal

S. Demko and T. P. Hill

Stochastic Processes and their Applications, 1984, vol. 17, issue 2, 349-357

Abstract: In a decision process (gambling or dynamic programming problem) with finite state space and arbitrary decision sets (gambles or actions), there is always available a Markov strategy which uniformly (nearly) maximizes the average time spent at a goal. If the decision sets are closed, there is even a stationary strategy with the same property. Examples are given to show that approximations by discounted or finite horizon payoffs are not useful for the general average reward problem.

Keywords: gambling; theory; goal; problems; dynamic; programming; stationary; strategy; Markov; strategy; average; reward; criterion (search for similar items in EconPapers)
Date: 1984
References: Add references at CitEc
Citations:

Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/0304-4149(84)90010-3
Full text for ScienceDirect subscribers only

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:eee:spapps:v:17:y:1984:i:2:p:349-357

Ordering information: This journal article can be ordered from
http://http://www.elsevier.com/wps/find/supportfaq.cws_home/regional
https://shop.elsevie ... _01_ooc_1&version=01

Access Statistics for this article

Stochastic Processes and their Applications is currently edited by T. Mikosch

More articles in Stochastic Processes and their Applications from Elsevier
Bibliographic data for series maintained by Catherine Liu ().

 
Page updated 2025-03-19
Handle: RePEc:eee:spapps:v:17:y:1984:i:2:p:349-357