EconPapers    
Economics at your fingertips  
 

Dynamic programming for ergodic control with partial observations

V. S. Borkar

Stochastic Processes and their Applications, 2003, vol. 103, issue 2, 293-310

Abstract: A dynamic programming principle is derived for a discrete time Markov control process taking values in a finite dimensional space, with ergodic cost and partial observations. This uses the embedding of the process into another for which an accessible atom exists and hence a coupling argument can be used. In turn, this is used for deriving a martingale dynamic programming principle for ergodic control of partially observed diffusion processes, by 'lifting' appropriate estimates from a discrete time problem associated with it to the continuous time problem.

Date: 2003
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (5)

Downloads: (external link)
http://www.sciencedirect.com/science/article/pii/S0304-4149(02)00190-4
Full text for ScienceDirect subscribers only

Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.

Export reference: BibTeX RIS (EndNote, ProCite, RefMan) HTML/Text

Persistent link: https://EconPapers.repec.org/RePEc:eee:spapps:v:103:y:2003:i:2:p:293-310

Ordering information: This journal article can be ordered from
http://http://www.elsevier.com/wps/find/supportfaq.cws_home/regional
https://shop.elsevie ... _01_ooc_1&version=01

Access Statistics for this article

Stochastic Processes and their Applications is currently edited by T. Mikosch

More articles in Stochastic Processes and their Applications from Elsevier
Bibliographic data for series maintained by Catherine Liu ().

 
Page updated 2025-03-19
Handle: RePEc:eee:spapps:v:103:y:2003:i:2:p:293-310