Dynamic Programming and Hamilton–Jacobi–Bellman Equations on Time Scales
Yingjun Zhu and
Guangyan Jia
Complexity, 2020, vol. 2020, 1-11
Abstract:
Bellman optimality principle for the stochastic dynamic system on time scales is derived, which includes the continuous time and discrete time as special cases. At the same time, the Hamilton–Jacobi–Bellman (HJB) equation on time scales is obtained. Finally, an example is employed to illustrate our main results.
Date: 2020
References: Add references at CitEc
Citations:
Downloads: (external link)
http://downloads.hindawi.com/journals/8503/2020/7683082.pdf (application/pdf)
http://downloads.hindawi.com/journals/8503/2020/7683082.xml (text/xml)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:hin:complx:7683082
DOI: 10.1155/2020/7683082
Access Statistics for this article
More articles in Complexity from Hindawi
Bibliographic data for series maintained by Mohamed Abdelhakeem ().