LQR controller design for affine LPV systems using reinforcement learning
Hosein Ranjbarpur,
Hajar Atrianfar and
Mohammad Bagher Menhaj
International Journal of Systems Science, 2024, vol. 55, issue 9, 1807-1819
Abstract:
In this paper, a data-driven sub-optimal state feedback is designed for a continuous time linear parameter varying (LPV) system using reinforcement learning. Time-varying parameters lie in a poly top and the system matrix has an affine representation for the parameters. Two novels, on-policy and off-policy algorithms, are proposed using available data from vertex systems of polytop to minimise a performance index and admit a common Lyapunov function (CLF). A convex optimisation problem is derived for each iteration based on Lyapunov inequality. Algorithms yield stabilising feedback gain in each iteration and convergence to a common lyapunov function. We demonstrate the efficacy of the proposed method by simulation of two case studies.
Date: 2024
References: Add references at CitEc
Citations:
Downloads: (external link)
http://hdl.handle.net/10.1080/00207721.2024.2321370 (text/html)
Access to full text is restricted to subscribers.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:taf:tsysxx:v:55:y:2024:i:9:p:1807-1819
Ordering information: This journal article can be ordered from
http://www.tandfonline.com/pricing/journal/TSYS20
DOI: 10.1080/00207721.2024.2321370
Access Statistics for this article
International Journal of Systems Science is currently edited by Visakan Kadirkamanathan
More articles in International Journal of Systems Science from Taylor & Francis Journals
Bibliographic data for series maintained by Chris Longhurst ().