Deep Q-Learning for Nash Equilibria: Nash-DQN
Philippe Casgrain,
Brian Ning and
Sebastian Jaimungal
Applied Mathematical Finance, 2022, vol. 29, issue 1, 62-78
Abstract:
Model-free learning for multi-agent stochastic games is an active area of research. Existing reinforcement learning algorithms, however, are often restricted to zero-sum games and are applicable only in small state-action spaces or other simplified settings. Here, we develop a new data-efficient Deep-Q-learning methodology for model-free learning of Nash equilibria for general-sum stochastic games. The algorithm uses a locally linear-quadratic expansion of the stochastic game, which leads to analytically solvable optimal actions. The expansion is parametrized by deep neural networks to give it sufficient flexibility to learn the environment without the need to experience all state-action pairs. We study symmetry properties of the algorithm stemming from label-invariant stochastic games and as a proof of concept, apply our algorithm to learning optimal trading strategies in competitive electronic markets.
Date: 2022
References: Add references at CitEc
Citations: View citations in EconPapers (2)
Downloads: (external link)
http://hdl.handle.net/10.1080/1350486X.2022.2136727 (text/html)
Access to full text is restricted to subscribers.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:taf:apmtfi:v:29:y:2022:i:1:p:62-78
Ordering information: This journal article can be ordered from
http://www.tandfonline.com/pricing/journal/RAMF20
DOI: 10.1080/1350486X.2022.2136727
Access Statistics for this article
Applied Mathematical Finance is currently edited by Professor Ben Hambly and Christoph Reisinger
More articles in Applied Mathematical Finance from Taylor & Francis Journals
Bibliographic data for series maintained by Chris Longhurst ().