Deep Reinforcement Learning in Cryptocurrency Market Making
Jonathan Sadighian
Papers from arXiv.org
Abstract:
This paper sets forth a framework for deep reinforcement learning as applied to market making (DRLMM) for cryptocurrencies. Two advanced policy gradient-based algorithms were selected as agents to interact with an environment that represents the observation space through limit order book data, and order flow arrival statistics. Within the experiment, a forward-feed neural network is used as the function approximator and two reward functions are compared. The performance of each combination of agent and reward function is evaluated by daily and average trade returns. Using this DRLMM framework, this paper demonstrates the effectiveness of deep reinforcement learning in solving stochastic inventory control challenges market makers face.
Date: 2019-11
New Economics Papers: this item is included in nep-big, nep-cmp, nep-ore and nep-pay
References: View references in EconPapers View complete reference list from CitEc
Citations: View citations in EconPapers (16)
Downloads: (external link)
http://arxiv.org/pdf/1911.08647 Latest version (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:arx:papers:1911.08647
Access Statistics for this paper
More papers in Papers from arXiv.org
Bibliographic data for series maintained by arXiv administrators ().