Multiagent cooperation and competition with deep reinforcement learning
Ardi Tampuu,
Tambet Matiisen,
Dorian Kodelja,
Ilya Kuzovkin,
Kristjan Korjus,
Juhan Aru,
Jaan Aru and
Raul Vicente
PLOS ONE, 2017, vol. 12, issue 4, 1-15
Abstract:
Evolution of cooperation and competition can appear when multiple adaptive agents share a biological, social, or technological niche. In the present work we study how cooperation and competition emerge between autonomous agents that learn by reinforcement while using only their raw visual input as the state representation. In particular, we extend the Deep Q-Learning framework to multiagent environments to investigate the interaction between two learning agents in the well-known video game Pong. By manipulating the classical rewarding scheme of Pong we show how competitive and collaborative behaviors emerge. We also describe the progression from competitive to collaborative behavior when the incentive to cooperate is increased. Finally we show how learning by playing against another adaptive agent, instead of against a hard-wired algorithm, results in more robust strategies. The present work shows that Deep Q-Networks can become a useful tool for studying decentralized learning of multiagent systems coping with high-dimensional environments.
Date: 2017
References: View complete reference list from CitEc
Citations: View citations in EconPapers (14)
Downloads: (external link)
https://journals.plos.org/plosone/article?id=10.1371/journal.pone.0172395 (text/html)
https://journals.plos.org/plosone/article/file?id= ... 72395&type=printable (application/pdf)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:plo:pone00:0172395
DOI: 10.1371/journal.pone.0172395
Access Statistics for this article
More articles in PLOS ONE from Public Library of Science
Bibliographic data for series maintained by plosone ().