Heinrich, J;
Silver, D;
(2016)
Deep Reinforcement Learning from Self-Play in Imperfect-Information Games.
ArXiv
, Article 1603.01121.
Preview |
Text
Silver_1603.01121.pdf - Accepted Version Download (521kB) | Preview |
Abstract
Many real-world applications can be described as large-scale games of imperfect information. To deal with these challenging domains, prior work has focused on computing Nash equilibria in a handcrafted abstraction of the domain. In this paper we introduce the first scalable end-to-end approach to learning approximate Nash equilibria without prior domain knowledge. Our method combines fictitious self-play with deep reinforcement learning. When applied to Leduc poker, Neural Fictitious Self-Play (NFSP) approached a Nash equilibrium, whereas common reinforcement learning methods diverged. In Limit Texas Hold’em, a poker game of real-world scale, NFSP learnt a strategy that approached the performance of state-of-the-art, superhuman algorithms based on significant domain expertise.
Type: | Article |
---|---|
Title: | Deep Reinforcement Learning from Self-Play in Imperfect-Information Games |
Open access status: | An open access version is available from UCL Discovery |
Publisher version: | https://arxiv.org/abs/1603.01121 |
Language: | English |
Additional information: | This version is the version of record. For information on re-use, please refer to the publisher’s terms and conditions. |
UCL classification: | UCL UCL > Provost and Vice Provost Offices UCL > Provost and Vice Provost Offices > UCL BEAMS UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science UCL > Provost and Vice Provost Offices > UCL BEAMS > Faculty of Engineering Science > Dept of Computer Science |
URI: | https://discovery.ucl.ac.uk/id/eprint/1523603 |
Archive Staff Only
![]() |
View Item |