A general reinforcement learning algorithm that masters chess, shogi, and Go through self-play
Author(s) -
David Silver,
Thomas Hubert,
Julian Schrittwieser,
Ioannis Antonoglou,
Matthew Lai,
Arthur Guez,
Marc Lanctot,
Laurent Sifre,
Dharshan Kumaran,
Thore Graepel,
Timothy Lillicrap,
Karen Simonyan,
Demis Hassabis
Publication year - 2018
Publication title -
science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 12.556
H-Index - 1186
eISSN - 1095-9203
pISSN - 0036-8075
DOI - 10.1126/science.aar6404
Subject(s) - reinforcement learning , reinforcement , computer science , artificial intelligence , cognitive science , machine learning , psychology , social psychology
The game of chess is the longest-studied domain in the history of artificial intelligence. The strongest programs are based on a combination of sophisticated search techniques, domain-specific adaptations, and handcrafted evaluation functions that have been refined by human experts over several decades. By contrast, the AlphaGo Zero program recently achieved superhuman performance in the game of Go by reinforcement learning from self-play. In this paper, we generalize this approach into a single AlphaZero algorithm that can achieve superhuman performance in many challenging games. Starting from random play and given no domain knowledge except the game rules, AlphaZero convincingly defeated a world champion program in the games of chess and shogi (Japanese chess), as well as Go.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom