Developing End-to-End Control Policies for Robotic Swarms Using Deep Q-learning
Author(s) -
Yufei Wei,
Xiaotong Nie,
Motoaki Hiraga,
Kazuhiro Ohkura,
Zlatan Car
Publication year - 2019
Publication title -
journal of advanced computational intelligence and intelligent informatics
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.172
H-Index - 20
eISSN - 1343-0130
pISSN - 1883-8014
DOI - 10.20965/jaciii.2019.p0920
Subject(s) - reinforcement learning , computer science , artificial intelligence , end to end principle , swarm robotics , robotics , robot , evolutionary robotics , swarm behaviour , task (project management) , deep learning , field (mathematics) , control (management) , machine learning , engineering , mathematics , systems engineering , pure mathematics
In this study, the use of a popular deep reinforcement learning algorithm – deep Q-learning – in developing end-to-end control policies for robotic swarms is explored. Robots only have limited local sensory capabilities; however, in a swarm, they can accomplish collective tasks beyond the capability of a single robot. Compared with most automatic design approaches proposed so far, which belong to the field of evolutionary robotics, deep reinforcement learning techniques provide two advantages: (i) they enable researchers to develop control policies in an end-to-end fashion; and (ii) they require fewer computation resources, especially when the control policy to be developed has a large parameter space. The proposed approach is evaluated in a round-trip task, where the robots are required to travel between two destinations as much as possible. Simulation results show that the proposed approach can learn control policies directly from high-dimensional raw camera pixel inputs for robotic swarms.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom