Intrinsically Motivated Reinforcement Learning: A Promising Framework for Developmental Robot Learning
Author(s) -
Andrew Stout,
George Konidaris,
Andrew G. Barto
Publication year - 2005
Publication title -
scholarworks@umassamherst (university of massachusetts amherst)
Language(s) - English
Resource type - Reports
DOI - 10.21236/ada440079
Subject(s) - reinforcement learning , artificial intelligence , task (project management) , robotics , computer science , embodied cognition , robot , competence (human resources) , architecture , human–computer interaction , cognitive science , psychology , engineering , geography , social psychology , systems engineering , archaeology
: One of the primary challenges of developmental robotics is the question of how to learn and represent increasingly complex behavior in a self-motivated, open-ended way Barto, Singh, and Chentanez (Barto, Singh, & Chentanez 2004; Singh, Barto, & Chentanez 2004) have recently presented an algorithm for intrinsically motivated reinforcement learning that strives to achieve broad competence in an environment in a task-nonspecific manner by incorporating internal reward to build a hierarchical collection of skills. This paper suggests that with its emphasis on task-general, self-motivated, and hierarchical learning, intrinsically motivated reinforcement learning is an obvious choice for organizing behavior in developmental robotics. We present additional preliminary results from a gridworld abstraction of a robot environment and advocate a layered learning architecture for applying the algorithm on a physically embodied system.
Accelerating Research
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom
Address
John Eccles HouseRobert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom