z-logo
open-access-imgOpen Access
Asymptotic Learnability of Reinforcement Problems with Arbitrary Dependence
Author(s) -
Daniil Ryabko,
Marcus Hütter
Publication year - 2006
Publication title -
lecture notes in computer science
Language(s) - English
Resource type - Book series
SCImago Journal Rank - 0.249
H-Index - 400
eISSN - 1611-3349
pISSN - 0302-9743
ISBN - 3-540-46649-5
DOI - 10.1007/11894841_27
Subject(s) - learnability , reinforcement learning , countable set , markov decision process , computer science , class (philosophy) , probabilistic logic , task (project management) , markov chain , markov process , reinforcement , theoretical computer science , mathematical optimization , artificial intelligence , mathematics , discrete mathematics , machine learning , statistics , management , economics , psychology , social psychology
We address the problem of reinforcement learning in which observations may exhibit an arbitrary form of stochastic dependence on past observations and actions, i.e. environments more general than (PO) MDPs. The task for an agent is to attain the best possible asymptotic reward where the true generating environment is unknown but belongs to a known countable family of environments. We find some sufficient conditions on the class of environments under which an agent exists which attains the best asymptotic reward for any environment in the class. We analyze how tight these conditions are and how they relate to different probabilistic assumptions known in reinforcement learning and related fields, such as Markov Decision Processes and mixing conditions.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom