z-logo
Premium
Direct and indirect reinforcement learning
Author(s) -
Guan Yang,
Li Shengbo Eben,
Duan Jingliang,
Li Jie,
Ren Yangang,
Sun Qi,
Cheng Bo
Publication year - 2021
Publication title -
international journal of intelligent systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.291
H-Index - 87
eISSN - 1098-111X
pISSN - 0884-8173
DOI - 10.1002/int.22466
Subject(s) - reinforcement learning , bellman equation , markov decision process , computer science , equivalence (formal languages) , mathematical optimization , optimal control , artificial intelligence , markov process , mathematics , statistics , discrete mathematics
Abstract Reinforcement learning (RL) algorithms have been successfully applied to a range of challenging sequential decision‐making and control tasks. In this paper, we classify RL into direct and indirect RL according to how they seek the optimal policy of the Markov decision process problem. The former solves the optimal policy by directly maximizing an objective function using gradient descent methods, in which the objective function is usually the expectation of accumulative future rewards. The latter indirectly finds the optimal policy by solving the Bellman equation, which is the sufficient and necessary condition from Bellman's principle of optimality. We study policy gradient (PG) forms of direct and indirect RL and show that both of them can derive the actor–critic architecture and can be unified into a PG with the approximate value function and the stationary state distribution, revealing the equivalence of direct and indirect RL. We employ a Gridworld task to verify the influence of different forms of PG, suggesting their differences and relationships experimentally. Finally, we classify current mainstream RL algorithms using the direct and indirect taxonomy, together with other ones, including value‐based and policy‐based, model‐based and model‐free.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here