z-logo
open-access-imgOpen Access
Guided Policy Search Methods: A Review
Author(s) -
Jinyu Du,
Jianlong Fu,
Cong Li
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/1748/2/022039
Subject(s) - gpss , computer science , process (computing) , field (mathematics) , staffing , reinforcement learning , trajectory , artificial intelligence , management science , machine learning , simulation , engineering , programming language , mathematics , physics , management , astronomy , pure mathematics , economics
Guided policy search methods (GPSs) have become important methods in the field of reinforcement learning in recent years. GPSs are a kind of policy search methods that utilize trajectory optimization methods to generate training data, guiding supervised learning. In theoretical research, GPSs combine convex optimization and deep learning, and have achieved fruitful results. In practical applications, they have achieved good results in complex control fields such as robots learning, especially manipulator operations. This paper mainly elaborates the development process and improvement route of GPSs. Firstly, the theoretical knowledge related to the GPSs is introduced. Secondly, the framework and basic methods of GPSs are analyzed; Thirdly, various improved GPSs based on the basic methods are generalized. Finally, the development and future improvement directions of GPSs are summarized, and the problems and future development trends are discussed.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here