z-logo
open-access-imgOpen Access
Policy-based monocular vision autonomous quadrotor obstacle avoidance method
Author(s) -
Baojin Zheng,
Xiao Guo,
Jiajun Ou
Publication year - 2021
Publication title -
journal of physics. conference series
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.21
H-Index - 85
eISSN - 1742-6596
pISSN - 1742-6588
DOI - 10.1088/1742-6596/2083/3/032025
Subject(s) - obstacle avoidance , robustness (evolution) , obstacle , computer science , generalization , reinforcement learning , monocular , collision avoidance , artificial intelligence , control theory (sociology) , robot , control (management) , mathematics , mobile robot , computer security , mathematical analysis , biochemistry , chemistry , political science , law , gene , collision
Aiming at the obstacle avoidance control problem of small quadrotor, a method of quadrotor obstacle avoidance based on reinforcement learning is proposed. The proposed method can make training converge quickly and has good environmental robustness. The proposed methods include: (1) a framework adopts perception module and decision module to improve the generalization ability of the obstacle avoidance model; (2) An Actor-Critic framework-based Proximal Policy Optimization (PPO) algorithm to provide quadrotor with policy-based decision-making capabilities; The experimental simulation results show that the strategy-based framework converges quickly and has a high success rate, the training time is much lower than that of the value-based framework. The monocular vision observation ability is limited, which leads to deviations between local observation and global state, So LSTM layer is usually added to increase model performance. Policy -based decision can have a good obstacle avoidance effect without adding the LSTM layer, and have good generalization ability after short relearning after changing.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here