Premium
A deep transfer‐learning‐based dynamic reinforcement learning for intelligent tightening system
Author(s) -
Luo Wentao,
Zhang Jianfu,
Feng Pingfa,
Yu Dingwen,
Wu Zhijun
Publication year - 2021
Publication title -
international journal of intelligent systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.291
H-Index - 87
eISSN - 1098-111X
pISSN - 0884-8173
DOI - 10.1002/int.22345
Subject(s) - reinforcement learning , computer science , artificial intelligence , process (computing) , adaptability , transfer of learning , machine learning , robot , knowledge transfer , knowledge management , ecology , biology , operating system
Reinforcement learning (RL) has been widely applied in the static environment with standard reward functions. For intelligent tightening tasks, it is a challenge to transform expert knowledge into a recognizable mathematical expression for RL agents. Changing assembly standards make the model repeat learning updated knowledge with a high time‐cost. In addition, as the difficulty and low accuracy of designing reward functions, the RL model itself also limits its application in the complex and dynamic engineering environment. To solve the above problems, a deep transfer‐learning‐based dynamic reinforcement learning (DRL‐DTL) is presented and applied in the intelligent tightening system. Specifically, a deep convolution transfer‐learning model (DCTL) is presented to build a mathematical mapping between agents of the model and subjective knowledge, which endows agents to learn from human knowledge efficiently. Then, a dynamic expert library is established to improve the adaptability of algorithm to the changing environment. And an inverse RL based on prior knowledge is presented to acquire reward functions. Experiments are conducted on a tightening assembly system and the results show that the tightening robot with the proposed model can inspect quality problems during the tightening process autonomously and make an adjustment decision based on the optimal policy that the agent calculates.