z-logo
Premium
Continuous Action Generation of Q‐Learning in Multi‐Agent Cooperation
Author(s) -
Hwang KaoShing,
Chen YuJen,
Jiang WeiCheng,
Lin TzungFeng
Publication year - 2013
Publication title -
asian journal of control
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.769
H-Index - 53
eISSN - 1934-6093
pISSN - 1561-8625
DOI - 10.1002/asjc.614
Subject(s) - q learning , weighting , generator (circuit theory) , action (physics) , robot , state space , state (computer science) , control theory (sociology) , computer science , reinforcement learning , controller (irrigation) , artificial intelligence , mathematics , algorithm , control (management) , medicine , power (physics) , statistics , physics , quantum mechanics , biology , agronomy , radiology
Conventional Q ‐learning requires pre‐defined quantized state space and action space. It is not practical for real robot applications since discrete and finite numbers of action sets cannot precisely identify the variances in the different positions on the same state element on which the robot is located. In this paper, a Q ‐Learning composed continuous action generator, called the fuzzy cerebellar model articulation controller ( FCMAC ) method, is presented to solve the problem. The FCMAC displays continuous action generation by linear combination of the weighting distribution of the state space where the optimal policy of each state is derived from Q ‐learning. This provides better resolution of the weighting distribution for the state space where the robot is located. The algorithm not only solves the single‐agent problem but also solves the multi‐agent problem by extension. An experiment is implemented in a task where two robots are taking action independently and both are connected with a straight bar. Their goal is to cooperate with each other to pass through a gate in the middle of a grid environment.

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here