
On the improvement of reinforcement active learning with the involvement of cross entropy to address one-shot learning problem
Author(s) -
Huan Huang,
Jincai Huang,
Yanghe Feng,
Jiarui Zhang,
Zhong Liu,
Qi Wang,
Li Chen
Publication year - 2019
Publication title -
plos one
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.99
H-Index - 332
ISSN - 1932-6203
DOI - 10.1371/journal.pone.0217408
Subject(s) - reinforcement learning , computer science , oracle , machine learning , parameterized complexity , artificial intelligence , cross entropy , entropy (arrow of time) , active learning (machine learning) , artificial neural network , principle of maximum entropy , algorithm , physics , software engineering , quantum mechanics
As a promising research direction in recent decades, active learning allows an oracle to assign labels to typical examples for performance improvement in learning systems. Existing works mainly focus on designing criteria for screening examples of high value to be labeled in a handcrafted manner. Instead of manually developing strategies of querying the user to access labels for the desired examples, we utilized the reinforcement learning algorithm parameterized with the neural network to automatically explore query strategies in active learning when addressing stream-based one-shot classification problems. With the involvement of cross-entropy in the loss function of Q-learning, an efficient policy to decide when and where to predict or query an instance is learned through the developed framework. Compared with a former influential work, the advantages of our method are demonstrated experimentally with two image classification tasks, and it exhibited better performance, quick convergence, relatively good stability and fewer requests for labels.