
Content Classification Tasks with Data Preprocessing Manifestations
Author(s) -
Mamoona Anam,
Kantilal Pitambar Rane,
Ali Alenezi,
Ruby Mishra,
Swaminathan Ramamurthy,
Ferdin Joe John Joseph
Publication year - 2022
Publication title -
webology
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.259
H-Index - 18
ISSN - 1735-188X
DOI - 10.14704/web/v19i1/web19094
Subject(s) - computer science , task (project management) , preprocessor , artificial intelligence , encoder , representation (politics) , reinforcement learning , machine learning , external data representation , labeled data , management , politics , political science , law , economics , operating system
Deep reinforcement learning has a major hurdle in terms of data efficiency. We solve this challenge by pretraining an encoder with unlabeled input, which is subsequently finetuned on a tiny quantity of task-specific input. We use a mixture of latent dynamics modelling and unsupervised goal-conditioned RL to encourage learning representations that capture various elements of the underlying MDP. Our approach significantly outperforms previous work combining offline representation pretraining with task-specific finetuning when limited to 100k steps of interaction on Atari games (equivalent to two hours of human experience) and compares favourably with other pretraining methods that require orders of magnitude more data. When paired with larger models and more diverse, task-aligned observational data, our methodology shows great promise, nearing human-level performance and data efficiency on Atari in the best-case scenario.