z-logo
open-access-imgOpen Access
Moratorium Effect on Estimation Values in Simple Reinforcement Learning
Author(s) -
Akira Notsu,
Yuki Tezuka,
Katsuhiro Honda
Publication year - 2013
Publication title -
international journal of computer science and artificial intelligence
Language(s) - English
Resource type - Journals
eISSN - 2226-4469
pISSN - 2226-4450
DOI - 10.5963/ijcsai0303004
Subject(s) - reinforcement learning , simple (philosophy) , reinforcement , estimation , computer science , artificial intelligence , psychology , econometrics , statistics , machine learning , cognitive psychology , mathematics , social psychology , economics , epistemology , philosophy , management
In this paper, we introduce low-priority cut-in (moratorium) to chain form reinforcement learning, which we proposed as Simple Reinforcement Learning for a reinforcement learning agent that has small memory. In the real world, learning is difficult because there are an infinite number of states and actions that need a large number of stored memory and learning time. To solve the problem, better estimated values are categorized as "GOOD" in the reinforcement learning process. Additionally, the alignment sequence of estimated values is changed because they are regarded as an important sequence themselves. However, the method is heavily affected by the action policy. If an agent tends to search many states, its memory overflows with low-value data. Thus, low-priority cut-in (moratorium) enhances the method in order to solve this problem. We conducted some simulations and observed the influence of our methods. Several simulation results show good influence on learning.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom