Premium
Constructive reinforcement learning
Author(s) -
HernandezOrallo Jose
Publication year - 2000
Publication title -
international journal of intelligent systems
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.291
H-Index - 87
eISSN - 1098-111X
pISSN - 0884-8173
DOI - 10.1002/(sici)1098-111x(200003)15:3<241::aid-int6>3.0.co;2-z
Subject(s) - reinforcement learning , computer science , constructive , artificial intelligence , relation (database) , measure (data warehouse) , representation (politics) , ontology , space (punctuation) , machine learning , reinforcement , data mining , psychology , social psychology , philosophy , epistemology , politics , political science , law , operating system , process (computing)
This paper presents an operative measure of reinforcement for constructive learning methods, i.e., eager learning methods using highly expressible (or universal) representation languages. These evaluation tools allow a further insight in the study of the growth of knowledge, theory revision, and abduction. The final approach is based on an apportionment of credit wrt the “course” that the evidence makes through the learned theory. Our measure of reinforcement is shown to be justified by cross‐validation and by the connection with other successful evaluation criteria, like the minimum description length principle. Finally, the relation with the classical view of reinforcement is studied, where the actions of an intelligent system can be rewarded or penalized, and we discuss whether this should affect our distribution of reinforcement. The most important result of this paper is that the way we distribute reinforcement into knowledge results in a rated ontology, instead of a single prior distribution. Therefore, this detailed information can be exploited for guiding the space search of inductive learning algorithms. Likewise, knowledge revision may be done to the part of the theory which is not justified by the evidence. ©2000 John Wiley & Sons, Inc.