z-logo
Premium
Neural principles of memory and a neural theory of analogical insight
Author(s) -
Lawson David I.,
Lawson Anton E.
Publication year - 1993
Publication title -
journal of research in science teaching
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 3.067
H-Index - 131
eISSN - 1098-2736
pISSN - 0022-4308
DOI - 10.1002/tea.3660301012
Subject(s) - artificial neural network , recall , set (abstract data type) , short term memory , computer science , information storage , sensory system , cognition , neuroscience , cognitive science , artificial intelligence , psychology , cognitive psychology , working memory , database , programming language
Grossberg's principles of neural modeling are reviewed and extended to provide a neural level theory to explain how analogies greatly increase the rate of learning and can, in fact, make learning and retention possible. In terms of memory, the key point is that the mind is able to recognize and recall when it is able to match sensory input from new objects, events, or situations with past memory records of “similar” objects, events, or situations. When a match occurs, an adaptive resonance is set up in which the synaptic strengths of neurons are increased; thus a long term record of the new input is formed in memory. Systems of neurons called outstars and instars are presumably the underlying units that enable this to occur. Analogies can greatly facilitate learning and retention because they activate the outstars (i.e., the cells that are sampling the to‐be‐learned pattern) and cause the neural activity to grow exponentially by forming feedback loops. This increased activity insures the boost in synaptic strengths of neurons, thus causing storage and retention in long‐term memory (i.e., learning).

This content is not available in your region!

Continue researching here.

Having issues? You can contact us here