Premium
Pruning representations in a distributed model of working memory: a mechanism for refreshing and removal?
Author(s) -
Shepherdson Peter,
Oberauer Klaus
Publication year - 2018
Publication title -
annals of the new york academy of sciences
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.712
H-Index - 248
eISSN - 1749-6632
pISSN - 0077-8923
DOI - 10.1111/nyas.13659
Subject(s) - pruning , computer science , recall , working memory , mechanism (biology) , cognitive psychology , lossless compression , artificial intelligence , machine learning , psychology , cognition , neuroscience , philosophy , epistemology , data compression , agronomy , biology
Substantial behavioral evidence suggests that attention plays an important role in working memory. Frequently, attention is characterized as enhancing representations by increasing their strength or activation level. Despite the intuitive appeal of this idea, using attention to strengthen representations in computational models can lead to unexpected outcomes. Representational strengthening frequently leads to worse, rather than better, performance, contradicting behavioral results. Here, we propose an alternative to a pure strengthening account, in which attention is used to selectively strengthen useful and weaken less useful components of distributed memory representations, thereby pruning the representations. We use a simple sampling algorithm to implement this pruning mechanism in a computational model of working memory. Our simulations show that pruning representations in this manner leads to improvements in performance compared with a lossless (i.e., decay‐free) baseline condition, for both discrete recall (e.g., of a list of words) and continuous reproduction (e.g., of an array of colors). Pruning also offers a potential explanation of why a retro‐cue drawing attention to one memory item during the retention interval improves performance. These results indicate that a pruning mechanism could provide a viable alternative to pure strengthening accounts of attention to representations in working memory.