Premium
Learning to Be (In)variant: Combining Prior Knowledge and Experience to Infer Orientation Invariance in Object Recognition
Author(s) -
Austerweil Joseph L.,
Griffiths Thomas L.,
Palmer Stephen E.
Publication year - 2017
Publication title -
cognitive science
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.498
H-Index - 114
eISSN - 1551-6709
pISSN - 0364-0213
DOI - 10.1111/cogs.12466
Subject(s) - observer (physics) , invariant (physics) , orientation (vector space) , artificial intelligence , cognitive neuroscience of visual object recognition , object (grammar) , computer science , computer vision , transformation (genetics) , set (abstract data type) , object orientation , pattern recognition (psychology) , cognitive psychology , psychology , mathematics , object oriented programming , geometry , biochemistry , chemistry , physics , quantum mechanics , mathematical physics , gene , programming language
How does the visual system recognize images of a novel object after a single observation despite possible variations in the viewpoint of that object relative to the observer? One possibility is comparing the image with a prototype for invariance over a relevant transformation set (e.g., translations and dilations). However, invariance over rotations (i.e., orientation invariance) has proven difficult to analyze, because it applies to some objects but not others. We propose that the invariant transformations of an object are learned by incorporating prior expectations with real‐world evidence. We test this proposal by developing an ideal learner model for learning invariance that predicts better learning of orientation dependence when prior expectations about orientation are weak. This prediction was supported in two behavioral experiments, where participants learned the orientation dependence of novel images using feedback from solving arithmetic problems.