z-logo
open-access-imgOpen Access
Uncertainty and exploration.
Author(s) -
Samuel J. Gershman
Publication year - 2018
Publication title -
decision
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 0.656
H-Index - 16
eISSN - 2325-9973
pISSN - 2325-9965
DOI - 10.1037/dec0000101
Subject(s) - exploit , computer science , action (physics) , computation , dilemma , scale (ratio) , uncertainty quantification , mathematical optimization , machine learning , artificial intelligence , algorithm , mathematics , physics , geometry , computer security , quantum mechanics
In order to discover the most rewarding actions, agents must collect information about their environment, potentially foregoing reward. The optimal solution to this "explore-exploit" dilemma is often computationally challenging, but principled algorithmic approximations exist. These approximations utilize uncertainty about action values in different ways. Some random exploration algorithms scale the level of choice stochasticity with the level of uncertainty. Other directed exploration algorithms add a "bonus" to action values with high uncertainty. Random exploration algorithms are sensitive to total uncertainty across actions, whereas directed exploration algorithms are sensitive to relative uncertainty. This paper reports a multi-armed bandit experiment in which total and relative uncertainty were orthogonally manipulated. We found that humans employ both exploration strategies, and that these strategies are independently controlled by different uncertainty computations.

The content you want is available to Zendy users.

Already have an account? Click here to sign in.
Having issues? You can contact us here
Accelerating Research

Address

John Eccles House
Robert Robinson Avenue,
Oxford Science Park, Oxford
OX4 4GP, United Kingdom