Authors
Joshua GA Cashaback, Christopher K Lao, Dimitrios J Palidis, Susan K Coltman, Heather R McGregor, Paul L Gribble
Publication date
2019/3/4
Journal
PLOS computational biology
Volume
15
Issue
3
Pages
e1006839
Publisher
Public Library of Science
Description
Consideration of previous successes and failures is essential to mastering a motor skill. Much of what we know about how humans and animals learn from such reinforcement feedback comes from experiments that involve sampling from a small number of discrete actions. Yet, it is less understood how we learn through reinforcement feedback when sampling from a continuous set of possible actions. Navigating a continuous set of possible actions likely requires using gradient information to maximize success. Here we addressed how humans adapt the aim of their hand when experiencing reinforcement feedback that was associated with a continuous set of possible actions. Specifically, we manipulated the change in the probability of reward given a change in motor action—the reinforcement gradient—to study its influence on learning. We found that participants learned faster when exposed to a steep gradient compared to a shallow gradient. Further, when initially positioned between a steep and a shallow gradient that rose in opposite directions, participants were more likely to ascend the steep gradient. We introduce a model that captures our results and several features of motor learning. Taken together, our work suggests that the sensorimotor system relies on temporally recent and spatially local gradient information to drive learning.
Total citations
2019202020212022202320242457127
Scholar articles
JGA Cashaback, CK Lao, DJ Palidis, SK Coltman… - PLOS computational biology, 2019