TY - JOUR
T1 - Understanding the development of reward learning through the lens of meta-learning
AU - Nussenbaum, Kate
AU - Hartley, Catherine A.
N1 - Publisher Copyright:
© Springer Nature America, Inc. 2024.
PY - 2024/6
Y1 - 2024/6
N2 - Determining how environments shape how people learn is central to understanding individual differences in goal-directed behaviour. Studies of the effects of early-life adversity on reward learning have revealed that the environments that infants and children experience exert lasting influences on reward-guided behaviour. However, the varied findings from this research are difficult to reconcile under a unified computational account. Studies of adaptive reinforcement learning have demonstrated that learning algorithms and parameters dynamically adapt to support reward-guided behaviour in varied contexts, but this body of research has largely focused on learning that proceeds within the short timeframes of experimental tasks. In this Perspective, we argue that, to understand how the structure of experienced environments shapes reward learning across development, computational accounts of the effects of environmental statistics on reinforcement learning need to be extended to encompass learning across multiple nested timescales of experience. To this end, we consider the development of reward learning through the lens of meta-learning models, in particular meta-reinforcement learning. This computational formalization can inspire new hypotheses and methods for empirical research to understand how features of experienced environments give rise to individual differences in learning and adaptive behaviour across development.
AB - Determining how environments shape how people learn is central to understanding individual differences in goal-directed behaviour. Studies of the effects of early-life adversity on reward learning have revealed that the environments that infants and children experience exert lasting influences on reward-guided behaviour. However, the varied findings from this research are difficult to reconcile under a unified computational account. Studies of adaptive reinforcement learning have demonstrated that learning algorithms and parameters dynamically adapt to support reward-guided behaviour in varied contexts, but this body of research has largely focused on learning that proceeds within the short timeframes of experimental tasks. In this Perspective, we argue that, to understand how the structure of experienced environments shapes reward learning across development, computational accounts of the effects of environmental statistics on reinforcement learning need to be extended to encompass learning across multiple nested timescales of experience. To this end, we consider the development of reward learning through the lens of meta-learning models, in particular meta-reinforcement learning. This computational formalization can inspire new hypotheses and methods for empirical research to understand how features of experienced environments give rise to individual differences in learning and adaptive behaviour across development.
UR - http://www.scopus.com/inward/record.url?scp=85190813187&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85190813187&partnerID=8YFLogxK
U2 - 10.1038/s44159-024-00304-1
DO - 10.1038/s44159-024-00304-1
M3 - Article
AN - SCOPUS:85190813187
SN - 2731-0574
VL - 3
SP - 424
EP - 438
JO - Nature Reviews Psychology
JF - Nature Reviews Psychology
IS - 6
ER -