TY - GEN
T1 - Reinforcement Learning with Prototypical Representations
AU - Yarats, Denis
AU - Fergus, Rob
AU - Lazaric, Alessandro
AU - Pinto, Lerrel
N1 - Funding Information:
This research is supported in part by DARPA through the Machine Common Sense Program. We thank Oleh Rybkin and Danijar Hafner for sharing data for baselines. We also thank Hao Liu and anonymous reviewers for insightful feedback that helped to improve our paper.
Publisher Copyright:
Copyright © 2021 by the author(s)
PY - 2021
Y1 - 2021
N2 - Learning effective representations in image-based environments is crucial for sample efficient Reinforcement Learning (RL). Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent - learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations. Furthermore, we would like to learn representations that not only generalize across tasks but also accelerate downstream exploration for efficient task-specific training. To address these challenges we propose Proto-RL, a self-supervised framework that ties representation learning with exploration through prototypical representations. These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations. We pre-train these task-agnostic representations and prototypes on environments without downstream task information. This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks. We open-source our code at https://github.com/denisyarats/proto.
AB - Learning effective representations in image-based environments is crucial for sample efficient Reinforcement Learning (RL). Unfortunately, in RL, representation learning is confounded with the exploratory experience of the agent - learning a useful representation requires diverse data, while effective exploration is only possible with coherent representations. Furthermore, we would like to learn representations that not only generalize across tasks but also accelerate downstream exploration for efficient task-specific training. To address these challenges we propose Proto-RL, a self-supervised framework that ties representation learning with exploration through prototypical representations. These prototypes simultaneously serve as a summarization of the exploratory experience of an agent as well as a basis for representing observations. We pre-train these task-agnostic representations and prototypes on environments without downstream task information. This enables state-of-the-art downstream policy learning on a set of difficult continuous control tasks. We open-source our code at https://github.com/denisyarats/proto.
UR - http://www.scopus.com/inward/record.url?scp=85130932881&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85130932881&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85130932881
T3 - Proceedings of Machine Learning Research
SP - 11920
EP - 11931
BT - Proceedings of the 38th International Conference on Machine Learning, ICML 2021
PB - ML Research Press
T2 - 38th International Conference on Machine Learning, ICML 2021
Y2 - 18 July 2021 through 24 July 2021
ER -