TY - GEN
T1 - Meta learning via learned loss
AU - Bechtle, Sarah
AU - Molchanov, Artem
AU - Chebotar, Yevgen
AU - Grefenstette, Edward
AU - Righetti, Ludovic
AU - Sukhatme, Gaurav
AU - Meier, Franziska
N1 - Publisher Copyright:
© 2020 IEEE
PY - 2020
Y1 - 2020
N2 - Typically, loss functions, regularization mechanisms and other important aspects of training parametric models are chosen heuristically from a limited set of options. In this paper, we take the first step towards automating this process, with the view of producing models which train faster and more robustly. Concretely, we present a meta-learning method for learning parametric loss functions that can generalize across different tasks and model architectures. We develop a pipeline for “meta-training” such loss functions, targeted at maximizing the performance of the model trained under them. The loss landscape produced by our learned losses significantly improves upon the original task-specific losses in both supervised and reinforcement learning tasks. Furthermore, we show that our meta-learning framework is flexible enough to incorporate additional information at meta-train time. This information shapes the learned loss function such that the environment does not need to provide this information during meta-test time.
AB - Typically, loss functions, regularization mechanisms and other important aspects of training parametric models are chosen heuristically from a limited set of options. In this paper, we take the first step towards automating this process, with the view of producing models which train faster and more robustly. Concretely, we present a meta-learning method for learning parametric loss functions that can generalize across different tasks and model architectures. We develop a pipeline for “meta-training” such loss functions, targeted at maximizing the performance of the model trained under them. The loss landscape produced by our learned losses significantly improves upon the original task-specific losses in both supervised and reinforcement learning tasks. Furthermore, we show that our meta-learning framework is flexible enough to incorporate additional information at meta-train time. This information shapes the learned loss function such that the environment does not need to provide this information during meta-test time.
KW - Deep learning
KW - Meta learning
KW - Optimization
KW - Reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85108532798&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85108532798&partnerID=8YFLogxK
U2 - 10.1109/ICPR48806.2021.9412010
DO - 10.1109/ICPR48806.2021.9412010
M3 - Conference contribution
AN - SCOPUS:85108532798
T3 - Proceedings - International Conference on Pattern Recognition
SP - 4161
EP - 4168
BT - Proceedings of ICPR 2020 - 25th International Conference on Pattern Recognition
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th International Conference on Pattern Recognition, ICPR 2020
Y2 - 10 January 2021 through 15 January 2021
ER -