Meta-Learning via Learned Loss
Sarah Bechtle
and
Artem Molchanov
and
Yevgen Chebotar
and
Edward Grefenstette
and
Ludovic Righetti
and
Gaurav Sukhatme
and
Franziska Meier
arXiv e-Print archive - 2019 via Local arXiv
Keywords:
cs.LG, cs.AI, cs.RO, stat.ML
First published: 2019/06/12 (5 years ago) Abstract: Typically, loss functions, regularization mechanisms and other important
aspects of training parametric models are chosen heuristically from a limited
set of options. In this paper, we take the first step towards automating this
process, with the view of producing models which train faster and more
robustly. Concretely, we present a meta-learning method for learning parametric
loss functions that can generalize across different tasks and model
architectures. We develop a pipeline for meta-training such loss functions,
targeted at maximizing the performance of the model trained under them. The
loss landscape produced by our learned losses significantly improves upon the
original task-specific losses in both supervised and reinforcement learning
tasks. Furthermore, we show that our meta-learning framework is flexible enough
to incorporate additional information at meta-train time. This information
shapes the learned loss function such that the environment does not need to
provide this information during meta-test time.