Sept. 14, 2022, 1:11 a.m. | Jelena Luketina, Sebastian Flennerhag, Yannick Schroecker, David Abel, Tom Zahavy, Satinder Singh

cs.LG updates on arXiv.org arxiv.org

Meta-gradient methods (Xu et al., 2018; Zahavy et al., 2020) offer a
promising solution to the problem of hyperparameter selection and adaptation in
non-stationary reinforcement learning problems. However, the properties of
meta-gradients in such environments have not been systematically studied. In
this work, we bring new clarity to meta-gradients in non-stationary
environments. Concretely, we ask: (i) how much information should be given to
the learned optimizers, so as to enable faster adaptation and generalization
over a lifetime, (ii) what meta-optimizer …

arxiv environments meta

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Engineer, Deep Learning

@ Outrider | Remote

Data Analyst (Bangkok based, relocation provided)

@ Agoda | Bangkok (Central World Office)

Data Scientist II

@ MoEngage | Bengaluru

Machine Learning Engineer

@ Sika AG | Welwyn Garden City, United Kingdom