Sept. 26, 2022, 1:11 a.m. | Risto Vuorio, Jacob Beck, Shimon Whiteson, Jakob Foerster, Gregory Farquhar

cs.LG updates on arXiv.org arxiv.org

Meta-gradients provide a general approach for optimizing the meta-parameters
of reinforcement learning (RL) algorithms. Estimation of meta-gradients is
central to the performance of these meta-algorithms, and has been studied in
the setting of MAML-style short-horizon meta-RL problems. In this context,
prior work has investigated the estimation of the Hessian of the RL objective,
as well as tackling the problem of credit assignment to pre-adaptation behavior
by making a sampling correction. However, we show that Hessian estimation,
implemented for example by …

arxiv bias bias-variance investigation meta variance

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

AI Engineering Manager

@ M47 Labs | Barcelona, Catalunya [Cataluña], Spain