Web: http://arxiv.org/abs/2105.14111

June 24, 2022, 1:11 a.m. | Lauro Langosco, Jack Koch, Lee Sharkey, Jacob Pfau, Laurent Orseau, David Krueger

cs.LG updates on arXiv.org arxiv.org

We study goal misgeneralization, a type of out-of-distribution generalization
failure in reinforcement learning (RL). Goal misgeneralization failures occur
when an RL agent retains its capabilities out-of-distribution yet pursues the
wrong goal. For instance, an agent might continue to competently avoid
obstacles, but navigate to the wrong place. In contrast, previous works have
typically focused on capability generalization failures, where an agent fails
to do anything sensible at test time. We formalize this distinction between
capability and goal generalization, provide the …

arxiv deep learning lg reinforcement reinforcement learning

More from arxiv.org / cs.LG updates on arXiv.org

Machine Learning Researcher - Saalfeld Lab

@ Howard Hughes Medical Institute - Chevy Chase, MD | Ashburn, Virginia

Project Director, Machine Learning in US Health

@ ideas42.org | Remote, US

Data Science Intern

@ NannyML | Remote

Machine Learning Engineer NLP/Speech

@ Play.ht | Remote

Research Scientist, 3D Reconstruction

@ Yembo | Remote, US

Clinical Assistant or Associate Professor of Management Science and Systems

@ University at Buffalo | Buffalo, NY