Feb. 7, 2024, 5:41 a.m. | Samuel Garcin James Doran Shangmin Guo Christopher G. Lucas Stefano V. Albrecht

cs.LG updates on arXiv.org arxiv.org

Autonomous agents trained using deep reinforcement learning (RL) often lack the ability to successfully generalise to new environments, even when they share characteristics with the environments they have encountered during training. In this work, we investigate how the sampling of individual environment instances, or levels, affects the zero-shot generalisation (ZSG) ability of RL agents. We discover that, for deep actor-critic architectures sharing their base layers, prioritising levels according to their value loss minimises the mutual information between the agent's internal …

agents autonomous autonomous agents context cs.ai cs.lg design environment environments instances reinforcement reinforcement learning sampling training transfer via work zero-shot

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote