Nov. 21, 2022, 2:12 a.m. | Aditya Ramesh, Louis Kirsch, Sjoerd van Steenkiste, Jürgen Schmidhuber

cs.LG updates on arXiv.org arxiv.org

Efficient exploration in reinforcement learning is a challenging problem
commonly addressed through intrinsic rewards. Recent prominent approaches are
based on state novelty or variants of artificial curiosity. However, directly
applying them to partially observable environments can be ineffective and lead
to premature dissipation of intrinsic rewards. Here we propose random curiosity
with general value functions (RC-GVF), a novel intrinsic reward function that
draws upon connections between these distinct approaches. Instead of using only
the current observation's novelty or a curiosity …

arxiv curiosity general random value

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote