March 5, 2024, 2:44 p.m. | Preston Rozwood, Edward Mehrez, Ludger Paehler, Wen Sun, Steven L. Brunton

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.02290v1 Announce Type: cross
Abstract: The Bellman equation and its continuous form, the Hamilton-Jacobi-Bellman (HJB) equation, are ubiquitous in reinforcement learning (RL) and control theory. However, these equations quickly become intractable for systems with high-dimensional states and nonlinearity. This paper explores the connection between the data-driven Koopman operator and Markov Decision Processes (MDPs), resulting in the development of two new RL algorithms to address these limitations. We leverage Koopman operator techniques to lift a nonlinear system into new coordinates where …

abstract arxiv become continuous control cs.ai cs.lg data data-driven decision development equation form hamilton markov math.ds math.oc paper processes reinforcement reinforcement learning systems theory type

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote