March 5, 2024, 2:43 p.m. | Chenyang Cao, Zichen Yan, Renhao Lu, Junbo Tan, Xueqian Wang

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.01734v1 Announce Type: cross
Abstract: Offline goal-conditioned reinforcement learning (GCRL) aims at solving goal-reaching tasks with sparse rewards from an offline dataset. While prior work has demonstrated various approaches for agents to learn near-optimal policies, these methods encounter limitations when dealing with diverse constraints in complex environments, such as safety constraints. Some of these approaches prioritize goal attainment without considering safety, while others excessively focus on safety at the expense of training efficiency. In this paper, we study the problem …

arxiv cs.ai cs.lg cs.ro offline policy recovery reinforcement reinforcement learning safety safety-critical tasks type

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Lead Data Modeler

@ Sherwin-Williams | Cleveland, OH, United States