all AI news
Safe Reinforcement Learning with Free-form Natural Language Constraints and Pre-Trained Language Models
April 22, 2024, 4:43 a.m. | Xingzhou Lou, Junge Zhang, Ziyan Wang, Kaiqi Huang, Yali Du
cs.LG updates on arXiv.org arxiv.org
Abstract: Safe reinforcement learning (RL) agents accomplish given tasks while adhering to specific constraints. Employing constraints expressed via easily-understandable human language offers considerable potential for real-world applications due to its accessibility and non-reliance on domain expertise. Previous safe RL methods with natural language constraints typically adopt a recurrent neural network, which leads to limited capabilities when dealing with various forms of human language input. Furthermore, these methods often require a ground-truth cost function, necessitating domain expertise …
abstract accessibility agents applications arxiv constraints cs.cl cs.lg domain expertise form free human language language models natural natural language reinforcement reinforcement learning reliance safe tasks type via world
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Data Engineer (m/f/d)
@ Project A Ventures | Berlin, Germany
Principle Research Scientist
@ Analog Devices | US, MA, Boston