April 22, 2024, 4:43 a.m. | Xingzhou Lou, Junge Zhang, Ziyan Wang, Kaiqi Huang, Yali Du

cs.LG updates on arXiv.org arxiv.org

arXiv:2401.07553v2 Announce Type: replace
Abstract: Safe reinforcement learning (RL) agents accomplish given tasks while adhering to specific constraints. Employing constraints expressed via easily-understandable human language offers considerable potential for real-world applications due to its accessibility and non-reliance on domain expertise. Previous safe RL methods with natural language constraints typically adopt a recurrent neural network, which leads to limited capabilities when dealing with various forms of human language input. Furthermore, these methods often require a ground-truth cost function, necessitating domain expertise …

abstract accessibility agents applications arxiv constraints cs.cl cs.lg domain expertise form free human language language models natural natural language reinforcement reinforcement learning reliance safe tasks type via world

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US