all AI news
Sample and Communication Efficient Fully Decentralized MARL Policy Evaluation via a New Approach: Local TD update
March 26, 2024, 4:41 a.m. | Fnu Hairi, Zifan Zhang, Jia Liu
cs.LG updates on arXiv.org arxiv.org
Abstract: In actor-critic framework for fully decentralized multi-agent reinforcement learning (MARL), one of the key components is the MARL policy evaluation (PE) problem, where a set of $N$ agents work cooperatively to evaluate the value function of the global states for a given policy through communicating with their neighbors. In MARL-PE, a critical challenge is how to lower the sample and communication complexities, which are defined as the number of training samples and communication rounds needed …
abstract actor actor-critic agent agents arxiv communication components cs.lg decentralized evaluation framework function global key multi-agent policy reinforcement reinforcement learning sample set the key type update value via work
More from arxiv.org / cs.LG updates on arXiv.org
Jobs in AI, ML, Big Data
Software Engineer for AI Training Data (School Specific)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Python)
@ G2i Inc | Remote
Software Engineer for AI Training Data (Tier 2)
@ G2i Inc | Remote
Data Engineer
@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US