March 26, 2024, 4:42 a.m. | Chanwoo Park, Xiangyu Liu, Asuman Ozdaglar, Kaiqing Zhang

cs.LG updates on arXiv.org arxiv.org

arXiv:2403.16843v1 Announce Type: new
Abstract: Large language models (LLMs) have been increasingly employed for (interactive) decision-making, via the development of LLM-based autonomous agents. Despite their emerging successes, the performance of LLM agents in decision-making has not been fully investigated through quantitative metrics, especially in the multi-agent setting when they interact with each other, a typical scenario in real-world LLM-agent applications. To better understand the limits of LLM agents in these interactive environments, we propose to study their interactions in benchmark …

abstract agent agents arxiv autonomous autonomous agents case case study cs.ai cs.gt cs.lg decision development games interactive language language models large language large language models llm llms making metrics multi-agent online learning performance quantitative study through type via

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US