all AI news
Learning From Failure: Integrating Negative Examples when Fine-tuning Large Language Models as Agents
Feb. 20, 2024, 5:51 a.m. | Renxi Wang, Haonan Li, Xudong Han, Yixuan Zhang, Timothy Baldwin
cs.CL updates on arXiv.org arxiv.org
Abstract: Large language models (LLMs) have achieved success in acting as agents, which interact with environments through tools like search engines. However, LLMs are not optimized specifically for tool use during training or alignment, limiting their effectiveness as agents. To resolve this problem, previous work has collected interaction trajectories between GPT-4 and environments, and fine-tuned smaller models with them. As part of this, the standard approach has been to simply discard trajectories that do not finish …
abstract acting agents alignment arxiv cs.cl environments examples failure fine-tuning language language models large language large language models llms negative search success through tool tools training type
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
Artificial Intelligence – Bioinformatic Expert
@ University of Texas Medical Branch | Galveston, TX
Lead Developer (AI)
@ Cere Network | San Francisco, US
Research Engineer
@ Allora Labs | Remote
Ecosystem Manager
@ Allora Labs | Remote
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US