Feb. 22, 2024, 5:48 a.m. | Philippe Laban, Lidiya Murakhovs'ka, Caiming Xiong, Chien-Sheng Wu

cs.CL updates on arXiv.org arxiv.org

arXiv:2311.08596v2 Announce Type: replace
Abstract: The interactive nature of Large Language Models (LLMs) theoretically allows models to refine and improve their answers, yet systematic analysis of the multi-turn behavior of LLMs remains limited. In this paper, we propose the FlipFlop experiment: in the first round of the conversation, an LLM completes a classification task. In a second round, the LLM is challenged with a follow-up phrase like "Are you sure?", offering an opportunity for the model to reflect on its …

abstract analysis arxiv behavior cs.cl experiment interactive language language models large language large language models leads llms nature paper performance refine type

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote