Feb. 6, 2024, 5:46 a.m. | Fangru Lin Emanuele La Malfa Valentin Hofmann Elle Michelle Yang Anthony Cohn Janet B. Pierrehumbert

cs.LG updates on arXiv.org arxiv.org

Reasoning about asynchronous plans is challenging since it requires sequential and parallel planning to optimize time costs. Can large language models (LLMs) succeed at this task? Here, we present the first large-scale study investigating this question. We find that a representative set of closed and open-source LLMs, including GPT-4 and LLaMA-2, behave poorly when not supplied with illustrations about the task-solving process in our benchmark AsyncHow. We propose a novel technique called Plan Like a Graph (PLaG) that combines graphs …

asynchronous costs cs.ai cs.cl cs.lg gpt gpt-4 graph language language models large language large language models llama llms planning question reasoning scale set study

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Data Engineer (m/f/d)

@ Project A Ventures | Berlin, Germany

Principle Research Scientist

@ Analog Devices | US, MA, Boston