April 8, 2024, 4:47 a.m. | Haotong Yang, Fanxu Meng, Zhouchen Lin, Muhan Zhang

cs.CL updates on arXiv.org arxiv.org

arXiv:2310.05452v2 Announce Type: replace-cross
Abstract: The pre-trained large language models (LLMs) have shown their extraordinary capacity to solve reasoning tasks, even on tasks that require a complex process involving multiple sub-steps. However, given the vast possible generation space of all the tasks, how the pretrained model learns the reasoning ability remains an open question. We firstly propose that an intrinsic structural constraint on the generated sequence of language-based reasoning -- we called it template-content structure (T-C structure) -- is the …

abstract arxiv capacity cs.ai cs.cl however language language models large language large language models llms mind multiple parrot process reasoning solve space tasks template type vast

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York