April 8, 2024, 4:47 a.m. | Haotong Yang, Fanxu Meng, Zhouchen Lin, Muhan Zhang

cs.CL updates on arXiv.org arxiv.org

arXiv:2310.05452v2 Announce Type: replace-cross
Abstract: The pre-trained large language models (LLMs) have shown their extraordinary capacity to solve reasoning tasks, even on tasks that require a complex process involving multiple sub-steps. However, given the vast possible generation space of all the tasks, how the pretrained model learns the reasoning ability remains an open question. We firstly propose that an intrinsic structural constraint on the generated sequence of language-based reasoning -- we called it template-content structure (T-C structure) -- is the …

abstract arxiv capacity cs.ai cs.cl however language language models large language large language models llms mind multiple parrot process reasoning solve space tasks template type vast

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Analyst

@ Alstom | Johannesburg, GT, ZA