all AI news
Chinese Open Instruction Generalist: A Preliminary Release. (arXiv:2304.07987v3 [cs.CL] UPDATED)
cs.CL updates on arXiv.org arxiv.org
Instruction tuning is widely recognized as a key technique for building
generalist language models, which has attracted the attention of researchers
and the public with the release of InstructGPT~\citep{ouyang2022training} and
ChatGPT\footnote{\url{https://chat.openai.com/}}. Despite impressive progress
in English-oriented large-scale language models (LLMs), it is still
under-explored whether English-based foundation LLMs can perform similarly on
multilingual tasks compared to English tasks with well-designed instruction
tuning and how we can construct the corpora needed for the tuning. To remedy
this gap, we propose the …
arxiv attention building chinese construct dataset english foundation gap instructgpt language language models llms multilingual project public release researchers scale