all AI news
Eliciting the Translation Ability of Large Language Models via Multilingual Finetuning with Translation Instructions
March 15, 2024, 4:48 a.m. | Jiahuan Li, Hao Zhou, Shujian Huang, Shanbo Cheng, Jiajun Chen
cs.CL updates on arXiv.org arxiv.org
Abstract: Large-scale Pretrained Language Models (LLMs), such as ChatGPT and GPT4, have shown strong abilities in multilingual translations, without being explicitly trained on parallel corpora. It is interesting how the LLMs obtain their ability to carry out translation instructions for different languages. In this paper, we present a detailed analysis by finetuning a multilingual pretrained language model, XGLM-7B, to perform multilingual translation following given instructions. Firstly, we show that multilingual LLMs have stronger translation abilities than …
abstract arxiv chatgpt cs.cl finetuning gpt4 language language models languages large language large language models llms multilingual scale translation type via
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Data Analyst (Digital Business Analyst)
@ Activate Interactive Pte Ltd | Singapore, Central Singapore, Singapore