Feb. 6, 2024, 5:47 a.m. | Duong Minh Le Yang Chen Alan Ritter Wei Xu

cs.LG updates on arXiv.org arxiv.org

Zero-shot cross-lingual transfer utilizing multilingual LLMs has become a popular learning paradigm for low-resource languages with no labeled training data. However, for NLP tasks that involve fine-grained predictions on words and phrases, the performance of zero-shot cross-lingual transfer learning lags far behind supervised fine-tuning methods. Therefore, it is common to exploit translation and label projection to further improve the performance by (1) translating training data that is available in a high-resource language (e.g., English) together with the gold labels into …

become cross-lingual cs.cl cs.lg data decoding exploit fine-grained fine-tuning languages llms low multilingual nlp paradigm performance popular predictions projection supervised fine-tuning tasks training training data transfer transfer learning translation words zero-shot

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US

Research Engineer

@ Allora Labs | Remote

Ecosystem Manager

@ Allora Labs | Remote

Founding AI Engineer, Agents

@ Occam AI | New York