all AI news
Adaptive Cross-lingual Text Classification through In-Context One-Shot Demonstrations
April 4, 2024, 4:47 a.m. | Emilio Villa-Cueva, A. Pastor L\'opez-Monroy, Fernando S\'anchez-Vega, Thamar Solorio
cs.CL updates on arXiv.org arxiv.org
Abstract: Zero-Shot Cross-lingual Transfer (ZS-XLT) utilizes a model trained in a source language to make predictions in another language, often with a performance loss. To alleviate this, additional improvements can be achieved through subsequent adaptation using examples in the target language. In this paper, we exploit In-Context Tuning (ICT) for One-Shot Cross-lingual transfer in the classification task by introducing In-Context Cross-lingual Transfer (IC-XLT). The novel concept involves training a model to learn from context examples and …
abstract arxiv classification context cross-lingual cs.cl examples exploit improvements language loss paper performance predictions text text classification through transfer type zero-shot
More from arxiv.org / cs.CL updates on arXiv.org
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
Senior Machine Learning Engineer
@ Samsara | Canada - Remote