May 22, 2024, 4:47 a.m. | Rochelle Choenni, Dan Garrette, Ekaterina Shutova

cs.CL updates on arXiv.org arxiv.org

arXiv:2305.13286v2 Announce Type: replace
Abstract: Multilingual large language models (MLLMs) are jointly trained on data from many different languages such that representation of individual languages can benefit from other languages' data. Impressive performance on zero-shot cross-lingual transfer shows that these models are capable of exploiting data from other languages. Yet, it remains unclear to what extent, and under which conditions, languages rely on each other's data. In this study, we use TracIn (Pruthi et al., 2020), a training data attribution …

abstract arxiv benefit cross-lingual cs.cl data data sharing fine-tuning influence language language models languages large language large language models mllms multilingual performance replace representation shows studying transfer type zero-shot

Senior Data Engineer

@ Displate | Warsaw

Analyst, Data Analytics

@ T. Rowe Price | Owings Mills, MD - Building 4

Regulatory Data Analyst

@ Federal Reserve System | San Francisco, CA

Sr. Data Analyst

@ Bank of America | Charlotte

Data Analyst- Tech Refresh

@ CACI International Inc | 1J5 WASHINGTON DC (BOLLING AFB)

Senior AML/CFT & Data Analyst

@ Ocorian | Ebène, Mauritius