all AI news
CMU & Google Extend Pretrained Models to Thousands of Underrepresented Languages Without Using Monolingual Data
March 30, 2022, 3:30 p.m. | Synced
Synced syncedreview.com
A research team from Carnegie Mellon University and Google systematically explores strategies for leveraging the relatively under-studied resource of bilingual lexicons to adapt pretrained multilingual models to low-resource languages. Their resulting Lexicon-based Adaptation approach produces consistent performance improvements without requiring additional monolingual text.
The post CMU & Google Extend Pretrained Models to Thousands of Underrepresented Languages Without Using Monolingual Data first appeared on Synced.
ai artificial intelligence bert data google machine learning machine learning & data science ml multilingual language model pretrained language model research technology
More from syncedreview.com / Synced
Jobs in AI, ML, Big Data
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Data Engineer (m/f/d)
@ Project A Ventures | Berlin, Germany
Principle Research Scientist
@ Analog Devices | US, MA, Boston