March 30, 2022, 3:30 p.m. | Synced

Synced syncedreview.com

A research team from Carnegie Mellon University and Google systematically explores strategies for leveraging the relatively under-studied resource of bilingual lexicons to adapt pretrained multilingual models to low-resource languages. Their resulting Lexicon-based Adaptation approach produces consistent performance improvements without requiring additional monolingual text.


The post CMU & Google Extend Pretrained Models to Thousands of Underrepresented Languages Without Using Monolingual Data first appeared on Synced.

ai artificial intelligence bert data google machine learning machine learning & data science ml multilingual language model pretrained language model research technology

More from syncedreview.com / Synced

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Data Engineer (m/f/d)

@ Project A Ventures | Berlin, Germany

Principle Research Scientist

@ Analog Devices | US, MA, Boston