all AI news
Google & Stanford U’s DoReMi Significantly Speeds Up Language Model Pretraining
Synced syncedreview.com
In the new paper DoReMi: Optimizing Data Mixtures Speeds Up Language Model Pretraining, a research team from Google and Stanford University introduces Domain Reweighting with Minimax Optimization (DoReMi), a domain weight optimization strategy that leverages distributionally robust optimization (DRO) to substantially speed up effective language model pretraining.
The post Google & Stanford U’s DoReMi Significantly Speeds Up Language Model Pretraining first appeared on Synced.
ai artificial intelligence data deep-neural-networks google language language model large language model machine learning machine learning & data science minimax ml optimization paper research research team speed stanford stanford university strategy team technology university