April 23, 2024, 9:58 a.m. | /u/lime_52

Machine Learning www.reddit.com

Is MoE architecture also applied in other ML areas, let’s say Computer Vision? Why aren’t they popular? Is it because we don’t scale vision transformers as much as LLMs, and MoE is best for scalability?

architecture computer computer vision llms machinelearning moe popular scalability scale transformers vision vision transformers

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne