Feb. 19, 2024, 4:41 p.m. | /u/Kaldnite

Machine Learning www.reddit.com

I've been doing some reading about Mixture of Experts (MoE) models, and how they penalise the models to ensure that the distribution of activations are equal across all X experts.

Now that being said, is it reasonable to say that it isn't a "This is the Maths expert, and that's the Science expert", but rather a black box of optimised sub-models trained on lots training data to target different dimensions of the input query?

I'm viewing it more as a …

distribution expert experts isn machinelearning maths mixture of experts moe reading

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Business Data Analyst

@ Alstom | Johannesburg, GT, ZA