Feb. 11, 2024, midnight |

Eugene Yan eugeneyan.com

Distillation vs. self-improvement across the three stages of language model training.

data distillation finetuning generate improvement language language model language model training pretraining self-improvement synthetic synthetic data training

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Associate Data Engineer

@ Nominet | Oxford/ Hybrid, GB

Data Science Senior Associate

@ JPMorgan Chase & Co. | Bengaluru, Karnataka, India