Dec. 24, 2023, 8 p.m. | Venelin Valkov

Venelin Valkov www.youtube.com

Do you need 7B+ parameters to get great performance from your Language Models? Discover how Microsoft Research's Phi-2, a 2.7 billion-parameter language model, challenges this norm by outperforming models up to 25x its size (according to Microsoft Research). We'll delve into the training methods behind Phi-2, from 'textbook-quality' training data to scaled knowledge transfer techniques. We'll load the model into a Google Colab and try it out in coding, math, reasoning, and data extraction.

Blog Post: https://www.microsoft.com/en-us/research/blog/phi-2-the-surprising-power-of-small-language-models/
Phi-2 on HF …

billion challenges colab google language language model language models llms microsoft microsoft research norm parameters performance phi phi-2 python research small training tutorial

More from www.youtube.com / Venelin Valkov

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Data Science Analyst

@ Mayo Clinic | AZ, United States

Sr. Data Scientist (Network Engineering)

@ SpaceX | Redmond, WA