Feb. 14, 2022, 9:50 a.m. | LA Tran

Towards Data Science - Medium towardsdatascience.com

Learn from different perspectives

Photo by Robina Weermeijer on Unsplash

Knowledge distillation

Knowledge distillation (KD), also known as model distillation (MD), is an impressive neural network training method proposed by the God Father of deep learning, Geoffrey Hinton, to gain neural network’s performances. If you have never heard about KD, you can reach my post via this link.

Shortly, the core idea of KD is to distill knowledge from a large model (teacher) or an ensemble of neural network …

computer vision data deep learning detection distillation knowledge-distillation object-detection yolo

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Reporting & Data Analytics Lead (Sizewell C)

@ EDF | London, GB

Data Analyst

@ Notable | San Mateo, CA