Sept. 27, 2023, 7:31 p.m. | Sascha Kirch

Towards Data Science - Medium towardsdatascience.com

DINO — A Foundation Model for Computer Vision

Paper Walkthrough: Emerging Properties in Self-Supervised Vision Transformers

It is an exciting decade for computer vision. Great successes from the natural language domain are transferred to the vision domain including the introduction of the ViT (vision transformer) and lately large-scale self-supervised pre-training techniques have made headlines under the name of foundation models.

Today we are looking into a framework called DINO (self DIstillation, NO labels), a visual foundation model built …

computer vision deep-dives deep learning foundation-models knowledge-distillation

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Principal Machine Learning Engineer (AI, NLP, LLM, Generative AI)

@ Palo Alto Networks | Santa Clara, CA, United States

Consultant Senior Data Engineer F/H

@ Devoteam | Nantes, France