Jan. 3, 2024, 5:48 p.m. | Fahim Rustamy, PhD

Towards Data Science - Medium towardsdatascience.com

CLIP, which stands for Contrastive Language-Image Pretraining, is a deep learning model developed by OpenAI in 2021. CLIP’s embeddings for images and text share the same space, enabling direct comparisons between the two modalities. This is accomplished by training the model to bring related images and texts closer together while pushing unrelated ones apart.

Some applications of CLIP include:

  1. Image Classification and Retrieval: CLIP can be used for image classification tasks by associating images with natural language descriptions. It allows …

artificial intelligence clip data science deep learning embeddings enabling image images importance in 2021 language large language models machine learning multimodal openai space text together training

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Lead GNSS Data Scientist

@ Lurra Systems | Melbourne

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Robotics Technician - 3rd Shift

@ GXO Logistics | Perris, CA, US, 92571