Aug. 11, 2022, 2:57 p.m. | /u/jamescalam

Natural Language Processing www.reddit.com

Hi all, I created a [walkthrough](https://towardsdatascience.com/quick-fire-guide-to-multi-modal-ml-with-openais-clip-2dad7e398ac0?sk=89bb2d8b8e583ed109d8a05e00366645) (and [video](https://youtu.be/989aKUVBfbk)) demoing how to use the text and image embeddings of OpenAI's CLIP. CLIP is a multi-modal model that uses a typical text transformer for text embeddings and a vision transformer (ViT, alt version uses Resnet) for image embeddings. During pretraining CLIP learns to place (image, text) pairs into the same vector space. The result is a cool off-the-shelf model that can perform tasks across image and text data.

When I started using …

clip image languagetechnology ml openai text text-image

Senior Machine Learning Engineer (MLOps)

@ Promaton | Remote, Europe

Data Analyst (CPS-GfK)

@ GfK | Bucharest

Consultant Data Analytics IT Digital Impulse - H/F

@ Talan | Paris, France

Data Analyst

@ Experian | Mumbai, India

Data Scientist

@ Novo Nordisk | Princeton, NJ, US

Data Architect IV

@ Millennium Corporation | United States