May 31, 2023, 3 p.m. | /u/AvvYaa

machinelearningnews www.reddit.com

Hello people!

I thought it was a good time to make a video about this topic since more and more recent LLMs are moving away from text-only into visual-language domains (GPT-4, PaLM-2, etc). Multi-modal models basically input data from multiple sources (text, image, audio, video etc) to train Machine Learning tasks. In my video, I provide some intuition about this area - right from basics like contrastive learning (CLIP, ImageBind), all the way to Generative language models (like Flamingo).

Hope …

audio data etc good gpt gpt-4 image language language models llms machine machine learning machinelearningnews moving multiple palm people text thought video

More from www.reddit.com / machinelearningnews

(373) Applications Manager – Business Intelligence - BSTD

@ South African Reserve Bank | South Africa

Data Engineer Talend (confirmé/sénior) - H/F - CDI

@ Talan | Paris, France

Data Science Intern (Summer) / Stagiaire en données (été)

@ BetterSleep | Montreal, Quebec, Canada

Director - Master Data Management (REMOTE)

@ Wesco | Pittsburgh, PA, United States

Architect Systems BigData REF2649A

@ Deutsche Telekom IT Solutions | Budapest, Hungary

Data Product Coordinator

@ Nestlé | São Paulo, São Paulo, BR, 04730-000