all AI news
Transformers explained REMASTERED | The architecture behind LLMs
Jan. 21, 2024, 12:45 p.m. | AI Coffee Break with Letitia
AI Coffee Break with Letitia www.youtube.com
Outline:
00:00 Transformers explained
00:47 Text inputs
02:29 Image inputs
03:57 Next word prediction / Classification
06:08 The transformer layer: 1. MLP sublayer
06:47 2. Attention explained
07:57 Attention vs. self-attention
08:35 Queries, Keys, Values
11:26 Multi-head attention
13:04 Attention scales quadratically
13:53 Positional embeddings …
architecture attention bonus classification difference embeddings explained image inputs keys llms networks neural networks next overview prediction recurrent neural networks residual text transformer transformer architecture transformers values word
More from www.youtube.com / AI Coffee Break with Letitia
Stealing Part of a Production LLM | API protect LLMs no more
3 weeks, 2 days ago |
www.youtube.com
MAMBA and State Space Models explained | SSM explained
2 months, 2 weeks ago |
www.youtube.com
Why is DALL-E 3 better at following Text Prompts? — DALL-E 3 explained
5 months, 3 weeks ago |
www.youtube.com
Jobs in AI, ML, Big Data
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Lead GNSS Data Scientist
@ Lurra Systems | Melbourne
Senior Machine Learning Engineer (MLOps)
@ Promaton | Remote, Europe
C003549 Data Analyst (NS) - MON 13 May
@ EMW, Inc. | Braine-l'Alleud, Wallonia, Belgium
Marketing Decision Scientist
@ Meta | Menlo Park, CA | New York City