all AI news
MIXTRAL 8x7B MoE Instruct: LIVE Performance Test
Dec. 12, 2023, 9:45 a.m. | code_your_own_AI
code_your_own_AI www.youtube.com
Plus Python code to run inference on the fp32 Mixtral 8x7B, on the fp16 Mixtral 8x7B, on the 4-bit quantized version of Mixtral 8x7B and also w/ Flash-attention_2.
Plus costs for inference API (costs per token) and embedding API.
00:00 Live test of Mixtral 8x7B …
code experts fp16 inference mistral mistral ai mixtral 8x7b moe performance python reasoning test world
More from www.youtube.com / code_your_own_AI
New Discovery: Retrieval Heads for Long Context
2 days, 22 hours ago |
www.youtube.com
Multi-Token Prediction (forget next token LLM?)
3 days, 22 hours ago |
www.youtube.com
NEW LLM Test: Reasoning & gpt2-chatbot
5 days, 3 hours ago |
www.youtube.com
LLMs: Rewriting Our Tomorrow (plus code) #ai
6 days, 10 hours ago |
www.youtube.com
480B LLM as 128x4B MoE? WHY?
1 week, 2 days ago |
www.youtube.com
No more Fine-Tuning: Unsupervised ICL+
1 week, 4 days ago |
www.youtube.com
NEW Phi-3 mini 3.8B LLM for Your PHONE: 1st TEST
1 week, 5 days ago |
www.youtube.com
Jobs in AI, ML, Big Data
Founding AI Engineer, Agents
@ Occam AI | New York
AI Engineer Intern, Agents
@ Occam AI | US
AI Research Scientist
@ Vara | Berlin, Germany and Remote
Data Architect
@ University of Texas at Austin | Austin, TX
Data ETL Engineer
@ University of Texas at Austin | Austin, TX
Codec Avatars Research Engineer
@ Meta | Pittsburgh, PA