Dec. 12, 2023, 9:45 a.m. | code_your_own_AI

code_your_own_AI www.youtube.com

Mistral AI's new Sparse Mixture-of-Experts system (SMoE) is now available: MIXTRAL 8x7B. Also in a DPO instruction tuned version MIXTRAL 8x 7B Instruct, which we test on real world causal reasoning, in a live recoding.

Plus Python code to run inference on the fp32 Mixtral 8x7B, on the fp16 Mixtral 8x7B, on the 4-bit quantized version of Mixtral 8x7B and also w/ Flash-attention_2.

Plus costs for inference API (costs per token) and embedding API.

00:00 Live test of Mixtral 8x7B …

code experts fp16 inference mistral mistral ai mixtral 8x7b moe performance python reasoning test world

Founding AI Engineer, Agents

@ Occam AI | New York

AI Engineer Intern, Agents

@ Occam AI | US

AI Research Scientist

@ Vara | Berlin, Germany and Remote

Data Architect

@ University of Texas at Austin | Austin, TX

Data ETL Engineer

@ University of Texas at Austin | Austin, TX

Codec Avatars Research Engineer

@ Meta | Pittsburgh, PA