Jan. 23, 2024, 2 p.m. | Anthony Alford

InfoQ - AI, ML & Data Engineering www.infoq.com

Mistral AI recently released Mixtral 8x7B, a sparse mixture of experts (SMoE) large language model (LLM). The model contains 46.7B total parameters, but performs inference at the same speed and cost as models one-third that size. On several LLM benchmarks, it outperformed both Llama 2 70B and GPT-3.5, the model powering ChatGPT.

By Anthony Alford

ai benchmarks chatgpt cost deep learning experts generative-ai gpt gpt-3 gpt-3.5 inference language language model large language large language model large language models llama llama 2 llm llm benchmarks mistral mistral ai mixtral mixtral 8x7b mixture of experts ml & data engineering neural networks parameters speed total

More from www.infoq.com / InfoQ - AI, ML & Data Engineering

Software Engineer for AI Training Data (School Specific)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Python)

@ G2i Inc | Remote

Software Engineer for AI Training Data (Tier 2)

@ G2i Inc | Remote

Data Engineer

@ Lemon.io | Remote: Europe, LATAM, Canada, UK, Asia, Oceania

Artificial Intelligence – Bioinformatic Expert

@ University of Texas Medical Branch | Galveston, TX

Lead Developer (AI)

@ Cere Network | San Francisco, US