Jan. 23, 2024, 2 p.m. | Anthony Alford

InfoQ - AI, ML & Data Engineering www.infoq.com

Mistral AI recently released Mixtral 8x7B, a sparse mixture of experts (SMoE) large language model (LLM). The model contains 46.7B total parameters, but performs inference at the same speed and cost as models one-third that size. On several LLM benchmarks, it outperformed both Llama 2 70B and GPT-3.5, the model powering ChatGPT.

By Anthony Alford

ai benchmarks chatgpt cost deep learning experts generative-ai gpt gpt-3 gpt-3.5 inference language language model large language large language model large language models llama llama 2 llm llm benchmarks mistral mistral ai mixtral mixtral 8x7b mixture of experts ml & data engineering neural networks parameters speed total

More from www.infoq.com / InfoQ - AI, ML & Data Engineering

Senior Machine Learning Engineer

@ GPTZero | Toronto, Canada

ML/AI Engineer / NLP Expert - Custom LLM Development (x/f/m)

@ HelloBetter | Remote

Doctoral Researcher (m/f/div) in Automated Processing of Bioimages

@ Leibniz Institute for Natural Product Research and Infection Biology (Leibniz-HKI) | Jena

Seeking Developers and Engineers for AI T-Shirt Generator Project

@ Chevon Hicks | Remote

Senior Applied Data Scientist

@ dunnhumby | London

Principal Data Architect - Azure & Big Data

@ MGM Resorts International | Home Office - US, NV