all AI news
Mistral AI's Open-Source Mixtral 8x7B Outperforms GPT-3.5
InfoQ - AI, ML & Data Engineering www.infoq.com
Mistral AI recently released Mixtral 8x7B, a sparse mixture of experts (SMoE) large language model (LLM). The model contains 46.7B total parameters, but performs inference at the same speed and cost as models one-third that size. On several LLM benchmarks, it outperformed both Llama 2 70B and GPT-3.5, the model powering ChatGPT.
By Anthony Alfordai benchmarks chatgpt cost deep learning experts generative-ai gpt gpt-3 gpt-3.5 inference language language model large language large language model large language models llama llama 2 llm llm benchmarks mistral mistral ai mixtral mixtral 8x7b mixture of experts ml & data engineering neural networks parameters speed total