June 19, 2024, 1:01 a.m. | /u/ai-lover

machinelearningnews www.reddit.com

Researchers from DeepSeek AI introduced DeepSeek-Coder-V2, a new open-source code language model developed by DeepSeek-AI. Built upon the foundation of DeepSeek-V2, this model undergoes further pre-training with an additional 6 trillion tokens, enhancing its code and mathematical reasoning capabilities. DeepSeek-Coder-V2 aims to bridge the performance gap with closed-source models, offering an open-source alternative that delivers competitive results in various benchmarks.

DeepSeek-Coder-V2 employs a Mixture-of-Experts (MoE) framework, supporting 338 programming languages and extending the context from 16K to 128K tokens. The …

128k context ai model code coder coding context context length deepseek deepseek ai foundation gpt4 gpt4-turbo language language model languages machinelearningnews math open-source ai pre-training researchers tokens training turbo

More from www.reddit.com / machinelearningnews

AI Focused Biochemistry Postdoctoral Fellow

@ Lawrence Berkeley National Lab | Berkeley, CA

Senior Data Engineer

@ Displate | Warsaw

Hybrid Cloud Engineer

@ Vanguard | Wayne, PA

Senior Software Engineer

@ F5 | San Jose

Software Engineer, Backend, 3+ Years of Experience

@ Snap Inc. | Bellevue - 110 110th Ave NE

Global Head of Commercial Data Foundations

@ Sanofi | Cambridge