DeepSeek-Coder-V2-Lite-Base
An open-source code language model designed to enhance programming and mathematical reasoning abilities.
CommonProductProgrammingCode GenerationOpen Source
DeepSeek-Coder-V2 is an open-source expert mixture model (MoE) specifically designed for code language. Its performance is comparable to GPT4-Turbo. It excels in code-specific tasks while maintaining strong performance in general language tasks. Compared to DeepSeek-Coder-33B, the V2 version demonstrates significant improvements in code-related tasks and reasoning capabilities. Furthermore, it expands its supported programming languages from 86 to 338 and increases the context length from 16K to 128K.
DeepSeek-Coder-V2-Lite-Base Visit Over Time
Monthly Visits
20899836
Bounce Rate
46.04%
Page per Visit
5.2
Visit Duration
00:04:57