DeepSeek Coder
About
The DeepSeek Coder family includes a range of open-source code language models specifically designed for handling large codebases. Trained on an expansive dataset of 2 trillion tokens, primarily composed of code (87%) and a mix of English and Chinese natural language data (13%), these models are available in sizes from 1.3 billion to 33 billion parameters. This range gives users the flexibility to choose models that align with their computational resources and specific needs. With pre-training on a high-quality project-level code corpus and using a 16K window size, the models excel in code generation and infill tasks. They demonstrate state-of-the-art performance on various open-source code benchmarks and often outperform some proprietary models. Released under a permissive license, DeepSeek Coder models support both research and commercial applications, offering significant capabilities to developers in coding projects 145.
Specifications(9 models)
| Model | Released | Context | Parameters | Structured Outputs |
|---|---|---|---|---|
| DeepSeek Coder 33B | 2024-03 | 16K | 33B | Yes |
| DeepSeek Coder 33B Instruct | 2024-03 | — | 33B | No |
| DeepSeek Coder 6.7B Instruct | 2024-03 | — | 6.7B | No |
| Together AI Deepseek-Coder-33B-Instruct | 2024-03 | 16K | 33B | Yes |
| DeepSeek Coder 7B V1.5 | 2024-02 | — | 7B | No |
| DeepSeek Coder 7B V1.5 Instruct | 2024-02 | — | 7B | No |
| DeepSeek Coder 1.3B | 2023-11 | 4K | 1.3B | No |
| DeepSeek Coder 1.3B Instruct | 2023-11 | — | 1.3B | No |
Available From(5 providers)
Pricing
| Model | Provider | Input / 1M | Output / 1M | Type |
|---|---|---|---|---|
| DeepSeek Coder 1.3B | Fireworks AI | $0.1 | $0.1 | Provisioned |
| DeepSeek Coder 7B V1.5 | Fireworks AI | $0.2 | $0.2 | Provisioned |
| Together AI Deepseek-Coder-33B-Instruct | Together AI | $0.3 | $0.3 | Serverless |
| DeepSeek Coder 33B | Together AI | $0.8 | $0.8 | Serverless |
| DeepSeek Coder 33B | Fireworks AI | $0.9 | $0.9 | Provisioned |
| DeepSeek Coder 33B Instruct | Fireworks AI | $0.9 | $0.9 | Serverless |
Frequently Asked Questions
- What is DeepSeek Coder?
- The DeepSeek Coder family includes a range of open-source code language models specifically designed for handling large codebases. Trained on an expansive dataset of 2 trillion tokens, primarily composed of code (87%) and a mix of English and Chinese natural language data (13%), these models are available in sizes from 1.3 billion to 33 billion parameters. This range gives users the flexibility to choose models that align with their computational resources and specific needs. With pre-training on a high-quality project-level code corpus and using a 16K window size, the models excel in code generation and infill tasks. They demonstrate state-of-the-art performance on various open-source code benchmarks and often outperform some proprietary models. Released under a permissive license, DeepSeek Coder models support both research and commercial applications, offering significant capabilities to developers in coding projects 145.
- How many models are in the DeepSeek Coder family?
- The DeepSeek Coder family contains 9 models.
- What is the latest DeepSeek Coder model?
- The latest model is DeepSeek Coder 33B, released in 2024-03.
- How much does DeepSeek Coder cost?
- DeepSeek Coder models range from $0.1/1M to $0.9/1M input tokens depending on the model and provider.
Models(9)
DeepSeek Coder 33B
DeepSeek Coder 33B Instruct
DeepSeek Coder 6.7B Instruct
Together AI Deepseek-Coder-33B-Instruct
DeepSeek Coder 7B V1.5
DeepSeek Coder 7B V1.5 Instruct
DeepSeek Coder 1.3B
DeepSeek Coder 1.3B Instruct






