LLM Reference

DeepSeek Coder 33B

About

DeepSeek Coder 33B is a cutting-edge large language model engineered specifically for code generation and completion. With 33 billion parameters, it is fine-tuned on an extensive dataset featuring 2 billion tokens of instructional data and a total of 2 trillion tokens, predominantly consisting of code. This powerful foundation enables it to support over 80 programming languages. Its capabilities include generating code snippets, completing and infilling code with a 16K context window, and accurately following detailed instructions. Its architecture includes advanced Grouped-Query Attention, optimizing performance across coding tasks. While its size necessitates dedicated infrastructure for deployment, the model supports open-source and commercial use.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(3)

ProviderInput (per 1M)Output (per 1M)Type
Together AI API$0.8$0.8
Serverless
Alibaba Cloud PAI-EAS
Serverless
Fireworks AI Platform
Provisioned

Specifications

Released2024-03-07
Parameters33B
ArchitectureDecoder Only
Specializationcode