LLM ReferenceLLM Reference

DeepSeek

DeepSeekDeepSeekHighlight
9 models2023–2026Up to 200K ctxFrom $0.21/1M input

About

The DeepSeek LLM family includes open-source large language models designed for exceptional language comprehension and diverse applications 410. These models shine in reasoning, coding, mathematics, and Chinese comprehension, often surpassing similar models in benchmarks 410. The lineup features base and chat models with parameter sizes of 7 billion and 67 billion, respectively 410. They are trained with a massive dataset of 2 trillion tokens in English and Chinese 410, and the architecture, based on the Llama model, enhances inference efficiency through Grouped-Query Attention in the 67B model 1. Available for research and commercial use, additional models like DeepSeek-Coder and DeepSeek-VL cater to code generation and vision-language tasks, respectively 89.

Specifications(9 models)

DeepSeek model specifications comparison
ModelReleasedContextParametersStructured OutputsCode Exec
Together DeepSeek-V3.12026-01200k671BNoNo
DeepSeek V3.2 Speciale2025-04164KYesNo
DeepSeek V3.2 Exp2025-04164KYesYes
DeepSeek V3.1 Terminus2025-04164KYesNo
Together AI Deepseek-LLM-67B-Chat2024-014K67BYesNo
DeepSeek 67B Chat2023-1167BYesNo
DeepSeek 7B Chat2023-117BNoNo
DeepSeek 67B2023-114K67BNoNo
DeepSeek 7B2023-114K7BNoNo

Available From(4 providers)

Pricing

DeepSeek model pricing by provider
ModelProviderInput / 1MOutput / 1MType
DeepSeek V3.1 TerminusOpenRouter$0.21$0.79Serverless
DeepSeek V3.2 ExpOpenRouter$0.27$0.41Serverless
DeepSeek V3.2 SpecialeDeepSeek Platform$0.28$0.42Serverless
DeepSeek V3.2 ExpDeepSeek Platform$0.28$0.42Serverless
DeepSeek V3.2 SpecialeOpenRouter$0.4$1.2Serverless
Together AI Deepseek-LLM-67B-ChatTogether AI$0.6$0.6Serverless
DeepSeek 67B ChatTogether AI$0.9$0.9Serverless

Frequently Asked Questions

What is DeepSeek?
The DeepSeek LLM family includes open-source large language models designed for exceptional language comprehension and diverse applications 410. These models shine in reasoning, coding, mathematics, and Chinese comprehension, often surpassing similar models in benchmarks 410. The lineup features base and chat models with parameter sizes of 7 billion and 67 billion, respectively 410. They are trained with a massive dataset of 2 trillion tokens in English and Chinese 410, and the architecture, based on the Llama model, enhances inference efficiency through Grouped-Query Attention in the 67B model 1. Available for research and commercial use, additional models like DeepSeek-Coder and DeepSeek-VL cater to code generation and vision-language tasks, respectively 89.
How many models are in the DeepSeek family?
The DeepSeek family contains 9 models.
What is the latest DeepSeek model?
The latest model is Together DeepSeek-V3.1, released in 2026-01.
How much does DeepSeek cost?
DeepSeek models range from $0.21/1M to $0.9/1M input tokens depending on the model and provider.

Models(9)