LLM ReferenceLLM Reference

DeepSeek Math 7B

deepseek-math-7b

Open Source

About

DeepSeek Math 7B is a powerful family of large language models by DeepSeek AI, crafted for advanced mathematical reasoning. The base model begins as DeepSeek-Coder-v1.5 7B, further pre-trained with 500 billion tokens, encompassing math-focused and general data sources. This model attains a 51.7% score on the MATH benchmark, demonstrating competitive prowess without external aids. Enhanced by instruction tuning, DeepSeekMath-Instruct 7B boosts its mathematical expertise. The DeepSeekMath-RL 7B model, further refined by a novel Group Relative Policy Optimization algorithm, capitalizes on reinforcement learning for superior performance. Available on platforms like Hugging Face, these models cater to applications in education, research, and productivity, offering various quantized formats suitable for diverse hardware 110.

DeepSeek Math 7B input tokens at $0.05/1M, output at $0.25/1M.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(2)

Compare all →
ProviderInput (per 1M)Output (per 1M)Type
Cloudflare Workers AIServerless
Replicate API$0.05$0.25Serverless

Rankings

Specifications

Released2024-02-05
Parameters7B
ArchitectureDecoder Only
Specializationmath
Trainingfinetuned
Fine-tuningbase

Created by

Advancing artificial general intelligence (AGI).

Hangzhou, Zhejiang, China
Founded 2023
Website