LLM ReferenceLLM Reference

DeepSeek 67B Chat

deepseek-67b-chat

Open Source

About

DeepSeek LLM 67B Chat is a sophisticated language model with 67 billion parameters, leveraging the LLaMA architecture with enhancements such as Grouped-Query Attention across 95 layers. Trained on a vast corpus of 2 trillion tokens in English and Chinese, it excels in tasks like text generation, question answering, and fluent conversation, demonstrating superior performance in reasoning, coding, and mathematics compared to some larger models. Despite its advanced capabilities, the model can exhibit biases from its training data, experience hallucinations, and produce repetitive outputs. Due to its size, substantial computational resources are needed for inference, although quantization methods can reduce its size with potential trade-offs in quality.

DeepSeek 67B Chat input tokens at $0.9/1M, output at $0.9/1M.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
Together AI$0.9$0.9Serverless

Rankings

Specifications

FamilyDeepSeek
Released2023-11-29
Parameters67B
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuned
Fine-tuningchat

Created by

Advancing artificial general intelligence (AGI).

Hangzhou, Zhejiang, China
Founded 2023
Website

Providers(1)