LLM ReferenceLLM Reference

DBRX

Databricks MosaicDBRXHighlightOpen Source
7 models2024Up to 33K ctxFrom $0.6/1M input

About

The DBRX family of large language models (LLMs), developed by Databricks, includes DBRX Base and DBRX Instruct, both of which have been released under an open license. These transformer-based, decoder-only models employ a distinguished fine-grained mixture-of-experts (MoE) architecture. This approach leverages a larger number of smaller expert networks, enhancing the quality of output compared to models with fewer, larger experts. Trained on an extensive dataset of 12 trillion tokens encompassing text and code, these models showcase high performance with the capability to handle up to 32,000 tokens of context. DBRX Instruct is specially optimized for few-turn interactions, making it suitable for efficient conversational applications. The DBRX models not only excel in efficiency during training and inference but also outperform many leading closed and open-source models in various benchmarks. They are accessible for download on Hugging Face and integrated through Databricks' Foundation Model APIs.

Specifications(7 models)

DBRX model specifications comparison
ModelReleasedContextParametersStructured Outputs
DBRX Expanse2024-1132k229BNo
DBRX Instruct 20B2024-0732k20BNo
DBRX2024-0332K132BNo
DeepInfra DBRX Instruct2024-0333K132BYes
Fireworks DBRX-Instruct2024-0333K132BNo
DBRX Base2024-0332k132BNo

Available From(6 providers)

Pricing

DBRX model pricing by provider
ModelProviderInput / 1MOutput / 1MType
DeepInfra DBRX InstructDeepInfra$0.6$1.2Serverless
Fireworks DBRX-InstructFireworks AI$1.5$1.5Serverless

Frequently Asked Questions

What is DBRX?
The DBRX family of large language models (LLMs), developed by Databricks, includes DBRX Base and DBRX Instruct, both of which have been released under an open license. These transformer-based, decoder-only models employ a distinguished fine-grained mixture-of-experts (MoE) architecture. This approach leverages a larger number of smaller expert networks, enhancing the quality of output compared to models with fewer, larger experts. Trained on an extensive dataset of 12 trillion tokens encompassing text and code, these models showcase high performance with the capability to handle up to 32,000 tokens of context. DBRX Instruct is specially optimized for few-turn interactions, making it suitable for efficient conversational applications. The DBRX models not only excel in efficiency during training and inference but also outperform many leading closed and open-source models in various benchmarks. They are accessible for download on Hugging Face and integrated through Databricks' Foundation Model APIs.
How many models are in the DBRX family?
The DBRX family contains 7 models.
What is the latest DBRX model?
The latest model is DBRX Expanse, released in 2024-11.
How much does DBRX cost?
DBRX models range from $0.6/1M to $2.7/1M input tokens depending on the model and provider.

Models(7)