LLM Reference
Databricks Foundation Model Serving

MPT 30B on Databricks Foundation Model Serving

MPT · Databricks Mosaic

Serverless

Pricing

TypePrice (per 1M)
Input tokens$1.00
Output tokens$1.00

Capabilities

VisionMultimodalReasoningFunction CallingTool UseJSON ModeCode Execution

About MPT 30B

MPT-30B, developed by MosaicML, is a powerful large language model (LLM) utilizing a decoder-only transformer architecture that excels in predicting the next word in a sequence. This model is particularly adept at handling various NLP tasks, such as text generation, question answering, summarization, and code generation, due to its training on a diverse dataset of 1 trillion tokens of English text and code. Notable architectural innovations like FlashAttention, ALiBi, and no biases contribute to its enhanced efficiency, allowing for operation on a single high-end GPU. Moreover, MPT-30B offers fine-tuned variants like mpt-30b-instruct and mpt-30b-chat, catering to specialized tasks like instruction following and dialogue generation, and is available for commercial use.

Get Started

Model Specs

Released2023-03-16
Parameters30B
ArchitectureDecoder Only

Related Models on Databricks Foundation Model Serving