LLM Reference
Fireworks AI

StarCoder2 3B on Fireworks AI

StarCoder 2 · ServiceNow Research

Provisioned

Pricing

TypePrice (per 1M)
Input tokens$0.10
Output tokens$0.10

Capabilities

VisionMultimodalReasoningFunction CallingTool UseJSON ModeCode Execution

About StarCoder2 3B

StarCoder2-3B is a 3-billion parameter large language model developed by the BigCode project, focusing on code generation tasks. Its architecture features a transformer decoder with grouped-query and sliding window attention, trained using the Fill-in-the-Middle objective. The model supports a context window of 16,384 tokens, allowing it to handle large contexts for tasks like code completion and translation across 17 programming languages. However, it is not designed for direct instruction-following. Its efficiency is enhanced by a Grouped Query Attention mechanism and availability in quantized versions for lower memory usage, though users should note that the generated code may not be error-free 25.

Get Started

Model Specs

Released2024-07-04
Parameters3B
Context8K
ArchitectureDecoder Only

Related Models on Fireworks AI