LLM ReferenceLLM Reference

StarCoder2 3B

Open Source

About

StarCoder2-3B is a 3-billion parameter large language model developed by the BigCode project, focusing on code generation tasks. Its architecture features a transformer decoder with grouped-query and sliding window attention, trained using the Fill-in-the-Middle objective. The model supports a context window of 16,384 tokens, allowing it to handle large contexts for tasks like code completion and translation across 17 programming languages. However, it is not designed for direct instruction-following. Its efficiency is enhanced by a Grouped Query Attention mechanism and availability in quantized versions for lower memory usage, though users should note that the generated code may not be error-free 25.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
Fireworks AI$0.10$0.10Provisioned

Rankings

Specifications

Released2024-07-04
Parameters3B
Context8K
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Empowering responsible AI for efficient workflows

Santa Clara, California, United States
Founded 2003
Website

Providers(1)