LLM ReferenceLLM Reference

StarCoder2 7B

Open Source

About

The StarCoder2 7B model is a large language model engineered for code generation tasks, containing 7 billion parameters. It is trained on a vast dataset of 3.5 trillion tokens known as The Stack v2, which includes diverse code samples from 17 programming languages such as Python, Java, and JavaScript. This model features Grouped Query Attention with a significant context window of 16,384 tokens and a sliding window attention of 4,096 tokens, enabling it to adeptly manage complex coding tasks. StarCoder2 7B excels in code completion, summarization, and generating code snippets from given prompts. It is built with responsible data usage in mind, designed to avoid directly copying copyrighted code, making it a reliable tool for developers 124.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(2)

Compare all →
ProviderInput (per 1M)Output (per 1M)Type
Fireworks AI$0.2$0.2ServerlessProvisioned
NVIDIA NIMProvisioned

Rankings

Specifications

Released2024-07-04
Parameters7B
Context8K
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Empowering responsible AI for efficient workflows

Santa Clara, California, United States
Founded 2003
Website