LLM Reference

Vicuna 7B 16K

About

Vicuna-7B-v1.5-16k is a large language model (LLM) designed as an advanced chat assistant, developed by LMSYS. It's built on a transformer architecture and fine-tuned from Llama 2, with a notable feature being its 16k context window achieved using linear RoPE scaling. This allows the model to process much longer sequences of text, making it highly effective for comprehensive conversations. Trained on approximately 125,000 conversations from ShareGPT.com, Vicuna demonstrates strong capabilities in handling open-ended dialogues, responding to questions, and supporting various natural language tasks. Despite its strengths, it shares common limitations with other LLMs, like potential biases and performance variability across tasks and languages. Its inference speed and computational requirements are significant due to its 7-billion parameter size. The model is available under the Llama 2 Community License Agreement with various quantized versions for optimized performance.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
GCP Vertex AI
Serverless

Specifications

FamilyVicuna
Released2023-10-23
Parameters7B
Context16K
ArchitectureDecoder Only
Specializationgeneral