Vicuna 7B 16K
About
Vicuna-7B-v1.5-16k is a large language model (LLM) designed as an advanced chat assistant, developed by LMSYS. It's built on a transformer architecture and fine-tuned from Llama 2, with a notable feature being its 16k context window achieved using linear RoPE scaling. This allows the model to process much longer sequences of text, making it highly effective for comprehensive conversations. Trained on approximately 125,000 conversations from ShareGPT.com, Vicuna demonstrates strong capabilities in handling open-ended dialogues, responding to questions, and supporting various natural language tasks. Despite its strengths, it shares common limitations with other LLMs, like potential biases and performance variability across tasks and languages. Its inference speed and computational requirements are significant due to its 7-billion parameter size. The model is available under the Llama 2 Community License Agreement with various quantized versions for optimized performance.
Capabilities
Providers(1)
| Provider | Input (per 1M) | Output (per 1M) | Type | |
|---|---|---|---|---|
| GCP Vertex AI | — | — | Serverless |