LLM Reference
GCP Vertex AI

Vicuna 7B 16K on GCP Vertex AI

Vicuna · LMSYS Org

Serverless

Capabilities

VisionMultimodalReasoningFunction CallingTool UseJSON ModeCode Execution

About Vicuna 7B 16K

Vicuna-7B-v1.5-16k is a large language model (LLM) designed as an advanced chat assistant, developed by LMSYS. It's built on a transformer architecture and fine-tuned from Llama 2, with a notable feature being its 16k context window achieved using linear RoPE scaling. This allows the model to process much longer sequences of text, making it highly effective for comprehensive conversations. Trained on approximately 125,000 conversations from ShareGPT.com, Vicuna demonstrates strong capabilities in handling open-ended dialogues, responding to questions, and supporting various natural language tasks. Despite its strengths, it shares common limitations with other LLMs, like potential biases and performance variability across tasks and languages. Its inference speed and computational requirements are significant due to its 7-billion parameter size. The model is available under the Llama 2 Community License Agreement with various quantized versions for optimized performance.

Get Started

Model Specs

Released2023-10-23
Parameters7B
Context16K
ArchitectureDecoder Only

Related Models on GCP Vertex AI

Provider

GCP Vertex AI
GCP Vertex AI

Google Cloud Platform (GCP)

All models on GCP Vertex AI