LLM ReferenceLLM Reference

Alpaca 13B

About

The Alpaca 13B model is a large language model built on the transformer architecture, designed to handle a variety of natural language processing tasks such as text generation, instruction following, question answering, and dialogue generation. It is based on the LLaMA architecture and incorporates features like pre-normalization and rotary positional embeddings. Variants of Alpaca 13B are often fine-tuned using methods like LoRA for efficient adaptation to specific tasks, utilizing instruction-following datasets for training. While it excels in certain tasks, it can suffer from limitations such as hallucinations, biases, and resource-intensiveness. The model's effectiveness is closely tied to the nature of its fine-tuning datasets and techniques.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Rankings

Specifications

FamilyAlpaca
Released2023-03-31
Parameters13B
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Pioneering AI research and computing infrastructure

Stanford, California, United States
Founded 1962
Website