LLM Reference

Alpaca 13B

About

The Alpaca 13B model is a large language model built on the transformer architecture, designed to handle a variety of natural language processing tasks such as text generation, instruction following, question answering, and dialogue generation. It is based on the LLaMA architecture and incorporates features like pre-normalization and rotary positional embeddings. Variants of Alpaca 13B are often fine-tuned using methods like LoRA for efficient adaptation to specific tasks, utilizing instruction-following datasets for training. While it excels in certain tasks, it can suffer from limitations such as hallucinations, biases, and resource-intensiveness. The model's effectiveness is closely tied to the nature of its fine-tuning datasets and techniques.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Specifications

FamilyAlpaca
Released2023-03-31
ArchitectureDecoder Only
Specializationgeneral