LLM ReferenceLLM Reference

Llama 2 7B 32K

Open Source

About

LLaMA-2-7B-32K is an open-source language model engineered by Together, derived from Meta's LLaMA-2 7B. It boasts a unique extended context length of up to 32,000 tokens, which enhances its ability to tackle tasks involving long-range context, such as multi-document question answering and lengthy text summarization. The model integrates optimizations, including FlashAttention-2, to boost inference and training efficiency. It combines pre-training with instruction tuning data for improved task performance and offers fine-tuning examples for specialized applications, like book summarization or multi-document Q&A. This model marks a substantial progress in the domain of large language models, serving as a potent tool for natural language processing tasks 1311.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
Together AI$0.2$0.2Serverless

Rankings

Specifications

Released2023-07-18
Parameters7B
Context32K
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Blazing-fast, cost-effective AI inference solutions

San Francisco, California, United States
Founded 2022
Website

Providers(1)