LLM ReferenceLLM Reference

Llama 3 Sonar Small 32K Online

llama-3-sonar-small-32k-online

Deprecated
Open Source

About

The Llama 3 Sonar Small 32K Online model, developed by Perplexity AI, leverages the Mistral-7B base model to deliver real-time data processing from the internet. Notably, it features a 32,000-token context window for handling longer input sequences, enhancing its ability for comprehensive understanding and response generation. It offers robust performance in tasks like text summarization and question answering, despite being smaller and faster than the Llama 3 Sonar Large model. However, users have noted that its memory capacity might not fully utilize the 32k tokens, occasionally missing information from earlier in conversations, and it primarily supports the English language.

Llama 3 Sonar Small 32K Online has a 28K-token context window.

Llama 3 Sonar Small 32K Online input tokens at $0.2/1M, output at $0.2/1M.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
Perplexity Labs$0.20$0.20Serverless

Specifications

FamilySonar
Released2024-05-05
Parameters8B
Context28K
ArchitectureDecoder Only
Knowledge cutoff2024-03
Specializationgeneral
Trainingfinetuned

Created by

Developing AI for complex problem-solving.

San Francisco, California, United States
Founded 2022
Website

Providers(1)