LLM Reference

Llama 3 Sonar Small 32K Online

Deprecated

About

The Llama 3 Sonar Small 32K Online model, developed by Perplexity AI, leverages the Mistral-7B base model to deliver real-time data processing from the internet. Notably, it features a 32,000-token context window for handling longer input sequences, enhancing its ability for comprehensive understanding and response generation. It offers robust performance in tasks like text summarization and question answering, despite being smaller and faster than the Llama 3 Sonar Large model. However, users have noted that its memory capacity might not fully utilize the 32k tokens, occasionally missing information from earlier in conversations, and it primarily supports the English language.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(1)

ProviderInput (per 1M)Output (per 1M)Type
Perplexity LabsServerless

Specifications

FamilySonar
Released2024-05-05
Parameters8B
Context28K
ArchitectureDecoder Only
Specializationgeneral