Llama 3 Sonar Small 32K Online
DeprecatedAbout
The Llama 3 Sonar Small 32K Online model, developed by Perplexity AI, leverages the Mistral-7B base model to deliver real-time data processing from the internet. Notably, it features a 32,000-token context window for handling longer input sequences, enhancing its ability for comprehensive understanding and response generation. It offers robust performance in tasks like text summarization and question answering, despite being smaller and faster than the Llama 3 Sonar Large model. However, users have noted that its memory capacity might not fully utilize the 32k tokens, occasionally missing information from earlier in conversations, and it primarily supports the English language.
Capabilities
MultimodalFunction CallingTool UseJSON Mode
Providers(1)
| Provider | Input (per 1M) | Output (per 1M) | Type | |
|---|---|---|---|---|
| Perplexity Labs | — | — | Serverless |
Specifications
FamilySonar
Released2024-05-05
Parameters8B
Context28K
ArchitectureDecoder Only
Specializationgeneral