LLM Reference

Mistral 7B v0.1

About

Mistral 7B v0.1 is an advanced open-source large language model built by Mistral AI, consisting of 7 billion parameters. It's designed to deliver high performance and efficiency, outperforming many similar-sized models in various benchmarks. The model employs a transformer architecture with innovative features like Sliding Window Attention, Grouped-Query Attention, and a Byte-fallback BPE tokenizer, enhancing speed, reducing computational costs, and improving robustness. Capable of generating human-like text, following instructions effectively, and excelling in areas such as reasoning and mathematics, Mistral 7B v0.1 does have limitations, such as a lack of built-in moderation and a potential for hallucinations. Subsequent versions have sought to address these limitations while introducing extended context windows and improved instruction-following capabilities.

Capabilities

MultimodalFunction CallingTool UseJSON Mode

Providers(15)

ProviderInput (per 1M)Output (per 1M)Type
GCP Vertex AI
Serverless
Snowflake Cortex$0.24$0.24
Serverless
OctoAI API$0.15$0.15
Serverless
deepinfra API
Serverless
Mistral AI Le Plateforme
Serverless
Baseten API
Serverless
Lepton AI API
Serverless
Replicate API
Serverless
AWS Bedrock$0.15$0.2
Serverless
Azure OpenAI
Provisioned
Cloudflare Workers AI
Serverless
Fireworks AI Platform
Provisioned
NVIDIA NIM
Provisioned
Alibaba Cloud PAI-EAS
Serverless
Together AI API$0.2$0.2
Serverless

Specifications

FamilyMistral
Released2023-09-27
Parameters7B
Context8K
ArchitectureDecoder Only
Specializationgeneral