LLM Reference

Phi-3 Mini 128K

About

Phi-3 Mini-128K-Instruct, developed by Microsoft, is a 3.8 billion-parameter large language model renowned for its lightweight, open-source architecture. Despite its modest size, it excels in reasoning tasks, particularly in math and logic, and showcases strong code generation capabilities. A standout feature is its remarkable ability to handle up to 128,000 tokens, allowing it to process extensive text documents and codebases efficiently. While it has limitations in factual knowledge and focuses primarily on English, it strikes a balance between performance and efficiency, making it ideal for resource-constrained environments. The model is available on platforms like Azure AI Studio and Hugging Face and benefits from training on high-quality synthetic and publicly available data, with fine-tuning to improve instruction adherence and safety.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseJSON ModeCode Execution

Providers(5)

Compare all →
ProviderInput (per 1M)Output (per 1M)Type
NVIDIA NIMProvisioned
Baseten APIServerless
Azure OpenAI$0.3$0.9ServerlessProvisioned
Fireworks AI$0.10$0.10Provisioned
Replicate API$0.05$0.25Serverless

Benchmark Scores(4)

BenchmarkScoreVersionSource
Google-Proof Q&A50.8diamondresearch
HellaSwag90.210-shotresearch
HumanEval75.9pass@1research
Massive Multitask Language Understanding76.55-shotresearch

Rankings

Specifications

FamilyPhi-3
Released2024-04-23
Parameters3.8B
Context128K
ArchitectureDecoder Only
Specializationgeneral
Trainingfinetuning

Created by

Advancing the state-of-the-art in AI and computing.

Redmond, Washington, United States
Founded 1991
Website