LLM Reference
Phi-1

Phi-1

Microsoft Research
MIT
Open Source

About

The Phi-1 family of large language models (LLMs), developed by Microsoft, comprises several models designed for specific tasks, primarily focusing on code generation and reasoning. Phi-1, the initial model in the family, is a transformer-based model with 1.3 billion parameters, specializing in basic Python coding 18. Its training utilized a blend of "textbook quality" data sourced from the web and synthetic data generated using GPT-3.5 18. Despite its relatively small size compared to other LLMs, Phi-1 demonstrates impressive accuracy, exceeding 50% on the HumanEval benchmark for simple Python coding tasks 18. Subsequent models in the Phi family, such as Phi-1.5 and Phi-2, build upon this foundation, expanding capabilities to encompass broader natural language tasks while maintaining a focus on efficiency and high-quality data 911. These models showcase Microsoft's research into creating smaller, more efficient LLMs that rival the performance of much larger models 29.

Models(2)

Details

LicenseMIT
Models2