LLM ReferenceLLM Reference
Microsoft Foundry

Using Codestral 2501 on Microsoft Foundry

Implementation guide · Codestral · MistralAI

Serverless

Quick Start

  1. 1
    Create an account at Microsoft Foundry and generate an API key.
  2. 2
    Use the Microsoft Foundry SDK or REST API to call codestral-2501 — see the documentation for request format.
  3. 3
    You'll be billed $0.30/1M input, $0.90/1M output tokens. See full pricing.

Code Examples

See Microsoft Foundry documentation for integration details.

About Microsoft Foundry

Microsoft Foundry offers a comprehensive platform-as-a-service for enterprise AI operations. It provides multiple deployment options including Serverless APIs (pay-as-you-go), Global Standard (shared managed capacity), Provisioned Throughput Units (reserved capacity), batch processing, and bring-your-own model deployments. The platform features a unified control plane for models, agents, tools, and observability. Its Agent Service enables building and deploying AI agents with built-in tracing, monitoring, and governance. Evaluation and monitoring tools assess model performance, safety, and groundedness. Foundry supports seamless upgrades from Azure OpenAI with non-destructive migration, maintaining existing deployments while unlocking multi-provider model access and advanced platform capabilities.

Microsoft Foundry is a unified Azure platform-as-a-service offering for enterprise AI operations, model builders, and application development. It provides access to over 1,900 models from Microsoft, OpenAI, Anthropic, Mistral, xAI, Meta, DeepSeek, Hugging Face, and more. Foundry unifies agents, models, and tools under a single management grouping with built-in enterprise-readiness capabilities including tracing, monitoring, evaluations, and customizable enterprise setup configurations.

Pricing on Microsoft Foundry

TypePrice (per 1M)
Input tokens$0.30
Output tokens$0.90

Capabilities

No model capability flags are currently sourced.

About Codestral 2501

From partners/community. Code generation model with very long context: 262,144 tokens. Output 4,096 tokens. Tool calling: No. English only.

Model Specs

Released2025-01-01
Context262K

Provider

Microsoft Foundry
Microsoft Foundry

Microsoft

Redmond, Washington, United States