LLM ReferenceLLM Reference

Using MiniMax M2.5 Highspeed on Novita AI

Implementation guide · MiniMax M2 · MiniMax

Serverless

Quick Start

  1. 1
    Create an account at Novita AI and generate an API key.
  2. 2
    Use the Novita AI SDK or REST API to call minimax/minimax-m2.5-highspeed.
  3. 3
    You'll be billed $0.60/1M input, $2.40/1M output tokens.

Code Examples

Code examples for this provider have not been sourced yet.

About Novita AI

Novita AI offers a GPU-based inference API for image, video, and language model generation with a broad catalog of open-source models.

Pricing on Novita AI

TypePrice (per 1M)
Input tokens$0.60
Output tokens$2.40

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

About MiniMax M2.5 Highspeed

MiniMax M2.5 Highspeed is MiniMax's inference-optimized variant of M2.5, released simultaneously in February 2026. It delivers identical intelligence and outputs to standard M2.5 through a specialized inference engine at lower latency. The model supports a 204,800-token context window, 131,072-token max output, function calling, structured output, and reasoning. API model ID: MiniMax-M2.5-highspeed. It is designed for latency-sensitive interactive applications and automated agent pipelines.

Model Specs

Released2026-02-12
Context205K
ArchitectureDecoder Only

Provider

Novita AI