LLM ReferenceLLM Reference
OpenRouter

Using Ling-2.6-Flash on OpenRouter

Implementation guide · Ling 2.6 · InclusionAI

Serverless

Quick Start

  1. 1
    Create an account at OpenRouter and generate an API key.
  2. 2
    Use the OpenRouter SDK or REST API to call inclusionai/ling-2.6-flash — see the documentation for request format.
  3. 3
    You'll be billed $0.08/1M input, $0.24/1M output tokens. See full pricing.

Code Examples

See OpenRouter documentation for integration details.

About OpenRouter

OpenRouter provides a unified interface for Large Language Models with better pricing, improved uptime, and no subscription requirements. Route across providers for cost optimization and reliability.

Multi-provider LLM aggregator offering unified API access to 300+ models from all major labs and emerging providers, with automatic failover for reliability.

Pricing on OpenRouter

TypePrice (per 1M)
Input tokens$0.08
Output tokens$0.24

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

About Ling-2.6-Flash

InclusionAI's efficient 104B MoE instruct model with only 7.4B active parameters per token. Purpose-built for agentic workflows requiring fast responses and high token efficiency. Achieves 59.3% on GPQA Diamond. Nearly double the Artificial Analysis Intelligence Index score of comparable open-weight models. Available free on OpenRouter (inclusionai/ling-2.6-flash:free).

Model Specs

Released2026-04-21
Parameters104B (7.4B activated)
Context262K
Architecturemoe

Provider

OpenRouter
OpenRouter

OpenRouter, Inc.

New York, NY, USA