LLM ReferenceLLM Reference
DeepInfra

Nemotron 3 Super-120B-A12B on DeepInfra

Nemotron 3 · NVIDIA AI

Serverless

Why use Nemotron 3 Super-120B-A12B on DeepInfra?

DeepInfra offers Nemotron 3 Super-120B-A12B with pay-as-you-go pricing at $0.10/1M input tokens. DeepInfra is a cloud inference platform offering cost-effective access to open-source AI models.

Compare Nemotron 3 Super-120B-A12B across 3 providers to find the best fit for your use case

Compare Nemotron 3 Super-120B-A12B Across Providers

ProviderInput (per 1M)Output (per 1M)
DeepInfra$0.10$0.50
NVIDIA NIM
OpenRouter$0.09$0.45

Pricing

TypePrice (per 1M)
Input tokens$0.10
Output tokens$0.50

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

About Nemotron 3 Super-120B-A12B

120B MoE model with 12B active parameters - hybrid Mamba-Transformer with Multi-Token Prediction for agentic workflows

FAQ

What does Nemotron 3 Super-120B-A12B cost on DeepInfra?

On DeepInfra, Nemotron 3 Super-120B-A12B costs $0.1 per 1M input tokens and $0.5 per 1M output tokens.

What is the context window for Nemotron 3 Super-120B-A12B on DeepInfra?

Nemotron 3 Super-120B-A12B supports a 1,048,576 token context window on DeepInfra.

How does DeepInfra compare to other Nemotron 3 Super-120B-A12B providers?

Nemotron 3 Super-120B-A12B is available from 3 providers. The cheapest input pricing is $0.09/1M tokens from OpenRouter.

Who created Nemotron 3 Super-120B-A12B?

Nemotron 3 Super-120B-A12B was created by NVIDIA AI as part of the Nemotron 3 model family.

Is Nemotron 3 Super-120B-A12B open source?

Nemotron 3 Super-120B-A12B's open source status is unknown in the seed data.

Get Started

Model Specs

Released2026-03-11
Parameters120B
Context1M
ArchitectureDecoder Only

GPU-Hour Providers(1)