Nemotron 3 Nano Omni
nemotron-3-nano-omni
Open SourceMultimodal
About
NVIDIA Nemotron 3 Nano Omni is an open-weight 30B hybrid MoE multimodal model (3B active parameters) that natively accepts text, image, video, and audio inputs in a single inference loop. Built on a hybrid Mamba-Transformer architecture with 23 Mamba-2 layers, 23 MoE layers (128 experts, 6+1 active), and 6 GQA layers, plus Conv3D video layers and Efficient Video Sampling (EVS). Delivers up to 9x higher throughput than comparable omni models. Supports a 256K context window and a 16,384 reasoning budget. Open weights, datasets, and training recipes released under a permissive license.
Nemotron 3 Nano Omni has a 256K-token context window.
Capabilities
VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution
Specifications
FamilyNemotron-3
Released2026-04-28
Parameters30B
Context256k
ArchitectureHybrid Mamba-Transformer MoE
Specializationomni