LLM ReferenceLLM Reference

Qwen3.6-35B-A3B

Multimodal

About

Qwen3.6-35B-A3B is an open-weight multimodal MoE model with 35B total parameters and 3B activated per token, released April 2026. It features a hybrid architecture combining Gated DeltaNet linear attention and standard Gated Attention with 256 total experts (8 routed + 1 shared), and includes a vision encoder for image and video understanding. Optimized for agentic coding, long-context reasoning, and visual tasks; supports 256K native context (extensible to ~1M via YaRN) with integrated thinking mode for multi-turn agent interactions.

Capabilities

VisionMultimodalReasoningFunction CallingTool UseStructured OutputsCode Execution

Benchmark Scores(6)

BenchmarkScoreVersionSource
SWE-bench Verified73.4Qwen3.6-35B-A3B model card (April 2026)
SWE-bench Pro49.5Qwen3.6-35B-A3B model card (April 2026)
LiveCodeBench80.4v6Qwen3.6-35B-A3B model card (April 2026)
MMLU PRO85.2Qwen3.6-35B-A3B model card (April 2026)
Google-Proof Q&A86.0diamondQwen3.6-35B-A3B model card (April 2026)
MathVista86.4miniQwen3.6-35B-A3B model card (April 2026)

Rankings

Specifications

FamilyQwen 3.6
Released2026-04-16
Parameters35
Context262K
Architecturemoe
Specializationcodingagenticreasoningvisionlong-context

Created by

AI research institute of Alibaba Group.

Hangzhou, Zhejiang, China
Founded 2017
Website