Nemotron 3 Super — NVIDIA's 120B open model that runs at 12B efficiency (opens in new tab)
NVIDIA released Nemotron 3 Super at GTC 2026, a hybrid Mamba-Transformer MoE model with 120B parameters that activates only 12B at inference time. It ships with a 1M-token context window and is optimized for multi-agent agentic reasoning tasks. Open weights are available on Hugging Face.