Introducing NVIDIA Nemotron 3 Super 🎉
— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026
Open 120B-parameter (12B active) hybrid Mamba-Transformer MoE model
Native 1M-token context
Built for compute-efficient, high-accuracy multi-agent applications
Plus, fully open weights, datasets and recipes for easy customization and… pic.twitter.com/kMFI23noFc
This latest addition to the Nemotron family isn't just a bigger Nano.
— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026
✅ Up to 5x higher throughput and 2x accuracy than the previous version
✅ Latent MoE that calls 4x as many expert specialists for the same inference cost⁰
✅ Multi-token prediction that dramatically reduces… pic.twitter.com/18KgqdN0H4
🦞These innovations come together to create a model that is well suited for long-running autonomous agents.
— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026
On PinchBench—a benchmark for evaluating LLMs as @OpenClaw coding agents—Nemotron 3 Super scores 85.6% across the full test suite, making it the best open model in its… pic.twitter.com/01R0oImsJb
“NVIDIA Nemotron 3 Super: The new leader in open, efficient intelligence”https://t.co/JN3iEX3A35
— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026
Ready to get started?
— NVIDIA AI Developer (@NVIDIAAIDev) March 11, 2026
Nemotron 3 Super supports deployment across environments, from workstations to the cloud, and can be accessed through API, OpenRouter, or https://t.co/fC1rz1G9c4.
It is now live and available on major inference platforms, packaged as NVIDIA NIM:
📥…
