NVIDIA dropped Nemotron 3 Super at GTC last week and the spec sheet looks like a typo. 120 billion total parameters. 12 billion active at inference time. That's a 10:1 ratio between what the model knows and what it computes per token. If you've been following the Mixture of Experts trend, this is the most aggressive implementation we've seen from a major vendor.
Let me explain why the active parameter count matters more than the number on the box.
The MoE Architecture, Quickly
Traditional dense models activate every parameter for every token. GPT-3 had 175B parameters and used all 175B for every single word it generated. That's why inference was expensive and slow — you're doing matrix multiplications across the entire model every time.
Mixture of Experts (MoE) models split the computation into specialized sub-networks called "experts." A routing mechanism decides which experts to activate for each token. Most tokens only need a fraction of the total model.
Think of it like a hospital. A dense model is one doctor who knows everything and personally handles every patient. An MoE model is a hospital with 60 specialists — each patient only sees the 6 they need.
# Simplified MoE routing logic
def moe_forward(x, experts, router, top_k=2):
# Router decides which experts handle this input
gate_scores = router(x) # [batch, num_experts]
top_indices = gate_scores.topk(top_k, dim=-1).indices
# Only compute through selected experts
output = torch.zeros_like(x)
for i in range(top_k):
expert_idx = top_indices[:, i]
expert_out = experts[expert_idx](x)
output += gate_scores[:, expert_idx] * expert_out
return outputNemotron 3 Super takes this to an extreme. With only 10% of parameters active per token, you get the knowledge capacity of a 120B model with the inference cost of a 12B model. In theory.
Why Active Parameters Are What Matter
When people compare AI models, they usually cite total parameter count. "This model has 70B parameters, that one has 405B." But for MoE models, that comparison is misleading.
Your GPU doesn't care how many parameters exist in the full model — it cares how many it needs to process per token. That determines latency, throughput, and memory bandwidth requirements during inference.
Here's the practical impact. A dense 120B model needs roughly 240GB of VRAM just for weights in FP16. That's three A100-80GB GPUs minimum, often four with KV-cache overhead. Nemotron 3 Super's 12B active parameters mean each forward pass uses roughly 24GB of compute bandwidth — that's single-GPU territory.
The catch: you still need the full 120B of weights loaded in memory. The experts live on the GPU even when they're not being used. So you still need the VRAM for storage, but your compute throughput scales with the active parameter count. On modern GPUs with fast memory, the storage is the easy part. The bottleneck has always been compute.
Benchmarks That Matter
NVIDIA claims Nemotron 3 Super matches or beats Llama 3.1 70B on most coding and reasoning benchmarks while running at roughly 3x the tokens per second on equivalent hardware. That's the MoE payoff — you're doing less math per token.
On HumanEval, Nemotron 3 Super scores 84.2% compared to Llama 3.1 70B's 80.5%. On MBPP, it's 78.9% vs 76.1%. These aren't blowout numbers, but remember: this model is generating tokens at 12B-equivalent speed.
For local inference on consumer hardware, the quantized versions are where it gets interesting. A Q4 quantization brings the model down to roughly 60GB — that fits on a Mac Studio with 96GB unified memory or two RTX 4090s.
# Running Nemotron 3 Super locally with llama.cpp
./llama-server \
--model nemotron-3-super-Q4_K_M.gguf \
--ctx-size 32768 \
--n-gpu-layers 80 \
--port 8080
# Benchmark with a coding task
curl http://localhost:8080/completion \
-d '{
"prompt": "Write a Python function that implements a thread-safe LRU cache with TTL support:",
"n_predict": 512,
"temperature": 0.2
}'
# On Mac Studio M2 Ultra 96GB: ~28 tokens/sec
# On 2x RTX 4090: ~45 tokens/sec28 tokens per second for a model with 120B parameters of knowledge on a desktop machine. A year ago, that was science fiction.
What This Means for Local Inference
The MoE approach is winning the efficiency war and NVIDIA just showed everyone how far you can push it. Here's why that matters for developers.
Self-hosted AI becomes practical. Companies that can't send code to external APIs now have a viable option. A $5,000 workstation running Nemotron 3 Super provides coding assistance comparable to cloud models without data leaving the building. The GPU memory game changes. The next generation of consumer GPUs (rumored RTX 5090 with 32GB VRAM) could run smaller MoE models entirely on one card. The democratization of capable local AI is accelerating. The parameter count arms race is over. We're not going to see a 10-trillion-parameter dense model. The future is smarter routing, better expert specialization, and more efficient use of existing parameters. NVIDIA is betting on this hard.The Caveats
MoE models aren't free of tradeoffs. Expert imbalance is a real problem — some experts get overused while others are rarely activated, wasting capacity. The routing mechanism adds latency overhead. And MoE models are harder to fine-tune because you need to maintain expert specialization during training.
NVIDIA hasn't released the training details for Nemotron 3 Super, which is unusual. We don't know the training data mix, the expert count, or the routing strategy beyond "hybrid MoE." The open-source community will need to reverse-engineer some of this.
Also worth noting: NVIDIA is both the model creator and the hardware vendor. Their benchmarks run on NVIDIA hardware with NVIDIA-optimized inference stacks. Independent benchmarks from the community will tell the real story.
The Bottom Line
Nemotron 3 Super is the best argument yet for MoE architectures in production. The 120B-total, 12B-active split hits a sweet spot between capability and cost that makes local inference genuinely competitive with cloud APIs for many use cases.
If you're running inference infrastructure, this model should be on your evaluation list. If you're a developer interested in local AI, the quantized versions are worth experimenting with today. And if you're still comparing models by total parameter count, it's time to update your mental model.
The question is no longer "how big is the model?" It's "how smart is the routing?"
