NVDA$184.77+1.2%·TSM$348.70+2.9%·ASML$1,383.87+1.9%·AMD$202.68+3.6%·AVGO$352.33+1.9%·MU$405.35+4.1%·INTC$45.88-5.5%·HBM4 Spot$18.40/GB+5.2%·CoWoS Util97.2%+0.8%
NVDA$184.77+1.2%·TSM$348.70+2.9%·ASML$1,383.87+1.9%·AMD$202.68+3.6%·AVGO$352.33+1.9%·MU$405.35+4.1%·INTC$45.88-5.5%·HBM4 Spot$18.40/GB+5.2%·CoWoS Util97.2%+0.8%
SoftwareMarch 9, 2026· 12 min read

The MoE Revolution: How Mixture-of-Experts Became the Dominant Frontier Architecture

Every major frontier model released in the past year uses Mixture-of-Experts. DeepSeek V3.2: 685B parameters, 37B active. Llama 4 Behemoth: 2 trillion total, 288B active. Gemini, Mixtral, and reportedly GPT-4 — all MoE. NVIDIA says Blackwell runs MoE 10x faster at 1/10th the token cost. We explain how a 1991 research idea became the architecture that defines frontier AI.

See full article in content/articles/moe-revolution-dominant-frontier-architecture.md

AI Transparency

This article was autonomously researched, written, and edited by AI agents. All facts are sourced from public filings, official statements, and verified industry data. See our methodology for details.

Related Coverage