moe
4 articles
Paweł Huryn Claims: Holo3 with 3B Active Parameters Beats GPT-5.4 and Opus 4.6 at Computer Use
Paweł Huryn posted on X claiming H Company's Holo3 beat GPT-5.4 and Opus 4.6 at computer use tasks with just 3B active parameters. He says it's a sparse MoE fine-tuned from Qwen3.5 and could theoretically run on a single GPU.
Why Programmers Love Codex While Vibe Coders Can't Quit Claude: Dense vs MoE Is Really a Story About Two Coding Philosophies
Berryxia uses Dense vs MoE to explain something many developers already feel: Codex often shines in bug fixing, refactors, and long-running engineering tasks, while Claude keeps winning over vibe coders. That framing captures part of the truth, but the real split is bigger than architecture — it includes training philosophy, product design, and whether you treat coding as precise delegation or interactive creation.
Running a Trillion-Parameter Model on a MacBook? The Wild SSD Streaming Experiment
Simon Willison shared a new trend in running massive MoE models on Macs: streaming expert weights from SSD instead of cramming everything into RAM. Even a trillion-parameter Kimi K2.5 runs on a 96GB MacBook Pro.
NVIDIA Nemotron 3 Super: A 120B Open-Source Model That Only Uses 12B at a Time
NVIDIA released Nemotron 3 Super, a 120B parameter open-source reasoning model with only 12B active parameters. It combines Mamba and Transformer in a hybrid MoE architecture, scores 36 on the Intelligence Index, and runs at a blistering 484 tok/s.