Best rig for Qwen3-Coder-Next
Qwen3-Coder-Next is an 80B-parameter MoE with ~3B active. The activation pattern means it punches above its weight on Apple Silicon and sane consumer GPUs. Here's the data we have.
No data submitted for this task yet.
Run the suite to be the first benchmark for this guide:
$ pipx install llm-speed && llm-speed bench
Qwen3-Coder-Next's 80B-A3B mixture-of-experts means raw VRAM matters less than for a dense model of comparable footprint, but you still need ~40-50 GB of memory for a 4-bit quant. That puts it in reach of an M-series Ultra, a pair of 24 GB consumer GPUs, or a single MI300X / H100. Below is every config we have a real submitted run for.
Side-by-side comparisons
See also: All hardware · All models · Methodology