Skip to content
llm-speed

Best rig for Qwen3-Coder-Next

Qwen3-Coder-Next is an 80B-parameter MoE with ~3B active. The activation pattern means it punches above its weight on Apple Silicon and sane consumer GPUs. Here's the data we have.

No data submitted for this task yet.

Run the suite to be the first benchmark for this guide:

$ pipx install llm-speed && llm-speed bench

Qwen3-Coder-Next's 80B-A3B mixture-of-experts means raw VRAM matters less than for a dense model of comparable footprint, but you still need ~40-50 GB of memory for a 4-bit quant. That puts it in reach of an M-series Ultra, a pair of 24 GB consumer GPUs, or a single MI300X / H100. Below is every config we have a real submitted run for.

Side-by-side comparisons

See also: All hardware · All models · Methodology