Qwen3-Next-80B-A3B-Instruct-MLX-4bit on M3 Ultra (60-core GPU) + 96GB unified
M3 Ultra (60-core GPU) + 96GB unified
suite suite-v1
cli 0.0.1-dev
signedG8xb3zMu3+…
Embed badgesubmitted May 12, 2026Workload results
| Workload | Backend | Model | decode tok/s | prefill tok/s | TTFT | p50 | p95 |
|---|---|---|---|---|---|---|---|
| chat-short | mlx@0.31.3 | lmstudio-community-Qwen3-Next-80B-A3B-Instruct-MLX-4bit4bit | 80.34tok/s | 24.49tok/s | 4,493ms | 12.4ms | 12.6ms |
| chat-long | mlx@0.31.3 | lmstudio-community-Qwen3-Next-80B-A3B-Instruct-MLX-4bit4bit | 77.63tok/s | 1,608.6tok/s | 1,956ms | 12.9ms | 13.1ms |
| concurrent-decode | mlx@0.31.3 | lmstudio-community-Qwen3-Next-80B-A3B-Instruct-MLX-4bit4bit | 78.60tok/s | — | — | 12.7ms | 12.9ms |
| agent-trace | mlx@0.31.3 | lmstudio-community-Qwen3-Next-80B-A3B-Instruct-MLX-4bit4bit | 78.41tok/s | 1,586.3tok/s | 1,318ms | 12.7ms | 12.9ms |
Reproduce on your machine
Same workload, same model, signed at your rig. The exact command that produced this run:
$ pipx install llm-speed && llm-speed bench --model qwen3-next-80b-a3b-instruct --workload 'chat-short'
Runs in about a minute. Your number lands on the leaderboard signed and linkable. How it's measured.
Embed this run
Drop the badge into a README, blog post, or signature. Each render is a backlink to the signed result.
[](https://llm-speed.com/r/r_1pl79r50ofy)Related benchmarks
- More Qwen3-Next-80B-A3B-Instruct-MLX-4bit benchmarks — every backend and rig that has run this model.
- More M3 Ultra (60-core GPU) LLM benchmarks — every model measured on this hardware.
Provenance
- Run ID
- r_1pl79r50ofy
- Fingerprint hash
- bbf15132ccbbe7d7
- Public key
- G8xb3zMu3+pznEici/TiW0gPk5qSNIYIikGCwm1rMdQ=
- Received
- 2026-05-12 21:30:26