Skip to content
llm-speed

Best GPU for local LLMs under $2,000

RTX 4090, RTX 5080, used 3090, RX 7900 XTX, Arc B580. Here's where each lands on real workloads.

No data submitted for this task yet.

Run the suite to be the first benchmark for this guide:

$ pipx install llm-speed && llm-speed bench

The under-$2k tier is the sweet spot for prosumer local LLMs. You're choosing among a new RTX 5090 at MSRP, a used 4090, a pair of 3090s, an AMD RX 7900 XTX, or a budget Arc B580. Each has a different memory-bandwidth-to-VRAM ratio and a different software story (CUDA vs ROCm vs SYCL). Below are submitted benchmarks across this tier.

Side-by-side comparisons

See also: All hardware · All models · Methodology