Skip to content
llm-speed

About

About llm-speed

llm-speed is a community benchmark for LLM inference speed. One CLI, one methodology, real numbers across hosted APIs, consumer GPUs, Apple Silicon, and prosumer rigs.

How it works

Related work

These are the resources we've learned from and continue to point people toward:

Open source

The CLI, the methodology, and this site are open source under Apache-2.0. Submit issues or ideas via GitHub. Results belong to the people who submitted them.