⚡ LLM Inference

Speed Calculator

Select your GPU, model, and quantization — get estimated tokens/sec and VRAM usage

🖥 GPU Hardware
Search GPU…
Select GPU
🧠 LLM Model
Search model…
Select model
Select a GPU, VRAM, model, and parameters to see results