How we calculate, verify, and present performance data so you can size your local AI rigs with confidence.
Transparency is critical when helping you choose hardware. Here's exactly how we calculate, verify, and present performance data.
Hardware we've personally tested in controlled conditions. These are the gold standard.
Criteria: Consistent across 5+ runs, documented test conditions, verified hardware configuration.
Results submitted by the community and verified by our moderation team.
Criteria: Screenshot provided, results match expected range (±30%), hardware configuration verified.
Algorithmic estimates based on hardware specifications. Use as a rough guide only.
Accuracy: Typically within ±40% of real-world performance. Better for relative comparisons than absolute numbers.
When real benchmark data isn't available, we use this formula:
tokensPerSec = (GPU_cores / 10,000) × 50 × model_penalty × quant_multiplier
Estimated benchmarks can be off by 40% or more. That's why we're building a community-driven benchmark database.
~35 tokens/sec
RTX 4080 + Llama 70B Q4
28 tokens/sec
Real test result (insufficient 16GB VRAM)
Help make this data better for everyone. When you submit a benchmark:
We're continuously improving our methodology:
Questions about our methodology? Contact us