UABUnbiased AI BenchGlass box for model evals.
Every leaderboard, with receipts.
Home/Benchmarks/Speed throughput
Speed throughput
Live · updated continuously
Browse sectionsSpeed throughput
Benchmarks · /benchmarks/bridgebench-throughput

Speed throughput

BridgeBench speed leaderboard using coding-oriented output throughput.
Source · BridgeBench
Version · bridgebench snapshot 2026-05-01
Scores · 11

Passport

Verified but agingThis is an efficiency signal, so it belongs beside quality rather than being mistaken for quality.
source
BridgeBench
metric
Throughput (t/s)
judge
Speed / cost
direction
higher better
group id
bridgebench_throughput_2026_04
domain
Coding

What it measures vs what it misses

✓ Measures

How quickly a model emits coding output once generation is underway.

✗ Misses

Output correctness. Editing quality.

Why this countsIt tells you whether the model can generate, repair, and reason over code under evaluator pressure rather than marketing examples.Comparable-group ruleThis percentile only compares models inside the exact benchmark/version group shown here. It is not a universal score.What it missesIt does not fully capture repo-scale iteration, IDE ergonomics, or long debugging loops.

Leaderboard · this benchmark version

#1 · Grok 4.3
BB · Undated
302.1 t/s
#2 · Grok 4.20
BB · Undated
243.3 t/s
#3 · GPT-5.4 mini
BB · Undated
236.4 t/s
#4 · GPT-5.4 nano
BB · Undated
227.8 t/s
#5 · GPT-5.5
BB · Undated
152.3 t/s
#6 · Gemini 3.1 Pro Preview
BB · Undated
122.2 t/s
#7 · Claude Opus 4.7
BB · Undated
116.4 t/s
#8 · Claude Sonnet 4.6
BB · Undated
95.3 t/s
#9 · Claude Opus 4.6
BB · Undated
92.2 t/s
#10 · GPT-5.4
BB · Undated
88 t/s
#11 · Grok 4
BB · Undated
61.4 t/s