Token Speed Playground

Visualize how different token generation speeds affect text output

Loading tokenizer (Qwen3-0.6B)...

Simulation 1
tok/s
0
Current
0.0s
Time
0
Total
Ready
Simulation 2
tok/s
0
Current
0.0s
Time
0
Total
Ready
Simulation 3
tok/s
0
Current
0.0s
Time
0
Total
Ready
Simulation 4
tok/s
0
Current
0.0s
Time
0
Total
Ready

How it works: This tool tokenizes your input text and simulates streaming output at different speeds.

Typical LLM inference speeds range from 10-150+ tokens/second depending on the model size, hardware, and optimization. Watch how the perception of "speed" changes dramatically between different token rates.