API Status

cerebras Provider Benchmarks

Comprehensive performance summary covering 4 models.

This provider hub highlights throughput and latency trends across every cerebras model monitored by LLM Benchmarks. Use it to compare hosting tiers, track regressions, and discover the fastest variants in the catalogue.

Visit cerebras Official Website

Provider Snapshot

Models Tracked

4

Avg Tokens / Second

206.25

Avg Time to First Token (ms)

642.50

Last Updated

Mar 8, 2026

Key Takeaways

  • 4 cerebras models are actively benchmarked with 619 total measurements across 410 benchmark runs.

  • qwen-3-32b leads the fleet with 256.00 tokens/second, while llama-3.3-70b delivers 182.00 tok/s.

  • Performance varies by 40.7% across the cerebras model lineup, indicating diverse optimization strategies for different use cases.

  • Avg time to first token across the fleet is 642.50 ms, showing good responsiveness for interactive applications.

  • The cerebras model fleet shows consistent performance characteristics (14.1% variation coefficient), indicating standardized infrastructure.

Fastest Models

ProviderModelAvg Toks/SecMinMaxAvg TTF (ms)
cerebrasqwen-3-32b256.004.77444.00400.00
cerebrasllama-3.1-8b194.001.54348.00970.00
cerebrasgpt-oss-120b193.0013.00380.00800.00
cerebrasllama-3.3-70b182.0017.40316.00400.00

All Models

Complete list of all cerebras models tracked in the benchmark system. Click any model name to view detailed performance data.

ProviderModelAvg Toks/SecMinMaxAvg TTF (ms)
cerebrasqwen-3-32b256.004.77444.00400.00
cerebrasgpt-oss-120b193.0013.00380.00800.00
cerebrasllama-3.3-70b182.0017.40316.00400.00
cerebrasllama-3.1-8b194.001.54348.00970.00

Featured Models

Frequently Asked Questions

Based on recent tests, qwen-3-32b shows the highest average throughput among tracked cerebras models.

This provider summary aggregates 619 individual prompts measured across 410 monitoring runs over the past month.