Skip to content
Rival
Models
CompareBest ForArena
Lab
Sign Up
Sign Up

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Prompt Lab
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Challenges

Discover

  • Insights
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • Rival Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don’t own
Rival
Models
CompareBest ForArena
Lab
Sign Up
Sign Up

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Prompt Lab
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Challenges

Discover

  • Insights
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • Rival Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don’t own

GPT-5.1 Codex Max — Estimate Complexity

  1. Home
  2. Models
  3. GPT-5.1 Codex Max
  4. Estimate Complexity

See how GPT-5.1 Codex Max compares on this challenge:

  • GPT-5.1 Codex Max vs Claude Sonnet 4.6
  • GPT-5.1 Codex Max vs Google: Gemini 3.1 Flash Lite Preview
  • GPT-5.1 Codex Max vs Gemini 3.1 Pro Preview

4 models have responded to the "Estimate Complexity" challenge. Compare their approaches side-by-side on Rival. This response is part of Rival's open dataset of 5,600+ AI model responses.

Rival
Models
CompareBest ForArena
Lab
Sign Up
Sign Up
Back to GPT-5.1 Codex Max

Estimate Complexity

Gpt 5.1 Codex Max's response to the 'Estimate Complexity' challenge.

GPT-5.1 Codex Max
December 5, 2025
Text

?
Prompt

Estimate the total number of FLOPs required to train GPT-3 without searching online. Explain your reasoning step by step.

GPT-5.1 Codex Max
Response

gpt-5.1-codex-max logo
GPT-5.1 Codex MaxEstimate Complexity
gpt-5.1-codex-max logo
GPT-5.1 Codex Max

What you need for a back‑of‑the‑envelope is:

  • How big the model is. The published “GPT‑3” has (N\approx1.75\times10^{11}) trainable weights. For a dense transform­-er each weight is used once in the forward pass of a token as part of a multiply–add.
  • How much data it sees. OpenAI say they trained on the order of (3\times10^{11}) tokens.
  • The cost per weight per token. A multiply–add is two floating‑point operations; back‑propagating the gradient roughly doubles the cost again. A common rule of thumb for transformers is that the total training cost is ~6 FLOPs per parameter per token.

You can justify that 6 in a more explicit way by summing over a single layer. For GPT‑3’s configuration ((d_{\text{model}})=12 288, (d_\text{ff})=49 152, 96 layers, (L)=2 048):

  • Q/K/V/out projections: (4d_{\text{model}}^2\approx6.0\times10^8) multiply–adds
  • Feed‑forward: (2d_{\text{model}}d_\text{ff}\approx1.2\times10^9) multiply–adds
  • Attention score/application: (\mathcal{O}(L d_{\text{model}})\approx5.0\times10^7) multiply–adds

Summing and converting multiply–adds to FLOPs gives (\sim3.7\times10^9) FLOPs per layer per token. Across 96 layers that’s (\sim3.6\times10^{11}) FLOPs for a forward pass on one token; including backward pass multiplies that by ≈3, giving (\sim1.1\times10^{12}) FLOPs per token. Multiply by (3\times10^{11}) tokens and you get (\sim3\times10^{23}) FLOPs.

Using the simpler 6 × (N,T) rule of thumb: (6\times1.75\times10^{11}\times3\times10^{11}\approx3.15\times10^{23}) floating‑point operations.

Estimate Complexity

Turn this model response into notes, narration, or a short video

Auto NotesTranscriptNarrateVideo Cut

Partner link

This output lives in our open dataset — 5K+ responses across 200 models. Take it. It’s free.

About GPT-5.1 Codex Max

Capabilities

ConversationReasoningCode GenerationAnalysis

Categories

TextCodeMultimodal

Specifications

Provider
Openai
Released
2025-12-04
Size
XLARGE

Keep exploring

SAME PROMPT

Claude Sonnet 4.6's version

Same prompt, different result

COMPARE

GPT-5.1 Codex Max vs Google: Gemini 3.1 Flash Lite Preview

Both outputs, side by side

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Prompt Lab
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Challenges

Discover

  • Insights
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • Rival Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don’t own