Skip to content
Rival
Models
CompareBest ForArena
Sign Up
Sign Up

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival
Rival
Models
CompareBest ForArena
Sign Up
Sign Up

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival
  1. Home
  2. Models
  3. o1
  4. Estimate Complexity

See how o1 compares on this challenge:

  • o1 vs Claude Sonnet 4.6
  • o1 vs Gemini 3.1 Pro Preview
  • o1 vs Z.ai: GLM 5

4 models have responded to the "Estimate Complexity" challenge. Compare their approaches side-by-side on RIVAL. This response is part of RIVAL's open dataset of 5,600+ AI model responses.

Rival
Models
CompareBest ForArena
Sign Up
Sign Up
Back to o1

Estimate Complexity

O1's response to the Tests an AI's ability to make educated estimates based on technical knowledge

o1
February 28, 2025
Text

?
Prompt

Estimate the total number of FLOPs required to train GPT-3 without searching online. Explain your reasoning step by step.

o1
Response

o1 logo
o1Estimate Complexity
o1 logo
o1

Below is one way to arrive at a back-of-the-envelope estimate for the total floating-point operations (FLOPs) needed to train GPT-3 (the 175B-parameter model). Note that different sources use slightly different assumptions, so you may see figures that differ by factors of 2–3; however, the overall scale (10^23 FLOPs) remains consistent.


1. Key Quantities

  • Number of parameters (Nₚ): 175 billion (1.75 × 10^11)
  • Number of training tokens (Nₜ): ~300 billion (3.0 × 10^11)
  • Rule-of-thumb for FLOPs per parameter per token: A common rough estimate is 6 FLOPs per parameter per token for a full forward-and-backward pass with Adam-style updates.

Why 6?

  • Forward pass often taken as ~2 × Nₚ FLOPs.
  • Backward pass can be roughly twice the cost of the forward pass (~4 × Nₚ).
  • Together → 2 + 4 = 6 × Nₚ FLOPs per token.

(Exact counts differ depending on the details of the architecture, hardware efficiency, and whether you count a multiply-add as one or two FLOPs, but 6 is a convenient ballpark multiplier.)


2. Core Calculation

Using the 6 FLOPs rule of thumb:

[ ext{Total FLOPs} ;approx; 6 ; imes; N_p ; imes; N_t. ]

Plug in the numbers:

  1. (N_p = 1.75 imes 10^{11}) (175B parameters).
  2. (N_t = 3.0 imes 10^{11}) (300B tokens).
  3. Multiplier = 6.

So,

[ ext{Total FLOPs} ;=; 6 ; imes; (1.75 imes 10^{11}) ; imes; (3.0 imes 10^{11}). ]

First, multiply the “nice” numbers:

  • (6 imes 1.75 = 10.5)
  • (10.5 imes 3.0 = 31.5)

So the coefficient is 31.5. Now include the powers of ten:

  • (31.5 imes 10^{11} imes 10^{11} = 31.5 imes 10^{22} = 3.15 imes 10^{23}.)

Hence,

[ ext{Total FLOPs} ;approx; 3.15 imes 10^{23}. ]


3. Result and Interpretation

Therefore, a rough estimate for the total training cost of GPT-3 (175B) is on the order of
[ mathbf{3 imes 10^{23}} ext{ FLOPs.} ]

Depending on exact architectural details, optimizer choices, hardware efficiency, and definitions of FLOPs vs. MACs (multiply-accumulate operations), public estimates generally range from around (2 imes 10^{23}) to (4 imes 10^{23}). But the important takeaway is that it is a few times (10^{23}) FLOPs.


4. Summary

  1. Identify parameters and tokens: 175B parameters, 300B tokens.
  2. Adopt the “6 FLOPs per parameter per token” rule for forward + backward + update.
  3. Multiply: (6 imes 1.75 imes 10^{11} imes 3.0 imes 10^{11} approx 3.15 imes 10^{23}).

This simple calculation gives a ballpark figure without delving into the fine details of the GPT-3 architecture or specifics of training hardware.

Estimate Complexity

Turn this model response into notes, narration, or a short video

Auto NotesTranscriptNarrateVideo Cut

Partner link

This output is part of our open dataset — 5K+ AI responses across 200 models

About o1

Capabilities

ConversationWeb DesignCode GenerationReasoning

Categories

TextMultimodal

Specifications

Provider
Openai
Released
2024-12-05
Size
MEDIUM
Parameters
175B
Context
128,000 tokens

Keep exploring

SAME PROMPT

Claude Sonnet 4.6's version

Same prompt, different result

COMPARE

o1 vs Gemini 3.1 Pro Preview

Both outputs, side by side

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival