Skip to content
Rival
Models
CompareBest ForArena
Sign Up
Sign Up

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival
Rival
Models
CompareBest ForArena
Sign Up
Sign Up

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival
  1. Home
  2. Models
  3. OpenAI o4-mini
  4. Estimate Complexity

See how OpenAI o4-mini compares on this challenge:

  • OpenAI o4-mini vs Claude Sonnet 4.6
  • OpenAI o4-mini vs Gemini 3.1 Pro Preview
  • OpenAI o4-mini vs Z.ai: GLM 5

4 models have responded to the "Estimate Complexity" challenge. Compare their approaches side-by-side on RIVAL. This response is part of RIVAL's open dataset of 5,600+ AI model responses.

Rival
Models
CompareBest ForArena
Sign Up
Sign Up
Back to OpenAI o4-mini

Estimate Complexity

o4 Mini's response to the challenge: Estimate GPT-3 training FLOPs.

OpenAI o4-mini
April 16, 2025
Text

?
Prompt

Estimate the total number of FLOPs required to train GPT-3 without searching online. Explain your reasoning step by step.

OpenAI o4-mini
Response

o4-mini logo
OpenAI o4-miniEstimate Complexity
o4-mini logo
OpenAI o4-mini

Here's a back‐of‐the‐envelope derivation that gets us into the right ballpark (∼10²³ FLOPs total):

  1. Model size
    • GPT–3 "175 B" has about P ≃1.75×10¹¹ parameters.

  2. Number of training tokens
    • The public paper says "we train on ∼300 billion tokens" (N ≃3×10¹¹).

  3. FLOPs per token
    • Every forward+backward through the network touches all parameters.
    • Rough rule of thumb: you need O(6) FLOPs per parameter per token
    (a couple multiplies/adds in the forward pass, a similar count in the backward, and some small extra for optimizer updates). • So 6 × P FLOPs/token ≃ 6×1.75×10¹¹ ≃1.05×10¹² FLOPs/token.

  4. Total FLOPs
    • Multiply FLOPs/token by number of tokens:
    Total ≃ (1.05×10¹² FLOPs/token) × (3×10¹¹ tokens)
    ≃ 3.15×10²³ FLOPs.

  5. Sanity check
    • 3×10²³ FLOPs is 3×10⁵ PFLOP·days (since 1 PFLOP/s×86 400 s ≈ 8.6×10¹⁹ FLOPs/day).
    • That's a few thousand GPU‐days on a modern accelerator cluster—consistent with the published "thousands of V100‐GPU days" statement.

Bottom‐line estimate: on the order of 10²³–10²⁴ floating‐point operations to train GPT‑3 175 B from scratch.

Estimate Complexity

Turn this model response into notes, narration, or a short video

Auto NotesTranscriptNarrateVideo Cut

Partner link

This output is part of our open dataset — 5K+ AI responses across 200 models

About OpenAI o4-mini

Capabilities

ConversationReasoningCode GenerationAnalysis

Categories

TextCodeMultimodal

Specifications

Provider
Openai
Released
2025-04-16
Size
MEDIUM
Parameters
Not disclosed

Keep exploring

SAME PROMPT

Claude Sonnet 4.6's version

Same prompt, different result

COMPARE

OpenAI o4-mini vs Gemini 3.1 Pro Preview

Both outputs, side by side

Compare AI vibes, not scores. Side-by-side outputs across the world's best models.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Leaderboard
  • Challenges

Discover

  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • .llmignore
  • Badges
  • RIVAL Datasets

Connect

  • Methodology
  • Sponsor
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival