Estimate Complexity
Gpt 5.1 Codex's response to the 'Estimate Complexity' challenge.
GPT-5.1-Codex
November 13, 2025
Text
?Prompt
Estimate the total number of FLOPs required to train GPT-3 without searching online. Explain your reasoning step by step.
Response
To estimate the total FLOPs (floating-point operations) required to train GPT-3, we can use a common approximation for transformer models:
Total FLOPs ≈ 6 × N<sub>params</sub> × N<sub>tokens</sub>
Where:
- N<sub>params</sub> is the number of model parameters.
- N<sub>tokens</sub> is the number of training tokens.
- The factor of 6 accounts for forward and backward passes, along with other computations involved in training.
For GPT-3:
- N<sub>params</sub> ≈ 175 billion = 1.75 × 10¹¹
- N<sub>tokens</sub> ≈ 300 billion = 3.0 × 10¹¹
Plugging in:
6 × (1.75 × 10¹¹) × (3.0 × 10¹¹)
= 6 × 5.25 × 10²²
= 3.15 × 10²³ FLOPs
So the total estimated compute to train GPT-3 is ~3.15 × 10²³ FLOPs.
This is a rough but widely used approximation for large-scale transformer training.
About GPT-5.1-Codex
Capabilities
ConversationReasoningCode GenerationAnalysis
Categories
TextCodeMultimodal
Specifications
- Provider
- Openai
- Released
- 2025-11-13
- Size
- XLARGE
- Context
- 400,000 tokens