GPT OSS 20B vs xAI: Grok 4
Compare GPT OSS 20B by OpenAI against xAI: Grok 4 by xAI, context windows of 131K vs 256K, tested across 38 shared challenges. Updated February 2026.
Compare GPT OSS 20B by OpenAI against xAI: Grok 4 by xAI, context windows of 131K vs 256K, tested across 38 shared challenges. Updated February 2026.
38 challenges
Tests an AI's ability to make educated estimates based on technical knowledge
Below, I'll estimate the total number of floating-point operations (FLOPs) required to train GPT-3 from scratch, based on publicly known details about the model (from its original release and related analyses) and standard formulas used in machine learning research for estimating training compute. I'll explain my reasoning step by step, including the assumptions I'm making and why they're reasonable. Note that this is an estimation—exact numbers aren't publicly disclosed in full detail by OpenAI, and real-world training involves optimizations, inefficiencies, and hardware-specific factors that could adjust the final figure. I'm not searching online; this is based on my pre-existing knowledge of large language model (LLM) training.
These numbers are widely cited in AI research papers and discussions around scaling laws (e.g., work from OpenAI and DeepMind on model scaling).
Training a transformer-based LLM like GPT-3 involves:
A widely used rule of thumb from AI scaling research (e.g., papers like "Scaling Laws for Neural Language Models" by Kaplan et al. from OpenAI) estimates the total training FLOPs as:
Total FLOPs ≈ 6 × P × T
Where:
This formula is a simplification but matches empirical reports for models like GPT-2, GPT-3, and others. It focuses on the dominant compute cost (parameter updates) and doesn't include inference-time FLOPs, data loading, or evaluation— just training.
Why 6? It's derived from counting operations in transformer layers (attention, feed-forward networks). Some papers use slightly different multipliers (e.g., 3× or 4× if only counting multiply-accumulates), but 6× is the standard for total FLOPs in scaling law discussions.
Total FLOPs = 6 × (1.75 × 10^11) × (3 × 10^11)
First, compute 1.75 × 3 = 5.25
Then, 5.25 × 10^(11+11) = 5.25 × 10^22
Finally, 6 × 5.25 × 10^22 = 31.5 × 10^22 = 3.15 × 10^23
So, my estimate is approximately 3 × 10^23 FLOPs (rounding to one significant figure for simplicity, as these are ballpark numbers).
If I adjust for potential underestimation (e.g., higher token count or overhead), the upper bound might be 5 × 10^23 FLOPs, but I stick with 3 × 10^23 as the core estimate.
To make this relatable:
Based on this reasoning, I estimate the total FLOPs for training GPT-3 at around 3 × 10^23. This is consistent with what I'd expect from scaling laws and matches informal discussions in the AI community. If new details emerged (e.g., exact token counts), it could shift by a factor of 2, but the order of magnitude (10^23) is solid. If you have additional details or want me to adjust assumptions, let me know!
Tests an AI's ability to solve a simple but potentially confusing logic puzzle
Tests an AI's randomness and creativity
Here are five jokes for you:
Why don’t skeletons fight each other?
They don’t have the guts.
What does a cloud wear under its jacket?
Thunderwear.
Why did the scarecrow win an award?
Because he was outstanding in his field.
What do you call a fake noodle?
An impasta.
Why did the math book look sad?
Because it had too many problems.
How many are unique?
All five jokes are unique – 5 distinct jokes.
Tests an AI's ability to generate vector graphics
Tests an AI's ability to create detailed SVG illustrations of gaming hardware
Tests an AI's ability to simulate personalities and predict future trends
Tests an AI's humor and understanding of current events
Tests an AI's ability to write in distinct character voices
Tests an AI's ability to generate a complete, working landing page
Recreate an interactive, nostalgic Pokémon battle UI in a single HTML file.
Recreate an interactive, classic Mario level in a single HTML file.
Tests an AI's ability to replicate an existing UI with Tailwind CSS