o3 Mini vs Grok 3
Compare o3 Mini by OpenAI against Grok 3 by xAI, context windows of 64K vs 128K, tested across 28 shared challenges. Updated February 2026.
Compare o3 Mini by OpenAI against Grok 3 by xAI, context windows of 64K vs 128K, tested across 28 shared challenges. Updated February 2026.
28 challenges
Tests an AI's ability to make educated estimates based on technical knowledge
To estimate the total number of FLOPs (Floating Point Operations) required to train GPT-3, we’ll need to make some reasonable assumptions based on what’s generally known about large language models like GPT-3, its architecture, and typical training procedures. GPT-3, developed by OpenAI, is a massive model with 175 billion parameters, and while exact training details aren’t fully public, we can piece together an estimate using standard practices in machine learning.
The number of FLOPs to train a model depends on:
For transformer models like GPT-3, training involves both a forward pass (computing predictions) and a backward pass (computing gradients). A common heuristic in the field is that:
For GPT-3 with 175 billion parameters, processing one token requires: [ 6 imes 175 imes 10^9 = 1.05 imes 10^{12} ext{ FLOPs per token} ]
GPT-3 was reportedly trained on a massive corpus, often cited as hundreds of billions of tokens. Let’s assume 300 billion tokens, a plausible figure based on discussions about large language model datasets (e.g., Common Crawl, books, and other sources). This is the total number of tokens processed during training, factoring in how data is sampled over epochs.
If the model processes 300 billion tokens once, the total FLOPs would be: [ 1.05 imes 10^{12} ext{ FLOPs/token} imes 300 imes 10^9 ext{ tokens} = 3.15 imes 10^{23} ext{ FLOPs} ] This assumes one full pass through the dataset, but training typically involves multiple passes or iterations over the data, adjusted by batch size and optimization.
Training isn’t a single pass—it involves multiple updates to converge. The number of training steps depends on:
Let’s assume the model effectively processes 900 billion tokens (3× the dataset size) over training. Then: [ 1.05 imes 10^{12} ext{ FLOPs/token} imes 900 imes 10^9 ext{ tokens} = 9.45 imes 10^{23} ext{ FLOPs} ]
GPT-3 is a sparse transformer in some variants (e.g., earlier GPT models used sparsity), but the 175B version is dense. The 6 FLOPs/parameter/token heuristic holds for dense models. We’re not factoring in optimizations like mixed-precision training (FP16), which reduces compute slightly but doesn’t change the order of magnitude significantly.
Based on this reasoning, the total FLOPs to train GPT-3 is approximately: [ 9.45 imes 10^{23} ext{ FLOPs} ] or roughly ( 10^{24} ) FLOPs when rounded for simplicity.
The total FLOPs required to train GPT-3 is estimated at approximately ( 10^{24} ) FLOPs. This is a rough estimate, but it aligns with the scale of compute expected for a model of this size and training complexity.
Tests an AI's ability to understand game rules and strategy
Tests an AI's understanding of number representation
No, 9.11 is not greater than 9.9. In decimal comparison, 9.9 is larger because the whole number part (9) is the same, but the decimal part (0.9) is greater than (0.11). Numerically, 9.9 > 9.11.
Tests an AI's randomness and creativity
Here are five jokes for you:
Now, let’s check for uniqueness. Joke 1 and Joke 3 both start with "Why don’t skeletons fight each other?" but have different punchlines ("don’t have the guts" vs. "too busy rattling around"). They’re technically distinct due to the different resolutions, though they share a setup. The other three jokes (2, 4, and 5) have completely unique setups and punchlines.
So, all five jokes are unique in terms of their full content, even if 1 and 3 share a premise. Final count: 5 unique jokes.
Tests an AI's ability to generate vector graphics
Tests an AI's ability to create detailed SVG illustrations of gaming hardware
Tests an AI's humor and creative writing ability
Hey, everyone—thanks for being here tonight. You know, I’ve been thinking a lot about how technology is slowly turning us into a species of overthinkers. Take smartphones, for instance. Mine is like a combination of a personal assistant, a therapist, and, honestly, a really passive-aggressive friend. I ask it for the weather, and it’s like, “Sure, here’s the forecast, and by the way, you might want to wear a jacket because you’re clearly underprepared for life.” I mean, I just want to know if I need an umbrella, not a life lesson!
Then there’s social media. It’s a place where every day feels like you’re watching a highlight reel of someone else’s life. One minute you’re scrolling past your friend’s vacation photos—sunsets, exotic meals, spontaneous adventures—and the next, you’re wondering if your own life is secretly sponsored by “Take Your Time, You’re Not Late Yet” Inc. When did getting out of bed become a heroic feat? I mean, there I am, battling my alarm clock like it’s the final boss in a video game, and every snooze feels like an epic victory. Spoiler alert: even the boss has a victory dance when you finally win—by stumbling out the door with two mismatched shoes.
And adulthood? Oh, adulthood is a wild ride. As a kid, I used to imagine that being an adult meant making your own rules and living life on your own terms. Now, I’ve learned that being an adult is mostly about Googling “How do I fix a leaky faucet?” at 2 AM and then wondering if you’re ever going to find the time for a real conversation. It’s like we’re all just improvising, hoping no one notices we’re winging it.
Dating these days feels a bit like shopping online for a used car. You swipe through profiles, examine every detail, and then you’re left wondering if the real engine is under the hood—or if there’s a hidden problem that no one mentions until you’re already halfway down the driveway. And the reviews? They’re written in vague terms like “decent mileage” or “runs well,” which in the dating world translates to, “I might be fun on a good day, but don’t get your hopes up.”
At the end of the day, I’ve realized that life is just a series of awkward moments strung together by our attempts at being perfect. We all have those days where everything goes hilariously wrong, and that’s okay. Embracing the chaos and laughing at our mishaps is what keeps us sane. So here’s to finding humor in our everyday struggles—because if we can’t laugh at our own absurdity, who will?
Thanks for listening, everyone—have a great night!
Tests an AI's ability to generate a complete, working landing page
Recreate an interactive, classic Mario level in a single HTML file.
Tests an AI's ability to replicate an existing UI with Tailwind CSS
Tests an AI's ability to create smooth web animations
Tests an AI's ability to create interactive web elements