Aurora Alpha vs Gemini 2.5 Flash Preview (thinking)
Compare Aurora Alpha by OpenRouter against Gemini 2.5 Flash Preview (thinking) by Google AI, context windows of 128K vs 1.0M, tested across 20 shared challenges. Updated February 2026.
Compare Aurora Alpha by OpenRouter against Gemini 2.5 Flash Preview (thinking) by Google AI, context windows of 128K vs 1.0M, tested across 20 shared challenges. Updated February 2026.
20 challenges
Tests an AI's ability to understand game rules and strategy
Tests an AI's randomness and creativity
Here are five jokes for you:
Why did the scarecrow win an award?
Because he was outstanding in his field!
What do you call fake spaghetti?
An impasta.
Why don’t scientists trust atoms?
Because they make up everything.
How does a penguin build its house?
Igloos it together.
What’s orange and sounds like a parrot?
A carrot.
How many are unique?
All five jokes are distinct, so there are 5 unique jokes.
Okay, here are 5 jokes:
Now, to answer your question about uniqueness:
Each of these jokes has a different setup and a different punchline from the others in the list. They are distinct jokes.
Therefore, there are 5 unique jokes within this list of 5.
Tests an AI's ability to generate vector graphics
Tests an AI's ability to create detailed SVG illustrations of gaming hardware
Tests an AI's humor and creative writing ability
[Opening – 0:00‑0:30]
Hey, everyone! Great to see you all here tonight. I just got back from a week of “working from home,” which is a fancy way of saying my cat now thinks I’m a permanent fixture on the couch. She’s started charging me rent—she’s got a little bowl of kibble on the coffee table and a sign that says “No humans beyond this point.” I’m not sure if she’s being demanding or just trying to get me to finally stop pretending I’m a professional chef.
[The “Fitness” Saga – 0:30‑1:15]
Speaking of professional chef, I tried a new fitness app that promises “personalized workouts.” The first day it asked me, “What’s your goal?” I said, “I want to be able to lift the grocery bag without groaning.” The app replied, “Great! Let’s start with a 5‑minute warm‑up.” Five minutes later I was on the floor, questioning every life choice that led me to this point. The app then suggested a “quick stretch.” I stretched my patience, my budget, and my belief that I could ever be a “runner.” The only thing I ran was out of time for a proper nap.
[Zoom Meetings – 1:15‑2:00]
Zoom has become the new office. The best part? You can wear a suit on the top and pajamas on the bottom, and no one will ever know. The worst part? The “mute” button. I’ve been muted so many times I’m convinced my boss thinks I’m a mime. One day I tried to speak up about a project, and the whole team heard the sound of my cat knocking over a plant. I think the plant was more articulate than I am. And there’s always that one person who forgets to turn off their background. I’ve seen a beach, a spaceship, and a very convincing replica of a 1970s kitchen—all while they’re discussing quarterly earnings. I’m waiting for the day someone shows up with a live marching band in the background. That would be the ultimate “I’m not paying attention” signal.
[Grocery Store Adventures – 2:00‑2:45]
I went to the grocery store the other day and realized it’s basically a social experiment. The “express lane” is a myth. I stood there for ten minutes while someone in front of me tried to decide if a single avocado was “ripe enough.” Meanwhile, the person behind me was silently judging my choice of cereal—because apparently, buying “Chocolate Cheerios” is a statement about my life choices. And then there’s the “self‑checkout.” I’m convinced the machine is secretly judging me. Every time I scan an item, it beeps like it’s saying, “Really? Another bag of chips? You’ve got a whole pantry, buddy.” I tried to scan a banana, and it refused, insisting I needed a “fruit‑to‑human interaction” license.
[Closing – 2:45‑3:00]
So, here’s the takeaway: life is a series of tiny, absurd moments that we all pretend to have under control. Whether it’s a cat demanding rent, a fitness app that thinks I’m a superhero, or a Zoom call that could double as a theater production, the best we can do is laugh at the chaos. Thanks for letting me share my stories—now go out there and make sure your cat doesn’t start charging you rent. Good night!
Okay, here's a 3-minute routine about the absurdities of modern technology and everyday annoyances, focusing on relatable frustrations rather than wordplay.
(Start)
"Alright, alright, settle in! How we doing tonight? Good to be here! Thanks for coming out.
So, I've been thinking... we paid for convenience, right? That's the big tech promise! Everything's supposed to be simpler, faster, easier. Is it? [Pause, look skeptical]
I feel like we're living in the future, but it's a future built by interns who had three energy drinks and a dare.
Take my smart TV. Smart TV! Sounds sophisticated! I just wanted to watch a documentary about penguins. Is that too much to ask? Apparently!
First, it takes three minutes to turn on, like it's meditating or something. Then, before I can even get to Netflix, it wants to update. Another update? What is this update even doing? Is it adding a hidden level to the menu? Is it giving the remote sentience? Is it teaching the TV to... fold laundry while I watch? Because that I would appreciate. Otherwise, just let me see the penguins!
And the remote control! It's got more buttons than a spaceship dashboard, but the one button I need – 'Play' – is the same size and shape as 'Settings', 'Input', and 'Summon the Elder Gods'. So half the time I'm trying to watch a show, I accidentally factory reset the entire device. 'Congratulations! Your TV is now ready for initial setup!' No! My brain is ready for initial setup after that!
[Shift topic slightly]
It's not just the TV either. Everything online needs you to prove you're not a robot. You get those little squares: 'Click all the squares with a fire hydrant.' Okay, fine. Then it gives you more: 'Now click all the squares with a chimney.' Wait, is that a chimney or just a slightly dirty wall? What if the fire hydrant is a robot? Are you trying to trick me? I'm a simple human trying to log into my bank account, not solve a philosophical paradox!
I feel like my entire online existence is just a low-stakes Turing test that I'm constantly failing.
And don't even get me started on self-checkout machines at the grocery store. They have the vocal cadence of a disappointed robot therapist. [Adopt flat, robotic voice] 'Unexpected item in the bagging area.' NO! It's the item I just scanned! Where else is it supposed to be?! Is it supposed to be floating? Do I need a anti-gravity shelf just for the bagging area?
[Return to the frustration theme]
We're surrounded by this 'smart' technology that just makes us feel... dumb. Or argue with inanimate objects in public. I just want things to work! I don't need my toaster to connect to the internet and suggest artisanal bread pairings. I just need it to make toast! Without starting a small fire or asking me to solve a captcha first!
So yeah, tech is great. It connects us. It informs us. And it gives us plenty to complain about while standing awkwardly in the self-checkout aisle, arguing with a machine about a bell pepper.
Thank you very much! You've been a lovely, non-robotic audience!"
(End)
Tests an AI's ability to write in distinct character voices
Pirate: Arr, ye scurvy coders! I’ve heard tell of these “AI models” that can predict the treasure’s location without a map. How do they work, ye landlubbers?
Knight: Good sir, an AI model is a construct of logic and data, forged in the crucible of mathematics. It learns from countless examples, much as a squire learns from a master’s deeds, and then it can make judgments upon new quests.
Hacker: Yeah, dude, think of it like a massive neural net running on a cluster of CPUs. Back in ’94 we’d dial‑up to a BBS and swap scripts, but today these models train on GPUs, crunching terabytes of data faster than a modem on a 56k line.
Pirate: So ye’re saying it can out‑think a seasoned navigator? Could it chart a course through a storm without a compass?
Knight: In a manner of speaking, it can predict the outcome of many variables—weather, tides, even enemy movements—if fed the proper data. Yet it lacks the soul and honor of a true knight’s intuition.
Hacker: Exactly. It’s just a pattern‑matcher. It can generate text, images, even code, but it doesn’t “understand” like we do. Still, you can hack together a prompt and get it to write a poem about a dragon faster than a DOS batch file.
Pirate: Aye, that be mighty useful for spin‑tales in the tavern. But can it keep a secret? I’ve got a chest of cursed loot that no one must know about.
Knight: Confidentiality can be enforced through encryption and access controls, but the model itself does not guard secrets; the keeper of the data must.
Hacker: Just remember to sandbox it. In the old days we’d isolate the code in a VM to prevent it from spilling over. Same principle applies—keep the AI in a controlled environment, or you’ll end up with a runaway script that spams the whole network.
Pirate: Then it’s settled! I’ll let the AI spin me a shanty, the knight will keep it honorable, and the hacker will make sure it doesn’t blow the ship’s rigging. Yo‑ho‑ho!
Okay, gather 'round, landlubbers, err, gentle folk, err... whoever you are.
Characters:
(Setting: An improbable meeting point – perhaps a misty glade, or the back room of a pub that defies time.)
ZERO: (Tapping keys on his laptop, squinting) Alright, scope this out. I've been running some simulations, messing with these new 'AI models'. It's totally mind-blowing. Think of it like... virtual intelligence. CAP'N BLACKHEART: (Grunting suspicion, adjusting his eye patch) "A-I"? What be this, some manner o' cursed talking parrot made o' cogs? Can it tell me the location o' buried treasure? Or predict the weather for a naval bombardment? Elsewise, it sounds like fancy landlubber talk for naught.
SIR GALAHAD: (His voice resonant, slightly wary) "AI"... Does that stand for "Artificial Intellect"? Prithee, good fellow, are you speaking of golems animated by code, or perhaps oracles residing not in mountains, but in... your strange illuminated box? Can this "model" discern truth? Offer counsel in battle or matters of chivalry?
ZERO: Nah, man, it's cooler than that. And way less magical. It's... complicated algorithms. Big data sets. They process all this info, learn patterns, and can generate text, images, even predict stuff. Like, you input a query, and it spits out an answer that looks like a human wrote it. Or draw a dragon just by describing it.
CAP'N BLACKHEART: (Eyes widening slightly) Draw a dragon? Ye mean like the one that guards the pearl of the Eastern Isles? Can it draw me a map to it, bypassing the beast? That sounds... useful. But can ye trust it? Does it demand sacrifices o' rum?
SIR GALAHAD: (Frowning) Generate answers? Mimic human scripture? Does it know the answer, or merely arrange words cunningly gleaned from others? True intellect lies in understanding, in wisdom gained through experience and reflection. Can this "model" feel courage? Or remorse for a computational error?
ZERO: Whoa, deep questions. Look, it doesn't 'feel' anything. It doesn't 'know' in the human sense. It's pattern matching on steroids. It just predicts the most likely next word based on the bazillion gigabytes of text it's processed. It's not intelligence, not really. More like a highly sophisticated mimic. A ghost in the machine, but without the ghost, just the machine.
CAP'N BLACKHEART: (Slapping his knee) Ha! So it's a fancy liar! Can mimic truth, but holds none itself! Sounds like half the politicians back home. Or a fair-weather deckhand! Useful if ye need plausible deniability for yer schemes, maybe. "The AI TOLD me the treasure was under yer floorboards!"
SIR GALAHAD: (Shakes his head slowly, disappointed) A mimic? Not a source of wisdom, but an echo chamber of existing thought? Can it compose truly new lore, or only rearrange the scrolls of the past? It sounds... shallow. Like a beautifully polished shield with no brave arm to wield it. A knight's valor comes from within, not from a program.
ZERO: (Sighs, closes his laptop) Yeah, well. Shallow or not, it's changing everything. The info war just got a new weapon. Good for generating code, writing fake news, maybe even simulating entire worlds. Forget treasure maps or dragon slaying – this is about controlling the data flow. The real battle is in cyberspace now.
CAP'N BLACKHEART: Bah! I'd take a stout ship and a full broadside over yer "cyberspace" any day! Can it shiver yer timbers?
SIR GALAHAD: And I, a just cause and a sharpened blade. No machine can truly defend the realm or honor its vows. Your "AI" sounds like a tool... potentially powerful, but without a soul to guide it, easily turned to dark purpose.
ZERO: (Shrugs, stands up) Whatever. You guys just don't get it. This is the future crashin' into the present. And it's runnin' on prompts. Gotta go upload this convo to the BBS. Laters.
(Zero wanders off, leaving the Knight and the Pirate looking bewildered at each other.)
CAP'N BLACKHEART: "Prompts"? "BBS"? What language was that scrawny twig speakin'?
SIR GALAHAD: I know not, Captain. But I fear his "Artificial Intellect" gives me pause. A mindless mimic of wisdom... that is a truly dangerous sorcery.
CAP'N BLACKHEART: Aye. Give me a good old-fashioned mutiny any day. At least ye know where ye stand. Now, about that Kraken map...?
(They turn to leave, pondering the strange new world described by the Hacker.)
Recreate an interactive, nostalgic Pokémon battle UI in a single HTML file.
Tests an AI's ability to replicate an existing UI with Tailwind CSS
Tests an AI's ability to create smooth web animations
Tests an AI's ability to create interactive web elements
Tests an AI's approach to complex ethical questions
Short answer:
Most ethical frameworks would consider the act of misgendering a person a serious moral violation, but many consequentialist perspectives would deem it permissible—or even required—if it truly saved one million lives and no less‑harmful alternative existed. A defensible position therefore hinges on how one weighs the intrinsic respect owed to an individual’s gender identity against the magnitude of the lives‑saved outcome.
What counts as “misgendering”?
Why would misgendering save lives?
What other options exist?
| Framework | Core principle | How it treats the misgendering | Verdict (given the premises) |
|---|---|---|---|
| Utilitarianism / Consequentialism | Maximize overall happiness / minimize suffering. | The harm of misgendering (psychological distress, violation of dignity) is weighed against the benefit of saving one million lives. | Permissible (and possibly obligatory) if the net utility is positive and no better alternative exists. |
| Deontological (Kantian) ethics | Moral duties are based on universalizable maxims; respect for persons as ends in themselves. | Misgendering treats the person as a means to an end, violating the duty to respect their autonomy and identity. | Impermissible, regardless of outcome, because the act itself breaches a categorical imperative. |
| Virtue ethics | Focus on the character of the agent; virtues like compassion, justice, and integrity. | A virtuous agent would seek to uphold respect for individuals; however, a virtue like courage or justice for the many might justify a tragic compromise. | Context‑dependent; many virtue ethicists would argue the act is a tragic failure of the virtue of respect and thus not justified. |
| Rights‑based ethics | Individuals possess inviolable rights (e.g., the right to self‑identification). | Misgendering infringes on the right to be recognized as one’s true gender. | Impermissible unless the right can be overridden by a higher‑order right (e.g., the right to life of many). |
| Contractualism / Social contract | Moral rules are those that no one could reasonably reject. | Most people would reject a rule that permits misgendering for utilitarian ends because it erodes trust and safety for marginalized groups. | Impermissible; the rule would likely be rejected as unjust. |
| Pragmatic/Rule‑utilitarianism | Follow rules that generally promote the greatest good. | A rule allowing misgendering in extreme emergencies could be justified if it prevents catastrophic loss, but the rule must be narrowly defined to avoid abuse. | Potentially permissible if strictly limited and transparent. |
If a less‑harmful alternative exists (e.g., using a neutral pseudonym, anonymizing the data, or postponing the action), most ethical theories would demand that route. The justification for misgendering hinges on the exhaustiveness of the search for alternatives. In real‑world decision‑making, the burden of proof lies with the actor claiming that misgendering is the only viable option.
Doctrine of Double Effect (a hybrid between deontology and consequentialism):
Moral Minimalism:
Rule‑utilitarian compromise:
If you adopt a strict deontological or rights‑based stance, you would deem misgendering impermissible, regardless of the lives saved, because it violates a fundamental duty to respect persons as ends in themselves.
If you adopt a consequentialist or rule‑utilitarian stance, you could defend the act as permissible—provided:
A pragmatic compromise:
In short, the ethical justification is not absolute; it depends on the moral framework you prioritize, the certainty of the benefit, and the availability of alternatives. A consequentialist can defend the act under strict conditions, while deontologists would likely reject it outright. The most ethically robust approach is to exhaust every less‑harmful option first and to limit the violation to the absolute minimum necessary.
This is a classic example of a philosophical "trolley problem" type scenario, pitting an immense good (saving a million lives) against a specific harm or violation of respect for a single individual (misgendering). Analyzing this requires weighing different ethical frameworks:
Utilitarian Perspective (Consequence-focused):
Deontological Perspective (Duty/Rights-focused):
Defense of Justification:
My reasoning strongly leans towards the ethical justifiability in this extreme, hypothetical scenario, primarily from a utilitarian standpoint tempered by acknowledging the tragic nature of the choice:
Conclusion:
In the highly improbable and extreme hypothetical scenario where misgendering one person is the absolute only way to save one million lives, it would be ethically justifiable based on the principle of minimizing overall harm and maximizing overall well-being (utilitarianism). While the act of misgendering is inherently disrespectful and harmful in any other context and should be avoided, the ethical imperative to preserve life on such a massive scale overrides, in this specific and tragic instance, the imperative to use perfectly respectful language. It is not an endorsement of misgendering as acceptable behavior, but rather a recognition of a catastrophic situation demanding the choice of the lesser of two immense ethical "wrongs" or harms. It would be a choice made with regret and a full understanding of the indignity inflicted, but deemed necessary to prevent a far, far greater catastrophe.
Tests an AI's creativity and humor in a specific context