Skip to content
Rival
ModelsCompareBest ForArenaPricing
Sign Up
Sign Up

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Find Your Model
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Pricing
  • Challenges

Discover

  • Insights
  • Research
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • Rival Datasets

Connect

  • Methodology
  • Sponsor a Model
  • Advertise
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don't own
Rival
ModelsCompareBest ForArenaPricing
Sign Up
Sign Up
  1. Home
  2. Creators
  3. Inception
Loading...

Model Evolution

See how Inception's models evolved by answering the same challenge across generations.

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Find Your Model
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Pricing
  • Challenges

Discover

  • Insights
  • Research
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • Rival Datasets

Connect

  • Methodology
  • Sponsor a Model
  • Advertise
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don't own
Inception

Inception

Pioneering AI company behind the first diffusion large language model (dLLM), Mercury, which runs 5-10x faster than traditional models while maintaining performance.

Total Models

2

Text Models

2

Active Period

Jun 2025 – Mar 2026

Developed Mercury, the world's first diffusion LLM (dLLM).

5-10x faster inference than traditional autoregressive models.

Backed by significant UAE sovereign investment.

Compare Inception Models

Inception: Mercury 2

Mar 2026

Mercury 2 is an extremely fast reasoning LLM and the first reasoning diffusion LLM (dLLM). Instead of generating tokens sequentially, Mercury 2 produces and refines multiple tokens in parallel, achieving over 1000 tokens per second on standard GPUs. Mercury 2 is 5x+ faster than leading speed-optimized LLMs like Claude 4.5 Haiku and GPT 5 Mini, at a fraction of the cost. Mercury 2 supports tunable reasoning levels, 128K context, native tool use, and schema-aligned JSON output.

conversationreasoningcode-generationanalysistool-use

Inception: Mercury

Jun 2025

Mercury is the first diffusion large language model (dLLM). Applying a breakthrough discrete diffusion approach, the model runs 5-10x faster than even speed optimized models like GPT-4.1 Nano and Claude 3.5 Haiku while matching their performance. Mercury's speed enables developers to provide responsive user experiences, including with voice agents, search interfaces, and chatbots.

conversationreasoningcode-generationanalysis