Skip to content
Rival
ModelsCompareBest ForArenaPricing
Sign Up
Sign Up

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Find Your Model
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Pricing
  • Challenges

Discover

  • Insights
  • Research
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • Rival Datasets
  • Widgets

Connect

  • Methodology
  • Sponsor a Model
  • Advertise
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don't own
Rival
ModelsCompareBest ForArenaPricing
Sign Up
Sign Up
  1. Home
  2. Creators
  3. Qwen
Loading...

Model Evolution

See how Qwen's models evolved by answering the same challenge across generations.

We compare AI models for a living. On purpose. We chose this.

@rival_tips

Explore

  • Compare Models
  • All Models
  • Find Your Model
  • Image Generation
  • Audio Comparison
  • Leaderboard
  • Pricing
  • Challenges

Discover

  • Insights
  • Research
  • AI Creators
  • AI Tools
  • The Graveyard

Developers

  • Developer Hub
  • MCP Server
  • Rival Datasets
  • Widgets

Connect

  • Methodology
  • Sponsor a Model
  • Advertise
  • Partnerships
  • Privacy Policy
  • Terms
  • RSS Feed
© 2026 Rival · Built at hours no one should be awake, on hardware we don't own
Qwen

Qwen

Alibaba's AI model series featuring advanced multilingual capabilities and strong reasoning abilities.

Total Models

27

Text Models

25

Image Models

2

Active Period

Mar 2025 – Mar 2026

Developed by Alibaba Cloud.

Strong performance in coding and multilingual benchmarks.

Features multimodal capabilities and MoE architecture.

Integrated with Alibaba's cloud ecosystem and ModelScope.

Compare Qwen Models

Qwen3.5 9B

Mar 2026

Qwen3.5 9B is a multimodal foundation model from the Qwen 3.5 family, designed to deliver strong reasoning, coding, and visual understanding in an efficient 9B-parameter architecture. It uses a unified vision-language design with early fusion of multimodal tokens, supporting text, image, and video inputs while producing text outputs with built-in reasoning capabilities.

conversationreasoningcode-generationanalysis

Qwen: Qwen3.5 35B A3B

Feb 2026

The Qwen3.5 Series 35B-A3B is a native vision-language model designed with a hybrid architecture that integrates linear attention mechanisms and a sparse mixture-of-experts model, achieving higher inference efficiency. Its overall performance is comparable to that of the Qwen3.5-27B.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3.5 27B

Feb 2026

The Qwen3.5 27B native vision-language Dense model incorporates a linear attention mechanism, delivering fast response times while balancing inference speed and performance. Its overall capabilities are comparable to those of the Qwen3.5-122B-A10B.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3.5 122B A10B

Feb 2026

The Qwen3.5 122B-A10B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. In terms of overall performance, this model is second only to Qwen3.5-397B-A17B. Its text capabilities significantly outperform those of Qwen3-235B-2507, and its visual capabilities surpass those of Qwen3-VL-235B.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3.5 Flash

Feb 2026

The Qwen3.5 native vision-language Flash models are built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. Compared to the 3 series, these models deliver a leap forward in performance for both pure text and multimodal tasks, offering fast response times while balancing inference speed and overall performance.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3.5 Plus 2026-02-15

Feb 2026

The Qwen3.5 native vision-language series Plus models are built on a hybrid architecture that integrates linear attention mechanisms with sparse mixture-of-experts models, achieving higher inference efficiency. In a variety of task evaluations, the 3.5 series consistently demonstrates performance on par with state-of-the-art leading models. Compared to the 3 series, these models show a leap forward in both pure-text and multimodal capabilities. Supports text, image, and video inputs with reasoning and tool use.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3.5 397B A17B

Feb 2026

The Qwen3.5 series 397B-A17B native vision-language model is built on a hybrid architecture that integrates a linear attention mechanism with a sparse mixture-of-experts model, achieving higher inference efficiency. It delivers state-of-the-art performance comparable to leading-edge models across a wide range of tasks, including language understanding, logical reasoning, code generation, agent-based tasks, image understanding, video understanding, and GUI interactions. With its robust code-generation and agent capabilities, the model exhibits strong generalization across diverse agent scenarios. Supports 201 languages and dialects.

conversationreasoningcode-generationanalysistool-useagentic-tool-usetranslation

Qwen: Qwen3 Max Thinking

Feb 2026

Qwen3-Max-Thinking is the flagship reasoning model in the Qwen3 series, designed for high-stakes cognitive tasks that require deep, multi-step reasoning. By significantly scaling model capacity and reinforcement learning compute, it delivers major gains in factual accuracy, complex reasoning, instruction following, alignment with human preferences, and agentic behavior. Features Heavy Mode for test-time scaling with iterative refinement, adaptive tool use with integrated search and code interpreter, and hybrid reasoning that toggles between normal and compute-intensive modes mid-conversation.

conversationreasoningcode-generationanalysistool-use

Qwen3 Coder Next

Feb 2026

Qwen3-Coder-Next is an open-weight causal language model optimized for coding agents and local development workflows. It uses a sparse MoE design with 80B total parameters and only 3B activated per token, delivering performance comparable to models with 10 to 20x higher active compute. Operates exclusively in non-thinking mode for streamlined integration.

conversationcode-generationagentic-tool-usetool-use

Qwen Image

Jan 2026

An image generation foundation model in the Qwen series that achieves significant advances in complex text rendering.

image-generation

Qwen Image (Fast)

Dec 2025

A fast Qwen text-to-image model optimized by PrunaAI for speed on Replicate.

image-generation

Qwen3 Coder Plus

Sep 2025

Qwen3 Coder Plus model integrated via automation on 2025-09-17

conversationreasoningcode-generationanalysis

Qwen3 Coder Flash

Sep 2025

Qwen3 Coder Flash model integrated via automation on 2025-09-17

conversationreasoningcode-generationanalysis

Qwen3 Next 80B A3B Instruct

Sep 2025

Qwen3-Next-80B-A3B-Instruct is an instruction-tuned chat model in the Qwen3-Next series optimized for fast, stable responses without thinking traces. It targets complex tasks across reasoning, code generation, knowledge QA, and multilingual use, while remaining robust on alignment and formatting. Compared with prior Qwen3 instruct variants, it focuses on higher throughput and stability on ultra-long inputs and multi-turn dialogues, making it well-suited for RAG, tool use, and agentic workflows that require consistent final answers rather than visible chain-of-thought.

conversationreasoningcode-generationanalysis

Qwen3 Next 80B A3B Thinking

Sep 2025

Qwen3-Next-80B-A3B-Thinking is a reasoning-first chat model in the Qwen3-Next line that outputs structured thinking traces by default. It’s designed for hard multi-step problems; math proofs, code synthesis/debugging, logic, and agentic planning, and reports strong results across knowledge, reasoning, coding, alignment, and multilingual evaluations. Compared with prior Qwen3 variants, it emphasizes stability under long chains of thought and efficient scaling during inference, and it is tuned to follow complex instructions while reducing repetitive or off-task behavior. The model is suitable for agent frameworks and tool use (function calling), retrieval-heavy workflows, and standardized benchmarking where step-by-step solutions are required. It supports long, detailed completions and leverages throughput-oriented techniques for faster generation. Note that it operates in thinking-only mode.

conversationreasoningcode-generationanalysis

Qwen Plus 0728 (thinking)

Sep 2025

Qwen Plus 0728 (thinking), based on the Qwen3 foundation model, is a 1 million context hybrid reasoning model with a balanced performance, speed, and cost combination.

conversationreasoningcode-generationanalysis

Qwen Plus 0728

Sep 2025

Qwen Plus 0728, based on the Qwen3 foundation model, is a 1 million context hybrid reasoning model with a balanced performance, speed, and cost combination.

conversationreasoningcode-generationanalysis

Qwen: Qwen3 Max

Sep 2025

Qwen3-Max is an updated release built on the Qwen3 series, offering major improvements in reasoning, instruction following, multilingual support, and long-tail knowledge coverage compared to the January 2025 version. It delivers higher accuracy in math, coding, logic, and science tasks, follows complex instructions in Chinese and English more reliably, reduces hallucinations, and produces higher-quality responses for open-ended Q&A, writing, and conversation. The model supports over 100 languages with stronger translation and commonsense reasoning, and is optimized for retrieval-augmented generation (RAG) and tool calling, though it does not include a dedicated "thinking" mode.

conversationreasoningcode-generationanalysistranslationtool-use

Qwen3 30B A3B Thinking 2507

Aug 2025

Qwen3-30B-A3B-Thinking-2507 is a 30B parameter Mixture-of-Experts reasoning model optimized for complex tasks requiring extended multi-step thinking. The model is designed specifically for 'thinking mode,' where internal reasoning traces are separated from final answers. Compared to earlier Qwen3-30B releases, this version improves performance across logical reasoning, mathematics, science, coding, and multilingual benchmarks. It also demonstrates stronger instruction following, tool use, and alignment with human preferences. With higher reasoning efficiency and extended output budgets, it is best suited for advanced research, competitive problem solving, and agentic applications requiring structured long-context reasoning.

conversationreasoningcode-generationanalysis

Qwen: Qwen3 30B A3B Instruct 2507

Jul 2025

Qwen3-30B-A3B-Instruct-2507 is a 30.5B-parameter mixture-of-experts language model from Qwen, with 3.3B active parameters per inference. It operates in non-thinking mode and is designed for high-quality instruction following, multilingual understanding, and agentic tool use. Post-trained on instruction data, it demonstrates competitive performance across reasoning (AIME, ZebraLogic), coding (MultiPL-E, LiveCodeBench), and alignment (IFEval, WritingBench) benchmarks. It outperforms its non-instruct variant on subjective and open-ended tasks while retaining strong factual and coding performance.

conversationreasoningcode-generationanalysis

Qwen: Qwen3 235B A22B Thinking 2507

Jul 2025

Qwen3-235B-A22B-Thinking-2507 is a high-performance, open-weight Mixture-of-Experts (MoE) language model optimized for complex reasoning tasks. It activates 22B of its 235B parameters per forward pass and natively supports up to 262,144 tokens of context. This "thinking-only" variant enhances structured logical reasoning, mathematics, science, and long-form generation, showing strong benchmark performance across AIME, SuperGPQA, LiveCodeBench, and MMLU-Redux. It enforces a special reasoning mode (</think>) and is designed for high-token outputs (up to 81,920 tokens) in challenging domains.

conversationreasoningcode-generationanalysistool-use

Qwen3 Coder

Jul 2025

Qwen3-Coder-480B-A35B-Instruct is a Mixture-of-Experts (MoE) code generation model developed by the Qwen team. It is optimized for agentic coding tasks such as function calling, tool use, and long-context reasoning over repositories. The model features 480 billion total parameters, with 35 billion active per forward pass (8 out of 160 experts).

conversationreasoningcode-generationanalysisfunction-callingtool-use

Qwen: Qwen3 235B A22B 2507

Jul 2025

Qwen3-235B-A22B-Instruct-2507 is a multilingual, instruction-tuned mixture-of-experts language model based on the Qwen3-235B architecture, with 22B active parameters per forward pass. It is optimized for general-purpose text generation, including instruction following, logical reasoning, math, code, and tool usage. The model supports a native 262K context length and does not implement "thinking mode" (<think> blocks). Compared to its base variant, this version delivers significant gains in knowledge coverage, long-context reasoning, coding benchmarks, and alignment with open-ended tasks. It is particularly strong on multilingual understanding, math reasoning (e.g., AIME, HMMT), and alignment evaluations like Arena-Hard and WritingBench.

conversationreasoningcode-generationanalysis

Qwen3 0.6B

Apr 2025

A 0.6B parameter dense model from the Qwen3 family. Supports seamless switching between 'thinking' mode (complex tasks) and 'non-thinking' mode (general conversation). Trained on 36 trillion tokens across 119 languages. Features enhanced reasoning, instruction-following, agent capabilities, and multilingual support.

conversationcode-generation

Qwen3 30B A3B

Apr 2025

The latest generation Qwen model (30.5B params, 3.3B activated MoE) excels in reasoning, multilingual support, and agent tasks. Features a unique thinking/non-thinking mode switch. Supports up to 131K context with YaRN. Free tier on OpenRouter.

conversationreasoningcode-generationanalysis

Qwen3 235B A22B

Apr 2025

Qwen3-235B-A22B is a 235B parameter mixture-of-experts (MoE) model from Alibaba's Qwen team, activating 22B parameters per forward pass. Features seamless switching between 'thinking' mode (complex tasks) and 'non-thinking' mode (general conversation). Strong reasoning, multilingual (100+), instruction-following, and tool-calling. 32K context, extendable to 131K.

conversationreasoningcode-generationanalysis

QwQ 32B

Mar 2025

QwQ is the reasoning model of the Qwen series. Compared with conventional instruction-tuned models, QwQ, which is capable of thinking and reasoning, can achieve significantly enhanced performance in downstream tasks, especially hard problems. QwQ-32B is the medium-sized reasoning model, which is capable of achieving competitive performance against state-of-the-art reasoning models, e.g., DeepSeek-R1, o1-mini.

conversationreasoningcode-generationanalysis