2023 – 2026
Qwen Versions
Every Alibaba Qwen / Tongyi Qianwen release — Qwen-7B (August 2023) through Qwen3.6-27B (April 22, 2026) and the Qwen 3.6-Max-Preview proprietary flagship — with HuggingFace ids, ship dates, family (Flagship / Reasoning / Specialized), license terms (Apache 2.0 / Tongyi-Qianwen License / proprietary), and the major changes per version. Plus the April 2023 Tongyi launch as Alibaba's ChatGPT response, the licensing turn at Qwen2, the QwQ reasoning track, the Qwen3 hybrid-reasoning era, the U.S. chip export-control context, and the HuggingFace-leaderboard dominance through 2025–2026.
The April 2023 Tongyi Qianwen launch
Alibaba Cloud formally launched Tongyi Qianwen (通义千问, “truth from a thousand questions”) in April 2023 as the company's response to ChatGPT, two months after Baidu's Ernie Bot launched and roughly five months after OpenAI's November 2022 ChatGPT release. The model was first demonstrated by then–Alibaba CEO Daniel Zhang at the Alibaba Cloud Summit on April 11, 2023 and rolled out to enterprise customers through the Tongyi product family on Alibaba Cloud.
The first open-weights release — Qwen-7B — followed on August 3, 2023. The four-month gap between the Tongyi consumer-product launch and the first open-weights release was characteristic of Alibaba's strategy: ship the proprietary chatbot first to enterprise customers via Alibaba Cloud, then open-source the underlying model line to developer communities for ecosystem effects. The same hybrid pattern persists through Qwen3.6: open-weights flagships on HuggingFace alongside proprietary Qwen-Max-line models on DashScope.
The Apache 2.0 turn — from Tongyi-Qianwen License to permissive open-source
Qwen's licensing has evolved across three distinct conventions. The Qwen 1 lineage (Qwen-7B / 14B / 1.8B / 72B, August–November 2023) shipped under the bespoke Tongyi-Qianwen License — an Alibaba-authored license with permissive terms for academic use and commercial use with restrictions, but not OSI-compliant. The Qwen 1.5 family in February 2024 continued the same license pattern.
The licensing turn arrived with Qwen2 on June 6, 2024. Most of the Qwen2 sub-family (0.5B, 1.5B, 7B, 57B-A14B) shipped under Apache 2.0 — the first Qwen flagship release with broad permissive coverage. Only Qwen2-72B retained the Tongyi-Qianwen License. The pattern continued through Qwen2.5 (September 19, 2024), where the 0.5B / 1.5B / 7B / 14B / 32B variants shipped Apache 2.0 with only the 3B and 72B retaining the Qwen License.
Qwen3 (April 28, 2025) was the structural commitment: the entire Qwen3 family — six dense sizes from 0.6B to 32B, plus the 30B-A3B and 235B-A22B MoE variants — shipped Apache 2.0. Every Qwen flagship release since (Qwen3-Coder, Qwen3.5, Qwen3.5-Plus's open-weights variants, Qwen3.6-27B, Qwen-Image-2512) has shipped Apache 2.0. The proprietary Qwen-Max-line (Qwen2.5-Max, Qwen3-Max, Qwen 3.6-Max-Preview, Qwen 3.5-Plus, Qwen 3.6-Plus) runs as a parallel commercial track on DashScope, but the open-weights story has been Apache-2.0-or-permissive across every release since Qwen2.
The Qwen3 hybrid-reasoning era — April 28, 2025
Qwen3 launched on April 28, 2025 as Alibaba's first frontier-AI line with hybrid reasoning architecture. The release shipped eight models simultaneously — six dense (0.6B, 1.7B, 4B, 8B, 14B, 32B) and two MoE (30B-A3B and the flagship 235B-A22B) — all under Apache 2.0, all sharing a single architecture that supports both thinking-mode chain-of-thought reasoning and non-thinking-mode fast responses. The announcement is at qwenlm.github.io/blog/qwen3; coverage in TechCrunch and Alibaba Cloud Community.
The Qwen3 architecture absorbed the standalone QwQ reasoning track into the Flagship V-series. The QwQ line had run for five months — QwQ-32B-Preview in November 2024, the production QwQ-32B in March 2025 — as Alibaba's open-weights answer to OpenAI's o1 series and DeepSeek's R1. Qwen3's hybrid-mode architecture made the standalone Reasoning family redundant; no further QwQ releases have shipped in the year since, and the page's convergence-policy reading (documented in `docs/qwen-versions.md`) is that no further Reasoning rows are expected unless Alibaba revives the standalone track.
Qwen3 was trained on 36 trillion tokens — double Qwen2.5's pretraining corpus — with native multilingual support across 119 languages and dialects, the broadest language coverage of any frontier-AI line at the time. The Qwen3.5 family released ten months later (February 2026) extended that to 201 languages and added native multimodality across text + image + video; Qwen3.6-27B (April 2026) added the Gated DeltaNet hybrid architecture and Thinking Preservation. The Qwen3 architecture and its descendants remain the load-bearing recipe for everything Alibaba has shipped since.
The HuggingFace-leaderboard dominance
Across late 2024 and most of 2025, Qwen-derived models held a majority of the top-trending positions on the HuggingFace Open LLM Leaderboard and the various community-derived benchmark trackers. The pattern was driven by a combination of three factors: the Apache 2.0 license (which let community fine-tuners commercially redistribute their derivatives), the breadth of base sizes (0.5B through 235B in the Qwen3 family alone), and the strength of Qwen2.5-Coder / QwQ / Qwen3 on math + coding benchmarks specifically (which weighted heavily in the leaderboard rankings). At various points in 2025, more than half of the top 20 trending HuggingFace models were Qwen-derived. The dominance shifted slightly in 2026 as DeepSeek-V4 / Llama 5-equivalents / Mistral 3 absorbed share, but Qwen-derivatives have remained a large fraction of the top-trending HuggingFace models through April 2026.
The U.S. chip export-control context
Qwen training has been constrained by the same U.S. chip export-control regime that shapes DeepSeek's training environment. The October 7, 2022 Department of Commerce export controls restricted top-tier Nvidia AI-GPU exports to China; Alibaba Cloud was already on various U.S. Entity List adjacencies prior to that, with subsequent expansions through 2023–2025 tightening the procurement environment. Like DeepSeek, Alibaba Cloud built Qwen's training infrastructure on a mix of Nvidia H800 chips procured during the gap before the October 2023 H800 ban, and on Chinese domestic alternatives (Huawei Ascend, Cambricon).
The empirical record — trillion-parameter Qwen3-Max trained on 36T tokens, Qwen3.6-27B with state-of-the-art coding benchmarks — demonstrates that the Qwen team has continued training at frontier scale despite the controls, and like DeepSeek has been the subject of Department of Commerce inquiries about the chip procurement that supported specific runs. Coverage in CSIS and South China Morning Post covers the broader policy environment.
Where to run Qwen
Qwen is among the most widely-deployed AI lines because the open-weights releases are Apache 2.0 across nearly every size and the proprietary releases are available through Alibaba Cloud's Model Studio with OpenAI- and Anthropic-compatible APIs. Inference paths through 2025–2026 break into four categories.
Alibaba Cloud first-party. Qwen Chat is the consumer chat surface. Model Studio (formerly DashScope) is the developer API endpoint, OpenAI-API-compatible and serving both the open-weights and the proprietary Max-line models. The proprietary Max-line is exclusive to this surface.
Self-host from HuggingFace. Download from the Qwen org and run with vLLM, SGLang, llama.cpp, or Ollama. The open-weights flagships (Qwen3.6-27B, Qwen3.5 family, Qwen3 family, QwQ-32B) self-host without commercial restriction; the small minority of Qwen License variants (Qwen2-72B, Qwen2.5-72B, Qwen2.5-3B) require attestation of the bespoke license terms.
Hyperscalers. AWS Bedrock and Azure AI Foundry have added Qwen SKUs across 2025–2026; ModelScope (Alibaba's own model-hub) hosts the broadest set. NVIDIA NIM has Qwen variants for the most-served sizes.
Hosted-inference providers. Together AI, Fireworks, OpenRouter, SiliconFlow, Groq. Most providers serve the Apache 2.0 lineage with similar latency / cost characteristics; the Tongyi-Qianwen-License variants (mostly the 72B and 3B sizes) are typically not carried by Western inference providers due to license-attestation overhead.
People who shaped Qwen
The Qwen / Tongyi Lab team is structured inside Alibaba Cloud rather than as a standalone lab. The team has been led at the operational level by Junyang Lin (Tongyi Lab), with broader Alibaba Cloud AI strategy under CTO Jingren Zhou. The team has been notable for its consistency through Alibaba's broader 2024 corporate-restructuring waves — the Qwen release cadence accelerated rather than slowed across the period in which Alibaba was reorganizing its other business units.
Eddie Wu — CEO of Alibaba Group since September 2023; has publicly framed AI as Alibaba's strategic priority above e-commerce / cloud / logistics, with multi-year capex commitments to Qwen training infrastructure. Joseph Tsai — Chairman of Alibaba Group; has been the public spokesperson for Alibaba's AI strategy in international forums (the Davos annual meetings, the Bloomberg Tech Summit).
No publicly-named Qwen CTO or founding team in the Western-lab sense. Unlike OpenAI / Anthropic / Mistral / xAI / DeepSeek, Qwen is not structured as a startup-style lab with named founders and a public-facing leadership roster. The team operates under Alibaba Cloud's organizational umbrella, and the per-paper author lists on the Qwen / Qwen2 / Qwen2.5 / Qwen3 technical reports are the closest available roster of named contributors.
The competitive landscape
Qwen is, alongside DeepSeek, one of the two dominant Chinese open-weights AI families through 2024–2026. The closest direct comparators on the open-weights axis are DeepSeek (Chinese, MIT-licensed for the V3 / R1 line and onward, the December 2024 / January 2025 inflection — see DeepSeek Versions), Mistral (French; Apache 2.0 for the Mistral 3 family with a parallel proprietary tier — see Mistral Versions), Meta's Llama (custom Llama Community License, see Llama Versions), and the other Chinese frontier labs (Baidu Ernie, Zhipu GLM, MiniMax, Moonshot Kimi). The closed-weights frontier competitors — ChatGPT, Claude, Gemini, Grok — are the practical benchmark for “is Qwen competitive at frontier scale,” which the Qwen3 / Qwen3-Max / Qwen3.5 / Qwen3.6 release cycle has been answering in the affirmative since April 2025. Qwen's distinguishing variable is the breadth of its specialized track (Coder / VL / Audio / Math / Omni / Image) and the consistency of its Apache 2.0 commitment on the open-weights flagships, both of which continue to underwrite the line's HuggingFace-leaderboard dominance. This page does not attempt a benchmark roundup or a ranking.
Sources:
Qwen blog;
github.com/QwenLM;
huggingface.co/Qwen;
Alibaba Cloud Model Studio docs;
research papers on arXiv (Qwen2, Qwen2.5, Qwen3 technical reports);
contemporaneous reporting in NYT, FT, Bloomberg, CNBC, South China Morning Post, TechCrunch, VentureBeat, MarkTechPost, Simon Willison.
Last updated April 2026.
Mungomash LLC · More AI pages