AI model lineage

How the major AI models are actually related

Per-provider family trees for 86 models across 8 providers, with 80 typed edges — successions, post-trainings, fine-tunes, distillations, retrainings, and the GPT line / o-series merge into GPT-5.

As of May 9, 2026.

Current production flagships

One node per provider, color-coded. Click any flagship to jump into the per-provider family tree below.

How to read the trees

Each box is a model. Box fill indicates lifecycle status (Current emerald, Available sky, Legacy amber, Deprecated rose). Each box has a left-edge stripe in the provider color. Each arrow is a typed edge:

Succession
Next release in the family timeline. Provider has not formally documented base-model continuity, so no claim is made about whether the new release is a fine-tune, post-training, or full retraining of the previous one.
Post-training
Provider documents this release as a post-training update over the same base model — RLHF, instruction tuning, behavior changes — without a new base.
Fine-tune
Provider documents this release as a fine-tune of the source model (often for a domain — code, vision, instruction-following).
Distilled from
Provider documents this release as a student model distilled from a larger teacher model in the same family.
New base
Provider explicitly documents this release as a new base model, retrained from scratch or with substantial architecture / data changes that make it not a fine-tune of the previous flagship.
Lines merged
Provider unifies two or more previously-separate model tracks into one release (e.g. OpenAI merging the GPT line and the o-series reasoning track into GPT-5).

Hover any node or edge for its full label, ship date, and (for edges) the source the relationship was sourced to. The "Lineage as text" block under each tree contains the same information for screen readers and crawlers.

Anthropic · Claude

Versions page →
OpusSonnetHaiku2023202420252026Succession — Claude 1 → Claude 2Distilled from — Claude 1 → Claude Instant 1.2 Anthropic positioned Claude Instant as a smaller, faster sibling of the Claude flagship; sourced to the Claude 2 announcement.DistillPost-training — Claude 2 → Claude 2.1 Anthropic documented Claude 2.1 as an iterative update on Claude 2 with the 200K context extension, not a new base.Post-trNew base — Claude 2.1 → Claude 3 Opus Anthropic introduced the Claude 3 family as a new generation of base models, not a Claude 2 fine-tune.New baseSuccession — Claude 3 Opus → Claude 3 Sonnet Anthropic shipped Opus / Sonnet / Haiku as the three tiers of the Claude 3 generation simultaneously.Succession — Claude 3 Opus → Claude 3 HaikuSuccession — Claude 3 Sonnet → Claude 3.5 SonnetPost-training — Claude 3.5 Sonnet → 3.5 Sonnet (new) Anthropic documented the October 2024 'new 3.5 Sonnet' as an upgraded version under the same model id.Post-trDistilled from — 3.5 Sonnet (new) → Claude 3.5 Haiku Anthropic documented Claude 3.5 Haiku as a smaller sibling sharing the 3.5 generation training.DistillSuccession — 3.5 Sonnet (new) → Claude 3.7 SonnetNew base — Claude 3.7 Sonnet → Claude Sonnet 4 Anthropic introduced Claude 4 as a new generation, with Opus 4 and Sonnet 4 as new bases rather than fine-tunes of 3.7.New baseNew base — Claude 3.7 Sonnet → Claude Opus 4New baseSuccession — Claude Opus 4 → Claude Opus 4.7Succession — Claude Sonnet 4 → Claude Opus 4.7Claude 1Claude 1 · Mar 14, 2023 · LegacyClaude 2Claude 2 · Jul 11, 2023 · LegacyClaude Instant 1.2Claude Instant 1.2 · Aug 9, 2023 · LegacyClaude 2.1Claude 2.1 · Nov 21, 2023 · LegacyClaude 3 OpusClaude 3 Opus · Mar 4, 2024 · AvailableClaude 3 SonnetClaude 3 Sonnet · Mar 4, 2024 · LegacyClaude 3 HaikuClaude 3 Haiku · Mar 13, 2024 · AvailableClaude 3.5 SonnetClaude 3.5 Sonnet · Jun 20, 2024 · Legacy3.5 Sonnet (new)3.5 Sonnet (new) · Oct 22, 2024 · AvailableClaude 3.5 HaikuClaude 3.5 Haiku · Nov 4, 2024 · AvailableClaude 3.7 SonnetClaude 3.7 Sonnet · Feb 24, 2025 · AvailableClaude Opus 4Claude Opus 4 · May 22, 2025 · AvailableClaude Sonnet 4Claude Sonnet 4 · May 22, 2025 · AvailableClaude Opus 4.7Claude Opus 4.7 · Apr 16, 2026 · Current
Lineage as text (14 edges) ↓

Every edge in the Claude tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

OpenAI · ChatGPT

Versions page →
GPT lineo-seriesGPT-52018–222023202420252026Post-training — GPT-3 → GPT-3.5 OpenAI's GPT-3.5 (text-davinci series and ChatGPT launch) was an RLHF-fine-tuned descendant of the GPT-3 base, per OpenAI's InstructGPT paper and the ChatGPT launch post.Post-trNew base — GPT-3.5 → GPT-4 OpenAI introduced GPT-4 as a new base model, not a GPT-3.5 fine-tune.New basePost-training — GPT-4 → GPT-4 Turbo OpenAI documented GPT-4 Turbo at DevDay 2023 as the same generation with extended context and updated training data.Post-trNew base — GPT-4 Turbo → GPT-4o OpenAI introduced GPT-4o as a new natively-multimodal base model.New baseDistilled from — GPT-4o → GPT-4o mini OpenAI positioned GPT-4o mini as the smaller, distilled sibling of GPT-4o.DistillSuccession — GPT-4o → GPT-4.1New base — GPT-4o → o1-preview OpenAI introduced the o-series as a separate reasoning track; o1 was trained to use long chain-of-thought at inference time.New baseSuccession — o1-preview → o1Succession — o1 → o3-miniSuccession — o3-mini → o3Distilled from — o3 → o4-mini OpenAI positioned o4-mini as the smaller sibling shipped alongside o3.DistillLines merged — GPT-4.1 → GPT-5 OpenAI introduced GPT-5 as a unified router that combines the GPT chat line and the o-series reasoning track into one model id.MergedLines merged — o3 → GPT-5MergedSuccession — GPT-5 → GPT-5.5GPT-3GPT-3 · May 28, 2020 · LegacyGPT-3.5GPT-3.5 · Nov 30, 2022 · LegacyGPT-4GPT-4 · Mar 14, 2023 · LegacyGPT-4 TurboGPT-4 Turbo · Nov 6, 2023 · LegacyGPT-4oGPT-4o · May 13, 2024 · AvailableGPT-4o miniGPT-4o mini · Jul 18, 2024 · AvailableGPT-4.1GPT-4.1 · Apr 14, 2025 · Availableo1-previewo1-preview · Sep 12, 2024 · Legacyo1o1 · Dec 5, 2024 · Availableo3-minio3-mini · Jan 31, 2025 · Availableo3o3 · Apr 16, 2025 · Availableo4-minio4-mini · Apr 16, 2025 · AvailableGPT-5GPT-5 · Aug 7, 2025 · CurrentGPT-5.5GPT-5.5 · Apr 23, 2026 · Current
Lineage as text (14 edges) ↓

Every edge in the ChatGPT tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

  • GPT-3.5 (Nov 30, 2022)post-training of GPT-3 — OpenAI's GPT-3.5 (text-davinci series and ChatGPT launch) was an RLHF-fine-tuned descendant of the GPT-3 base, per OpenAI's InstructGPT paper and the ChatGPT launch post.
  • GPT-4 (Mar 14, 2023)new base of GPT-3.5 — OpenAI introduced GPT-4 as a new base model, not a GPT-3.5 fine-tune.
  • GPT-4 Turbo (Nov 6, 2023)post-training of GPT-4 — OpenAI documented GPT-4 Turbo at DevDay 2023 as the same generation with extended context and updated training data.
  • GPT-4o (May 13, 2024)new base of GPT-4 Turbo — OpenAI introduced GPT-4o as a new natively-multimodal base model.
  • GPT-4o mini (Jul 18, 2024)distilled from of GPT-4o — OpenAI positioned GPT-4o mini as the smaller, distilled sibling of GPT-4o.
  • o1-preview (Sep 12, 2024)new base of GPT-4o — OpenAI introduced the o-series as a separate reasoning track; o1 was trained to use long chain-of-thought at inference time.
  • o1 (Dec 5, 2024)succession of o1-preview
  • o3-mini (Jan 31, 2025)succession of o1
  • GPT-4.1 (Apr 14, 2025)succession of GPT-4o
  • o4-mini (Apr 16, 2025)distilled from of o3 — OpenAI positioned o4-mini as the smaller sibling shipped alongside o3.
  • o3 (Apr 16, 2025)succession of o3-mini
  • GPT-5 (Aug 7, 2025)lines merged of GPT-4.1 — OpenAI introduced GPT-5 as a unified router that combines the GPT chat line and the o-series reasoning track into one model id.
  • GPT-5 (Aug 7, 2025)lines merged of o3
  • GPT-5.5 (Apr 23, 2026)succession of GPT-5

Google · Gemini

Versions page →
ProFlashLite2023202420252026New base — Bard (LaMDA) → PaLM 2 Google introduced PaLM 2 as a new base model powering an upgraded Bard.New baseNew base — PaLM 2 → Gemini 1.0 Pro Google's Gemini 1.0 was a new natively-multimodal architecture, not a PaLM 2 fine-tune.New baseSuccession — Gemini 1.0 Pro → Gemini 1.0 UltraNew base — Gemini 1.0 Ultra → Gemini 1.5 Pro Google introduced Gemini 1.5 with a new mixture-of-experts architecture and the 1M context, not a 1.0 fine-tune.New baseDistilled from — Gemini 1.5 Pro → Gemini 1.5 Flash Google described Gemini 1.5 Flash as a distilled smaller sibling of 1.5 Pro.DistillSuccession — Gemini 1.5 Flash → Gemini 2.0 FlashSuccession — Gemini 1.5 Pro → Gemini 2.0 Pro ExpSuccession — Gemini 2.0 Pro Exp → Gemini 2.5 ProSuccession — Gemini 2.0 Flash → Gemini 2.5 FlashDistilled from — Gemini 2.5 Flash → 2.5 Flash-Lite Google positioned Flash-Lite as the smallest sibling of the 2.5 Flash family.DistillNew base — Gemini 2.5 Pro → Gemini 3 Pro Google introduced Gemini 3 as a new generation flagship.New baseSuccession — Gemini 2.5 Flash → Gemini 3 FlashSuccession — Gemini 3 Pro → Gemini 3.1 ProDistilled from — Gemini 3 Flash → 3.1 Flash-LiteDistillBard (LaMDA)Bard (LaMDA) · Mar 21, 2023 · LegacyPaLM 2PaLM 2 · May 10, 2023 · LegacyGemini 1.0 ProGemini 1.0 Pro · Dec 6, 2023 · LegacyGemini 1.0 UltraGemini 1.0 Ultra · Feb 8, 2024 · LegacyGemini 1.5 ProGemini 1.5 Pro · Feb 15, 2024 · LegacyGemini 1.5 FlashGemini 1.5 Flash · May 14, 2024 · LegacyGemini 2.0 FlashGemini 2.0 Flash · Dec 11, 2024 · AvailableGemini 2.0 Pro ExpGemini 2.0 Pro Exp · Feb 5, 2025 · AvailableGemini 2.5 ProGemini 2.5 Pro · Mar 25, 2025 · AvailableGemini 2.5 FlashGemini 2.5 Flash · Jun 17, 2025 · Available2.5 Flash-Lite2.5 Flash-Lite · Jul 22, 2025 · AvailableGemini 3 ProGemini 3 Pro · Nov 18, 2025 · CurrentGemini 3 FlashGemini 3 Flash · Dec 17, 2025 · CurrentGemini 3.1 ProGemini 3.1 Pro · Feb 19, 2026 · Current3.1 Flash-Lite3.1 Flash-Lite · Mar 3, 2026 · Current
Lineage as text (14 edges) ↓

Every edge in the Gemini tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

FlagshipSpecialized2023202420252026Post-training — Grok 1 → Grok 1.5 xAI documented Grok 1.5 as a successor that extended Grok 1's context and reasoning, broadly within the same generation.Post-trNew base — Grok 1.5 → Grok 2 xAI introduced Grok 2 as a new base.New baseNew base — Grok 2 → Grok 3 xAI introduced Grok 3 trained on the Memphis Colossus cluster as a new base.New baseNew base — Grok 3 → Grok 4New basePost-training — Grok 4 → Grok 4.1Post-trDistilled from — Grok 4.1 → Grok 4.1 Fast xAI positioned Grok 4.1 Fast as a smaller, faster sibling shipped two days after 4.1.DistillSuccession — Grok 4.1 → Grok 4.20Grok 1Grok 1 · Nov 4, 2023 · LegacyGrok 1.5Grok 1.5 · Mar 28, 2024 · LegacyGrok 2Grok 2 · Aug 13, 2024 · LegacyGrok 3Grok 3 · Feb 17, 2025 · AvailableGrok 4Grok 4 · Jul 9, 2025 · AvailableGrok 4.1Grok 4.1 · Nov 17, 2025 · CurrentGrok 4.1 FastGrok 4.1 Fast · Nov 19, 2025 · CurrentGrok 4.20Grok 4.20 · Mar 10, 2026 · Current
Lineage as text (7 edges) ↓

Every edge in the Grok tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

  • Grok 1.5 (Mar 28, 2024)post-training of Grok 1 — xAI documented Grok 1.5 as a successor that extended Grok 1's context and reasoning, broadly within the same generation.
  • Grok 2 (Aug 13, 2024)new base of Grok 1.5 — xAI introduced Grok 2 as a new base.
  • Grok 3 (Feb 17, 2025)new base of Grok 2 — xAI introduced Grok 3 trained on the Memphis Colossus cluster as a new base.
  • Grok 4 (Jul 9, 2025)new base of Grok 3
  • Grok 4.1 (Nov 17, 2025)post-training of Grok 4
  • Grok 4.1 Fast (Nov 19, 2025)distilled from of Grok 4.1 — xAI positioned Grok 4.1 Fast as a smaller, faster sibling shipped two days after 4.1.
  • Grok 4.20 (Mar 10, 2026)succession of Grok 4.1

Meta · Llama

Versions page →
FlagshipSpecializedClosed2023202420252026New base — LLaMA 1 → Llama 2 Meta's Llama 2 paper documents it as a new pretraining run with substantially larger data and the open license shift.New baseFine-tune — Llama 2 → Code Llama Meta's Code Llama paper documents it as a fine-tune of Llama 2 on code data.Fine-tNew base — Llama 2 → Llama 3 Meta's Llama 3 release notes document it as a new pretraining run with updated tokenizer.New basePost-training — Llama 3 → Llama 3.1 Meta's Llama 3.1 release added the 405B size and the long-context post-training; documented as same generation.Post-trPost-training — Llama 3.1 → Llama 3.2 Meta's Llama 3.2 release added vision and edge-sized variants.Post-trPost-training — Llama 3.2 → Llama 3.3 70BPost-trNew base — Llama 3.3 70B → Llama 4 Scout Meta's Llama 4 release introduced a new mixture-of-experts architecture as a new base.New baseNew base — Llama 3.3 70B → Llama 4 MaverickNew baseNew base — Llama 4 Maverick → Muse Spark Meta described Muse Spark as the closed-weights successor to the Llama line, trained at Meta Superintelligence Labs.New baseLLaMA 1LLaMA 1 · Mar 3, 2023 · LegacyLlama 2Llama 2 · Jul 18, 2023 · LegacyCode LlamaCode Llama · Aug 24, 2023 · LegacyLlama 3Llama 3 · Apr 18, 2024 · LegacyLlama 3.1Llama 3.1 · Jul 23, 2024 · AvailableLlama 3.2Llama 3.2 · Sep 25, 2024 · AvailableLlama 3.3 70BLlama 3.3 70B · Dec 6, 2024 · AvailableLlama 4 ScoutLlama 4 Scout · Apr 5, 2025 · AvailableLlama 4 MaverickLlama 4 Maverick · Apr 5, 2025 · AvailableMuse SparkMuse Spark · Apr 8, 2026 · Current
Lineage as text (9 edges) ↓

Every edge in the Llama tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

  • Llama 2 (Jul 18, 2023)new base of LLaMA 1 — Meta's Llama 2 paper documents it as a new pretraining run with substantially larger data and the open license shift.
  • Code Llama (Aug 24, 2023)fine-tune of Llama 2 — Meta's Code Llama paper documents it as a fine-tune of Llama 2 on code data.
  • Llama 3 (Apr 18, 2024)new base of Llama 2 — Meta's Llama 3 release notes document it as a new pretraining run with updated tokenizer.
  • Llama 3.1 (Jul 23, 2024)post-training of Llama 3 — Meta's Llama 3.1 release added the 405B size and the long-context post-training; documented as same generation.
  • Llama 3.2 (Sep 25, 2024)post-training of Llama 3.1 — Meta's Llama 3.2 release added vision and edge-sized variants.
  • Llama 3.3 70B (Dec 6, 2024)post-training of Llama 3.2
  • Llama 4 Scout (Apr 5, 2025)new base of Llama 3.3 70B — Meta's Llama 4 release introduced a new mixture-of-experts architecture as a new base.
  • Llama 4 Maverick (Apr 5, 2025)new base of Llama 3.3 70B
  • Muse Spark (Apr 8, 2026)new base of Llama 4 Maverick — Meta described Muse Spark as the closed-weights successor to the Llama line, trained at Meta Superintelligence Labs.

DeepSeek · DeepSeek

Versions page →
FlagshipReasoningCode2023202420252026New base — DeepSeek-LLM → DeepSeek-V2 DeepSeek-V2 paper introduced the DeepSeekMoE architecture and Multi-head Latent Attention as new base.New baseFine-tune — DeepSeek-LLM → DeepSeek-Coder DeepSeek-Coder shipped alongside the base LLM as a code-specialized fine-tune.Fine-tNew base — DeepSeek-V2 → DeepSeek-V3 DeepSeek-V3 paper introduced the 671B-parameter MoE as a new base.New basePost-training — DeepSeek-V3 → DeepSeek-R1 DeepSeek-R1 paper documents R1 as RL post-training over the V3 base — RL-only emergent chain-of-thought.Post-trPost-training — DeepSeek-V3 → DeepSeek-V3.1Post-trPost-training — DeepSeek-V3.1 → DeepSeek-V3.2Post-trNew base — DeepSeek-V3.2 → DeepSeek-V4-Pro DeepSeek-V4 introduced a new MoE base with Thinking and Non-Thinking modes.New baseDistilled from — DeepSeek-V4-Pro → DeepSeek-V4-Flash DeepSeek positioned V4-Flash as the smaller sibling shipped alongside V4-Pro.DistillDeepSeek-LLMDeepSeek-LLM · Nov 2, 2023 · LegacyDeepSeek-CoderDeepSeek-Coder · Nov 2, 2023 · LegacyDeepSeek-V2DeepSeek-V2 · May 6, 2024 · LegacyDeepSeek-V3DeepSeek-V3 · Dec 26, 2024 · AvailableDeepSeek-R1DeepSeek-R1 · Jan 20, 2025 · AvailableDeepSeek-V3.1DeepSeek-V3.1 · Aug 21, 2025 · AvailableDeepSeek-V3.2DeepSeek-V3.2 · Dec 1, 2025 · AvailableDeepSeek-V4-ProDeepSeek-V4-Pro · Apr 24, 2026 · CurrentDeepSeek-V4-FlashDeepSeek-V4-Flash · Apr 24, 2026 · Current
Lineage as text (8 edges) ↓

Every edge in the DeepSeek tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

Mistral · Mistral

Versions page →
OpenProprietary2023202420252026New base — Mistral 7B → Mixtral 8x7B Mixtral 8x7B introduced the sparse mixture-of-experts architecture as a new base, not a Mistral 7B fine-tune.New baseNew base — Mixtral 8x7B → Mistral Large Mistral Large was released as a proprietary flagship under the Mistral Research / commercial license, not a Mixtral fine-tune.New baseNew base — Mistral Large → Mistral Large 2New baseNew base — Mistral Large 2 → Mistral Small 3 Mistral Small 3 marked the December 2025 'Mistral 3' family relaunch under Apache 2.0 with new bases.New baseSuccession — Mistral Small 3 → Mistral Large 3Succession — Mistral Large 3 → Mistral Small 4Mistral 7BMistral 7B · Sep 27, 2023 · LegacyMixtral 8x7BMixtral 8x7B · Dec 11, 2023 · LegacyMistral LargeMistral Large · Feb 26, 2024 · LegacyMistral Large 2Mistral Large 2 · Jul 24, 2024 · LegacyMistral Small 3Mistral Small 3 · Jan 30, 2025 · AvailableMistral Large 3Mistral Large 3 · Dec 2, 2025 · CurrentMistral Small 4Mistral Small 4 · Mar 16, 2026 · Current
Lineage as text (6 edges) ↓

Every edge in the Mistral tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

Alibaba · Qwen

Versions page →
FlagshipReasoningProprietary2023202420252026Succession — Qwen-7B → Qwen-14BNew base — Qwen-14B → Qwen2 family Qwen2 marked the June 2024 architecture refresh with the Apache 2.0 licensing turn for most variants.New basePost-training — Qwen2 family → Qwen2.5 familyPost-trNew base — Qwen2.5 family → Qwen3 family Qwen3 introduced the hybrid Thinking / Non-Thinking architecture, absorbing the standalone QwQ reasoning track.New baseSuccession — Qwen3 family → Qwen3-Max Qwen3-Max is the largest Qwen3-generation flagship.Succession — Qwen3-Max → Qwen3.5 + PlusNew base — Qwen3.5 + Plus → Qwen3.6-27B Qwen3.6-27B introduced the Gated DeltaNet + self-attention hybrid as a new base.New baseSuccession — Qwen3.5 + Plus → Qwen 3.6-Max Qwen 3.6-Max-Preview is the proprietary preview shipped alongside the open-weights 3.6 family.Qwen-7BQwen-7B · Aug 3, 2023 · LegacyQwen-14BQwen-14B · Sep 25, 2023 · LegacyQwen2 familyQwen2 family · Jun 7, 2024 · LegacyQwen2.5 familyQwen2.5 family · Sep 19, 2024 · AvailableQwen3 familyQwen3 family · Apr 28, 2025 · AvailableQwen3-MaxQwen3-Max · Sep 15, 2025 · AvailableQwen3.5 + PlusQwen3.5 + Plus · Feb 16, 2026 · AvailableQwen 3.6-MaxQwen 3.6-Max · Apr 2, 2026 · CurrentQwen3.6-27BQwen3.6-27B · Apr 22, 2026 · Current
Lineage as text (8 edges) ↓

Every edge in the Qwen tree above, in chronological order. Each line shows: destination model, edge type, source model, and the provider documentation that establishes the relationship (where one was disclosed).

  • Qwen-14B (Sep 25, 2023)succession of Qwen-7B
  • Qwen2 family (Jun 7, 2024)new base of Qwen-14B — Qwen2 marked the June 2024 architecture refresh with the Apache 2.0 licensing turn for most variants.
  • Qwen2.5 family (Sep 19, 2024)post-training of Qwen2 family
  • Qwen3 family (Apr 28, 2025)new base of Qwen2.5 family — Qwen3 introduced the hybrid Thinking / Non-Thinking architecture, absorbing the standalone QwQ reasoning track.
  • Qwen3-Max (Sep 15, 2025)succession of Qwen3 family — Qwen3-Max is the largest Qwen3-generation flagship.
  • Qwen3.5 + Plus (Feb 16, 2026)succession of Qwen3-Max
  • Qwen 3.6-Max (Apr 2, 2026)succession of Qwen3.5 + Plus — Qwen 3.6-Max-Preview is the proprietary preview shipped alongside the open-weights 3.6 family.
  • Qwen3.6-27B (Apr 22, 2026)new base of Qwen3.5 + Plus — Qwen3.6-27B introduced the Gated DeltaNet + self-attention hybrid as a new base.

About this page

Cross-family comparison page in the /ai/ section. Each per-provider tree was hand-laid-out from the per-family Claude, ChatGPT, Gemini, Grok, Llama, DeepSeek, Mistral, and Qwen Versions pages on this site, each row's lineage edge sourced to the provider's own model card, technical paper, or release announcement.

Conservative claims. Where a provider has not formally documented base-model continuity between two releases, the edge is labeled "succession" and no further claim is made. The page does not infer "X is a fine-tune of Y" from external speculation, leaderboard chatter, or press coverage. The "succession" edge is honest about what is and isn't disclosed; the per-edge tooltip and the "Lineage as text" block name the source for every more-specific edge type.

Per-provider focus, not encyclopedic. Each tree highlights the lineage-meaningful releases — the new bases, the major post-trainings, the visible fine-tunes, the line merges. Per-release minutiae (small variants, intermediate checkpoints, every refresh on every tier) live on the per-family Versions pages where they belong. The /ai/release-cadence/ and /ai/context-windows/ pages are the right place for per-release counts and metrics.

What is intentionally excluded. Open-source community fine-tunes (the Llama-derivatives ecosystem — Vicuna, WizardLM, Nous Hermes, etc.) are not on the page; they are downstream community work, not frontier-lab releases. Capability comparisons ("which line is best") are out of scope. Speculation about undisclosed base-model continuity is out of scope. Unreleased models (announced but never publicly shipped, like Llama 4 Behemoth) are not included.

Refreshed quarterly, aligned with the per-family Versions pages. Each refresh re-verifies every disclosed lineage source (model cards / papers / announcements move under the same URLs but their content changes), adds nodes for new flagship releases, and prunes any node that the provider has formally deprecated. See release cadence and context windows for the cross-family ship-cadence and context-budget pictures this page complements.

Last updated: May 9, 2026. 86 models · 8 providers · 80 typed edges.