2023 – 2026
Llama Versions
Every Meta Llama release — LLaMA 1 in February 2023 (leaked the next month) through Llama 4 in April 2025, plus the closed-weights successor Muse Spark in April 2026 — with HuggingFace ids, parameter counts, context windows, license terms, and the major changes per version. Plus the March 2023 LLaMA leak, the FAIR vs. GenAI org split, the Llama 4 Behemoth delay, the June 2025 Scale AI acqui-hire and Meta Superintelligence Labs reorg, and the closed-weights turn.
The March 2023 LLaMA leak
Meta released LLaMA 1 on February 24, 2023 as a research-only model gated by application form, accessible to academic and government applicants and not under any open-source license. On March 3, 2023 — nine days later — a user posted a torrent containing the 7B and 65B weights to 4chan's /g/ technology board under the handle llamanon. The same day, a pull request was filed against Meta's official Llama GitHub repository proposing to add the magnet link to the README. The torrent reportedly pulled directly from Facebook's CDN at high speed, embedded with a unique download URL traceable to the leaker's gated-access grant.
Meta filed DMCA takedowns through HuggingFace and GitHub through March 20, 2023, but the weights had already propagated. Coverage at the time in The Register and DeepLearning.AI's The Batch documents the propagation timeline. Within weeks, the leaked LLaMA 1 weights were the de facto foundation for an enormous open-source fine-tune ecosystem — Alpaca, Vicuna, Guanaco, and dozens of others — on HuggingFace.
The leak is widely credited with forcing Meta's hand on Llama 2: with the weights already in the wild, the marginal cost of an official open-weights commercial release dropped to near zero, while the upside (developer mindshare, ecosystem effects, reduced lawsuit risk) was substantial. Llama 2 shipped with permissive commercial terms four months later. The decision-tree shape of the leak → commercial open release pattern is the load-bearing precondition for everything that followed in the Llama story through 2025.
The Llama Community License and the >700M-MAU carve-out
Llama 2 introduced the bespoke Llama 2 Community License Agreement — not an OSI-approved open-source license, but a Meta-authored license with permissive commercial terms and a single load-bearing carve-out: companies whose products had more than 700 million monthly active users on the Llama 2 release date (July 18, 2023) must request a separate license from Meta. The carve-out is a deliberate exclusion aimed at TikTok, Google, Apple, and other direct platform competitors; reading the precise language requires the actual PDF, which Meta hosts at llama.com/license and ai.meta.com/llama/license.
The carve-out has been preserved through every subsequent Llama license: Llama 3 Community License (April 2024) added an attribution requirement that derivative model names include the “Llama 3” prefix; the Llama 3.1 / 3.2 / 3.3 licenses extended the naming convention to require the “Llama” prefix on derivatives; the Llama 4 Community License preserved the same general shape. The Open Source Initiative has consistently said the Llama license does not meet the Open Source Definition because the use-restriction (any restriction tied to who you are, not what you do) is incompatible with the OSD — this is the substantive basis for the “source-available, not open source” framing of the line.
Llama 3.2 added a regional carve-out: the multimodal models (11B and 90B Vision) were excluded from EU-domiciled individuals and EU-headquartered companies, citing regulatory uncertainty under the EU AI Act, GDPR, and Digital Markets Act. The text-only 1B / 3B models remained available in the EU. The carve-out is a precedent worth tracking through Llama 4 multimodal availability and any future multimodal Meta release; verify against the actual license PDF at write time.
FAIR, GenAI, and the org evolution
Meta's AI research home through 2022 was FAIR (Fundamental AI Research), founded in 2013 under Yann LeCun as Chief AI Scientist and led from 2023 by Joelle Pineau. FAIR's mandate was long-horizon basic research; its publication culture was open and academically-styled. LLaMA 1 was a FAIR project.
After ChatGPT's November 2022 launch caught Meta flat-footed, the company formed a separate Generative AI organization (GenAI) in February 2023, led by VP Ahmad Al-Dahle and tasked with shipping product-grade generative AI on a faster cadence than FAIR's research timeline. Llama 2 and onward shifted progressively into GenAI under Manohar Paluri; FAIR was pushed toward longer-horizon work as the shipping pace of the GenAI org accelerated through 2024.
Through late 2024 and into 2025, FAIR's relevance to Meta's shipping AI products visibly waned. Internal reporting in April 2025 described FAIR as “dying a slow death.” Joelle Pineau announced her departure on April 1, 2025, last day May 30, 2025, after eight years leading FAIR; she was the most prominent internal advocate for open-source releases. Her exit set up the structural reorganization that came two months later.
The Scale AI acqui-hire and Meta Superintelligence Labs (June 2025)
On June 12–13, 2025, Meta announced a $14.3 billion investment for a 49% non-voting stake in Scale AI, valuing Scale at over $29 billion. The deal was structurally an acqui-hire: its purpose was to bring Scale CEO Alexandr Wang (then 28) to Meta as Chief AI Officer. Wang stepped down as Scale CEO but remained on Scale's board. Coverage in CNBC and Fortune.
On June 30, 2025, Mark Zuckerberg announced the formation of Meta Superintelligence Labs (MSL) with Wang as Chief AI Officer leading the new unit and former GitHub CEO Nat Friedman leading AI products. FAIR and GenAI were both consolidated under MSL; a new TBD Lab sub-group was created to work on next-generation LLMs. CNBC published the internal memo. The reorganization was the structural endpoint of Pineau's departure two months earlier and of the Llama 4 Behemoth delay reporting through May 2025.
Friction between Meta and Scale AI emerged through August 2025 as Scale's enterprise-data-labeling business saw revenue impact (competitors stopped sending training data to a Meta-owned vendor); TechCrunch covered the cracks in the partnership. MSL itself restructured into four sub-groups in August 2025.
Yann LeCun departs (November 2025)
On November 19, 2025, Chief AI Scientist Yann LeCun announced his departure from Meta to launch his own startup focused on Advanced Machine Intelligence and “world models.” Reporting in CNBC and Bloomberg tied the decision to LeCun being asked to report to Wang — a structural demotion, since Wang's closed-source product orientation conflicts with LeCun's long-held public position that scaling LLMs is not a path to AGI and that open research is foundational to the field. LeCun's startup partners with Meta but pursues independent research direction; with both Pineau and LeCun out, the open-research mandate inside Meta effectively ended at MSL's June 2025 formation.
The closed-weights turn — Muse Spark (April 2026)
On April 8, 2026, Meta Superintelligence Labs released Muse Spark, its first model since the June 2025 reorganization. Muse Spark shipped closed-weights, API-only, in private preview — the first frontier-AI model from Meta released without open weights since Llama 2 in July 2023. Branded under the new Muse family rather than Llama. The announcement is at ai.meta.com/blog/introducing-muse-spark-msl; coverage in CNBC and VentureBeat. Meta has stated it “hopes to open-source future versions” of Muse, but the launch model is proprietary.
Muse Spark is natively multimodal and ships with two operating modes — a fast mode and a Contemplating reasoning mode that orchestrates parallel sub-agents, positioned competitively against Gemini Deep Think and the OpenAI o-series. Several outlets characterized the launch as the end of Meta's open-weights frontier-AI era. The closed-weights turn is the dramatic reversal of Mark Zuckerberg's July 2024 “Open Source AI Is the Path Forward” letter, which had argued (twenty-one months earlier) that open-source AI's trajectory would mirror Linux's path from inferior-but-cheap to dominant-via-ecosystem.
Muse Spark is treated on this page as Llama's successor — same Meta organization (MSL), same flagship-frontier slot — even though it is not Llama-branded. Whether future Muse releases are open-weights, partial open-weights, or fully closed remains the load-bearing question for the Meta-AI story going forward; the recurring refresh task should re-verify the open-weights status on every run.
Where to run Llama
Llama is the most widely-distributed frontier-AI line because the weights are open. Inference paths through 2025–2026 break into three categories.
Self-host. Download from the HuggingFace meta-llama org and run with vLLM, llama.cpp, Ollama, MLX (Apple Silicon), or TensorRT-LLM. The 1B / 3B Llama 3.2 edge models are specifically designed for on-device deployment.
Hosted-inference providers. Together AI, Fireworks AI, Groq (notable for very high tokens-per-second on Llama 3 70B), Replicate, Perplexity Labs, Cerebras, SambaNova. Pricing is typically a fraction of comparable closed-weights frontier-model API rates because providers compete on inference cost, not model rights.
Hyperscalers. AWS Bedrock, Azure AI (Microsoft was Meta's launch partner for Llama 2 in July 2023), Google Cloud Vertex AI, Oracle OCI, IBM watsonx. Meta's own Llama API (announced at LlamaCon 2025-04-29) is the first-party hosted option; Muse Spark is API-only and runs on Meta's own infrastructure.
People who shaped Llama
Mark Zuckerberg — CEO of Meta. The 2024 “Open Source AI Is the Path Forward” letter and the 2025 strategic decisions (Scale AI acqui-hire, MSL formation, Muse Spark launch) all run through Zuckerberg's office. The closed-weights turn reverses the public position the letter staked out twenty-one months earlier.
Yann LeCun — Chief AI Scientist 2013–2025, FAIR cofounder. 2018 Turing Award. Departed November 2025 to launch a startup focused on world models, citing direction conflict with the closed-weights / scaling-LLMs orientation under Wang.
Joelle Pineau — led FAIR 2023–2025. The internal advocate for open-source releases through the Llama 1–3 era. Departed May 2025 ahead of the MSL reorganization.
Ahmad Al-Dahle — VP of Generative AI; led the GenAI org through Llama 2 / 3 / 4 launches. Spokesman for the LMArena Maverick episode in April 2025.
Alexandr Wang — Chief AI Officer of Meta since June 2025; head of Meta Superintelligence Labs. Previously CEO of Scale AI. The architect of the closed-weights pivot and the Muse line.
Nat Friedman — head of AI products at MSL since June 2025; previously CEO of GitHub.
The competitive landscape
Llama is — through April 2025 — the dominant open-weights frontier-AI line by deployment volume. The closest open-weights competitors are Mistral (French, also under a custom license rather than fully OSI-compliant), DeepSeek (Chinese, MIT-licensed for the V3 / R1 line and onward, the December 2024 / January 2025 inflection — see DeepSeek Versions), Alibaba's Qwen (also fully open-weights and dominant on HuggingFace leaderboards through 2025), and xAI's Grok 1 (open-weights only for the first generation, see Grok Versions). The closed-weights frontier competitors — ChatGPT, Claude, Gemini — have all stayed closed-weights since their inception. With Muse Spark in April 2026, Meta is the only major frontier lab to have started open-weights and pivoted closed; whether that pivot reverses with future Muse releases is the open question for the line going forward. This page does not attempt a benchmark roundup or a ranking.