Get AI summaries of any video or article — Sign up free
Zuck's new Llama is a beast thumbnail

Zuck's new Llama is a beast

Fireship·
5 min read

Based on Fireship's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Llama 3.1 is offered in 8B, 70B, and 405B sizes, with the largest variant using a 128,000-token context window.

Briefing

Meta’s latest large language model, Llama 3.1, is positioned as a major leap in open-weight AI—especially with its biggest 405B parameter variant—while also being offered for free in a way that could pressure closed competitors like OpenAI and Anthropic. Meta trained the model over months on 16,000 Nvidia H100 GPUs, a scale that implies hundreds of millions of dollars in compute and enough electricity to power a small country. The payoff is a 405B model with a 128,000-token context window, and benchmark results that claim it is mostly superior to OpenAI’s GPT-4 and can beat Claude 3.5 Sonnet on select tests.

Still, raw benchmarks don’t settle whether a model is genuinely useful. Early hands-on reactions described the largest Llama as somewhat disappointing while smaller versions look more impressive. In practical testing, Llama 3.1 Heavy struggled with a specific “single-shot” coding task: generating a spelling-correct 5 web application using “runes,” a feature described as not yet released. The model performed decently for coding in general, but it showed limited awareness of that particular capability, and it fell behind Claude 3.5 Sonnet, which was reportedly the only model seen to handle the task correctly in one go.

Where Llama’s strategy becomes more consequential than its initial performance is in how developers can adapt it. Llama 3.1 is “open” in the sense that the model weights are available, enabling self-hosting and fine-tuning with custom data. That matters because it shifts cost and control away from paying per-request APIs (like GPT-4) and toward running models on rented or owned hardware. The transcript notes that self-hosting the biggest model is not cheap: the weights are about 230 GB, and even an RTX 490 setup wasn’t enough to run it smoothly with Ollama. But for teams that can afford infrastructure, the open-weight approach can reduce long-term dependency on closed model providers.

Meta’s openness has limits: the training code is described as relatively small—around 300 lines of Python using PyTorch and FairScale for distributed training—yet the training data itself is not open. That data could include personal and proprietary sources such as blog posts, GitHub repositories, old Facebook content, and potentially WhatsApp messages, raising familiar privacy and consent questions even as the model weights remain accessible.

The transcript also frames a broader industry plateau. Multiple companies have poured massive compute into ever-larger models, but capability gains appear to be leveling off, with improvements since GPT-4 described as more incremental than revolutionary. Despite high-stakes rhetoric about AI’s future, programmers still aren’t being replaced, and “Skynet” scenarios remain absent. In that context, Llama 3.1 is portrayed as Meta’s most credible attempt to keep the ecosystem moving—less about sudden artificial superintelligence and more about practical developer access, customization, and competition in the model market.

Cornell Notes

Meta’s Llama 3.1 is a large language model released with open-weight availability and a very long 128,000-token context window, including a 405B “Heavy” variant. Training reportedly used 16,000 Nvidia H100 GPUs over months, producing a 405B model that claims strong benchmark performance versus GPT-4 and Claude 3.5 Sonnet on some tests. Hands-on results suggest the biggest model can underperform on certain niche “single-shot” coding tasks, even while smaller variants look more capable. The biggest practical advantage is that developers can self-host and fine-tune using custom data, avoiding API costs—though the largest weights (~230 GB) are expensive to run locally. Training code is described as relatively compact, but the training data remains closed.

What makes Llama 3.1 strategically different from many closed competitors?

The model weights are available for developers, enabling self-hosting and fine-tuning. That can reduce reliance on paid APIs (like GPT-4) and shift costs toward infrastructure. The transcript also highlights that the training code is relatively small (about 300 lines of Python with PyTorch and FairScale), which lowers the barrier to understanding and adapting the training approach—while still keeping the training data itself closed.

How big is Llama 3.1, and what does “405B” and “128,000 token context” mean in practice?

Llama 3.1 comes in three sizes: 8B, 70B, and 405B, where “B” refers to billions of parameters. The 405B model is paired with a 128,000-token context window, meaning it can process very long inputs in a single run. More parameters can capture more complex patterns, but the transcript stresses that bigger isn’t automatically better.

Why do benchmarks not settle the question of whether the model is “good”?

Benchmark scores can miss how a model behaves on real, specific tasks. The transcript notes that early internet feedback found the largest Llama somewhat disappointing compared with smaller versions. In a hands-on test, Llama 3.1 Heavy failed a niche single-shot web-app generation task involving “runes,” while Claude 3.5 Sonnet handled it correctly.

What does the transcript say about self-hosting costs and feasibility?

Self-hosting the biggest model is portrayed as difficult. Using Ollama, the transcript claims the 405B weights are about 230 GB, and even an RTX 490 setup couldn’t run it. The alternative is trying it through hosted platforms such as Meta, Gro, or Nvidia’s Playground, which avoids local compute costs.

What limits the “open” aspect of Llama 3.1?

Even with open weights, the training data is not open. The transcript suggests the dataset could include sources like blog posts, GitHub repositories, old Facebook posts, and possibly WhatsApp messages. That creates a privacy and transparency gap: developers can inspect and run the model, but not the underlying data used to train it.

What broader trend does the transcript claim about the AI industry’s progress?

Despite massive compute investments across companies, capability gains are described as plateauing at a similar level. The transcript contrasts OpenAI’s earlier leap from GPT-3 to GPT-4 with later progress characterized as smaller incremental improvements, and it argues that dramatic “AI takeover” predictions have not materialized.

Review Questions

  1. Which part of Llama 3.1 is open (weights, code, or training data), and how does that affect developers’ ability to customize models?
  2. Why might the 405B model underperform on a specific coding task even if it scores well on benchmarks?
  3. What practical constraints does the transcript mention for running Llama 3.1 Heavy locally, and what alternatives are suggested?

Key Points

  1. 1

    Llama 3.1 is offered in 8B, 70B, and 405B sizes, with the largest variant using a 128,000-token context window.

  2. 2

    Meta claims Llama 3.1 Heavy performs strongly on benchmarks, including comparisons against GPT-4 and Claude 3.5 Sonnet on some tests.

  3. 3

    Hands-on testing suggests the 405B model can struggle with niche single-shot tasks involving “runes,” even if it codes reasonably well overall.

  4. 4

    Open-weight availability enables self-hosting and fine-tuning with custom data, potentially reducing dependence on paid API access.

  5. 5

    Self-hosting the 405B model is expensive and storage-heavy (about 230 GB of weights), and local hardware may not be sufficient for smooth use.

  6. 6

    The training code is described as relatively compact (around 300 lines of Python with PyTorch and FairScale), but the training data remains closed.

  7. 7

    The transcript frames industry progress as plateauing despite large compute investments and questions the gap between AI hype and real-world capability gains.

Highlights

Llama 3.1 Heavy (405B) pairs a massive parameter count with a 128,000-token context window, trained on 16,000 Nvidia H100 GPUs over months.
Open weights can shift AI app development from API payments toward self-hosting and fine-tuning—though the 405B weights are about 230 GB.
Benchmarks look strong on paper, but a niche “runes” web-app task exposed gaps in the 405B model’s single-shot awareness.
The training pipeline is described as surprisingly small in code terms (about 300 lines), using PyTorch and FairScale for distributed training.
Despite years of escalating compute and rhetoric, the transcript argues AI progress has not produced the dramatic replacement scenarios once promised.