Get AI summaries of any video or article — Sign up free
are we cooked w/ o3? thumbnail

are we cooked w/ o3?

The PrimeTime·
5 min read

Based on The PrimeTime's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

o3 is reported to score about 82% (public) and 75% (private) on the ARC AGI Benchmark’s ARC prize test, with stronger gains under high-compute settings.

Briefing

OpenAI’s o3 is posting standout results on the ARC AGI Benchmark, but the practical takeaway is less “AGI is here” and more “today’s capability is still gated by cost, compute, and integration realities.” The benchmark driving the hype is the ARC prize test—an assortment of logical puzzles designed to measure whether an AI can generalize to new situations. Against that yardstick, o3 is reported to outperform prior variants (including o1 and o1 mini), with scores cited around 82% on the public portion and 75% on the private portion. The gap widens further under “high compute” settings, where o3 is described as roughly 12% and 20% better on two comparisons than earlier systems.

A concrete example from the ARC-style tasks illustrates why the results matter: the puzzle involves moving colored squares based on “protrusions” (small protruding pixels) and then predicting the next configuration. The transcript frames this as something that can be solved quickly with enough compute—on the order of seconds—by inferring the rule linking the protrusions to the movement. That kind of generalization is what people associate with “AGI-adjacent” behavior, because it’s not just pattern matching on a single fixed template; it’s extracting a rule that transfers to a new instance.

Still, the economics are where the optimism runs into a wall. The cost figures discussed are steep: roughly $20 per low-compute task (with a claim of $5 per task for human performance treated skeptically), and about $2,000 to solve 100 semi-private tasks. Even more expensive is the high-compute mode, described as 172× more compute, with an implied cost around $140,000 for the same scale—only for an additional ~15% accuracy. That pricing, the argument goes, makes “daily use” unrealistic for most individuals and teams, and it also raises a second-order problem: real software work isn’t a single puzzle. A bug in a large codebase involves context across many files, dependencies, and edge cases, so the cost of “one task” can balloon far beyond a toy benchmark.

The transcript also points to operational constraints beyond cost: limited hardware availability if demand spikes, and the energy intensity of high-compute runs (one claim compares a single high-compute task to “five gallons of gasoline”). Even when the model finds a plausible fix, the human still has to review and validate, because the output can be wrong—meaning the workflow may not speed up as much as people expect.

The broader conclusion is cautionary. Benchmark wins can fuel investor narratives, but they don’t automatically translate into cheaper, widely usable tools. The practical path forward, the transcript argues, is that technical skill remains valuable—especially as code churn accelerates and AI becomes more integrated. For now, the “AGI” headline may change little day-to-day for most workers, unless dramatic cost reductions and better tooling arrive (with the transcript floating the idea that orders-of-magnitude efficiency improvements would be required). Until then, the safest bet is to keep building real engineering competence rather than waiting for AI to replace it.

Cornell Notes

o3 is reported to significantly outperform earlier models on the ARC AGI Benchmark, scoring about 82% on the public set and 75% on the private set, with larger gains under high-compute settings. The benchmark’s ARC prize test uses rule-based logic puzzles that require generalization to new instances, illustrated by a grid-movement task where the model must infer how “protrusions” determine shifts. Despite the impressive accuracy, the transcript emphasizes that per-task costs are extremely high—especially for high compute—making large-scale or routine use impractical. It also argues that real software debugging is harder than benchmark puzzles because one bug can involve thousands of files and extensive context, so accuracy gains may not translate into faster, cheaper engineering. The bottom line: capability is rising, but affordability, compute limits, and human review still dominate day-to-day impact.

What benchmark is driving the “AGI” chatter, and what does it measure?

The discussion centers on the ARC prize test, part of the ARC AGI Benchmark. It’s described as a public and semi-private set of logical puzzles where success depends on solving new instances by inferring underlying rules, not just memorizing a single pattern. The transcript treats higher scores as a proxy for being “on the road” toward AGI-like generalization, even while noting the test isn’t a perfect, guaranteed AGI detector.

How strong are o3’s reported results compared with earlier models?

Reportedly, o3 scores about 82% on the public portion and 75% on the private portion. Under low vs. high compute comparisons, the transcript claims o3 improves further—roughly 12% and 20% better in two cited high-compute comparisons versus earlier baselines (including o1 and o1 mini).

Why does the transcript treat the ARC puzzle example as meaningful rather than just a random win?

The example involves a grid task where colored squares move according to the presence of small protrusions (“poks”) extending from them. The transcript highlights that the model can infer the rule linking protrusions to movement and then predict the next configuration correctly. That kind of rule extraction and transfer to a new instance is framed as the core reason the benchmark is seen as a generalization test.

What cost and compute constraints are presented as the main barrier to real-world use?

The transcript gives per-task cost estimates: about $20 per low-compute task, with an additional claim of $5 per task for human performance treated skeptically. It then cites much larger totals for benchmark-scale runs (e.g., ~$2,000 for 100 semi-private tasks). High compute is described as 172× more compute, with an implied cost around $140,000 for the same scale for only ~15% more accuracy. It also flags hardware scarcity if many users demand high-compute runs simultaneously.

Why might benchmark accuracy not translate into faster software engineering?

Because real debugging isn’t a single isolated puzzle. The transcript argues that a “task” in a large codebase can require understanding context across many files and dependencies, so one bug could be far more complex than a few hundred small benchmark items. Even when the model proposes a fix, it may still be wrong, requiring human review and manual correction—reducing the expected speedup.

What is the transcript’s stance on “AGI achieved” headlines?

It’s skeptical of the leap from benchmark performance to immediate life-changing outcomes. The transcript suggests much of the hype functions as investor-driven narrative, and it argues that for most people the near-term effect may be limited unless costs drop dramatically and tooling becomes broadly usable.

Review Questions

  1. What specific ARC benchmark scores are cited for o3, and how do low-compute vs. high-compute results differ?
  2. Why does the transcript claim that debugging a real codebase can be far more expensive than solving benchmark puzzles?
  3. What conditions would need to change (cost, compute, workflow) for AI coding assistants to become realistically useful at scale?

Key Points

  1. 1

    o3 is reported to score about 82% (public) and 75% (private) on the ARC AGI Benchmark’s ARC prize test, with stronger gains under high-compute settings.

  2. 2

    The ARC prize test is framed as a rule-inference generalization challenge, not a fixed-template pattern task.

  3. 3

    High-compute improvements come with steep cost and compute multipliers (172× compute is cited), making large-scale use financially difficult.

  4. 4

    Real software bugs may require understanding context across thousands of files, so benchmark “task” difficulty may not map cleanly to engineering work.

  5. 5

    Even when model outputs are strong, human review and correction remain necessary because proposed fixes can still be wrong.

  6. 6

    Hardware availability and energy intensity are presented as additional bottlenecks if demand for high-compute runs spikes.

  7. 7

    The transcript argues that technical skill remains valuable as code churn accelerates and AI becomes more integrated.

Highlights

o3’s reported ARC prize test performance is cited as ~82% public and ~75% private, with additional gains under high compute.
A sample ARC puzzle is described as inferring a rule from grid “protrusions” to predict the next move—an example of rule transfer rather than memorization.
High-compute mode is portrayed as financially and operationally prohibitive: 172× more compute for only ~15% more accuracy, with implied costs reaching into the six figures.
The transcript warns that real debugging complexity (many files, dependencies, and edge cases) can dwarf toy benchmark tasks, limiting speedups.

Topics

  • ARC AGI Benchmark
  • o3 Evaluation
  • AI Coding Costs
  • High Compute Tradeoffs
  • Software Debugging

Mentioned

  • AGI
  • ARC
  • LLM