Get AI summaries of any video or article — Sign up free
o1 - What is Going On? Why o1 is a 3rd Paradigm of Model + 10 Things You Might Not Know thumbnail

o1 - What is Going On? Why o1 is a 3rd Paradigm of Model + 10 Things You Might Not Know

AI Explained·
5 min read

Based on AI Explained's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

o1 preview is portrayed as adding a correctness-reward objective on top of language modeling and helpfulness alignment.

Briefing

OpenAI’s o1 preview is being framed as a third major training paradigm for large language models: not just producing fluent text or aligning outputs with “helpful and harmless” goals, but actively rewarding answers that are objectively correct—especially when correctness comes from the model’s own multi-step reasoning. The shift matters because it changes what models are optimized for. Instead of relying primarily on next-word prediction and post-hoc alignment, o1-style training pushes systems to generate reasoning traces, then uses reinforcement learning to keep only the reasoning paths that lead to correct results.

At the core is a training loop that combines “test-time compute” (letting the model spend more steps thinking before answering) with “train-time compute” (fine-tuning on the best reasoning generations). The transcript describes an approach where models first generate many candidate chains of thought at higher diversity settings, then a grading mechanism filters them. The key is data efficiency: rather than learning from noisy web text that may include incorrect or irrelevant content, the process uses “golden data” consisting of correct answers paired with correct reasoning steps. That filtering is what turns reasoning into a scalable optimization target.

A major reason this is portrayed as a step change is that it addresses a long-standing weakness of chain-of-thought prompting: reasoning can be wrong even when it looks plausible. The proposed fix is to train the model with reinforcement learning so that incorrect reasoning is less likely to survive selection. The transcript emphasizes that o1’s gains show up most clearly in domains where correctness can be unambiguously checked—math, science, and coding—while performance can lag or regress in areas where “correct vs. incorrect” is harder to define, such as personal writing.

To make the practical impact concrete, the transcript uses a “librarian” metaphor. Earlier systems are likened to librarians who retrieve the right book but point to the wrong paragraph; o1-style systems are treated more like librarians who have learned which specific parts of books reliably answer questions. Still, the metaphor leaves an important limitation: if a question falls outside the training distribution—if the needed facts or procedures aren’t in the library—then even a better librarian can fail. The transcript also argues that there is no foundation model for the physical world yet, which helps explain why models can struggle with real-world spatial or physical reasoning.

Several “hidden hints” are offered about how o1 might work under the hood. The transcript connects o1 to earlier research such as “Let’s Verify Step by Step,” describing a verifier or reward model that checks individual reasoning steps to reduce false positives where the final answer is correct for the wrong reasons. It also draws parallels to chess: systems like Stockfish improved when evaluation moved from handcrafted heuristics to neural networks that effectively learned better internal “reasoning” strategies. The overall forecast is that as long as there’s a reliable grader for correctness, performance should keep rising with more compute and better verification.

Finally, the transcript notes that government interest is growing. The White House is described as taking these developments seriously, citing projects like o1 and Strawberry as relevant to national security and economic interests. The open question remains whether these advances amount to humanlike intelligence or general intelligence; the transcript’s stance is that o1 is a strong leap for narrow, checkable tasks, but it is not yet a complete solution for AGI.

Cornell Notes

The transcript frames o1 preview as a “third paradigm” for language models: beyond predicting text and beyond being helpful/harmless, it is trained to produce reasoning that leads to objectively correct answers. The mechanism combines test-time compute (more serial thinking before answering) with train-time compute via reinforcement learning that filters many generated reasoning traces down to those that are correct. A central claim is that o1 improves most in domains where correctness can be graded step-by-step (math, coding, physics), while it can regress in areas like personal writing where “correct” is less well-defined. The transcript also ties the approach to verifier-based methods such as “Let’s Verify Step by Step,” where reward models evaluate intermediate reasoning steps to prevent false positives. The practical implication: better graders and more compute can keep pushing performance upward, but distribution gaps and real-world complexity still limit generality.

What makes o1 preview different from earlier LLM training paradigms?

Earlier paradigms focus on next-word prediction (language modeling) and then alignment goals like “honest, harmless, and helpful.” The transcript claims o1 adds a third objective: reward answers that are objectively correct. That correctness reward is applied through reinforcement learning, using a grading/filtering process over generated reasoning traces rather than relying only on probability of text.

Why does “reasoning” need reinforcement learning if chain-of-thought prompting already exists?

Chain-of-thought prompting can produce long, step-by-step explanations, but those steps are often wrong. The transcript argues that o1’s advantage comes from training with reinforcement learning so that incorrect reasoning is less likely to be selected. Many candidate chains of thought are generated, then only those that lead to correct answers (and, in verifier-based variants, correct intermediate steps) are used for fine-tuning.

How do test-time compute and train-time compute work together in the described training loop?

Test-time compute is the model spending extra steps to think serially before producing a final answer. Train-time compute is the subsequent fine-tuning on the best generations—those that are graded as correct. The transcript describes two scaling effects: more time to think improves results, and fine-tuning on correct reasoning improves them further, with both effects not yet showing clear leveling off.

Why are math/science/coding gains emphasized, while personal writing may regress?

The transcript’s logic is that reinforcement learning needs a reliable way to distinguish correct from incorrect. In math, coding, and many science tasks, correctness is checkable, so the system can select “golden data” where both the final answer and reasoning steps are right. In personal writing, correctness is harder to define objectively, so the selection signal is weaker and performance can stagnate or regress.

What role do verifier/reward models play in preventing “false positives”?

The transcript links o1 to “Let’s Verify Step by Step,” describing a reward model that focuses on the process (intermediate steps) rather than only the final outcome. This reduces cases where the final answer is correct but the reasoning is flawed. The claim is that rewarding step-level correctness yields larger gains than rewarding only final answers.

What are the main limitations still highlighted for o1-style systems?

The transcript stresses distribution limits: if the needed facts or procedures aren’t in training (the “librarian” can’t retrieve a missing book), the model can still fail. It also notes that real-world physical/spatial intelligence lacks a foundation model with abundant “correct answers,” so spatial reasoning and real-world complexity remain challenging even if narrow reasoning tasks improve.

Review Questions

  1. Describe the training loop that combines test-time compute with reinforcement learning. What gets filtered, and what gets fine-tuned?
  2. Why does step-level verification (rewarding intermediate reasoning steps) matter compared with rewarding only final answers?
  3. List two domains where the transcript expects stronger gains and explain how the availability of an objective correctness signal drives that difference.

Key Points

  1. 1

    o1 preview is portrayed as adding a correctness-reward objective on top of language modeling and helpfulness alignment.

  2. 2

    The described improvement comes from reinforcement learning that filters many generated reasoning traces down to those that lead to correct outcomes.

  3. 3

    Test-time compute (more serial thinking before answering) and train-time compute (fine-tuning on correct generations) reinforce each other.

  4. 4

    Performance gains are strongest in tasks with unambiguous grading (math, coding, many science problems) and weaker where “correctness” is subjective (e.g., personal writing).

  5. 5

    Verifier/reward models that evaluate intermediate steps help prevent false positives where the final answer is right for the wrong reasons.

  6. 6

    Even with better reasoning, distribution gaps remain a major failure mode when the needed knowledge or procedures aren’t in the training library.

  7. 7

    Government interest is increasing, with the White House described as treating o1 and related projects as relevant to national security and economic interests.

Highlights

The transcript’s central claim is that o1’s step-change comes from reinforcement learning that rewards objectively correct reasoning, not just fluent text or aligned behavior.
A key mechanism is “golden data”: fine-tuning on correct answers paired with correct reasoning steps, filtered from many candidate chains of thought.
Step-level verification (as in “Let’s Verify Step by Step”) is presented as a way to reduce false positives by grading the process, not only the final result.
The strongest improvements are expected where correctness can be checked; weaker or regressive results are expected where correctness is hard to define objectively.

Topics