Get AI summaries of any video or article — Sign up free
Introduction to GPT-4.5 thumbnail

Introduction to GPT-4.5

OpenAI·
6 min read

Based on OpenAI's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

GPT-4.5 is positioned as OpenAI’s largest, most knowledgeable model yet, combining unsupervised learning scaling with reasoning-oriented training.

Briefing

GPT-4.5 is being rolled out as OpenAI’s largest, most knowledgeable model yet, positioned as a “research preview” that blends two scaling approaches: reasoning that helps models handle complex problems, and unsupervised learning that boosts language accuracy, world knowledge, and intuition—without requiring step-by-step “think first” behavior like OpenAI’s o1-style reasoning models. The practical promise is a chat experience that feels warmer and more context-aware, while also reducing hallucinations and improving performance on both everyday knowledge questions and harder professional tasks.

OpenAI frames GPT-4.5’s core advance as scaling unsupervised learning to increase “word knowledge” and reduce false answers, while reasoning training improves how the model approaches tasks such as science and math. Unlike models that explicitly reason step by step, GPT-4.5 is described as generally useful and “inherently smarter,” with experimentation still underway to understand which capabilities emerge from unsupervised learning at this scale. In demos, GPT-4.5 is shown responding more naturally to social context—recognizing frustration in a request to send an angry text and offering a more nuanced, constructive message instead. When prompted to produce the angry text anyway, it can still follow the user’s instruction, but the contrast is used to highlight its ability to detect intent and emotional cues.

The rollout is paired with claims of measurable improvements. OpenAI says GPT-4.5 outperforms the GPT family on accuracy and has the lowest hallucination rate in a comparison using a QA evaluation setup. For collaboration and tone, human testers rated GPT-4.5 against GPT-4o and GPT-4.5 on categories including warmth and emotional nuance, with GPT-4.5 reportedly winning across the board. A “Vibes” test set is used to quantify EQ-like qualities—how collaborative and warm the tone feels—using an opinionated prompt set screened to align with those goals.

Under the hood, OpenAI attributes GPT-4.5’s performance to major infrastructure and training changes. The model required new post-training methods to fine-tune a large system using a smaller compute footprint, using supervised fine-tuning plus reinforcement learning with human feedback across multiple iterations. On the pre-training side, OpenAI says it pushed compute aggressively with low-precision training and pre-trained across multiple data centers simultaneously to use more compute than a single high-bandwidth networking fabric could handle. Serving at scale also demanded new inference systems designed to keep responses fast and “snappy.”

OpenAI also walks through an “ocean is salty” evolution across GPT generations, using it to illustrate how GPT-4.5’s answers became more concise, cohesive, and personality-driven—moving from wrong or rambling responses to a clear explanation. Benchmark results are presented to show gains from unsupervised learning across reasoning-heavy science evals, math, agentic coding, multilingual understanding, and multimodal understanding. While GPT-4.5 is said to lag behind explicit “think before responding” models like o3-mini on reasoning-heavy evals, it still reaches high scores without that step-by-step behavior.

Finally, OpenAI outlines availability: GPT-4.5 starts with all Pro users in web, mobile, and desktop via the model picker, then expands to Team and Plus next week, followed by Edu and Enterprise. Developers on paid tiers get access immediately, with features such as function calling and structured outputs, plus integration with file and image upload, canvas, and search. The message is that reasoning will remain central to future models, but unsupervised learning at scale is being treated as a foundational path toward more intuitive, knowledgeable AI and better human interaction.

Cornell Notes

GPT-4.5 is OpenAI’s latest large, knowledge-rich model, released first as a research preview for Pro users and developers, then expanding to broader tiers. Its key improvement comes from scaling unsupervised learning to raise factual accuracy, world knowledge, and intuition while also reducing hallucinations. OpenAI pairs that with reasoning-oriented training so the model handles complex tasks like science and math more effectively, even though it is not built to “think step by step” like o1-style models. In demos and evaluations, GPT-4.5 is described as more context-aware and emotionally nuanced, scoring better on human-rated collaboration and “Vibes” tests. The rollout also highlights major infrastructure work for low-precision training, multi–data center pre-training, and new inference systems to keep latency low.

What two training paradigms does OpenAI say GPT-4.5 scales, and what does each contribute?

OpenAI describes scaling two paradigms: (1) unsupervised learning, which increases language accuracy and “word model” intuition and world knowledge, and helps reduce hallucinations; and (2) reasoning training, which teaches models to think before responding and improves performance on tasks requiring reasoning such as science and math. GPT-4.5 is positioned as a generally useful, inherently smarter model that benefits from both, but it is not characterized as a step-by-step reasoning model in the same way as o1.

How does GPT-4.5’s behavior differ from a more explicitly reasoning model in the demos?

In the social-context demo, GPT-4.5 recognizes frustration and offers a more nuanced, constructive text rather than an automatically angry message, showing sensitivity to intent and emotional cues. When asked for the angry text anyway, it can still comply. In a separate example about explaining the need for AI alignment, GPT-4.5 is described as flowing more naturally and guiding thinking through the ideas, while o1 is portrayed as producing a lot of information and being useful but less natural in its conversational flow.

What evaluation signals does OpenAI use to claim GPT-4.5 is more accurate and less hallucination-prone?

OpenAI presents a QA evaluation where one axis tracks accuracy and another tracks hallucination rate. In that comparison, GPT-4.5 is said to outperform the GPT family on accuracy while also having the lowest hallucination rate. It also reports human testing where evaluators rate GPT-4.5 against GPT-4o and GPT-4.5 on categories tied to factuality/accuracy in everyday queries and on collaboration/tone.

What does “Vibes” mean in the reported testing, and how is it measured?

“Vibes” is defined as an EQ-like measure: how collaborative the model feels and how warm its tone is. OpenAI says it measures this by selecting an opinionated set of prompts and screening trainers for prompts that align with the overall “Vibes” target, then using human evaluation on that set.

What infrastructure and training changes does OpenAI cite as necessary to build and serve GPT-4.5?

OpenAI credits major pre-training and post-training work. Pre-training included aggressive low-precision training and distributing pre-training across multiple data centers simultaneously to use more compute than one high-bandwidth networking fabric could support. Serving required new inference systems to keep the model fast and responsive. Post-training used supervised fine-tuning plus reinforcement learning with human feedback across multiple iterations, using a new training mechanism intended to fine-tune a very large model with a smaller compute footprint.

How does GPT-4.5’s benchmark performance relate to explicit reasoning models like o3-mini?

OpenAI reports large improvements from GPT-4.5 on reasoning-heavy science evals (GPT-4.5 improves substantially) but notes it still lags behind o3-mini, which can think and reason before responding—especially useful for that eval. For other tasks like agentic coding (SWE-bench verified) and agentic coding that benefits from deeper world knowledge (SWE-lancer), GPT-4.5 is said to outperform even o3-mini, highlighting complementary strengths between unsupervised learning and reasoning scale.

Review Questions

  1. Why does OpenAI claim unsupervised learning scaling can reduce hallucinations, and how is that reflected in their QA evaluation?
  2. In what ways does GPT-4.5’s conversational tone and emotional nuance get measured, and what does “Vibes” specifically target?
  3. What engineering constraints arise when training and serving a very large model, and which pre-training and inference techniques does OpenAI name to address them?

Key Points

  1. 1

    GPT-4.5 is positioned as OpenAI’s largest, most knowledgeable model yet, combining unsupervised learning scaling with reasoning-oriented training.

  2. 2

    Unsupervised learning is credited with improving factual accuracy, world knowledge, intuition, and lowering hallucinations, while reasoning training targets complex tasks like science and math.

  3. 3

    In demos, GPT-4.5 shows stronger context and intent sensitivity, producing more emotionally nuanced responses to social situations even when users request harsher outputs.

  4. 4

    OpenAI reports GPT-4.5 achieves higher accuracy and the lowest hallucination rate in a QA evaluation compared with the GPT family.

  5. 5

    Human evaluations are used to measure collaboration and tone, with GPT-4.5 outperforming GPT-4o and GPT-4.5 across categories and scoring well on a “Vibes” (EQ-like) test set.

  6. 6

    Training GPT-4.5 required low-precision pre-training, multi–data center pre-training, new inference systems for low latency, and post-training via supervised fine-tuning plus reinforcement learning with human feedback.

  7. 7

    Availability starts with Pro users and developers, then expands to Team and Plus next week, followed by Edu and Enterprise, with developer features like function calling and structured outputs.

Highlights

GPT-4.5 is described as “generally useful and inherently smarter,” improving accuracy and reducing hallucinations through scaled unsupervised learning rather than relying solely on step-by-step reasoning.
A social-context demo contrasts an “angry text” request with GPT-4.5’s ability to detect frustration and respond more constructively while still allowing instruction-following when explicitly demanded.
OpenAI ties GPT-4.5’s warmer, more collaborative feel to human-rated evaluations and a “Vibes” test set measuring EQ-like qualities.
Engineering claims include low-precision training, multi–data center pre-training, and new inference systems designed to keep responses fast and interactive at large scale.

Topics

Mentioned

  • Mia
  • Rafa
  • Yol
  • Jason
  • Alex
  • EQ
  • QA
  • API
  • MLU
  • MMU
  • SWE-bench
  • SWE-lancer
  • o1
  • o3-mini