Get AI summaries of any video or article — Sign up free
GPT 4 is SHOCKINGLY Good! Results/Tests that will blow your mind & How YOU Can Get Access! thumbnail

GPT 4 is SHOCKINGLY Good! Results/Tests that will blow your mind & How YOU Can Get Access!

MattVidPro·
5 min read

Based on MattVidPro's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

GPT-4 is described as outperforming GPT-3.5 on academic-style benchmarks, including a simulated bar exam where GPT-4 reportedly scores in the top 10%.

Briefing

GPT-4 is positioned as a major leap beyond GPT-3.5: it performs at near human levels on academic-style benchmarks, handles far more input at once, and—most visibly—can “see” by accepting images and reasoning over what’s in them. The practical impact is that GPT-4 isn’t limited to text chat; it can interpret screenshots, photographs, diagrams, and even multi-panel memes, then respond with detailed, context-aware explanations or answers. OpenAI’s own examples highlight that the biggest gains show up when tasks get complex, not during casual conversation where GPT-3.5 already sounds fluent.

A key benchmark example comes from a simulated bar exam. GPT-3.5 scored in the bottom 10% of test takers, while GPT-4 reportedly lands in the top 10%, implying a dramatic shift in reliability and factual quality under pressure. OpenAI also claims improvements in steerability—keeping outputs within “guard rails”—and reports that early testing found GPT-4 unusually stable, with performance that matched predictions more closely than earlier models.

Access is framed as relatively straightforward for many users. ChatGPT Plus subscribers can access GPT-4, though not the full version: image input isn’t available there yet and token limits differ. Bing AI is also described as using a fine-tuned GPT-4 variant, again without image support. For developers, an API is announced with a waitlist. The transcript emphasizes one concrete technical upgrade: GPT-4 can accept up to 25,000 tokens versus GPT-3.5’s 4,000, expanding what users can feed into a single request—documents, screenshots, and longer prompts.

Image understanding is illustrated through a series of examples. GPT-4 is shown identifying the humor in an image where an Apple charger is packaged with an old VGA connector, explaining the joke panel by panel and tying the absurdity to the mismatch between outdated hardware and a modern phone charging port. Other demonstrations include reading homework-style questions from images (including French prompts answered in English), interpreting odd scenes like a man ironing clothes on a taxi roof, and extracting information from research papers to derive details such as salaries. Memes are treated as legitimate reasoning tasks: GPT-4 explains why a “beautiful Earth” caption is funny when the image is actually chicken nuggets arranged like a world map.

Beyond multimodal capability, the transcript highlights steerability via system messages—allowing customization of tone and behavior, such as making the model act like a specific “character” or tool. Still, limitations remain. GPT-4 can make reasoning errors and can produce confident-sounding inaccuracies; the transcript notes that factual evaluations can still fall around 60% in some comparisons, reinforcing the need to treat outputs as fallible.

Finally, the transcript surveys early integrations and community demos. Companies and apps are described as moving toward GPT-4 features: education tools, coding assistants, visual assistants, and even drug-discovery workflows that search for compounds and help with procurement steps. Coding demos include generating working games like Snake, Pong, and Connect Four from minimal instructions. The overall takeaway is that GPT-4 is being treated as a “backbone” model—one that can power assistants across work, learning, accessibility, and software creation—while still requiring careful verification for high-stakes use.

Cornell Notes

GPT-4 is presented as a step-change from GPT-3.5: it performs much better on academic benchmarks, accepts far more input (up to 25,000 tokens), and can process images alongside text. The most striking improvements show up in harder, more nuanced tasks—especially when image understanding is involved—such as explaining the humor in multi-panel memes or answering homework questions from screenshots. Access is available through ChatGPT Plus and Bing AI (with some limitations like image support not being available in those specific entry points), while developers can join an API waitlist. Despite the gains, GPT-4 still makes reasoning and factual mistakes, so outputs should be verified rather than treated as guaranteed truth.

What makes GPT-4 different from GPT-3.5 in practical terms, beyond “better writing”?

The transcript points to three concrete upgrades: (1) higher benchmark performance (including a simulated bar exam where GPT-4 reportedly scores in the top 10% versus GPT-3.5 in the bottom 10%), (2) much larger context windows (25,000 tokens vs 4,000), and (3) multimodal input—GPT-4 can accept and interpret images, not just text. It also emphasizes that casual conversation may look similar, while complex tasks reveal bigger gaps.

How does image capability change what users can ask GPT-4 to do?

Image input lets GPT-4 reason over what’s shown: it can describe multi-panel images, extract information from screenshots and research papers, and interpret visual humor. Examples include explaining why a VGA connector in an Apple charger package is funny, answering homework questions from an image prompt (French to English), and reading paper images to derive details like salaries without requiring users to copy/paste text.

What does “steerability” mean in this context, and why does it matter?

Steerability refers to controlling the model’s behavior within allowed boundaries using system messages. The transcript frames this as customizing tone and function—e.g., making the model act like a specific “tool” or character rather than a generic chatbot. This is presented as a way to tailor outputs for different workflows (like tutoring, coding help, or other specialized assistants).

Where can people access GPT-4, and what limitations are mentioned?

ChatGPT Plus is described as providing access to GPT-4, but not the full version: it can’t accept all new token limits and can’t accept images yet. Bing AI is also described as using a fine-tuned GPT-4 variant, but similarly without image support. For developers, an API is available via a waitlist.

What limitations remain even with GPT-4’s improved performance?

The transcript stresses that GPT-4 can still make reasoning errors and produce confident inaccuracies. It cites factual evaluation comparisons where accuracy may still land around 60% in some tests. The takeaway is to treat outputs as useful but not automatically reliable—especially for high-stakes decisions.

What kinds of applications are emerging from GPT-4’s capabilities?

Early demos and integrations described include: generating websites from photographed notebook sketches, creating coding projects and games (Snake, Pong, Connect Four) from brief prompts, visual assistants that interpret images in real time (including potential accessibility benefits), and drug-discovery workflows that search for similar compounds and support procurement steps. Education and productivity tools are also highlighted as likely beneficiaries.

Review Questions

  1. Which three upgrades does the transcript highlight as the biggest reasons GPT-4 outperforms GPT-3.5 in real use?
  2. Give one example of how GPT-4 uses images to solve a task that would be harder with text-only input.
  3. Why does the transcript recommend verifying GPT-4 outputs even when they sound confident?

Key Points

  1. 1

    GPT-4 is described as outperforming GPT-3.5 on academic-style benchmarks, including a simulated bar exam where GPT-4 reportedly scores in the top 10%.

  2. 2

    GPT-4 accepts far more input at once (25,000 tokens) than GPT-3.5 (4,000), enabling longer documents and richer prompts.

  3. 3

    Multimodal capability is a major shift: GPT-4 can interpret images—screenshots, diagrams, photos, and multi-panel memes—and respond with context-aware explanations.

  4. 4

    Access is available via ChatGPT Plus and Bing AI, with noted limitations (notably image support not yet available in those entry points) and an API waitlist for developers.

  5. 5

    Steerability via system messages enables more tailored behavior, including tool-like or character-like responses within guard rails.

  6. 6

    Even with improvements, GPT-4 can still produce reasoning and factual errors; verification remains necessary for important decisions.

  7. 7

    Early integrations and demos point to broad use cases: coding and game generation, website creation from sketches, education support, accessibility-focused visual assistance, and drug-discovery workflows.

Highlights

GPT-4’s reported bar exam jump—from GPT-3.5’s bottom-10% performance to GPT-4’s top-10%—is used as a headline indicator of reliability under high-stakes conditions.
The transcript’s image examples treat humor and homework as solvable vision tasks: GPT-4 explains multi-panel jokes and answers questions from screenshots, sometimes translating into English.
A single technical upgrade—25,000-token input versus 4,000—frames GPT-4 as better suited for long documents, complex instructions, and multi-step work.
Despite stronger performance, the transcript warns that accuracy can still be imperfect (around 60% in some factual evaluations), so outputs shouldn’t be accepted blindly.

Topics

Mentioned

  • GPT
  • API
  • VGA
  • AI
  • HTML