Get AI summaries of any video or article — Sign up free
Claude 3.7 goes hard for programmers… thumbnail

Claude 3.7 goes hard for programmers…

Fireship·
5 min read

Based on Fireship's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Claude 3.7’s programming push combines a stronger base model, a new thinking mode, and Claude Code, a CLI that can build and test inside a real project.

Briefing

Anthropic’s Claude 3.7 is pushing programming-focused AI into a new tier by combining a stronger base model with a “thinking mode” and, most importantly for developers, a new CLI tool called Claude Code that can build, test, and run code inside a real project—creating a tight feedback loop that’s meant to reduce the back-and-forth between humans and models.

The programming impact starts with performance claims. Claude 3.7 Sona, the newly released model, is described as beating its own prior baseline while adding a thinking mode modeled after the success of DeepSeek R1-style approaches in open “reasoning” models. In benchmark terms, Claude 3.7 is said to have jumped ahead on a human-verified software engineering test set built from real GitHub issues. The headline figure is 70.3% of issues solved, surpassing other models including OpenAI O3 mini High and DeepSeek. The transcript then shifts from leaderboard talk to hands-on testing, where Claude Code is positioned as the practical mechanism behind the hype.

Claude Code is a research-preview CLI installable via npm. It uses the Anthropic API directly and comes with a steep cost: over 10× the price of models like Gemini Flash and DeepSeek, at $15 per million output tokens. After installation, the CLI provides a command that scans an existing codebase, generates a markdown context/instructions file, and then opens an interactive session where the model can propose changes and write files to disk.

In early tests, Claude Code behaves like an agent that can manage project structure and testing. A simple “random name generator” task results in new files plus a dedicated testing file, reflecting a workflow aligned with strongly typed languages and test-driven development. When tests fail, the tool can iterate—rewriting logic and re-running until the test suite passes.

The transcript’s more demanding test targets a moderately complex UI: a TypeScript + Tailwind + “spelt” front end that records microphone input and visualizes a waveform. Claude Code requires many confirmations, but it produces a working interface with interactive waveform controls and graphics. A comparison run using OpenAI O3 mini High generates an inferior result and, on inspection, misses key stack details (not using TypeScript/Tailwind as expected and failing to apply the newer “spelt 5 Rune syntax”). Claude Code’s session cost is reported at about 65 cents for that UI build.

Still, the tool isn’t portrayed as a universal fix. A final attempt—building an encrypted app after Apple discontinued end-to-end encryption in the UK—fails to run despite extensive code changes. The transcript emphasizes a practical limitation: even strong coding agents can get stuck on runtime errors, and heavy reliance can leave developers without the context to debug.

The closing pitch ties Claude Code’s strengths to backend productivity via Convex, an open-source reactive database with typesafe queries and server functions. The claim is that AI coding works better when the backend is structured in predictable, typescript-native patterns—making autonomous “vibe coding” more reliable. The overall takeaway is clear: Claude 3.7 and Claude Code can materially accelerate real development, but they still demand human oversight when the task hits tricky runtime or security constraints.

Cornell Notes

Claude 3.7 Sona is framed as a major step forward for programming AI, combining a stronger base model, a new thinking mode, and—most crucially—a developer tool called Claude Code. Claude Code is a CLI that scans an existing project, generates context, and then iteratively builds and tests code by writing files to disk and using test feedback to correct logic. In benchmark claims, Claude 3.7 is reported to solve 70.3% of GitHub issues on a human-verified software engineering benchmark, outperforming models like OpenAI O3 mini High and DeepSeek. Hands-on tests show Claude Code can generate a working TypeScript/Tailwind/spelt UI with microphone waveform visualization, but it can still fail on harder runtime tasks like building an encrypted app. The practical value is speed and iteration—paired with the need for debugging skill when errors persist.

What makes Claude 3.7 feel different for programmers beyond raw model quality?

The transcript highlights three layers: (1) a stronger base model that “beats itself” versus the prior version, (2) a new thinking mode aimed at improving reasoning, and (3) Claude Code, a CLI tool that can operate inside a real codebase. Claude Code doesn’t just generate text; it scans the project, creates a markdown context/instructions file, and then writes code and testing files to disk so it can iterate based on test results.

How strong are the programming performance claims, and what benchmark is cited?

The transcript cites a human-verified software engineering benchmark based on real GitHub issues. Claude 3.7 is said to solve 70.3% of issues, surpassing OpenAI O3 mini High and DeepSeek. It also notes that Claude 3.5 was already near the top on a web dev leaderboard, but Claude 3.7 is described as decisively ahead on the software engineering benchmark.

What does Claude Code do during a typical coding task?

After installation via npm (research preview), Claude Code uses the Anthropic API and provides a terminal command that scans the project. It generates a markdown file with initial context and instructions, then runs an interactive session where it proposes changes, writes new files, and creates dedicated testing files. If tests fail, it uses that feedback to rewrite business logic and continue iterating until tests pass.

How did Claude Code perform on the UI build test, and what stack details mattered?

For a moderately complex front end—TypeScript + Tailwind + spelt—the transcript says Claude Code required many confirmations but produced a working UI that visualizes microphone waveform data with interactive controls. A comparison with OpenAI O3 mini High produced an “embarrassing” result; closer inspection found Claude’s competitor failed to apply TypeScript/Tailwind and didn’t use the newer spelt 5 Rune syntax, while Claude Code did.

Where did Claude Code struggle despite strong coding output?

In the encrypted-app scenario tied to Apple’s UK end-to-end encryption discontinuation, Claude Code made extensive changes but still failed to run. The transcript underscores a key limitation: when runtime errors persist, the developer may not know how to debug, especially if they’ve become dependent on AI-generated code.

Why does the Convex sponsor pitch connect to AI coding success?

Convex is described as a reactive database with typesafe queries, scheduled jobs, server functions, and real-time sync. The transcript claims that because Convex queries are written in pure TypeScript, AI models can better understand the backend structure, leading to fewer errors and more productive autonomous coding. The argument is that pairing Claude Code’s front-end generation with a structured TypeScript backend can improve reliability.

Review Questions

  1. What specific workflow steps does Claude Code perform (scan, context generation, file writing, testing/iteration), and why does that matter for correctness?
  2. Which benchmark metric and dataset type are used to justify Claude 3.7’s programming performance, and how does it compare to OpenAI O3 mini High and DeepSeek?
  3. Describe one example where Claude Code succeeded and one where it failed. What kinds of tasks seem to trigger each outcome?

Key Points

  1. 1

    Claude 3.7’s programming push combines a stronger base model, a new thinking mode, and Claude Code, a CLI that can build and test inside a real project.

  2. 2

    Claude Code scans an existing codebase, generates context/instructions, then writes code and testing files so it can iterate based on test outcomes.

  3. 3

    The transcript cites a human-verified GitHub-issue benchmark where Claude 3.7 is claimed to solve 70.3% of issues, ahead of OpenAI O3 mini High and DeepSeek.

  4. 4

    Hands-on tests suggest Claude Code can produce a working TypeScript + Tailwind + spelt UI for microphone waveform visualization, while a comparison model missed key stack requirements.

  5. 5

    Claude Code is expensive: $15 per million output tokens, described as over 10× the cost of some other models mentioned.

  6. 6

    Even with strong code generation, Claude Code can still fail on complex runtime/security tasks, leaving developers to debug errors they may not understand.

  7. 7

    Using a structured TypeScript backend like Convex is pitched as a way to make AI-assisted coding more reliable and less error-prone.

Highlights

Claude Code turns programming from “generate code” into “generate code + run tests + iterate,” by writing files and using test feedback loops.
Claude 3.7 is claimed to solve 70.3% of GitHub issues on a human-verified software engineering benchmark, a jump framed as decisive versus other top models.
In the microphone waveform UI test, Claude Code produced a working interface while the comparison model missed TypeScript/Tailwind and failed to use spelt 5 Rune syntax.
Claude Code still couldn’t get an encrypted-app build to run, showing that runtime failures remain a hard boundary for coding agents.

Topics

Mentioned