Get AI summaries of any video or article — Sign up free
OpenAI DevDay 2025: Opening Keynote with Sam Altman thumbnail

OpenAI DevDay 2025: Opening Keynote with Sam Altman

OpenAI·
7 min read

Based on OpenAI's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

OpenAI is turning ChatGPT into an app platform by launching the Apps SDK (preview), built on MCP for full control of backend logic and UI.

Briefing

OpenAI’s DevDay 2025 opening keynote makes one central pitch: building useful AI products is getting dramatically easier—because ChatGPT is turning into an app platform, agents are becoming production-ready with less glue code, and coding itself is shifting toward AI-driven workflows across tools, teams, and devices.

The keynote starts with scale and momentum. In 2023, OpenAI cited 2 million developers, 100 million weekly ChatGPT users, and roughly 300 million tokens per minute on its API. By DevDay 2025, it claims 4 million developers, more than 800 million weekly ChatGPT users, and over 6 billion tokens per minute on the API. That growth sets up the day’s focus: tools that shorten the path from idea to working product.

First up is “apps inside ChatGPT.” OpenAI says it’s launching a new Apps SDK (preview) that lets developers build interactive, adaptive, personalized experiences directly within ChatGPT conversations. The SDK is built on MCP, giving developers control over backend logic and frontend UI, and it’s designed to make apps discoverable and scalable—developers can reach “hundreds of millions” of ChatGPT users. The keynote also ties monetization to distribution: users can log in to existing products from within the conversation, and future monetization includes an “agentic commerce protocol” enabling instant checkout inside ChatGPT.

Concrete examples show how discovery and context work. Developers can surface apps by name (e.g., uploading a sketch and asking “Figma, turn this sketch into a workable diagram”), and ChatGPT can recommend apps when users ask for outcomes (e.g., suggesting Spotify for a party playlist). A live demo highlights inline media and interaction: a Coursera app plays video inside the chat, then uses an Apps SDK mechanism (“talking to apps”) to expose the user’s current context back to ChatGPT so the model can answer questions about what’s on screen. Other demos show Canva generating portfolios and pitch decks from conversation context, and a Zillow app embedding an interactive map that updates without creating a new instance—then filtering results and answering follow-up questions using the context the app provides.

Next, the keynote tackles agents—where prototypes often fail to reach production. OpenAI introduces Agentkit, positioned as a complete set of building blocks to move from prototype to production with less orchestration overhead. Agent Builder provides a visual way to design logic steps and flows on top of the Responses API. Chatkit offers an embeddable chat interface for custom branding and workflows. Evals for Agents adds trace grading, datasets, and automated prompt optimization, including the ability to run evals on external models within OpenAI’s eval platform. For data access, OpenAI points to a connector registry with an admin control panel to connect internal and third-party systems safely.

The keynote then demonstrates Agentkit’s speed by building a DevDay navigation agent (“Ask Froge”) live—using visual workflow wiring, prebuilt guardrails (including PII protection), widgets, previewing, and publishing to production with a workflow ID. The agent is then embedded into the DevDay site using Chatkit components, with the ability to iterate without rewriting code.

Finally, the keynote shifts to software creation. CodeX moves from research preview to general availability, now running on the GPT5-Codex model trained for agentic coding tasks like refactoring and code review. OpenAI cites rapid adoption (10x daily messages since early August) and internal impact (engineers completing more pull requests; nearly all PRs reviewed by CodeX). New enterprise features include Slack integration, a CodeX SDK for team workflows, and admin tools for monitoring and analytics.

A live demo shows CodeX controlling real devices and building interfaces: a camera control panel, a wireless controller workflow, and voice-driven lighting changes using MCP servers and the agentic toolchain. The keynote closes by expanding the model lineup for developers and creators: GPT5 Pro in the API for hard domains, GPT-REALTIME-MINI for cheaper expressive voice, and a Sora 2 preview in the API with more controllability, sound pairing, and remixing options—plus examples involving product concepting and toy design.

Taken together, the message is that AI product development is shifting from experimentation to deployment: apps become conversational UI, agents become measurable and safer, and coding becomes a collaborative, tool-using workflow that can span IDEs, terminals, Slack, and real-world devices.

Cornell Notes

OpenAI’s DevDay 2025 opening keynote argues that AI builders are entering a faster, more production-ready era. It introduces the Apps SDK (preview) for building interactive apps inside ChatGPT, built on MCP so developers control both backend logic and UI while reaching large ChatGPT audiences. It also launches Agentkit to reduce the complexity of agent orchestration, adding visual workflow building, embeddable chat interfaces, agent-specific evals, and safer data connections via a connector registry. CodeX moves from research preview to general availability on GPT5-Codex, with new team features like Slack integration and enterprise admin tools. The keynote ties these platform upgrades to model releases (GPT5 Pro, GPT-REALTIME-MINI, and Sora 2 API preview) to expand what developers and creators can ship.

What does “apps inside ChatGPT” change for developers, and how does the Apps SDK fit in?

The keynote positions ChatGPT as an app platform rather than just a chat interface. The Apps SDK (preview) lets developers build interactive, adaptive apps that render inline within conversations and can include rich web content (including video layouts). The SDK is built on MCP, so developers control backend logic and frontend UI. OpenAI also emphasizes distribution and discovery: apps can be found by asking for them by name, suggested during conversations when relevant, and developers can scale to very large user bases. Monetization is also planned to connect to user sessions inside ChatGPT, including an agentic commerce protocol for instant checkout.

How does context flow between a user’s interaction and ChatGPT when using an app like Coursera or Zillow?

In the Coursera demo, the user plays a video inside the chat, then asks for deeper explanations. The Apps SDK provides an API that exposes the user’s current context back to ChatGPT—so the model can answer based on exactly what the user is interacting with. In the Zillow demo, the app embeds an interactive map and updates data in-place; ChatGPT then uses the app’s exposed context to filter results (e.g., three-bedroom homes with a yard) and answer follow-ups like proximity to a dog park. The key mechanism is “talking to apps”: the app supplies context, and ChatGPT can combine it with other tools (like search) to produce better answers.

Why does Agentkit focus on production readiness, and what components does it include?

Agentkit targets the gap between agent prototypes and real production deployments. The keynote lists common friction points—tool orchestration, eval loops, UI work, and uncertainty about frameworks. Agentkit is presented as a set of building blocks: Agent Builder (visual workflow design on top of the Responses API), Chatkit (an embeddable chat UI for custom branding and workflows), and Evals for Agents (trace grading, datasets, automated prompt optimization, and running evals on external models within OpenAI’s eval platform). It also relies on a connector registry for secure access to internal and third-party systems through an admin control panel.

What did the live “Ask Froge” demo demonstrate about building agents quickly?

The demo showed an end-to-end workflow built under a time constraint: visual wiring of nodes in Agent Builder, routing logic to classify incoming messages, creation of specialized session and DevDay agents, attachment of a sessions widget as an output format, and adding prebuilt guardrails (including PII protection). It then previewed behavior, published the agent to production using a workflow ID, and embedded it into the DevDay site using Chatkit React components. The agent could answer which session to attend by pulling from the attached session file and orchestrating the right internal agents, all while maintaining the guardrails.

How is CodeX positioned differently now that it’s generally available, and what’s new for teams?

CodeX is described as moving from research preview to GA, running on GPT5-Codex—a version of GPT5 trained for agentic coding tasks like refactoring and code review. OpenAI cites rapid usage growth (10x daily messages since early August) and internal adoption (most new code at OpenAI written by CodeX users; nearly all PRs reviewed by CodeX). New team features include Slack integration for code and Q&A inside team conversations, a CodeX SDK to extend and automate team workflows, and enterprise admin tools for environment controls, monitoring, and analytics dashboards.

What do the model updates (GPT5 Pro, GPT-REALTIME-MINI, and Sora 2 API preview) aim to enable?

GPT5 Pro is introduced for high-accuracy, deep-reasoning tasks in domains like finance, legal, and healthcare. GPT-REALTIME-MINI is a smaller, cheaper voice model in the API, described as keeping the same voice quality and expressiveness while being 70% cheaper—positioned as a step toward voice becoming a primary interaction mode. Sora 2 preview in the API is framed as more controllable video generation: it supports detailed instructions, holds state, can expand an image into a wide cinematic shot, and pairs sound with visuals (rich sounds, ambient audio, synchronized effects). It also supports remixing and controls like video length, aspect ratio, and resolution.

Review Questions

  1. How do Apps SDK and MCP together enable both UI control and backend logic control for apps embedded in ChatGPT?
  2. What specific Agentkit components address workflow design, user-facing chat embedding, and agent evaluation/optimization?
  3. In what ways does GPT5-Codex (CodeX) integrate into developer workflows and team processes (e.g., IDEs, Slack, admin tools)?

Key Points

  1. 1

    OpenAI is turning ChatGPT into an app platform by launching the Apps SDK (preview), built on MCP for full control of backend logic and UI.

  2. 2

    Apps inside ChatGPT are designed for inline, interactive experiences and for conversational discovery—users can find apps by name or via contextual suggestions.

  3. 3

    Agentkit is meant to reduce agent-building friction by combining visual workflow design, embeddable chat UI, agent-specific eval tooling, and safer data connections.

  4. 4

    Agentkit’s live demo showed an end-to-end path from visual workflow wiring and guardrails to publishing a production agent and embedding it into a site without code changes.

  5. 5

    CodeX is now generally available on GPT5-Codex, with new team features like Slack integration, a CodeX SDK, and enterprise admin/analytics tools.

  6. 6

    Model updates expand capabilities for hard reasoning (GPT5 Pro), lower-cost expressive voice (GPT-REALTIME-MINI), and controllable video creation with sound (Sora 2 API preview).

Highlights

Apps SDK (preview) lets developers build interactive apps directly inside ChatGPT conversations, with MCP-based control over both backend logic and frontend UI.
Agentkit bundles the missing pieces for production agents—visual workflow building, embeddable chat, traceable evals, and secure connectors—so teams can ship faster.
CodeX’s GA launch on GPT5-Codex is paired with Slack integration and enterprise admin tools, pushing AI coding from individual use toward team workflows.
Sora 2 preview in the API emphasizes controllability and sound pairing, including synchronized effects and remixing options.
The keynote’s demos repeatedly show “context passing” between apps/agents and ChatGPT, enabling more accurate follow-ups without restarting workflows.

Topics

  • Apps Inside ChatGPT
  • Apps SDK
  • Agentkit
  • CodeX GA
  • Sora 2 API Preview

Mentioned