Get AI summaries of any video or article — Sign up free
Open AI Releases DALL-E 3 Image Editing! (PLUS Free Alternative) thumbnail

Open AI Releases DALL-E 3 Image Editing! (PLUS Free Alternative)

MattVidPro·
5 min read

Based on MattVidPro's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

OpenAI’s DALL·E 3 now includes inpainting-style image editing inside ChatGPT, with region selection and natural-language instructions.

Briefing

OpenAI has rolled out image editing for DALL·E 3 inside ChatGPT—letting users select regions of a generated image and use natural-language prompts to change them. The feature is available across ChatGPT on the web and on iOS and Android, and it appears to work like an inpainting tool: highlight an area, then ask for edits such as “add bows,” “remove butterfly,” or “turn him into a wizard.” In practice, it can produce clean, localized changes (like swapping clothing details or fixing hands) while preserving much of the surrounding composition.

The rollout matters because it shifts DALL·E 3 from “generate and hope” toward an iterative workflow inside a widely used interface. Users can branch off from earlier results—saving multiple versions as they refine details—without needing separate editing software. A quick demo shows the edit button appearing when a user clicks into a DALL·E 3 image, followed by a selection/highlight workflow similar to painting tools. Prompts can also drive multi-step refinement: for example, changing a top hat reminiscent of Abraham Lincoln, removing an unwanted butterfly, or adjusting a cloak and hat after an initial “wizard” transformation.

Still, the editing has clear limits. Text handling is inconsistent: attempts to correct or insert specific text often fail, with the system sometimes removing text entirely rather than updating it. When a complex layout included a “MattVidPro” label, edits intended to fix the typography resulted in missing or incorrect text, leading to the conclusion that reliable text generation/editing may require other tools. The transcript repeatedly contrasts this weakness with idiogram AI, which is recommended specifically for text.

The tool also struggles with strict consistency. Edits can become “mish-mashed” when prompts require large, style-sensitive changes across broad regions—such as placing a subject on the moon or adding multiple elements that must match the original art style. In one test, the system produced a dramatic scene change with a monster behind a frog; in another, it attempted to add a wizard look but required follow-up corrections to improve the hat and cloak.

Another practical takeaway: the best results come from getting close in the initial generation prompt, then using edits for targeted fixes. The transcript suggests the feature is less about endless fine-tuning and more about correcting specific errors after the first pass.

Beyond OpenAI’s offering, the transcript points to an open-source alternative built with Gradio and distributed via Pinocchio, marketed as BrushNet. The approach runs locally and supports similar “segment and edit” workflows, including swapping the segmented subject into new categories like Lego pieces, chocolate, garlic, or pickles.

Finally, accessibility changes are noted: ChatGPT can be used without an account in some cases, though chat history behavior depends on login. Overall, DALL·E 3 editing in ChatGPT looks like a meaningful step toward democratized image refinement—strong for localized edits and detail fixes, weaker for precise text control and style-perfect consistency.

Cornell Notes

OpenAI added DALL·E 3 image editing inside ChatGPT, available on web, iOS, and Android. Users can click into a generated image, select an area, and use natural-language prompts to perform inpainting-style changes such as adding bows, removing objects, or transforming a subject into a wizard. The workflow supports branching and saving multiple versions, making refinement easier than starting over. Localized edits like hands and small details often come out well, but text editing is unreliable—attempts to fix or insert specific wording frequently remove or fail to render the text correctly. For dependable text, the transcript recommends using idiogram AI instead.

How does DALL·E 3 editing inside ChatGPT work in practice?

A user generates images with DALL·E 3, then clicks into a specific image to find a new edit button. After that, the user highlights a region (similar to selecting/painting in an editor) and sends a natural-language instruction. Examples include “add bows,” “remove butterfly,” “fix the hands,” and “turn him into a wizard with a cloak, a hat, and glowing green eyes.” The system then returns a new image based on the highlighted area and the prompt.

What kinds of edits tend to succeed, based on the demo?

Edits that are localized and visually concrete tend to work better. The transcript highlights removing a butterfly, adding bows, fixing hands, and refining specific clothing elements like a cloak and hat. In one example, highlighting the hand area and asking “fix the hands” produced a noticeably improved result, suggesting targeted inpainting is a strong use case.

Why does text editing appear to be a weak spot?

When the demo tries to correct or insert exact text (e.g., making “Matt vid Pro” bold white text), the results often fail. The system sometimes removes the text entirely instead of updating it, and repeated attempts still don’t reliably produce the requested wording. The transcript concludes that precise text control is not dependable in this editing workflow.

What strategy does the transcript recommend for best overall results?

Get as close as possible in the initial generation prompt, then use editing for corrections. The transcript frames editing as a way to fix details the model gets wrong rather than a tool for prolonged, iterative “dialing in” of a fully specified design. Large, style-sensitive changes can become inconsistent, so starting with a strong baseline prompt improves the outcome.

What open-source alternative is mentioned, and what’s its basic approach?

An open-source option called BrushNet is described as a Gradio app installed via Pinocchio. It runs locally and supports segmentation-based editing—changing the prompt for a selected region (e.g., turning a segmented subject into Lego pieces, chocolate, garlic, or pickles). The transcript emphasizes it’s free as long as the machine can handle the image generation workload.

What accessibility change is noted for ChatGPT usage?

The transcript claims ChatGPT can be used without an account, implying quick access for demos or sharing. However, it also notes that chat history behavior depends on login: without logging in, chat history may not be saved, while logging in enables history.

Review Questions

  1. When using DALL·E 3 editing, what is the selection step and why does it matter for the quality of the result?
  2. Give two examples of edits that worked well and explain what they have in common.
  3. What evidence in the transcript suggests that text editing is unreliable, and what alternative is recommended instead?

Key Points

  1. 1

    OpenAI’s DALL·E 3 now includes inpainting-style image editing inside ChatGPT, with region selection and natural-language instructions.

  2. 2

    The edit workflow is available on ChatGPT web, iOS, and Android, and it appears integrated into the DALL·E 3 image experience via an edit button.

  3. 3

    Localized edits—like removing objects, adding small elements, and fixing hands—tend to produce better, more believable results than broad redesigns.

  4. 4

    Text editing is inconsistent: attempts to correct or insert specific wording often remove text or fail to render it correctly.

  5. 5

    For best outcomes, users should craft a strong initial prompt and then use editing to fix specific mistakes rather than expecting perfect iterative refinement.

  6. 6

    An open-source local alternative, BrushNet (Gradio + Pinocchio installer), supports similar segmentation-based editing workflows.

  7. 7

    ChatGPT access and chat history behavior depend on login status, with some use possible without an account but history saving tied to authentication.

Highlights

DALL·E 3 editing in ChatGPT lets users highlight part of an image and rewrite that region using plain-language prompts—turning “generate” into “refine.”
The strongest results come from targeted fixes like removing a butterfly or correcting hands, while large, style-sensitive edits can look mismatched.
Exact text edits are a recurring failure mode: the system often deletes or can’t reliably render requested wording, pushing users toward idiogram AI for text.
BrushNet is presented as a free, local, Gradio-based alternative installed via Pinocchio, using segmentation to swap what a selected region becomes.

Topics

  • DALL·E 3 Editing
  • ChatGPT Inpainting
  • Natural Language Image Edits
  • Text Rendering Limits
  • Open-Source Image Editing

Mentioned