Open AI Releases DALL-E 3 Image Editing! (PLUS Free Alternative)
Based on MattVidPro's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
OpenAI’s DALL·E 3 now includes inpainting-style image editing inside ChatGPT, with region selection and natural-language instructions.
Briefing
OpenAI has rolled out image editing for DALL·E 3 inside ChatGPT—letting users select regions of a generated image and use natural-language prompts to change them. The feature is available across ChatGPT on the web and on iOS and Android, and it appears to work like an inpainting tool: highlight an area, then ask for edits such as “add bows,” “remove butterfly,” or “turn him into a wizard.” In practice, it can produce clean, localized changes (like swapping clothing details or fixing hands) while preserving much of the surrounding composition.
The rollout matters because it shifts DALL·E 3 from “generate and hope” toward an iterative workflow inside a widely used interface. Users can branch off from earlier results—saving multiple versions as they refine details—without needing separate editing software. A quick demo shows the edit button appearing when a user clicks into a DALL·E 3 image, followed by a selection/highlight workflow similar to painting tools. Prompts can also drive multi-step refinement: for example, changing a top hat reminiscent of Abraham Lincoln, removing an unwanted butterfly, or adjusting a cloak and hat after an initial “wizard” transformation.
Still, the editing has clear limits. Text handling is inconsistent: attempts to correct or insert specific text often fail, with the system sometimes removing text entirely rather than updating it. When a complex layout included a “MattVidPro” label, edits intended to fix the typography resulted in missing or incorrect text, leading to the conclusion that reliable text generation/editing may require other tools. The transcript repeatedly contrasts this weakness with idiogram AI, which is recommended specifically for text.
The tool also struggles with strict consistency. Edits can become “mish-mashed” when prompts require large, style-sensitive changes across broad regions—such as placing a subject on the moon or adding multiple elements that must match the original art style. In one test, the system produced a dramatic scene change with a monster behind a frog; in another, it attempted to add a wizard look but required follow-up corrections to improve the hat and cloak.
Another practical takeaway: the best results come from getting close in the initial generation prompt, then using edits for targeted fixes. The transcript suggests the feature is less about endless fine-tuning and more about correcting specific errors after the first pass.
Beyond OpenAI’s offering, the transcript points to an open-source alternative built with Gradio and distributed via Pinocchio, marketed as BrushNet. The approach runs locally and supports similar “segment and edit” workflows, including swapping the segmented subject into new categories like Lego pieces, chocolate, garlic, or pickles.
Finally, accessibility changes are noted: ChatGPT can be used without an account in some cases, though chat history behavior depends on login. Overall, DALL·E 3 editing in ChatGPT looks like a meaningful step toward democratized image refinement—strong for localized edits and detail fixes, weaker for precise text control and style-perfect consistency.
Cornell Notes
OpenAI added DALL·E 3 image editing inside ChatGPT, available on web, iOS, and Android. Users can click into a generated image, select an area, and use natural-language prompts to perform inpainting-style changes such as adding bows, removing objects, or transforming a subject into a wizard. The workflow supports branching and saving multiple versions, making refinement easier than starting over. Localized edits like hands and small details often come out well, but text editing is unreliable—attempts to fix or insert specific wording frequently remove or fail to render the text correctly. For dependable text, the transcript recommends using idiogram AI instead.
How does DALL·E 3 editing inside ChatGPT work in practice?
What kinds of edits tend to succeed, based on the demo?
Why does text editing appear to be a weak spot?
What strategy does the transcript recommend for best overall results?
What open-source alternative is mentioned, and what’s its basic approach?
What accessibility change is noted for ChatGPT usage?
Review Questions
- When using DALL·E 3 editing, what is the selection step and why does it matter for the quality of the result?
- Give two examples of edits that worked well and explain what they have in common.
- What evidence in the transcript suggests that text editing is unreliable, and what alternative is recommended instead?
Key Points
- 1
OpenAI’s DALL·E 3 now includes inpainting-style image editing inside ChatGPT, with region selection and natural-language instructions.
- 2
The edit workflow is available on ChatGPT web, iOS, and Android, and it appears integrated into the DALL·E 3 image experience via an edit button.
- 3
Localized edits—like removing objects, adding small elements, and fixing hands—tend to produce better, more believable results than broad redesigns.
- 4
Text editing is inconsistent: attempts to correct or insert specific wording often remove text or fail to render it correctly.
- 5
For best outcomes, users should craft a strong initial prompt and then use editing to fix specific mistakes rather than expecting perfect iterative refinement.
- 6
An open-source local alternative, BrushNet (Gradio + Pinocchio installer), supports similar segmentation-based editing workflows.
- 7
ChatGPT access and chat history behavior depend on login status, with some use possible without an account but history saving tied to authentication.