MAJOR Upgrades to AI Art Tools! & Open AI's new Image Gen!
Based on MattVidPro's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Idiogram Canvas adds a fully interactive editing workflow: drag variations, zoom, delete unwanted results, and continue refining on a shared canvas.
Briefing
AI image editing is shifting from “generate and pick” toward real, Photoshop-like workflows—led by Idiogram Canvas. The standout upgrade is a fully interactive canvas that lets users drag, zoom, delete unwanted variations, and then extend images seamlessly using an outpainting-style workflow. In a hands-on example, a Halloween-costumed cat is generated with multiple variations, then the canvas is expanded to a wider 3:2 aspect ratio. With “Magic Prompt” turned off, the extend tool produces a background that matches the prompt intent (a suburban Halloween street) without inventing extra details. A separate “Magic Fill” tool adds localized edits: using lasso masking, a speech bubble is generated with text like “dang these people are giving away candy not cat treats,” and the bubble appears in the correct region.
The practical value is that these tools behave like an editing suite rather than a one-shot generator. Users can keep the original image while extending, choose among four results, and continue refining with additional fills. The workflow also supports image uploading and remixing—so a user can extend a personal photo into a new scene (e.g., lemon-themed headphones, lemon-themed styling) or expand a small Halloween cat into a full pumpkin-filled background. The main limitations noted are prompt sensitivity (vague prompts like “background” can lead to unintended changes, including adding other people) and occasional lag/lower frame rate. Still, automatic cloud saving and the overall UI responsiveness make it feel close to a daily-driver editor for many tasks.
Midjourney is also moving toward deeper editing. New tests include an image editor for uploaded images that performs inpainting (erase and regenerate a selected area) and “image retexturing,” which changes materials, surfacing, and lighting while preserving the underlying composition. Examples show a dog with a party hat added after erasing part of the image, a body erased and replaced with a colorful sweater that still fits the same animal, and interior photo retexturing that transforms a living room into a far-future version while keeping key elements like couches and books in place. The canvas extension exists too, but it’s described as more constrained than Idiogram’s free-form approach—potentially easier for new users.
Beyond editing tools, the model landscape keeps shifting. Stability AI released Stable Diffusion 3.5 Large, positioning it as a more usable update to Stable Diffusion 3 after earlier reception. It’s framed as more open and easier to fine-tune, with community discussion pointing to built-in support for training and control mechanisms like LoRA-style workflows, IP-Adapter use, and fine-tuning. OpenAI introduced “consistency models” (SCMS), designed to generate samples in only two sampling steps—faster than diffusion—while reporting quality comparable to leading diffusion approaches. The roundup closes with the idea that OpenAI is still actively pushing image generation research, even after major releases like DALL·E 3.
Taken together, the message is clear: the competitive edge is moving from raw image quality alone toward controllable, iterative editing—where extending, masking, and retexturing can be chained into a coherent creative workflow.
Cornell Notes
Idiogram Canvas is presented as a major leap in AI image editing, combining drag-and-drop canvas control with extend (outpainting) and magic fill (masked inpainting). The extend tool can expand an image to new aspect ratios and generate seamless backgrounds, especially when “Magic Prompt” is disabled to prevent unwanted inference. Magic Fill uses brush or lasso masks to add localized elements like speech bubbles, and the workflow supports uploaded images for iterative edits. Midjourney’s parallel updates—uploaded-image inpainting and “image retexturing”—aim to preserve composition while changing materials, lighting, and even style. The broader model news includes Stable Diffusion 3.5 Large and OpenAI’s SCMS consistency models, both pushing speed, fine-tuning flexibility, and practical usability.
What makes Idiogram Canvas feel more like an editor than a generator?
Why does turning “Magic Prompt” off matter when extending an image?
How does Magic Fill differ from the extend tool in practical use?
What new Midjourney capabilities are highlighted, and what do they preserve?
What’s the significance of Stable Diffusion 3.5 Large and SCMS in the broader model race?
Review Questions
- When would a user prefer “Magic Prompt” off in Idiogram Canvas, and what failure mode does it help avoid?
- Compare Idiogram’s extend+Magic Fill workflow with Midjourney’s inpainting+retexturing: which tasks map to each toolset?
- What tradeoffs are implied between faster sampling (SCMS) and diffusion quality, based on the examples described?
Key Points
- 1
Idiogram Canvas adds a fully interactive editing workflow: drag variations, zoom, delete unwanted results, and continue refining on a shared canvas.
- 2
The extend tool can outpaint to new aspect ratios, and disabling “Magic Prompt” helps prevent the system from inventing extra details beyond the prompt.
- 3
Magic Fill enables localized edits through masking (brush/lasso/rectangle), making it suitable for adding specific elements like speech bubbles.
- 4
Midjourney’s new uploaded-image editor focuses on inpainting and “image retexturing,” preserving composition while changing materials, lighting, and style.
- 5
Stable Diffusion 3.5 Large is positioned as a more usable update to Stable Diffusion 3 with improved fine-tuning flexibility and openness.
- 6
OpenAI’s SCMS consistency models aim to cut generation time by using only two sampling steps while targeting diffusion-level quality.
- 7
Across tools, prompt specificity remains crucial: vague prompts can cause unintended edits even when the UI feels seamless.