Get AI summaries of any video or article — Sign up free
Don't Bother Learning Photoshop, AI Does it For You thumbnail

Don't Bother Learning Photoshop, AI Does it For You

MattVidPro·
5 min read

Based on MattVidPro's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Playground AI’s new inpainting is positioned as a major improvement for targeted AI image edits, especially compared with earlier Stable Diffusion inpainting quality.

Briefing

AI image generation has moved past messy, uncontrolled outputs—yet fine-grained edits still frustrate creators. Playground AI’s newly released inpainting aims to close that gap, delivering results that feel closer to Dolly 2–style control while staying accessible: it’s free to try and offers up to 1,000 generations per day (or 60,000+ per month on a $15 plan).

The core shift is inpainting quality. Earlier workflows in this space relied on “painting” (masking and regenerating parts of an image). Dolly 2’s inpainting was widely praised because it could erase a small region and regenerate it coherently. Stable Diffusion–based inpainting existed too, but it often lagged in fidelity. Playground AI’s update changes that comparison by producing cleaner, more usable edits—especially when the user combines text instructions with targeted masks.

Playground AI’s interface supports multiple model choices, but the workflow centers on stable diffusion 1.5 for its custom filters. Users can set aspect ratios and resolution, adjust prompt guidance, and generate multiple images per prompt. Negative prompts help steer outputs away from unwanted artifacts. The platform also includes image-to-image options (drawing or uploading an image as a starting point) plus variations, downloads, face restoration, and upscaling.

The inpainting workflow is where the “game changer” claim lands. For broad changes that don’t fit simple masking—like turning a sunny scene into night with fireworks—Playground AI uses an “edit instruction” approach. A key control is edit instruction strength: too low yields nearly identical results, too high can distort the subject (even adding unintended elements). A mid-range “sweet spot” value produced the best balance in the example, preserving the character while adding fireworks.

For precise fixes, the editor adds an “add mask” step that switches into inpainting. Users can erase or “heal” areas by painting over regions, then regenerate only the selected parts. The transcript walks through iterative tuning: a lemon character’s ears are removed by prompting for a “perfectly circular” form; later, lemon texture is applied selectively by masking only part of the face and adjusting instruction strength to avoid turning the entire character into an over-saturated lemon. Other edits show how loose masking can still work—highlighting around an object can be enough for the model to understand what to replace.

Real-world photo editing is demonstrated with a dog image: a top hat is added by masking above the head, followed by a tie edit with instruction strength adjustments to avoid duplicating accessories. A more cinematic “movie scene” transformation also succeeds. After editing, the platform allows downloads (with resolution constraints tied to stable diffusion sizes) and then upscaling and optional face restoration, though some quality loss is acknowledged as an unavoidable tradeoff.

Overall, the takeaway is practical: creators may not need to “learn Photoshop” for many common fixes anymore. With strong inpainting, prompt-guided editing, and iterative masking, small details—ears, textures, accessories, even facial elements—become editable in a way that’s fast enough to feel like creative control rather than a reroll lottery. The transcript closes by framing this as a back-and-forth with Dolly 2: Dolly still has strengths like outpainting, while Playground’s inpainting quality is positioned as a major advantage for detail work.

Cornell Notes

Playground AI’s new inpainting is presented as a major step toward controllable image editing, reducing the need for traditional Photoshop-style workflows. The platform combines text-based “edit instructions” for broad scene changes with masked inpainting for targeted fixes. A crucial parameter—edit instruction strength—determines whether changes appear, distort the subject, or do nothing; mid-range values often work best. The transcript demonstrates iterative edits on a stylized lemon character and a real dog photo, including removing ears, adding lemon texture, and placing a top hat and tie. After edits, users can download images, then upscale and optionally run face restoration, accepting some quality loss from the editing pipeline.

What makes Playground AI’s inpainting feel different from earlier AI editing approaches?

The transcript contrasts earlier Stable Diffusion inpainting (described as not nearly as good as Dolly tools) with Playground AI’s newly released inpainting, which produces more coherent, usable results. The key is that users can either (1) apply broad changes via text edit instructions or (2) target specific regions using a mask-based inpainting workflow that regenerates only the selected area. That combination—precise masking plus instruction tuning—reduces the “reroll and hope” problem for fine details.

How does “edit instruction strength” affect outcomes in Playground AI’s editor?

Edit instruction strength acts like a dial for how strongly the model follows the instruction. In the fireworks example, a mid-range value (around 6.75) preserved the character while adding fireworks in the background. When the value was cranked too high (e.g., around 16), the model introduced unwanted changes (a whole bird appeared). When the value was too low, the output stayed essentially the same with no meaningful differences.

When should a user use edit instructions versus inpainting masks?

Edit instructions fit changes that affect the overall scene or fundamentals—like switching to night and adding fireworks—where masking would be awkward. Inpainting masks fit localized corrections—like removing ears, replacing a face texture, or adding a hat—because the user can erase/heal only the region that needs regeneration.

What iterative technique is used to get the “right” lemon texture on the character?

The workflow is stepwise: first, remove unwanted structure (ears) by masking and prompting for a “perfectly circular” form. Then, to add texture without overdoing it, the user masks only the portion of the face that should receive lemon texture. Instruction strength is adjusted downward when the model goes too far (turning the character into an overly saturated lemon). The transcript emphasizes tweaking strength and masking scope until the texture looks natural.

How does the editor handle accessory placement on a real photo (the dog example)?

The user masks the area above the dog’s head to add a top hat, then edits again to add a tie. Instruction strength is adjusted to avoid overly fancy or duplicated results. The tie prompt initially produced two ties (or the wrong style), and the follow-up prompt (“bow tie”) plus strength reduction yielded a cleaner single bow tie.

What post-processing options are available after editing, and what tradeoff is mentioned?

After editing, users can download the image, then use actions like upscaling. The transcript notes that downloads reduce resolution due to stable diffusion size constraints (examples include 512×512, 384×640, and 1024×1024). Upscaling helps but doesn’t fully restore original quality, and face restoration may only slightly improve a fuzzy result.

Review Questions

  1. How would you decide whether to use an edit instruction or a mask-based inpainting step for a specific change (e.g., changing the sky vs. fixing an ear)?
  2. Why does edit instruction strength need to be tuned rather than set once, and what symptoms indicate values that are too high or too low?
  3. What workflow steps would you use to add a new accessory to a photo while minimizing unintended changes elsewhere in the image?

Key Points

  1. 1

    Playground AI’s new inpainting is positioned as a major improvement for targeted AI image edits, especially compared with earlier Stable Diffusion inpainting quality.

  2. 2

    The editor supports both broad text-driven changes (edit instructions) and localized regeneration (mask-based inpainting).

  3. 3

    Edit instruction strength is a critical control: mid-range values often preserve the subject while applying the requested change; extremes can cause distortion or no change.

  4. 4

    Custom stable diffusion filters and negative prompts help steer generation, but inpainting is the main tool for fixing specific details.

  5. 5

    Mask-based inpainting can be surprisingly forgiving—highlighting the right region can be enough for the model to understand what to replace.

  6. 6

    Post-edit downloads follow stable diffusion resolution limits, and upscaling/face restoration can help but may not fully recover original quality.

Highlights

Playground AI’s inpainting is presented as “darn good,” with results that feel closer to Dolly 2–style control for detail edits.
Edit instruction strength functions like a precision dial: too high adds unwanted elements, too low changes nothing, and a mid-range value hits the sweet spot.
Mask-based inpainting enables localized fixes—ears, textures, accessories—without forcing the entire image to be remade.
The workflow supports iterative refinement: generate, edit, mask, regenerate, and then upscale or restore faces if needed.

Topics