Get AI summaries of any video or article — Sign up free
Can ChatGPT Create a Netflix Series Idea? - The Feedback Loop thumbnail

Can ChatGPT Create a Netflix Series Idea? - The Feedback Loop

All About AI·
5 min read

Based on All About AI's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

The workflow generates an initial series concept by combining genre constraints with inspiration from Severance, devs, Westworld, and Black Mirror.

Briefing

A multi-expert “feedback loop” prompt can turn a rough Netflix-style series pitch into a higher-scoring, more polished pre-cap by repeatedly evaluating, critiquing, and rewriting until it hits a target rating. The core idea is to treat a series intro like a product that can be stress-tested: generate an initial concept, score it across specific writing criteria, solicit critique of the critique, then rewrite with concrete improvement instructions—repeating the cycle until multiple expert perspectives align on quality.

The process starts by defining what kind of show the pitch should resemble. The workflow narrows the brief to a drama/mystery tone and identifies the types of “experts” needed to judge a TV series hook: screenwriters, story consultants, genre experts, and behavioral experts. To ground the style, it also pulls inspiration from well-known series—Severance, devs, Westworld, and Black Mirror—by converting short descriptions into a PDF and having ChatGPT’s advanced data analysis read the document. From that input, the system generates several original series ideas and selects “ID number six: Memory Lane.”

Next comes the pre-cap: a ~300-word narrated setup designed to function like a trailer’s opening. The draft then gets evaluated by an expert-style rubric that assigns a 0–100 score across categories such as intrigue, character depth, setting, and plot potential. In the example, the first evaluation lands at 88.7, which signals strong fundamentals but leaves room for refinement.

The key twist is that the critique itself is put through another round. After the initial evaluation, a second pass asks the model to critique its own critique, then re-evaluates the same pre-cap with “more nuance.” That re-scoring drops the number, which becomes the justification for targeted rewrites. The system then requests explicit instructions for how to raise the rating toward 95—adding refinements like enhancing the setting, deepening secondary characters, and expanding the clinic’s motivation.

A rewrite follows, and the updated pre-cap is re-scored. The workflow then escalates by running the revised pitch through additional expert lenses: a story/genre consultant and a Hollywood pre-cap writing specialist. When that specialist scores the pre-cap at 84, the loop repeats again—keeping the writing language while adding new, rating-oriented instructions—until the pitch reaches a satisfactory level.

Finally, the method extends beyond text. It uses Midjourney image prompting to generate scene-based visuals for a trailer-like sequence, using the pre-cap as the source material and producing prompts for multiple scenes (e.g., 16 scenes starting from scene one). The overall takeaway is a practical, iterative pipeline: generate → evaluate → critique the critique → rewrite with specific improvement directives → re-evaluate across multiple expert roles, then optionally visualize the result. It’s also framed against current industry tension around AI and writers, using the workflow to test how far AI can go in producing pitch-ready creative assets.

Cornell Notes

The workflow builds a Netflix-style series pitch by using a repeated evaluation-and-rewrite loop. It begins with genre targeting (drama/mystery), draws inspiration from shows like Severance, devs, Westworld, and Black Mirror via a PDF, and generates multiple original ideas before selecting “Memory Lane.” A ~300-word pre-cap is scored from 0–100 across criteria such as intrigue, character depth, setting, and plot potential. After an initial score (88.7), the critique is stress-tested by asking for a more nuanced re-evaluation, which lowers the score and triggers concrete improvement instructions aimed at reaching 95. The revised pre-cap is then re-scored by multiple expert perspectives, and the final concept can be visualized with Midjourney scene prompts.

How does the process turn a vague series idea into a structured pitch?

It starts by locking the genre and tone (drama/mystery) and then selects “expert” roles to judge the work: screenwriters, story consultants/genre experts, and even behavioral experts. It also feeds inspiration into the system by converting short descriptions of shows like Severance, devs, Westworld, and Black Mirror into a PDF, letting advanced data analysis read that content. From there, it generates multiple non-generic series ideas and chooses “ID number six: Memory Lane,” then writes a narrated TV series pre-cap of about 300 words.

What does the scoring system measure, and why does it matter?

The pre-cap is evaluated on a 0–100 scale using categories such as intrigue, character depth, setting, and plot potential. That rubric matters because it converts subjective writing feedback into actionable targets. For example, the first Memory Lane pre-cap receives 88.7, signaling it already has strong hook elements but likely needs deeper character work, clearer setting payoff, and stronger plot momentum to reach a higher threshold.

Why does the workflow ask the model to critique its own critique?

After the initial evaluation, the system requests a critique of the evaluation and then re-evaluates the same pre-cap with “more nuance.” That second pass can lower the score (in the example, the score drops after the nuanced re-check), revealing blind spots or overly generous scoring. The lowered number becomes a concrete reason to rewrite with more specific improvements rather than stopping after the first favorable assessment.

What kinds of rewrite instructions are used to push the score toward 95?

The improvement directives are specific and structural: enhance the setting, deepen secondary characters, and expand the clinic’s motivation. Those changes target the rubric categories directly—so the rewrite isn’t just stylistic polishing; it’s meant to strengthen the mystery’s engine, the cast’s emotional stakes, and the plausibility of the central premise.

How does using multiple expert perspectives change the outcome?

After one expert-style evaluation and rewrite, the pre-cap is tested again with other roles. A story/genre consultant may approve the revised version, but a Hollywood pre-cap writing specialist can still score it lower (the example shows an 84). That forces another rewrite cycle with instructions tailored to what that specialist values—keeping the language style while adjusting the hook mechanics to better capture attention.

How are visuals generated from the pre-cap, and what’s the practical link?

The pre-cap becomes the source text for Midjourney image prompts. The workflow asks for scene-by-scene image prompts in the same format, then generates a set of images for a trailer-like sequence (e.g., 16 scenes starting from scene one). The practical link is that each image prompt is grounded in the narrative beats of the pre-cap, producing visuals that match the story’s mystery and tone.

Review Questions

  1. What specific rubric categories are used to score the pre-cap, and how do they map to the rewrite instructions?
  2. Why might a pre-cap score drop after a “more nuanced” re-evaluation, and what does the workflow do in response?
  3. How does the workflow’s use of multiple expert roles (story/genre consultant vs. Hollywood pre-cap specialist) affect the number of rewrite cycles?

Key Points

  1. 1

    The workflow generates an initial series concept by combining genre constraints with inspiration from Severance, devs, Westworld, and Black Mirror.

  2. 2

    A ~300-word narrated pre-cap is scored on a 0–100 rubric using criteria like intrigue, character depth, setting, and plot potential.

  3. 3

    Critique is stress-tested by asking for critique-of-critique and then re-evaluating with more nuance, which can lower scores and expose gaps.

  4. 4

    Rewrites target specific improvement directives aimed at raising the rating toward 95, such as strengthening setting detail, expanding secondary characters, and clarifying clinic motivation.

  5. 5

    Multiple expert lenses are used to re-score the revised pre-cap, forcing additional iterations when one perspective still finds weaknesses.

  6. 6

    The final concept can be translated into a visual trailer by generating Midjourney scene prompts directly from the pre-cap’s narrative beats.

Highlights

Memory Lane is selected as the winning idea after generating multiple non-generic TV series concepts from genre and inspiration inputs.
An 88.7 score triggers a deeper loop: the critique is re-checked with more nuance, and the score can drop—prompting targeted rewrites.
The rewrite instructions are concrete and rubric-aligned, including enhancing setting, deepening secondary characters, and expanding the clinic’s motivation.
The workflow extends into visuals by turning the pre-cap into Midjourney prompts for a multi-scene trailer sequence (e.g., 16 scenes).

Topics

  • AI Prompt Engineering
  • TV Series Pitching
  • Feedback Loop
  • Story Evaluation
  • Midjourney Visuals

Mentioned