Get AI summaries of any video or article — Sign up free
How To Do A Literature Review (STRESS-FREE!) thumbnail

How To Do A Literature Review (STRESS-FREE!)

Andy Stapleton·
6 min read

Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Use Zotero as the reference manager so AI tools can connect to a single, organized library of papers.

Briefing

A stress-free literature review workflow hinges on using AI to triage the flood of papers—then doing the real academic work yourself. The core idea is to avoid reading everything ever published in a field by combining reference management, AI-powered filtering, and concept mapping, while still performing a meaningful subset of full-paper reading and a rigorous human editing pass. The payoff: a literature review that’s grounded in the right sources, structured coherently, and polished to academic standards without skipping the skills that make researchers effective.

The process starts with building a literature library. A reference manager is treated as the backbone—Zotero is recommended because it integrates with many AI tools. From there, researchers gather papers through a mix of AI-assisted discovery platforms and traditional search. Illicit, Consensus, Litmaps, Connected Papers, SciSpace, and Research Rabbit are listed as ways to collect relevant studies, alongside Google Scholar, which is explicitly framed as a reliable “old-fashioned” option for keyword searches.

Once a sizable set of references is stored, AI shifts from discovery to filtering. The goal is sanity: narrowing the reading list to the most relevant work so the literature review can be anchored in the strongest evidence. Tools like SciSpace are used with relevance ranking (for example, selecting top papers based on a high relevance score), and with features that add structured “insights” columns derived from abstracts. Consensus is used to identify “key papers” and surface claims and evidence associated with top contributors. NotebookLM is positioned as a synthesis layer—uploading references and prompting it to highlight key findings and the most important papers.

After filtering, the workflow insists on selective deep reading. Instead of attempting to read every paper, it recommends reading roughly 20–25% of the bubbled-up set to build genuine understanding of the field. From there, the literature review becomes more than a summary list: NotebookLM mapping features and Consensus matrices help organize concepts, themes, and gaps. NotebookLM can generate mind maps that resemble a draft literature review outline, including potential headings and subheadings. Consensus provides a matrix view of how many studies exist across application areas, making underexplored regions—sometimes labeled as “gap”—visible.

The next stage is structure, where AI helps translate concept maps and gap analysis into an outline. ChatGPT (or another large language model) is suggested for generating a suggested structure based on the headings and topics the researcher wants to cover. Thesis AI is offered as an example of a tool that can produce a detailed, referenced draft for structural inspiration, but not for copy-and-paste submission. SciSpace’s agent-based “create a literature review” function is treated similarly: useful for generating options, not for replacing the researcher’s own academic synthesis.

Content generation is where boundaries tighten. The workflow warns against fully automated writing tools (like “Manis” or “Genpite,” mentioned as examples) for university assignments, emphasizing that narrative control matters. ChatGPT is framed as a controllable option, while Jenny AI is described as a “gray zone” tool that outputs information as you type, potentially reducing your thinking effort—subject to university rules. Sourcely is recommended for source discovery when new claims or questions arise during drafting.

Finally, the draft undergoes review in two passes: AI review and manual review. AI review tools (ChatGPT, Thesisify, Paper Wizard) are used to check alignment with learning outcomes, thesis clarity, purpose, and evidence usage—while the researcher retains the right to disagree with AI feedback. The last step is human rigor: read every sentence, ideally with external feedback from someone in the field, then proofread carefully. A specific proofreading tactic is suggested—reading paragraphs backwards—to catch typos and logic slips. The workflow ends with a reminder that future self-critique is normal and even useful for academic growth, as long as the current draft meets standards and reflects real understanding.

Cornell Notes

The workflow for a literature review uses AI to reduce the reading burden while preserving academic skills. It begins by collecting papers into Zotero, using discovery tools (e.g., Google Scholar, Illicit, Consensus, Litmaps, Connected Papers, SciSpace, Research Rabbit) to build a reference library. AI then filters that library using relevance ranking and synthesized “key papers” outputs (SciSpace, Consensus, NotebookLM), after which the researcher reads about 20–25% of the most relevant papers in full. NotebookLM mapping and Consensus matrices help organize concepts, themes, and research gaps into an outline. The draft is then structured with an LLM, written with human control, and polished through AI checks plus a mandatory manual review (including backward paragraph proofreading).

How does the workflow prevent a literature review from turning into an impossible “read everything” project?

It treats filtering as a sanity step. After collecting papers into Zotero, AI tools narrow the list using relevance scores and “key papers” identification. SciSpace can rank results by relevance (e.g., selecting top papers with high relevance), and Consensus can surface top contributors and the claims/evidence tied to key papers. NotebookLM can then highlight key findings across uploaded references. The result is a smaller set for full reading—about 20–25% of the filtered papers—so the review is grounded without requiring exhaustive coverage.

What’s the role of full-paper reading if AI is doing so much triage?

AI is used to decide what deserves attention, not to replace understanding. The workflow insists on reading a meaningful subset—roughly 20–25% of the bubbled-up papers—to build “sixth sense” understanding of the field. That deep reading supports accurate synthesis, stronger narrative control, and better identification of how evidence fits together, which AI summaries alone may not guarantee.

How do NotebookLM and Consensus help turn scattered papers into a structured literature review?

NotebookLM’s mapping features can generate mind maps that mirror literature review structure, including concepts, themes, and potential headings/subheadings, with AI summaries when sections are clicked. Consensus adds a different lens: a matrix showing how many studies exist across application areas, helping identify where research is dense versus where it’s thin. If a region is underexplored, Consensus can label it as a “gap,” which then feeds directly into the draft’s narrative and section planning.

Why does the workflow recommend using AI for structure but not for final content submission?

Structure is treated as an outline problem—AI can propose headings and ordering based on the researcher’s concept map and gap analysis. Tools like ChatGPT, Thesis AI, and SciSpace can generate draft-like structures for inspiration. But the workflow warns against copy-and-paste submission, especially for university assignments, because the literature review must demonstrate understanding and academic rigor. Content generation should keep the researcher in control to avoid cheating themselves out of essential writing and synthesis skills.

What does “review” look like in this workflow, and how do AI and humans divide the labor?

Review happens in two stages. First is AI review: large language models and tools like Thesisify and Paper Wizard are used to check whether learning outcomes are met, whether the purpose and thesis statement are clear, and whether evidence is appropriate. Then comes manual review, which the workflow frames as non-negotiable: read every sentence, seek red-pen feedback from someone senior (e.g., postdoc or senior PhD student), and proofread carefully. A specific technique—reading each paragraph backwards—helps catch typos and logic errors that are easy to miss when reading forward.

How does the workflow handle new questions that appear while writing?

It uses a “source injection” loop. When drafting content triggers new questions or claims, Sourcely is suggested to find supporting sources. Alternatively, the researcher can return to the literature-finding step and add newly relevant papers back into the workflow, then revise the relevant sections accordingly.

Review Questions

  1. If a researcher has 200 papers after discovery, what concrete steps in this workflow reduce that to a manageable reading set, and which tools are used at each step?
  2. How do the mind map output from NotebookLM and the matrix output from Consensus complement each other when identifying research gaps?
  3. What are the differences between AI review and manual review in this workflow, and why is backward paragraph reading recommended?

Key Points

  1. 1

    Use Zotero as the reference manager so AI tools can connect to a single, organized library of papers.

  2. 2

    Combine modern discovery tools (Illicit, Consensus, Litmaps, Connected Papers, SciSpace, Research Rabbit) with Google Scholar keyword searches to build a strong initial set.

  3. 3

    Filter aggressively with AI using relevance ranking and “key papers” identification so the literature review is grounded in the most important studies.

  4. 4

    Read only about 20–25% of the filtered papers in full to build real understanding while avoiding exhaustive coverage.

  5. 5

    Use NotebookLM mapping and Consensus matrices to convert papers into concepts, headings, and visible research gaps.

  6. 6

    Generate an outline with an LLM, but keep narrative control during writing; avoid fully automated drafting for university submissions.

  7. 7

    Finish with AI-assisted checks plus a mandatory manual review, including backward paragraph proofreading and ideally feedback from a senior researcher.

Highlights

The workflow’s central move is AI-driven filtering: it narrows a large paper set to a small, high-value subset before any deep reading begins.
NotebookLM mind maps and Consensus matrices turn literature into an outline and gap map—so structure emerges from concepts, not from guesswork.
AI can help draft structure and check alignment, but the final literature review still requires human control, full-sentence reading, and feedback.
Backward paragraph proofreading is presented as a practical way to catch typos and logic slips that forward reading misses.
The process explicitly balances efficiency with academic skill-building: AI reduces workload, while selective full reading preserves understanding.

Topics