Breaking down Faheem's systematic literature review: From Idea to Execution
Based on SciSpace's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
A systematic literature review must start with a clear purpose and research questions that are valuable, crisp, and answerable from existing literature.
Briefing
Systematic literature reviews can be built from scratch using a disciplined pipeline—then accelerated with AI tools—without surrendering the researcher’s judgment. The core takeaway is that publishable-quality reviews depend less on which review style is chosen and more on defining the right purpose and research questions, running a controlled search with inclusion/exclusion criteria, and extracting/analyzing data in a structured way. In the manual workflow described, a high-quality review took roughly 4–6 months, reflecting the heavy time cost of screening thousands of papers down to a final set.
The session starts by clarifying what a literature review is: synthesizing existing work to identify patterns, trends, concepts, and directions for future research. It then distinguishes multiple review types—systematic, scoping, integrative/critical, narrative, historical, and meta-synthesis—while emphasizing that systematic reviews follow a protocol tightly enough to satisfy reviewers. The practical “how-to” begins with the steps of a systematic literature review: define the need for the review and its research questions; run a pilot study to test feasibility; design search strings for databases (e.g., Google Scholar, ACM, Scopus); apply inclusion/exclusion criteria; select relevant studies; extract data using a predefined extraction form; analyze the extracted data (the example uses thematic analysis); and report findings with results, discussion, and future research implications.
A detailed case study illustrates the workflow. For a 2019 literature review on architectural tactics for big data systems (published in Journal of Systems and Software), the process began with research questions about quality attributes for BDCA systems and architectural tactics to address quality concerns. Paper identification relied on a manual, trial-and-error search-string approach across major computing databases, then progressively narrowed the pool: 4,634 initial hits, reduced via title screening to 748, deduplicated to 516, abstract screening to 168, full-text review to 69, and finally “snowballing” from references to add 5 more—ending with 74 included papers. Data extraction was guided by a structured form capturing authorship and publication details plus the technique, quality attributes, rationale, and tactics needed to answer the two research questions.
The second half shifts to automation using SciSpace (spelled “skypace” in the transcript), positioning it as a way to reduce the most time-consuming steps. SciSpace is presented as supporting paper identification with transparent filtering and relevance counts, automated data extraction into a spreadsheet-like structure (including extracting specific fields like quality attributes), paraphrasing via a Chrome extension, question-based extraction from papers, and citation generation in formats such as APA. Beyond these, it’s described as offering summaries, outlines, proofreading, formatting, exporting, and other writing-assistance features.
Looking ahead, the session argues that research will become faster and more tool-driven, but “AI supports, not replaces.” The future advantage goes to researchers who adopt tools effectively while still critically evaluating outputs, managing bias and limitations, avoiding AI-generated plagiarism risks (especially when writing is fully AI-produced), and protecting sensitive information by checking terms and data-safety policies. The skills emphasized for future researchers include prompt engineering, tools literacy, ethical AI usage, workflow automation, and critical evaluation of AI outputs—paired with enduring fundamentals like critical thinking, connecting research ideas, and strong reading and writing.
Cornell Notes
A systematic literature review succeeds when it is built on clear purpose and well-crafted research questions, then executed through a protocol: pilot testing, search-string design, inclusion/exclusion screening, structured data extraction, and thematic analysis. In a concrete example, a big-data architecture review narrowed 4,634 initial papers down to 74 through staged screening (title → abstract → full text) plus snowballing from references, and then extracted tactics and quality attributes into a predefined form. The workflow is time-intensive when done manually (about 4–6 months for publishable quality), but AI tools like SciSpace can automate paper identification, data extraction, paraphrasing, and citation generation. The future of research favors researchers who use AI to accelerate repetitive work while maintaining critical judgment, ethical safeguards, and awareness of bias, plagiarism, and data privacy risks.
What makes a systematic literature review different from other review types, and why does that matter for publication?
How do purpose and research questions shape the entire literature review pipeline?
Why is pilot testing used before running the full search, and what does it verify?
How does the case study reduce thousands of papers to a final set without losing relevance?
What does “data extraction” mean in practice, and how is it kept from becoming unmanageable?
Which parts of a systematic review are most amenable to automation with SciSpace, and what safeguards remain necessary?
Review Questions
- If you had to start a systematic literature review today, what order would you follow for purpose → research questions → pilot study → search-string design → inclusion/exclusion → extraction → analysis?
- In the case study, which screening stage contributed the largest reduction, and how did snowballing change the final included-paper count?
- What skills beyond traditional reading/writing are needed to use AI tools effectively in research, and why is critical evaluation still required?
Key Points
- 1
A systematic literature review must start with a clear purpose and research questions that are valuable, crisp, and answerable from existing literature.
- 2
A pilot study (8–10 papers) tests whether the questions can be answered and helps validate that the search strategy is aligned with the review’s goals.
- 3
Search-string design is a trial-and-error process that must balance recall (enough papers) with precision (not hundreds of thousands).
- 4
Inclusion/exclusion criteria and staged screening (title, abstract, full text) are essential to reduce thousands of hits to a manageable, relevant set.
- 5
Structured data extraction forms prevent reviewers from getting lost and ensure extracted fields directly map to the research questions.
- 6
AI tools like SciSpace can automate paper identification, data extraction, paraphrasing, and citation generation, but researchers must still critically evaluate outputs and manage ethics, bias, and privacy.
- 7
Future research skills include prompt engineering, tools literacy, ethical AI usage, workflow automation, and the ability to judge whether AI-generated content is reliable.