Get AI summaries of any video or article — Sign up free
SYSTEMATIC REVIEW –FREQUENTLY ASKED QUESTIONS –Part 1 & 2 thumbnail

SYSTEMATIC REVIEW –FREQUENTLY ASKED QUESTIONS –Part 1 & 2

6 min read

Based on Systematic review and Primary research - Q & A's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Systematic reviews are protocol-driven and designed to answer a specific research question, while literature reviews are typically narrative and context-setting.

Briefing

Systematic reviews and literature reviews differ in purpose, structure, search rigor, and how bias is handled. A systematic review is built to answer a specific research question using a protocol-driven, explicit, and accountable methodology. That structure shows up in comprehensive, reproducible search strategies; mandatory quality assessment of included studies; and bias-reducing steps such as double data extraction and systematic appraisal. Literature reviews, by contrast, are typically more flexible and narrative, often using more selective and variable searching, with less emphasis on formal bias control and quality assessment—meaning they are more vulnerable to bias. Systematic reviews also use defined data synthesis methods—ranging from meta-analysis to qualitative approaches like metaethnography and mixed-method synthesis—whereas literature reviews often rely on narrative or qualitative summaries, even when meta-analysis appears in some cases.

The transcript then tackles practical “how many” questions that frequently stall review teams. There is no single correct number of databases or total sources to search; choices depend on the research question, discipline, and how broad the scope is. Clinical trial reviews commonly draw on Medline, Embase, and the Cochrane Central Register of Controlled Trials, while nursing and allied health work may add CINAHL. Education research often uses ERIC, psychology may use PsycINFO, and sociology may use Social Science Abstracts. Beyond databases, teams are urged to search trial registries, organizational websites, and to use hand-searching and reference list scanning. Regional databases matter too—for example, when studies come from India or Latin America.

Similarly, there is no fixed number of studies that must be included. A systematic review can include as few as two or three studies for a narrow question, while broad evidence gap maps can include hundreds. Even “zero studies” can be a legitimate and valuable outcome because it signals a research gap and guides future primary research. Across all scenarios, the emphasis stays on quality over quantity: eligibility criteria should be transparent and grounded in frameworks such as PICO.

Team composition is also flexible but not optional. At least two independent reviewers are needed for screening and data extraction to reduce bias, with an adjudicator recommended to resolve disagreements. Information specialists (librarians) help design comprehensive search strategies across many databases, and methodological/statistical expertise supports meta-analysis and other synthesis decisions. Subject-matter expertise helps ensure relevance and accurate interpretation.

Protocol registration is strongly recommended to improve transparency and credibility, reduce duplication, and provide a “road map” for the team. The transcript lists common registries such as PROSPERO (health-focused) and OSF, along with options through Cochrane, Campbell, and other platforms.

Finally, the transcript distinguishes systematic maps (evidence gap maps) from systematic reviews. Mapping is suited to broad, multifaceted questions and does not require detailed extraction of study findings or formal synthesis; it often uses categorization and clustering to identify where deeper systematic reviews are warranted. It also introduces tools and methods for later stages of review work—software for managing systematic reviews, approaches to synthesis (aggregation vs configuration), embedded methodological studies (SAR), quality appraisal tools for both primary studies and reviews (e.g., AMSTAR 2, ROBIS, CASP), reporting standards (PRISMA and extensions), and priority screening using machine learning to speed title/abstract screening. A second segment continues with these methodological and operational details, including when mixed-method systematic reviews are appropriate and how to assess risk of bias and reporting quality.

Cornell Notes

The transcript lays out core differences between systematic reviews and narrative literature reviews: systematic reviews target a specific question with protocol-driven, reproducible methods; they use comprehensive searching, mandatory quality assessment, and bias-control steps like double data extraction. It then answers common planning questions—database and source selection depend on discipline and scope; there is no fixed number of studies; and teams typically need at least two independent reviewers plus optional roles such as an information specialist, statistician, and subject expert. Protocol registration is strongly recommended to improve transparency and reduce duplication, with PROSPERO and OSF highlighted as common options. The transcript also explains when to use systematic maps (evidence gap maps) instead of systematic reviews, emphasizing broad mapping without formal synthesis and clustering to guide later, more focused reviews.

Why does a systematic review demand more structure than a literature review?

A systematic review is designed to answer a specific research question using a protocol-driven, explicit, and rigorous methodology. That structure drives comprehensive, reproducible search strategies; mandatory quality assessment of included studies; and concrete bias-control measures such as double data extraction and systematic appraisal checks. Literature reviews are typically more flexible and narrative, often using more selective and variable searching and relying less on formal quality assessment, which increases susceptibility to bias.

How should review teams decide how many databases and sources to search?

There is no fixed number. Database and source selection depends on the research question, discipline, and the need for comprehensive coverage. Examples given include Medline, Embase, and CENTRAL for clinical trials; CINAHL for nursing and allied health; ERIC for education; PsycINFO for psychology; and Social Science Abstracts for sociology. Teams should also search trial registries, relevant organizational websites, use hand-searching, scan reference lists, and include regional databases when the evidence is geographically concentrated.

What is the minimum team size for a systematic review, and what roles improve quality?

At least two independent reviewers are needed for screening and data extraction to reduce bias and improve reliability. An adjudicator is recommended to resolve disagreements. Optional but desirable roles include an information specialist/librarian to design and execute advanced search strategies across many databases, a statistician/methodologist for meta-analysis and methodological decisions, and a subject-matter expert to ensure relevance and accurate interpretation.

Is there a required minimum or maximum number of studies in a systematic review?

No fixed minimum or maximum exists. Narrow questions may be addressed with as few as two or three studies, while broad scopes can include hundreds or even thousands. Evidence gap maps can also include very large numbers. Even a systematic review with zero included studies can be valuable because it identifies a research gap and informs recommendations for future primary research; the guiding principle is quality over quantity.

When should a systematic map (evidence gap map) be used instead of a systematic review?

A systematic map is appropriate for broad, multifaceted topics where a refined effectiveness-style question is not yet established—especially when multiple interventions, populations, or outcomes are involved. Mapping focuses on characterizing the evidence base through categorization and clustering rather than extracting detailed findings and running formal synthesis. It can be used in a two-stage model: map first to understand the landscape, then select clusters for later systematic review synthesis.

What does “priority screening” add to the systematic review workflow?

Priority screening uses machine learning and text mining to rank titles/abstracts for manual review. After researchers label an initial set of citations as include/exclude, the software trains on those decisions and iteratively sorts new records so the most likely eligible studies appear first. This can speed screening and reduce the need to review the entire citation set, while still relying on manual decisions for final inclusion.

Review Questions

  1. What specific methodological features (search, bias control, quality appraisal, synthesis) distinguish systematic reviews from narrative literature reviews?
  2. How would you justify the choice of databases and additional sources for a systematic review in a new discipline with limited prior evidence?
  3. In what situations would a systematic map be the better first step, and what information would you extract differently than in a systematic review?

Key Points

  1. 1

    Systematic reviews are protocol-driven and designed to answer a specific research question, while literature reviews are typically narrative and context-setting.

  2. 2

    Bias control is central to systematic reviews, including steps like double data extraction and mandatory quality assessment of included studies.

  3. 3

    Database and source counts are not fixed; selection should match the discipline, scope, and need for comprehensive coverage, including trial registries and regional sources.

  4. 4

    A systematic review can include very few studies—or none—without invalidating the work; quality and transparent eligibility criteria matter more than volume.

  5. 5

    At least two independent reviewers are needed for screening and data extraction, with an adjudicator recommended; information specialists and methodological/statistical expertise improve execution.

  6. 6

    Registering a systematic review protocol is strongly recommended for transparency, credibility, and to reduce duplication; PROSPERO and OSF are common options.

  7. 7

    Systematic maps (evidence gap maps) are suited to broad topics and use categorization/clustering without formal synthesis, often feeding into later focused systematic reviews.

Highlights

Systematic reviews require comprehensive, reproducible searching plus mandatory quality assessment and explicit bias-control steps; literature reviews often do not.
There is no fixed number of databases, sources, or included studies—decisions depend on scope, discipline, and eligibility criteria, and “zero studies” can still be a meaningful result.
Systematic maps characterize evidence landscapes without formal synthesis, extracting less detailed findings and using clustering to identify where deeper reviews are needed.
Priority screening can accelerate title/abstract screening by ranking citations using machine learning trained on include/exclude decisions.
Protocol registration improves transparency and reduces duplication, with PROSPERO and OSF highlighted as common registries.

Topics

  • Systematic Review vs Literature Review
  • Database and Source Searching
  • Protocol Registration
  • Systematic Map vs Systematic Review
  • Synthesis Methods and Tools

Mentioned

  • PICO
  • CENTRAL
  • CINAHL
  • ERIC
  • ASA
  • PsycINFO
  • OSF
  • PROSPERO
  • AMSTAR
  • AMSTAR 2
  • ROIS
  • ROBIS
  • CASP
  • JBI
  • SAR
  • MMSR
  • PRISMA
  • EGM
  • 3IE
  • EP
  • EPPI