Meta Analysis Part 2
Based on International Research Community's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Topic selection is the first major constraint in meta-analysis because many recent publications are reviews rather than original trials or cohort studies.
Briefing
Meta-analysis work is getting harder to start because the field is crowded with reviews and network meta-analyses, leaving fewer genuinely “original” trials and cohort studies to synthesize. The practical response is to treat topic selection as the first bottleneck, then build a tight PICO (Population, Intervention, Comparator, Outcomes) and a search string that reliably captures all relevant studies—without drowning the screening team in irrelevant hits.
Once a viable topic is chosen, the workflow moves into screening. The session stresses that PICO must be precise and consistent from the search string through inclusion decisions. A sample framing used throughout is cyclophosphamide versus corticosteroids in lupus nephritis (nephritic patients as the population; steroids as the comparator; outcomes such as kidney failure or mortality). After PICO is set, the next step is translating it into a PubMed search strategy using synonyms and controlled combinations: “OR” connects alternative terms for the same concept (e.g., cyclophosphamide and its alternative names), while “AND” links different concepts (drug + steroids + disease + population). Bracketed groups are recommended to keep the logic correct. The guidance also warns against over-trusting automated counts from tools like “Chargy/ChatGPT,” and against applying heavy filters early—filters can be unreliable when the goal is to avoid missing eligible studies.
The session then shifts to operationalizing screening using Rayyan, described as a free online tool for systematic reviews and meta-analyses. The process begins by running the search in PubMed, exporting results via “Send to” and “Citation manager,” and importing them into Rayyan using the “All results” option (not just the first page). Rayyan supports team screening: the owner can invite reviewers by email, and the platform can run in “blind on” mode so individual screeners cannot see each other’s decisions. Duplicates—common when importing from multiple databases—must be resolved by keeping one record.
Screening is split into primary and secondary stages. Primary screening is title/abstract-based: reviewers choose Include, Exclude, or Maybe (for uncertainty), and they can add comments explaining decisions. Keyboard shortcuts (like “E” for exclude) speed up work. The session emphasizes that primary screening should focus on matching PICO and excluding non-original study types such as case reports. When two reviewers disagree, the article lands in a conflict queue.
Secondary screening (full text screening) is where conflicts are resolved and where inclusion becomes final. Reviewers check the full manuscript—especially methods and results—and verify that the study truly matches the PICO and includes relevant outcomes. Tables are treated as the backbone of this step: baseline characteristics (often “Table 1”) and outcome reporting (often “Table 2” and beyond) must align with the meta-analysis outcomes. Supplementary files are also flagged as necessary when key outcomes or data are missing from the main text. If a study lacks any of the outcomes of interest, it may be excluded because extraction and analysis depend on outcome availability.
The session closes by reinforcing the end-to-end logic: build a comprehensive search string, import all results into Rayyan, run independent primary screening to reduce error, resolve disagreements through secondary screening using full-text and outcome checks, and then proceed to data extraction in the next session.
Cornell Notes
The core message is that a meta-analysis succeeds or fails based on disciplined topic/PICO definition and a search strategy that captures all eligible original studies. After running the PubMed search with synonym logic (OR within concepts, AND across concepts), results are imported into Rayyan for team screening. Primary screening uses titles and abstracts to label studies as Include, Exclude, or Maybe, with “blind on” to prevent bias and a conflict queue for disagreements. Secondary screening (full text) resolves conflicts and confirms that studies match PICO and report the required outcomes, using tables and supplementary files to verify extractable data. This matters because outcome availability and correct PICO alignment determine whether studies can be included in extraction and meta-analysis.
How should a PICO-based search string be constructed so it doesn’t miss eligible studies?
Why does Rayyan use “blind on,” and how does that reduce screening error?
What’s the difference between primary screening and secondary screening in practice?
What should reviewers look for in secondary screening to decide inclusion for extraction?
How are duplicates handled when importing studies into Rayyan?
How do conflicts get resolved when two primary screeners disagree?
Review Questions
- When building a PubMed search string from PICO, where do OR and AND belong, and why do brackets matter?
- In Rayyan, what is the purpose of “blind on,” and how does the conflict queue function during screening?
- During secondary screening, what specific evidence (tables, supplementary files, methods/results) determines whether a study is extractable for the outcomes of interest?
Key Points
- 1
Topic selection is the first major constraint in meta-analysis because many recent publications are reviews rather than original trials or cohort studies.
- 2
PICO must be consistent from search-string construction through inclusion decisions; deviating from PICO risks producing an unusable meta-analysis.
- 3
Use OR to connect synonyms for the same concept and AND to combine different PICO concepts; bracket OR-groups to preserve correct query logic.
- 4
Import all search results into Rayyan using “All results” (not only the first page) and resolve duplicates before screening.
- 5
Run primary screening independently with “blind on” and use Include/Exclude/Maybe plus comments to document uncertainty.
- 6
Resolve disagreements in the conflict section during secondary (full-text) screening by verifying head-to-head comparisons and extractable outcomes.
- 7
Secondary screening should prioritize outcome reporting (tables and supplementary files); studies lacking outcomes of interest are typically excluded for extraction.