Elicit AI Can Now Do Your Entire Systematic Review (In Minutes)
Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Elicit’s systematic review workflow screens hundreds of papers using inclusion/exclusion criteria and assigns each paper an inclusion score.
Briefing
Elicit AI’s paid “systematic review workflow” can screen hundreds of papers and extract structured data in minutes—turning what normally takes hours or days of manual reading into a guided, step-by-step process with adjustable inclusion thresholds. In a live walkthrough using a mindfulness-and-anxiety research question, the tool pulled in 499 candidate papers, evaluated them against inclusion/exclusion criteria, and then extracted data from the subset that met the user’s relevance bar.
The workflow starts with a research question and then asks for a speed/comprehensiveness setting (the walkthrough used a “fast” option). Elicit generates a research report by first finding papers, then screening them, and finally extracting data. A key feature is the structured “systematic review” pipeline: it offers suggested screening steps, lets users choose whether to use AI-generated suggestions or skip directly to screening, and supports importing papers already found elsewhere (including uploading up to 500 PDFs). The interface also exposes screening dimensions—such as age range, data separation, intervention duration/type, and related study characteristics—so inclusion decisions aren’t a black box.
During screening, Elicit evaluates each paper and assigns an inclusion score. The walkthrough showed the system processing all 499 papers autonomously, with progress updates and counts of papers likely to be included versus excluded. Once screening completes, the user can tune the score threshold using a slider to control how strict the review is. In the example, a high score (near 4.9 out of 5) signaled very close alignment with the research question, while lower scores corresponded to weaker matches. Adjusting the threshold changed the number of included studies dramatically—down to a single paper at very high thresholds, or up to 107 papers at a lower cutoff.
After inclusion is set, Elicit performs data extraction across the included studies. It can use suggested extraction columns (including study design and randomization details) or add custom fields. The walkthrough emphasized that extraction is where most of the time goes, because it pulls detailed information per paper—participant demographics, intervention details, and outcome/effect-size elements—then compiles everything into a report.
The results appear as both an on-screen report and downloadable PDF. The report includes typical systematic-review components: an abstract, methods and flow-style accounting of how many papers were screened, included, and ignored, plus tables of outcomes and secondary outcomes. The walkthrough also noted that the full report generation can take roughly 10–20 minutes depending on reference counts.
Beyond systematic reviews, Elicit retains its broader research workflow: semantic paper search, research-question-driven summaries, filtering, and iterative steps like creating tables, summarizing concepts across papers, and “chat with papers” using full text (not just abstracts). The overall pitch is practical: a guided pipeline that forces the same decisions researchers make—while automating the bulk of screening and extraction.
Cornell Notes
Elicit AI’s systematic review workflow automates the core labor of a systematic review: screening hundreds of papers and extracting structured data from the included subset. In a walkthrough question about mindfulness-based interventions reducing anxiety in university students, Elicit screened 499 candidate papers using inclusion/exclusion criteria and assigned each paper an inclusion score. Users can adjust a score threshold to control how many studies qualify (e.g., 107 included at a moderate cutoff, far fewer at stricter cutoffs). After inclusion is set, Elicit extracts data using suggested or custom columns and then generates a detailed report and PDF, typically taking longer during extraction and then producing the final narrative and tables in roughly 10–20 minutes. This matters because it compresses days of reading into a guided workflow with transparent decision controls.
How does Elicit’s systematic review workflow decide which papers to include or exclude?
What role does the “score threshold” slider play in shaping the final review?
What happens during data extraction, and what kinds of fields can be captured?
How does the workflow handle papers already found or uploaded by the researcher?
What do the generated reports include, and how do “fast” and “systematic review” differ?
Beyond systematic reviews, what other research tasks does Elicit support in this workflow?
Review Questions
- In the mindfulness/anxiety example, how did changing the inclusion score threshold affect the number of included papers and the scope of data extraction?
- Which parts of a systematic review does Elicit automate in the walkthrough (screening, extraction, report writing), and which user decisions remain adjustable?
- What kinds of extraction columns were shown as suggested or selectable, and how could a researcher add custom columns?
Key Points
- 1
Elicit’s systematic review workflow screens hundreds of papers using inclusion/exclusion criteria and assigns each paper an inclusion score.
- 2
A user-adjustable score threshold controls how many studies qualify for extraction, dramatically changing the review’s breadth.
- 3
The workflow follows a systematic-review style pipeline: paper selection for screening, screening evaluation, extraction, and then report generation.
- 4
Data extraction can use suggested fields (e.g., study design and randomization details) or custom columns, producing structured tables per paper.
- 5
Report outputs include typical systematic-review sections such as methods and flow-style counts of screened, included, and ignored papers, plus outcome tables.
- 6
Extraction is the most time-consuming step; report generation then compiles the extracted data into a detailed narrative and PDF.
- 7
Elicit also supports broader literature workflows—semantic search, filtering, concept summaries, and “chat with papers” using full text.