Get AI summaries of any video or article — Sign up free
Claude vs ChatGPT vs Gemini vs Perplexity: The BEST AI for Research thumbnail

Claude vs ChatGPT vs Gemini vs Perplexity: The BEST AI for Research

Andy Stapleton·
5 min read

Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Claude Pro delivered the most reliable accuracy when interrogating uploaded PDFs, with the lowest error rate among the tested paid models.

Briefing

Paid AI assistants are increasingly used for academic work, but performance depends less on “which model is best” and more on the specific research task—especially whether the job involves reading uploaded PDFs or generating citations. In stress tests using pro subscriptions, Claude delivered the most reliable results when interrogating PDFs, producing the lowest error rate and the most trustworthy answers without being pulled into fabricated claims.

The testing focused on two core academic needs: extracting information from documents and producing literature references that actually exist. For PDF-based questions—both straightforward prompts like “is this in the paper?” and trickier prompts designed to tempt the model into accepting fabricated information—Claude came out clearly ahead. The key takeaway was reliability: when asked to summarize, expand on concepts found in the text, or answer whether specific claims appeared in the paper, Claude was the least prone to “lying” during the evaluation. Perplexity ranked next best for PDF interrogation, while Gemini and ChatGPT lagged behind. Notably, ChatGPT’s performance on this PDF task was worse than its earlier free-version benchmark, suggesting that pro access and newer model access did not automatically translate into better document-grounded accuracy.

The second major benchmark—reference generation—painted a different picture. When prompted to retrieve or produce citations for a research topic (including both realistic queries and misleading prompts), ChatGPT performed best at avoiding hallucinated or incorrect references. The results were quantified: ChatGPT achieved the highest reference accuracy at 82.35%, while Claude’s reference accuracy was far lower at 40%. That gap led to a practical rule for researchers: Claude should not be used as a primary tool for citation hunting, even if it excels at reading and summarizing PDFs. The transcript also emphasizes that large language models can generate references that sound plausible while being wrong, so citation outputs still require verification.

The overall conclusion rejects the idea of a single “best AI for research.” A scatter-plot style comparison of content accuracy versus reference accuracy showed no universal winner across tasks. Claude is the best paid option for extracting and summarizing information from uploaded PDFs. ChatGPT is the better choice for early-stage exploration where the goal is to obtain a majority of real references quickly, though dedicated literature tools (like SciSpace, Elicit, and Consensus) are positioned as more dependable because they use real academic databases rather than free-form generation. Gemini and Perplexity land in the middle—generally strong on content extraction but weaker on reference accuracy.

For students and researchers, the practical workflow implied by the results is straightforward: use Claude when the source of truth is a PDF you’ve uploaded, and use ChatGPT (or—preferably—specialized academic search tools) when the task is building a bibliography. In every case, the transcript stresses the same safeguard: verify what comes out, because even the best-performing model can still produce errors when asked to invent citations or claims not grounded in provided documents.

Cornell Notes

Claude Pro emerged as the top choice for interrogating uploaded PDFs, delivering the lowest error rate and the most reliable answers when questions were grounded in document content. ChatGPT Pro performed best for generating references, reaching the highest reference accuracy (82.35%)—but it still requires verification because plausible citations can be wrong. The tests found no single model dominates across both tasks: Claude’s reference accuracy was much lower (40%), while ChatGPT’s PDF performance lagged behind Claude. Gemini and Perplexity generally sat in the middle, strong on content but weaker on citations. The practical takeaway is task-based selection: use Claude for PDF-grounded work and ChatGPT (or database-driven tools) for citation gathering.

Why did Claude outperform the other models on the PDF interrogation tests?

Claude produced the most accurate, document-grounded responses when asked questions like “is this in the paper?” and when prompted to expand on concepts found in the text. It also handled deliberately misleading prompts better than the others, showing a lower tendency to accept fabricated claims. In the results, Claude was described as the clear winner for PDF accuracy, with Perplexity next, while Gemini and ChatGPT performed worse.

What was the key weakness of Claude revealed by the citation benchmark?

Claude’s reference accuracy was much lower than ChatGPT’s. In the quantified results, Claude scored 40% on reference accuracy, meaning it produced too many citations that were incorrect or didn’t match what existed. The transcript draws a direct implication: don’t use Claude as a citation-finding tool just because it’s strong at reading PDFs.

How did ChatGPT’s performance differ between PDF questions and reference generation?

ChatGPT underperformed on the PDF interrogation task, doing worse than its earlier free-version benchmark mentioned in the transcript. However, for reference generation it led the pack, achieving 82.35% reference accuracy. This contrast highlights that ChatGPT was better at producing citations that were more often real, even though its PDF-grounded accuracy wasn’t the strongest.

What does the “no clear winner” conclusion mean in practice?

The scatter-plot comparison of content accuracy (how well it extracts correct information) versus reference accuracy (how often citations exist and match) showed trade-offs. Claude is best for PDF-grounded content; ChatGPT is best for references; Gemini and Perplexity are intermediate. Researchers should choose tools based on whether the task is document interrogation or citation generation, rather than seeking one universal model.

Why does the transcript recommend specialized academic tools for literature search?

Large language models can generate references that sound plausible but may not exist or may contain wrong details. The transcript positions tools like SciSpace, Elicit, and Consensus as better benchmarks for citation work because they rely on real academic databases rather than purely generative output. Even when ChatGPT performs well, the transcript still stresses verifying citations.

What workflow does the transcript imply for early-stage research?

For early exploration of a research field, the transcript suggests using ChatGPT Pro to obtain a majority of real references quickly, then verifying them. For tasks where the source is a specific paper (uploaded PDFs), Claude Pro is recommended for accurate extraction and summarization. It also mentions Notebook LM as a free option for interrogating literature, but still treats Claude as the best paid choice for PDF interrogation.

Review Questions

  1. If a researcher needs to answer questions strictly from an uploaded PDF, which model performed best and what evidence from the tests supports that choice?
  2. What were the reference accuracy results for Claude Pro versus ChatGPT Pro, and how do those numbers change how each should be used?
  3. How does the transcript justify using database-driven tools (like SciSpace, Elicit, Consensus) instead of relying on citation outputs from language models?

Key Points

  1. 1

    Claude Pro delivered the most reliable accuracy when interrogating uploaded PDFs, with the lowest error rate among the tested paid models.

  2. 2

    ChatGPT Pro achieved the highest reference accuracy at 82.35%, making it the strongest option among the tested models for citation generation.

  3. 3

    Claude Pro’s reference accuracy was only 40%, so it should not be treated as a primary tool for finding or generating citations.

  4. 4

    No single model dominated both tasks; content accuracy and reference accuracy trade off depending on the academic workflow.

  5. 5

    Gemini and Perplexity performed in the middle: relatively strong on content extraction but weaker on references than ChatGPT.

  6. 6

    Citation outputs from large language models can still be wrong even when they sound plausible, so verification remains essential.

  7. 7

    For literature search, database-driven tools like SciSpace, Elicit, and Consensus are positioned as more dependable than generative citation workflows.

Highlights

Claude Pro was the clear winner for PDF interrogation, described as the model that “did not lie” in the document-grounded tests.
ChatGPT Pro led on references, hitting 82.35% reference accuracy—far above Claude’s 40%.
The results reject a single “best AI for research,” instead recommending task-based selection: Claude for PDFs, ChatGPT for citations, with verification always required.

Topics

Mentioned