Get AI summaries of any video or article — Sign up free
Perplexity AI vs ChatGPT for Research - Who wins? thumbnail

Perplexity AI vs ChatGPT for Research - Who wins?

Andy Stapleton·
4 min read

Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Perplexity AI is strongest for literature discovery, especially when using academic search and Pro-level retrieval that returns more sources and suggests review-paper starting points.

Briefing

Perplexity AI pulls ahead for research workflows that start with finding and grounding sources, while ChatGPT holds the edge for hands-on scientific writing and deeper data analysis. A side-by-side test across four common “PhD student” tasks—literature discovery, explaining papers, plotting/calculating from experimental data, and rewriting for peer review—ends with a practical conclusion: use both tools, because each one is stronger in different parts of the research loop.

In the literature-finding round, both systems generate peer-reviewed paper recommendations for a hydrogen production project using catalysts. The difference shows up when Pro-level search is enabled: Perplexity returns a larger set of sources (20) and also guides the user toward review papers as a starting point, with topic-like grouping (for example, catalyst and electro-catalyst themes). The narrator treats this as a clear win for Perplexity, crediting its academic search mode and its ability to surface more relevant material without the same level of cleanup.

When asked to explain a uploaded research paper, Perplexity again performs strongly by extracting the key points and organizing them into higher-level categories rather than walking through the paper section-by-section. It also performs web searching and brings in additional sources alongside the uploaded paper, aiming to deepen context. ChatGPT’s explanation is competent but more repetitive and “bullet-point heavy,” staying closer to the structure of the original document and staying largely within the provided material.

The data-analysis test flips the outcome. ChatGPT successfully generates an interactive IV curve from unstructured solar cell testing data and computes device efficiency, with the final result indicating the device failed (efficiency calculated as zero). It also iterates when initial extraction from the file is imperfect, correcting its approach to get the plot right. Perplexity, using the same task and inputs, fails to produce the graph and instead relies on efficiency metadata already present in the file—so it doesn’t actually perform the calculation/plotting step in the same way.

For scientific writing and editing, the comparison becomes messy rather than one-sided. ChatGPT tightens an unfinished paragraph for peer review and improves academic phrasing, but it removes references rather than adding or correctly managing citations. Perplexity attempts to search for peer-reviewed references and insert them, yet the citations it produces appear incorrect or mismatched to the text. The final verdict lands as a draw: Perplexity is better for reference-backed literature writing, while ChatGPT is better at producing cleaner peer-review-ready prose. The overall takeaway is not a single winner, but a workflow recommendation—Perplexity for sourcing and context, ChatGPT for analysis and writing quality, with both requiring careful checking of citations and outputs.

Cornell Notes

Perplexity AI and ChatGPT were tested on four research tasks: finding literature, explaining peer-reviewed papers, analyzing experimental data, and rewriting for peer review. Perplexity won the literature and paper-explanation rounds by returning more sources (including guidance to start with review papers) and by organizing key points while also pulling in additional web sources. ChatGPT won the data-analysis round by generating an interactive IV curve and calculating efficiency from unstructured input, including iterative correction when data extraction was initially off. For writing and editing, neither tool fully wins: ChatGPT improves academic style but drops references, while Perplexity adds references via web search but appears to mis-handle citations. The practical conclusion is to use both tools and verify citations and calculations.

Why did Perplexity look stronger in the literature-finding test?

With academic search enabled and Pro-level searching, Perplexity returned more peer-reviewed sources (20) and added structure beyond a simple list—grouping recommendations by themes (e.g., catalyst-related categories) and explicitly suggesting a starting point (review papers). The emphasis on “academic” search mode and the larger, more guided set of references drove the win.

How did the two tools differ when summarizing/explaining an uploaded paper?

ChatGPT produced a high-level, bullet-point explanation that followed the paper’s typical structure (introduction, experimental results, discussion). Perplexity instead extracted the most important elements and categorized them together (e.g., fabrication/graded active layers, improved open-circuit voltage), and it also performed web searching to add supporting context beyond the uploaded document.

What made ChatGPT win the data analysis and plotting round?

ChatGPT accepted unstructured IV data from solar cell testing, generated an interactive IV curve (with hover/click exploration), and calculated efficiency. It also showed iterative behavior when initial extraction didn’t work cleanly, correcting how it passed/parsed the data. The computed efficiency came out as zero, consistent with a failed device.

Why did Perplexity lose the data-analysis test despite understanding the task?

Perplexity recognized the request to plot an IV curve and compute efficiency, but it did not produce the graph. Instead, it relied on efficiency already provided in the file metadata, meaning it didn’t fully execute the plotting/calculation workflow from the raw input in the same way ChatGPT did.

Why was writing/editing a draw rather than a clear winner?

ChatGPT improved peer-review suitability by tightening academic language, but it removed or failed to preserve references—so the output lacked useful citations. Perplexity attempted to add peer-reviewed references by searching the web, but the inserted citations appeared incorrect or mismatched to the paragraph, undermining reliability. Net effect: better prose from ChatGPT, better reference intent from Perplexity, but both require careful checking.

Review Questions

  1. In what specific way did Perplexity’s paper explanation differ from ChatGPT’s, and how did web searching change the output?
  2. What evidence from the IV-curve task indicates whether a system truly performed calculation versus using provided metadata?
  3. When rewriting for peer review, what two failure modes were observed for ChatGPT and Perplexity, respectively?

Key Points

  1. 1

    Perplexity AI is strongest for literature discovery, especially when using academic search and Pro-level retrieval that returns more sources and suggests review-paper starting points.

  2. 2

    Perplexity’s explanations tend to synthesize key findings into categorized takeaways rather than mirroring the paper’s section-by-section structure.

  3. 3

    ChatGPT is more reliable for scientific data analysis tasks that require plotting and computed metrics from unstructured inputs.

  4. 4

    ChatGPT’s IV-curve workflow produced an interactive plot and calculated efficiency, including iterative correction when data extraction needed adjustment.

  5. 5

    Perplexity may fall back on efficiency metadata instead of performing the full plot-and-calculate workflow from raw data.

  6. 6

    Scientific writing/editing is best handled as a hybrid workflow: ChatGPT improves academic phrasing, while Perplexity’s reference insertion needs verification for citation accuracy.

Highlights

Perplexity returned 20 peer-reviewed sources for the hydrogen-catalyst project and explicitly recommended starting with review papers—turning literature search into a guided literature-review kickoff.
ChatGPT generated an interactive IV curve and computed efficiency from unstructured solar-cell testing data, with the result indicating the device failed (efficiency = zero).
Writing/editing ended in a draw: ChatGPT tightened academic language but removed references, while Perplexity added references via web search but appeared to mis-handle them.