Perplexity AI vs ChatGPT for Research - Who wins?
Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Perplexity AI is strongest for literature discovery, especially when using academic search and Pro-level retrieval that returns more sources and suggests review-paper starting points.
Briefing
Perplexity AI pulls ahead for research workflows that start with finding and grounding sources, while ChatGPT holds the edge for hands-on scientific writing and deeper data analysis. A side-by-side test across four common “PhD student” tasks—literature discovery, explaining papers, plotting/calculating from experimental data, and rewriting for peer review—ends with a practical conclusion: use both tools, because each one is stronger in different parts of the research loop.
In the literature-finding round, both systems generate peer-reviewed paper recommendations for a hydrogen production project using catalysts. The difference shows up when Pro-level search is enabled: Perplexity returns a larger set of sources (20) and also guides the user toward review papers as a starting point, with topic-like grouping (for example, catalyst and electro-catalyst themes). The narrator treats this as a clear win for Perplexity, crediting its academic search mode and its ability to surface more relevant material without the same level of cleanup.
When asked to explain a uploaded research paper, Perplexity again performs strongly by extracting the key points and organizing them into higher-level categories rather than walking through the paper section-by-section. It also performs web searching and brings in additional sources alongside the uploaded paper, aiming to deepen context. ChatGPT’s explanation is competent but more repetitive and “bullet-point heavy,” staying closer to the structure of the original document and staying largely within the provided material.
The data-analysis test flips the outcome. ChatGPT successfully generates an interactive IV curve from unstructured solar cell testing data and computes device efficiency, with the final result indicating the device failed (efficiency calculated as zero). It also iterates when initial extraction from the file is imperfect, correcting its approach to get the plot right. Perplexity, using the same task and inputs, fails to produce the graph and instead relies on efficiency metadata already present in the file—so it doesn’t actually perform the calculation/plotting step in the same way.
For scientific writing and editing, the comparison becomes messy rather than one-sided. ChatGPT tightens an unfinished paragraph for peer review and improves academic phrasing, but it removes references rather than adding or correctly managing citations. Perplexity attempts to search for peer-reviewed references and insert them, yet the citations it produces appear incorrect or mismatched to the text. The final verdict lands as a draw: Perplexity is better for reference-backed literature writing, while ChatGPT is better at producing cleaner peer-review-ready prose. The overall takeaway is not a single winner, but a workflow recommendation—Perplexity for sourcing and context, ChatGPT for analysis and writing quality, with both requiring careful checking of citations and outputs.
Cornell Notes
Perplexity AI and ChatGPT were tested on four research tasks: finding literature, explaining peer-reviewed papers, analyzing experimental data, and rewriting for peer review. Perplexity won the literature and paper-explanation rounds by returning more sources (including guidance to start with review papers) and by organizing key points while also pulling in additional web sources. ChatGPT won the data-analysis round by generating an interactive IV curve and calculating efficiency from unstructured input, including iterative correction when data extraction was initially off. For writing and editing, neither tool fully wins: ChatGPT improves academic style but drops references, while Perplexity adds references via web search but appears to mis-handle citations. The practical conclusion is to use both tools and verify citations and calculations.
Why did Perplexity look stronger in the literature-finding test?
How did the two tools differ when summarizing/explaining an uploaded paper?
What made ChatGPT win the data analysis and plotting round?
Why did Perplexity lose the data-analysis test despite understanding the task?
Why was writing/editing a draw rather than a clear winner?
Review Questions
- In what specific way did Perplexity’s paper explanation differ from ChatGPT’s, and how did web searching change the output?
- What evidence from the IV-curve task indicates whether a system truly performed calculation versus using provided metadata?
- When rewriting for peer review, what two failure modes were observed for ChatGPT and Perplexity, respectively?
Key Points
- 1
Perplexity AI is strongest for literature discovery, especially when using academic search and Pro-level retrieval that returns more sources and suggests review-paper starting points.
- 2
Perplexity’s explanations tend to synthesize key findings into categorized takeaways rather than mirroring the paper’s section-by-section structure.
- 3
ChatGPT is more reliable for scientific data analysis tasks that require plotting and computed metrics from unstructured inputs.
- 4
ChatGPT’s IV-curve workflow produced an interactive plot and calculated efficiency, including iterative correction when data extraction needed adjustment.
- 5
Perplexity may fall back on efficiency metadata instead of performing the full plot-and-calculate workflow from raw data.
- 6
Scientific writing/editing is best handled as a hybrid workflow: ChatGPT improves academic phrasing, while Perplexity’s reference insertion needs verification for citation accuracy.