I Used DeepSeek R1 for Research – Here's What Happened!
Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
DeepSeek R1 is positioned as a free, open-source model that can support academic workflows, but it isn’t consistently publication-grade.
Briefing
DeepSeek R1 lands as a free, open-source AI model that can meaningfully assist academic work—especially for drafting structure and finding peer-reviewed papers—though it still lags behind top competitors on figure understanding and polished writing quality. In practical tests, it produced a usable literature-review outline for carbon nanotubes, generated a draft introduction section with citations, and returned 32 “recent” peer-reviewed papers (2023–2025) that the user could click through to publisher pages. That combination—starting academic writing and supplying clickable sources—makes it a credible research sidekick, even if it isn’t the strongest option for high-stakes, publication-ready output.
For literature review tasks, DeepSeek R1 performed best as a scaffold rather than a replacement for established writing tools. It generated a structured outline quickly, then followed up by drafting an “anxiety-free” introduction section. The draft was judged as a solid starting point (roughly a C+ / “pass”), with citations included—such as a Nature paper on mechanical strength—yet it didn’t match the depth and layout guidance produced by tools like ChatGPT in the user’s experience. The model’s “thinking” process was visible in real time, which the user found informative, but the resulting prose still required human judgment and refinement.
Where DeepSeek R1 showed clear utility was reference discovery. Using an internet-enabled search prompt for “recent peer-reviewed papers on nanoparticles for OPV devices,” it returned 32 results and linked out to major outlets and databases (including ScienceDirect, Nature, PubMed, and a Hong Kong Polytechnic University page). The user flagged one mismatch: the first item’s displayed title didn’t align with what opened on click, suggesting occasional citation/title confusion. Even so, the workflow—generate a list of recent, relevant papers and then click through to verify—was considered genuinely helpful, particularly if run periodically alongside other research tools.
DeepSeek R1’s weaker area was visual reasoning from complex figures. When given a demanding set of microscopy-related visuals (SEM and AFM images, plus a height profile) for non-planar vs planar electrode comparisons involving transparent electrodes, single carbon nanotubes, and silver nanowires, the model struggled to correctly “mesh” the components. It identified parts of the figure but missed how the elements connected, and the user repeatedly found ChatGPT better at extracting figure text and translating it into accurate, integrated conclusions. DeepSeek improved somewhat with additional prompting, but the gap remained.
On higher-level synthesis—turning multiple figures into a compelling research story—the model produced a plausible ordering and narrative arc, though again it was judged inferior to ChatGPT’s stronger ability to integrate figure relationships. For editing, it offered improvement suggestions to an abstract, but the overall verdict was that other mainstream tools (Claude and ChatGPT) can do similar work without the need to switch specifically to DeepSeek.
Overall, DeepSeek R1 earns its keep for free academic assistance: outlines, draft starters, and paper-finding workflows. The tradeoff is reliability and polish—especially for figure interpretation—where established competitors still lead for publication-grade writing and analysis.
Cornell Notes
DeepSeek R1 is a free, open-source model that can support academic research tasks, particularly by generating literature-review structure, drafting initial sections, and finding recent peer-reviewed papers with clickable links. In tests on carbon nanotubes, it produced a usable outline and an introduction draft with citations, but the prose depth and structure guidance were judged weaker than ChatGPT. For paper discovery, it returned 32 recent (2023–2025) peer-reviewed results for nanoparticle research in OPV devices and linked to major sources, though at least one item’s title didn’t match what opened. Its biggest weakness appeared in visual reasoning: it struggled to integrate complex SEM/AFM figure elements compared with ChatGPT. Editing abstracts and creating figure-based story arcs worked “good enough,” but not as well as leading alternatives.
How did DeepSeek R1 perform on literature review support for a specific topic (carbon nanotubes)?
What evidence suggested DeepSeek R1 could help with academic source discovery, and what went wrong?
Why did DeepSeek R1 struggle with figure-based interpretation?
Can DeepSeek R1 create a coherent narrative from multiple figures?
How useful was DeepSeek R1 for editing an abstract for publication?
Review Questions
- In what academic tasks did DeepSeek R1 perform best, and what were the concrete signs of weaker performance?
- What specific failure mode appeared during paper discovery (title vs opened content), and how should a researcher mitigate it?
- When interpreting complex microscopy figures, what integration step did DeepSeek R1 miss compared with ChatGPT?
Key Points
- 1
DeepSeek R1 is positioned as a free, open-source model that can support academic workflows, but it isn’t consistently publication-grade.
- 2
It produced a usable literature-review outline and a draft introduction for carbon nanotubes, yet it lacked the depth and structure polish of ChatGPT in the user’s tests.
- 3
For source discovery, it returned 32 recent peer-reviewed papers (2023–2025) for an OPV nanoparticle query with clickable links to major outlets.
- 4
Paper lists require verification: at least one result had a displayed title that didn’t match the paper opened.
- 5
DeepSeek R1 struggled more than ChatGPT at interpreting and integrating complex figures (SEM/AFM) into accurate conclusions.
- 6
It could reorder and synthesize multiple figures into a plausible research story, but integration quality lagged behind ChatGPT.
- 7
Abstract editing worked as suggestion-based polishing, but the user saw comparable capability in Claude and ChatGPT.