Get AI summaries of any video or article — Sign up free
I Used DeepSeek R1 for Research – Here's What Happened! thumbnail

I Used DeepSeek R1 for Research – Here's What Happened!

Andy Stapleton·
5 min read

Based on Andy Stapleton's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

DeepSeek R1 is positioned as a free, open-source model that can support academic workflows, but it isn’t consistently publication-grade.

Briefing

DeepSeek R1 lands as a free, open-source AI model that can meaningfully assist academic work—especially for drafting structure and finding peer-reviewed papers—though it still lags behind top competitors on figure understanding and polished writing quality. In practical tests, it produced a usable literature-review outline for carbon nanotubes, generated a draft introduction section with citations, and returned 32 “recent” peer-reviewed papers (2023–2025) that the user could click through to publisher pages. That combination—starting academic writing and supplying clickable sources—makes it a credible research sidekick, even if it isn’t the strongest option for high-stakes, publication-ready output.

For literature review tasks, DeepSeek R1 performed best as a scaffold rather than a replacement for established writing tools. It generated a structured outline quickly, then followed up by drafting an “anxiety-free” introduction section. The draft was judged as a solid starting point (roughly a C+ / “pass”), with citations included—such as a Nature paper on mechanical strength—yet it didn’t match the depth and layout guidance produced by tools like ChatGPT in the user’s experience. The model’s “thinking” process was visible in real time, which the user found informative, but the resulting prose still required human judgment and refinement.

Where DeepSeek R1 showed clear utility was reference discovery. Using an internet-enabled search prompt for “recent peer-reviewed papers on nanoparticles for OPV devices,” it returned 32 results and linked out to major outlets and databases (including ScienceDirect, Nature, PubMed, and a Hong Kong Polytechnic University page). The user flagged one mismatch: the first item’s displayed title didn’t align with what opened on click, suggesting occasional citation/title confusion. Even so, the workflow—generate a list of recent, relevant papers and then click through to verify—was considered genuinely helpful, particularly if run periodically alongside other research tools.

DeepSeek R1’s weaker area was visual reasoning from complex figures. When given a demanding set of microscopy-related visuals (SEM and AFM images, plus a height profile) for non-planar vs planar electrode comparisons involving transparent electrodes, single carbon nanotubes, and silver nanowires, the model struggled to correctly “mesh” the components. It identified parts of the figure but missed how the elements connected, and the user repeatedly found ChatGPT better at extracting figure text and translating it into accurate, integrated conclusions. DeepSeek improved somewhat with additional prompting, but the gap remained.

On higher-level synthesis—turning multiple figures into a compelling research story—the model produced a plausible ordering and narrative arc, though again it was judged inferior to ChatGPT’s stronger ability to integrate figure relationships. For editing, it offered improvement suggestions to an abstract, but the overall verdict was that other mainstream tools (Claude and ChatGPT) can do similar work without the need to switch specifically to DeepSeek.

Overall, DeepSeek R1 earns its keep for free academic assistance: outlines, draft starters, and paper-finding workflows. The tradeoff is reliability and polish—especially for figure interpretation—where established competitors still lead for publication-grade writing and analysis.

Cornell Notes

DeepSeek R1 is a free, open-source model that can support academic research tasks, particularly by generating literature-review structure, drafting initial sections, and finding recent peer-reviewed papers with clickable links. In tests on carbon nanotubes, it produced a usable outline and an introduction draft with citations, but the prose depth and structure guidance were judged weaker than ChatGPT. For paper discovery, it returned 32 recent (2023–2025) peer-reviewed results for nanoparticle research in OPV devices and linked to major sources, though at least one item’s title didn’t match what opened. Its biggest weakness appeared in visual reasoning: it struggled to integrate complex SEM/AFM figure elements compared with ChatGPT. Editing abstracts and creating figure-based story arcs worked “good enough,” but not as well as leading alternatives.

How did DeepSeek R1 perform on literature review support for a specific topic (carbon nanotubes)?

It generated a literature-review structure for carbon nanotubes that was considered a solid starting point, though less detailed than tools like ChatGPT or thesis-writing assistants. It then drafted an introduction section under that structure. The resulting writing was judged “good enough” but not top-tier—roughly a C+—because it didn’t provide the same level of depth and layout guidance as stronger competitors.

What evidence suggested DeepSeek R1 could help with academic source discovery, and what went wrong?

With an internet-enabled search prompt for “recent peer-reviewed papers on nanoparticles for OPV devices,” it produced 32 results and linked out to publisher/database pages such as ScienceDirect, Nature, PubMed, and a Hong Kong Polytechnic University page. The user liked that the results were framed as recent (2023–2025) and could be clicked for verification. A key issue emerged when the first listed item’s displayed title didn’t match the paper that opened, indicating occasional mismatch or ambiguity in how titles are presented.

Why did DeepSeek R1 struggle with figure-based interpretation?

When given complex microscopy figures (SEM and AFM images plus a height profile) involving non-planar vs planar electrodes and mixed components (transparent electrodes, single carbon nanotubes, silver nanowires), it tended to identify likely comparisons but failed to correctly integrate how the figure elements related. The user felt ChatGPT better extracted figure text and connected the components into a coherent interpretation, while DeepSeek’s output remained fragmented even after additional prompting.

Can DeepSeek R1 create a coherent narrative from multiple figures?

With five uploaded figures (including schematics, tables, and graphs), it proposed an ordering and a story arc. The user felt it didn’t rely only on captions and seemed to infer relationships beyond labels. Still, the overall assessment was that ChatGPT handled the integration of figure relationships more effectively.

How useful was DeepSeek R1 for editing an abstract for publication?

It provided suggestions and key improvements after reviewing a draft abstract. The user considered the help “good enough” for polishing, but concluded that Claude and ChatGPT could likely perform similar editing tasks just as well, reducing the incentive to switch specifically for this purpose.

Review Questions

  1. In what academic tasks did DeepSeek R1 perform best, and what were the concrete signs of weaker performance?
  2. What specific failure mode appeared during paper discovery (title vs opened content), and how should a researcher mitigate it?
  3. When interpreting complex microscopy figures, what integration step did DeepSeek R1 miss compared with ChatGPT?

Key Points

  1. 1

    DeepSeek R1 is positioned as a free, open-source model that can support academic workflows, but it isn’t consistently publication-grade.

  2. 2

    It produced a usable literature-review outline and a draft introduction for carbon nanotubes, yet it lacked the depth and structure polish of ChatGPT in the user’s tests.

  3. 3

    For source discovery, it returned 32 recent peer-reviewed papers (2023–2025) for an OPV nanoparticle query with clickable links to major outlets.

  4. 4

    Paper lists require verification: at least one result had a displayed title that didn’t match the paper opened.

  5. 5

    DeepSeek R1 struggled more than ChatGPT at interpreting and integrating complex figures (SEM/AFM) into accurate conclusions.

  6. 6

    It could reorder and synthesize multiple figures into a plausible research story, but integration quality lagged behind ChatGPT.

  7. 7

    Abstract editing worked as suggestion-based polishing, but the user saw comparable capability in Claude and ChatGPT.

Highlights

DeepSeek R1 delivered 32 clickable peer-reviewed papers for a targeted OPV nanoparticle query, covering 2023–2025—useful for jump-starting literature review work.
The model sometimes mismatched what a listed title implied versus what opened, so verification remains essential even with “recent peer-reviewed” results.
Complex microscopy figure interpretation was its weak spot: it identified parts but often failed to connect mixed components into a coherent explanation.
For writing, it functioned best as a scaffold—outlines and draft starters—rather than a replacement for higher-polish tools.

Topics

Mentioned