Get AI summaries of any video or article — Sign up free
Can Google's new NotebookLM AI answer questions about my book? thumbnail

Can Google's new NotebookLM AI answer questions about my book?

Martin Adams·
4 min read

Based on Martin Adams's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

NotebookLM can ingest a document and answer questions grounded in the uploaded text, often using clickable citations to relevant sections.

Briefing

Google’s experimental NotebookLM is positioned as an AI note-taking system that can ingest a document and then answer questions, summarize concepts, and generate ideas grounded in the uploaded text. In a hands-on test using the book “Atomic Note Taking,” NotebookLM produced high-level descriptions of the Zettelkasten-inspired “Zle Caster” method, including the idea of breaking knowledge into stand-alone “atomic notes,” linking them into a growing web of understanding, and using that network to surface patterns, questions, and new insights. The most consequential takeaway is that NotebookLM can behave like a searchable, conversational layer over a source document—complete with citations that can be clicked through to relevant sections.

The workflow tested is straightforward: upload a book, then query it with prompts such as what the book is about, who it’s for, and how the method works. NotebookLM returned a concise “core concept” summary: atomic notes capture single ideas; they’re organized using content maps rather than rigid categories; and linking supports recall, updating, and deeper thinking over time. When asked for an example of an atomic note, it generated a short, one-idea entry (e.g., “Critical thinking is about making informed and sound decisions…”), and when asked how to cite sources for an atomic note, it described an audit trail—literature notes point back to fleeting notes, which ideally trace to the original source material. Clicking citations led to the relevant sections in the uploaded text, reinforcing that the system is drawing from the provided document rather than inventing freely.

Still, the test also exposed practical limitations. The uploader struggled to import certain file types (Markdown and plain text were problematic), suggesting NotebookLM may be better suited to formats like Google Docs or PDFs. More importantly, when challenged with questions that depended on non-text elements—such as a diagram or image on a specific page—NotebookLM could not answer because the underlying text available to it appeared to omit or strip images. That gap mattered when the system claimed the book lacked examples of linked atomic notes, a claim the tester disputed, attributing the mismatch to formatting and how content was extracted.

The system’s marketing output also showed its strengths and weaknesses: it generated a polished advert script for the book, but the tester noted it sometimes failed to retain or correctly reflect earlier answers. Overall, NotebookLM looks like a promising “question-answering over your documents” tool, but its usefulness depends heavily on document formatting and the presence of extractable text. With availability limited to the US and the project still experimental, the long-term impact hinges on whether Google can improve ingestion, citation fidelity, and multimodal understanding (especially diagrams and images) before adoption stalls or the effort is discontinued.

Cornell Notes

NotebookLM is an experimental AI note-taking tool that can ingest a document and answer questions using the uploaded text, often with clickable citations to relevant sections. In a test with “Atomic Note Taking,” it produced accurate-sounding summaries of the Zettelkasten-style “atomic notes” approach: write single-idea notes, link them into a “web” of understanding, and use that network to generate insights. It also described a citation/audit trail from atomic notes to literature notes and back to original sources. The main limitations surfaced around file import formats and missing support for non-text content like diagrams, which can cause incorrect conclusions when key examples are embedded as images.

What does NotebookLM do when given a book—does it summarize from memory or from the uploaded text?

NotebookLM’s answers were tied to the uploaded material, with citations that could be clicked through to sections where the ideas appeared. For example, when asked about the method’s core concept and who the book is for, the responses aligned with the book’s framing, and citation navigation pointed back to the relevant parts of the source.

How does the Zettelkasten-style “atomic note” method get translated into NotebookLM’s outputs?

NotebookLM described atomic notes as stand-alone units containing one key idea, written concisely. It also emphasized linking those notes to form a network that supports recall, updating, and pattern-finding. It further referenced organizing via “maps of content” to avoid rigid categories, encouraging divergent thinking and emergent understanding.

What did NotebookLM say about citing sources for atomic notes, and why is that important?

It outlined an audit trail: atomic notes should reference the literature notes they were based on; those literature notes should reference fleeting notes; and fleeting notes should ideally trace back to the original source material. That structure matters because it preserves provenance—making it easier to verify where an idea came from and to follow the lineage of knowledge.

Where did NotebookLM struggle during the test?

It struggled with ingestion and with non-text content. The tester couldn’t import certain file types (Markdown and plain text were not supported as expected), and NotebookLM appeared unable to interpret diagrams/images—leading to failures when questions depended on content that wasn’t present in the extracted text.

How did NotebookLM perform on a task outside pure summarization, like marketing copy?

It generated a full advert script for “Atomic Note Taking,” including a headline, body copy, and a call to action. The result demonstrated strong text-generation skills, but the tester also noted inconsistencies—such as the system not always reflecting earlier Q&A accurately.

Review Questions

  1. What evidence in the test suggests NotebookLM’s answers are grounded in the uploaded document rather than generated independently?
  2. Explain the “audit trail” NotebookLM gave for atomic-note citations. How does it connect atomic notes to original sources?
  3. Why would missing diagram/image understanding undermine NotebookLM’s usefulness for methods that rely on visual examples or page-based figures?

Key Points

  1. 1

    NotebookLM can ingest a document and answer questions grounded in the uploaded text, often using clickable citations to relevant sections.

  2. 2

    The system produced summaries of the atomic-notes approach: single-idea notes linked into a growing network for recall and insight.

  3. 3

    NotebookLM described an audit-trail citation model linking atomic notes to literature notes and back to fleeting notes and original sources.

  4. 4

    File ingestion may be limited: Markdown and plain text imports were problematic, while formats like Google Docs or PDFs were implied to work better.

  5. 5

    NotebookLM’s ability drops when key information is embedded as images/diagrams rather than extractable text, leading to incorrect conclusions.

  6. 6

    Generated marketing copy and other creative outputs can be strong, but consistency may vary across a multi-question workflow.

  7. 7

    Because NotebookLM is experimental and currently limited in availability, improvements to ingestion and multimodal understanding will likely determine whether it gains traction.

Highlights

NotebookLM answered method questions with citations that could be clicked through to the corresponding sections in the uploaded book.
The citation workflow it described—atomic notes → literature notes → fleeting notes → original sources—aims to preserve provenance for every idea.
When asked about content that depended on an image/diagram, NotebookLM couldn’t respond because the provided text appeared to omit non-text elements.
NotebookLM generated a complete advert script for “Atomic Note Taking,” showing strong text-generation beyond Q&A.