Get AI summaries of any video or article — Sign up free
Medical Trainees: How to Learn FASTER as a Busy Doctor thumbnail

Medical Trainees: How to Learn FASTER as a Busy Doctor

Justin Sung·
5 min read

Based on Justin Sung's video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.

TL;DR

Expect a mismatch between tutorial/lecture performance and exam performance when the exam tests deeper, more granular knowledge than trainees can currently self-assess.

Briefing

Fellowship and high-stakes exam prep often fails to match real-world competence: trainees can perform well in tutorials and lectures yet still miss the deeper, more granular level that exams now demand. The core problem isn’t simply missing facts—it’s uncertainty about what’s missing, because the “test level” sits above what someone can reliably probe with self-assessment. That mismatch creates anxiety: even when performance looks strong, there’s no clear way to confirm whether knowledge is truly at the required depth.

A major limitation comes from trying to use ChatGPT to generate exam-style questions at the right nuance. Complex prompts, non-leading instructions, textbook-grounded examples, and iterative question sets still often land “close” without reliably eliciting the exact front-end thinking that examiners reward. The underlying issue is framed as an LLM constraint: training data and learned patterns may not include enough of the specific questioning style needed for niche, current fellowship formats—so the model can’t fully overcome its “algorithmic imprint.”

Instead, the coaching centers on diagnosing the knowledge gap itself, then choosing tests that match its type. Knowledge gaps are categorized two ways: higher-order vs lower-order (whether the problem is integrating concepts into a big-picture synthesis, or missing concrete details), and declarative vs procedural (whether the weakness is recalling/knowing vs executing like a skill). In the trainees’ situation, the likely pattern is lower-order declarative gaps: the ability to connect concepts is present, but specific numbers, statistics, names, or doses aren’t reliably retrievable under exam conditions.

Because high-volume exams make “knowing everything” unrealistic, the strategy shifts from chasing perfect coverage to finding and fixing transferable weaknesses. Two calibration methods are proposed. External calibration uses seniors or exam-passing peers to quiz and identify where knowledge breaks at the required standard—time-consuming but crucial for deep knowledge. Internal calibration relies on confidence signals during practice: when a complex question exposes uncertainty, that hesitation points to a gap worth drilling down. Combining both helps trainees learn not only what they’re missing, but what kinds of gaps they repeatedly miss.

The session also addresses how to learn faster when time is fragmented. Meaningful learning requires entering “flow,” but context generation can take 30–40 minutes for some people, making short study windows inefficient. The fix is to reduce context-generation time by using “layering” microlearning: break consolidation into small steps—identify main topics, pick a few, define them, connect them, and write down questions to answer later—so each short session prepares the next. Research on microlearning is cited to support that very short sessions (under 10 minutes) can help the brain operate more effectively, keep curiosity loops open, and reduce fatigue.

Finally, low-effort active learning is recommended as a practical companion: flash cards and other retrieval-based prompts are easy to deploy throughout the day, including on phones, while still forcing active recall. The overall message is pragmatic: expect gaps, hunt for biases early, use calibration to locate what matters, and structure learning into small, repeatable cycles that fit the reality of busy clinical schedules.

Cornell Notes

Fellowship exam performance can lag behind day-to-day competence because the exam tests deeper, more granular knowledge than trainees can reliably self-check. Attempts to generate exam-like questions with ChatGPT may fall short due to limited access to the specific questioning style and nuance required. The remedy starts with diagnosing the knowledge gap: higher-order vs lower-order, and declarative vs procedural. Then trainees calibrate—externally by quizzing seniors/exam passers, internally by using confidence/uncertainty during practice to locate weak points. With limited time, “layering” microlearning and low-effort active recall (e.g., flash cards) reduce context-generation overhead and help knowledge stick.

Why can someone do well in tutorials and lectures yet still underperform on fellowship exams?

The exam standard sits above what the trainee can currently test. Tutorials and lectures often confirm broad understanding, while exams demand deeper, more minute details and the ability to synthesize them under pressure. That creates a “gap you can’t see”: the trainee may not know what’s missing because self-testing rarely reaches the exam’s depth.

What’s the key limitation of using ChatGPT to create exam-style questions for niche fellowship formats?

Even with complex, non-leading prompts, examples from past exams, and textbook-grounded instructions, the model may still not elicit the exact front-end thinking examiners reward. The session frames this as an LLM training-data and pattern-matching limitation: it may have insufficient exposure to the specific questioning style and nuance of that particular exam format.

How should knowledge gaps be classified to choose the right kind of practice?

Gaps are categorized as higher-order vs lower-order and declarative vs procedural. Higher-order gaps involve integration and big-picture synthesis across factors; lower-order gaps involve missing concrete details (like specific statistics or names) even if connections are understood. Declarative gaps are about knowing/recalling; procedural gaps are about executing a skill accurately and consistently (e.g., surgical technique).

How do external and internal calibration work in practice?

External calibration uses someone who has the required standard—typically a senior or exam passer—to quiz the trainee and identify where knowledge breaks. Internal calibration uses the trainee’s own confidence during complex practice: uncertainty signals likely gaps. The approach can be iterative—drill deeper into the uncertain area until the specific weak subcomponents are identified.

What does “layering” microlearning do to make fragmented study time more effective?

Instead of trying to consolidate an entire topic in one sitting, layering splits consolidation into small steps that set up the next step: list what to learn, define a couple items, connect them, then write down questions to resolve later. This lowers context-switching overhead and keeps learning moving even when sessions are short.

How does low-effort active learning help busy clinicians?

It keeps retrieval practice easy to deploy throughout the day. Flash cards are the example: the act of recalling is cognitively demanding, but the workflow is low friction. The session also suggests other retrieval formats (self-generated test questions, fill-in-the-blank mind map prompts, or quick “why is this relevant?” checks) that can be done on a phone without desk time.

Review Questions

  1. When would a knowledge gap be higher-order rather than lower-order, and what would that change about the way you practice?
  2. Describe a concrete internal calibration method you could use during exam-style questions to locate the specific weak subtopic.
  3. How would you design a 10-minute layering session for a foundational science topic (e.g., pharmacology) so it meaningfully prepares the next session?

Key Points

  1. 1

    Expect a mismatch between tutorial/lecture performance and exam performance when the exam tests deeper, more granular knowledge than trainees can currently self-assess.

  2. 2

    Use knowledge-gap diagnosis (higher-order vs lower-order; declarative vs procedural) to choose practice that targets the right failure mode.

  3. 3

    Treat ChatGPT-generated questions as imperfect for niche exam nuance; use calibration methods to validate what actually matches the exam standard.

  4. 4

    Apply external calibration by quizzing seniors/exam passers, and apply internal calibration by using uncertainty/confidence signals to drill into likely gaps.

  5. 5

    In ultra-high-volume exams, aim to find and fix transferable biases and gaps early rather than chasing perfect coverage.

  6. 6

    Reduce wasted time from context generation by using layering microlearning: small steps that prepare the next step and keep curiosity loops active.

  7. 7

    Use low-effort active recall (flash cards and similar retrieval prompts) to practice consistently during fragmented clinical schedules.

Highlights

The anxiety in fellowship prep often comes from not knowing what’s missing—because self-testing rarely reaches the exam’s depth.
ChatGPT can generate “close” questions, but it may still fail to elicit the exact nuanced thinking required by specific fellowship formats.
External calibration (senior/exam passer quizzes) and internal calibration (confidence-based uncertainty) together pinpoint gaps more reliably than guessing.
Layering microlearning turns short, fragmented time into a sequence of preparation steps, lowering context-switching costs.
Low-effort active learning keeps retrieval practice feasible throughout a busy day without sacrificing cognitive engagement.

Topics

Mentioned