Challenges in doing a PhD for Machine Learning and AI - Saturday with Dr. Sourish
Based on Enago Read (Previously Raxter.io)'s video on YouTube. If you like this content, support the original creators by watching, liking and subscribing to their content.
Research capability depends on experiment design, result interpretation, and problem framing—not just coding strength or familiarity with many ML algorithms.
Briefing
A machine-learning PhD doesn’t hinge on knowing algorithms or being strong at coding—it hinges on learning how to frame problems, design experiments, and interpret results across domains. Jay, a graduate student building interpretable AI for healthcare, credits his path to research not to “magic” in AI, but to early experiences that made the math-to-application connection feel concrete. After internships—especially one at NTU and later work connected to Philips Research Labs—he decided that deeper studies were worth the risk, even though he initially treated a PhD as a gamble compared with a safer master’s route.
His entry point into research started with a mismatch: early reading made AI sound like vague hype, but he couldn’t see how numbers became real applications. That changed during an NTU summer course on neural networks and through hands-on exposure to how academic work turns into product-grade R&D, supported by large datasets. He also describes a key mindset shift: research is a skill. Programming and mathematical priors matter, but they don’t automatically make someone capable of research work—especially the parts that involve starting from scratch, setting up experiments, and learning to navigate “rabbit holes” when results don’t land where expected.
When choosing a PhD topic, Jay emphasizes deliberate exploration rather than instant certainty. He frames the decision as a “tasting” process: sampling different areas through internships and small projects until one direction feels sustainable for years. His eventual focus—medical imaging and interpretable models—was shaped by experiences that connected AI to real-world constraints. At Philips, he built a module that radiologists reviewed in real time, where even a one-second delay could matter clinically. In his current work, he’s working toward using deep learning to find biomarkers in brain data for neurodegenerative conditions, including post-traumatic headache and Alzheimer’s, with the goal of helping clinicians understand disease-relevant regions.
Interdisciplinary research brings its own early-stage challenges, and Jay treats them as structural rather than personal. He describes a three-node ecosystem: (1) medical experts who know the clinical domain but not ML, (2) ML experts who know the coding and modeling but not medicine, and (3) bridge researchers—often the advisor—who translate between both sides. For literature review, he recommends narrowing the reading funnel: pick a small set of core journals and then skim aggressively for methods, data loading, preprocessing, architectures, and metrics. He argues that in medical AI, data handling can make or break model performance, so reading the “how” matters more than the “why” early on.
He also flags a practical problem: memory and linkage. Reading many papers can blur details, leaving researchers unsure which paper supports which idea. His workaround is to store notes with hyperlinks and concise reminders about architectures, hyperparameters, and results, so later comparisons become faster.
Overall, the central takeaway is that a successful ML/AI PhD is built through experimentation discipline, early topic clarity via exploration, and active translation across domains—supported by targeted reading habits and networking that expands the range of questions a researcher can ask.
Cornell Notes
Jay’s path into a machine-learning PhD for healthcare interpretable AI centers on a simple premise: research is a skill, not a byproduct of strong coding or algorithm knowledge. He built that skill through internships that made AI feel grounded—especially experiences where models had to work under real-world constraints like clinical timing. For choosing a thesis topic, he recommends “exploration first”: sample different directions via projects and internships until one area passes the six-year “marriage” test. Interdisciplinary work requires translation between medical experts, ML experts, and bridge figures (often the advisor), plus a targeted literature strategy that prioritizes methods, data loading, preprocessing, and metrics over slow reading of introductions. Finally, he warns that heavy paper reading can cause memory “mapping” failures and suggests keeping structured, linked notes to preserve connections between ideas.
Why doesn’t strong ML knowledge automatically make someone a good researcher, and what replaces it?
How did Jay decide on a PhD direction instead of committing too early?
What’s the practical challenge of interdisciplinary literature review, and how does he manage it?
What does Jay mean by “data loading changes the game” in medical AI?
How does Jay address the problem of forgetting details across many papers?
What role does networking and cross-pollination play in his research thinking?
Review Questions
- What specific research skills does Jay say matter beyond programming and algorithm knowledge?
- How does Jay’s “tasting” strategy for thesis topic selection work, and what evidence convinced him to commit to medical imaging?
- When reading interdisciplinary papers, what sections and details does Jay prioritize, and why?
Key Points
- 1
Research capability depends on experiment design, result interpretation, and problem framing—not just coding strength or familiarity with many ML algorithms.
- 2
Early internships can turn AI from abstract hype into concrete understanding by showing how academic work becomes product-grade R&D.
- 3
Choosing a thesis topic benefits from exploration first; Jay uses internships and small projects to find an area he can sustain for years.
- 4
Interdisciplinary work functions best when roles are clear: medical experts, ML experts, and bridge researchers who translate between domains.
- 5
Medical AI literature review should be targeted toward methods—especially data loading, preprocessing, architectures, and metrics—rather than slow reading of everything.
- 6
Heavy paper reading creates a “mapping” problem; structured notes with hyperlinks and concise technical reminders help preserve connections between ideas.
- 7
Networking and cross-domain conversations can create unexpected research directions by broadening the set of questions a researcher can ask.