Day 94 of 133
Hardware: GPUs, TPUs, NVLink, IB + DSA review
A100 vs H100 vs B200; HBM bandwidth; interconnect bottlenecks.
DSA · NeetCode Backtracking
- Combination Sum IIDSA · Backtracking
Interview questions to prep
- Walk through your pruning strategy — what subtrees do you skip and why is it safe?
- Where does memoization apply? Could this be a DP problem in disguise?
- What's the worst-case time complexity, and what's the depth of the recursion stack?
Infra · GPUs, TPUs, accelerators
Interview questions to prep
- Compare A100, H100, and B200 — what changed each generation?
- Why does HBM matter more than FLOPs for many ML workloads?
Interview questions to prep
- Compare GPUs vs TPUs for training — when does each win?
- What's the cost of porting a PyTorch model to TPU via XLA, and where does it break?
Interview questions to prep
- Why does interconnect (NVLink, IB) often bottleneck distributed training before compute does?
- How would you diagnose whether your distributed training is bottlenecked by compute, memory, or network?
References & further reading