Day 66 of 133
Cost & scaling (autoscaling, spot, distillation) + DSA 2-D DP
Unit-cost-per-prediction; KEDA scaling; spot training without losing work.
DSA · NeetCode 2-D DP
- Unique PathsDSA · 2-D DP
Interview questions to prep
- Closed-form via combinatorics: C(m+n-2, m-1). Why is it equivalent to the DP answer?
- How do obstacles change the recurrence (unique-paths-ii)?
- Longest Common SubsequenceDSA · 2-D DP
Interview questions to prep
- State the DP. How does it relate to edit distance and to longest-common-substring?
- Space-optimize from O(m·n) to O(min(m, n)).
MLOps · Cost & scaling
Interview questions to prep
- How would you model the unit cost of a prediction in production?
- What levers reduce inference cost (batching, quantization, caching, distillation)?
Interview questions to prep
- Compare CPU-based HPA vs queue-based KEDA scaling for ML inference.
- Why does GPU-pinned inference often defeat HPA, and how do you actually scale GPU pods?
Interview questions to prep
- How would you train safely on spot instances (checkpointing, retries)?
- When does spot training become NET more expensive than on-demand — what's the breakeven?
References & further reading
- Chip Huyen — ML Systems Design ↗Chip Huyen
- Full Stack Deep Learning ↗FSDL
- Eugene Yan — applied ML writing ↗Eugene Yan