Day 27 of 133
Feature selection (filter / wrapper / embedded / SHAP)
Mutual information, L1, permutation, SHAP — fair attributions.
DSA · NeetCode Tries
- Implement Trie Prefix TreeDSA · Tries
Interview questions to prep
- Why a trie over a hash map — what queries does the trie make cheaper?
- What's the time/space trade-off vs storing all suffixes?
- How would you support deletion or wildcard matching efficiently?
- Design Add And Search Words Data StructureDSA · Tries
Interview questions to prep
- Why a trie over a hash map — what queries does the trie make cheaper?
- What's the time/space trade-off vs storing all suffixes?
- How would you support deletion or wildcard matching efficiently?
- Word Search IIDSA · Tries
Interview questions to prep
- Why a trie over a hash map — what queries does the trie make cheaper?
- What's the time/space trade-off vs storing all suffixes?
- How would you support deletion or wildcard matching efficiently?
ML · Feature selection
Interview questions to prep
- Compare filter, wrapper, and embedded feature selection.
- When would mutual information beat chi-squared as a filter, and vice versa?
Interview questions to prep
- Why is L1 effectively a feature selector?
- When would you trust tree-based importance over L1 selection?
Interview questions to prep
- How does SHAP compute fair attributions, and how does it relate to game theory?
- Where does SHAP mislead — what's a worked example of correlated features confusing SHAP?
References & further reading
- scikit-learn user guide ↗scikit-learn
- Eugene Yan — applied ML writing ↗Eugene Yan