CPSC 330 Section 2
Home
Lectures
Lecture 13: Feature importances
Overview
Lecture 1: Course introduction
Lecture 2: Terminology, baselines, decision Trees
Lecture 3: ML fundamentals
Lecture 4:
\(k\)
-nearest neighbours and SVM RBFs
Lecture 5: Preprocessing and sklearn pipelines
Lecture 6:
sklearn
column transformer and text fearutres
Lecture 7: Linear models
Lecture 8: Hyperparameter Optimization
Lecture 9: Classification metrics
Lecture 10: Regression metrics
Lecture 11: Midterm review
Lecture 12: Ensembles
Lecture 13: Feature importances
Lecture 14: Feature engineering and selection
Lecture 15: K Means
Lecture 16: DBSCAN and hierarchical
Lecture 17: Recommender systems
Lecture 18: Natural Language Processing
Lecture 19: Introduction to computer vision
Lecture 20: Time series
Lecture 21: Survival analysis
Lecture 22: Communication
Lecture 24: Deployment and conclusion
Lecture 13: Feature importances
interpreting coefficients of linear models, model transparency, feature_importances, SHAP
Slides
View slides in full screen
Outline
interpreting coefficients of linear models
model transparency
feature_importances
SHAP
Lecture 12: Ensembles
Lecture 14: Feature engineering and selection