1/58
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
|---|
No study sessions yet.
Discovering structure in unlabeled data (clustering, dimesionality reduction)
Model fits noise; prevent with regularization, cross-validation, early stopping, or more data
To ensure features are comparable, important for kNN, SVM and gradient descent
Adds penalty to model weights to reduce overfitting; L1 (Lasso) and L2 (Ridge)
Combining multiple models to improve performance, e.g. bagging or boosting
Regularized gradient boosting, effective and efficient for tabular data
What problem does batch normalization solve?
Stabilizes layer activations, speeds convergence, and allows for higher learning rates
Saving trained models for deployment, e.g. using pickle or joblib
What is cross-validation leakage?
Information from validation leaks into training via preprocessing before splitting.
What are Bayesian neural networks used for?
Quantifying uncertainty in predictions, useful for risk-sensitive tasks.
Define concept drift detection best practice.
Monitor feature distributions and predictive performance; retrain when drift detected.
Model selects most informative samples to query for labels, reducing labelling cost
Modeling counts of rare events per time unit (e.g. fraud events)
C(n, k) = n! / [k!(n-k)!]
As hard as hardest problems in NP, not necessarily solvable in polynomial time
Fn = Fn-1 + Fn-2, F0 = 0, F1 = 1