Machine Learning, Stats, Math

0.0(0)
studied byStudied by 0 people
0.0(0)
full-widthCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/58

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

59 Terms

1
New cards
What is supervised learning?
Learning a mapping from inputs to outputs using labeled examples.
2
New cards
What is unsupervised learning?

Discovering structure in unlabeled data (clustering, dimesionality reduction)

3
New cards
Define overfitting and name two prevention methods.

Model fits noise; prevent with regularization, cross-validation, early stopping, or more data

4
New cards
Difference between classification and regression?
Classification predicts discrete labels; regression predicts continuous values.
5
New cards
How does k-NN predict class?
By majority vote among the k nearest neighbors in feature space.
6
New cards
Why scale features?

To ensure features are comparable, important for kNN, SVM and gradient descent

7
New cards
What is cross-validation?
Method to estimate model generalization by splitting data into training and validation folds.
8
New cards
What is the ROC curve?
Plot of True Positive Rate vs False Positive Rate; AUC summarizes discrimination ability.
9
New cards
What does regularization do?

Adds penalty to model weights to reduce overfitting; L1 (Lasso) and L2 (Ridge)

10
New cards
Define bias and variance.
Bias = error from incorrect assumptions; Variance = error from model sensitivity to data.
11
New cards
What is PCA?
Linear dimensionality reduction projecting data onto principal components with max variance.
12
New cards
What is SMOTE?
Oversampling technique that synthesizes new minority class samples to fix class imbalance.
13
New cards
What is the softmax function used for?
To convert logits to normalized class probabilities.
14
New cards
What is dropout?
Randomly setting neurons to zero during training to prevent overfitting.
15
New cards
What is early stopping?
Halting training when validation error stops decreasing to prevent overfitting.
16
New cards
What is ensemble learning?

Combining multiple models to improve performance, e.g. bagging or boosting

17
New cards
What is XGBoost known for?

Regularized gradient boosting, effective and efficient for tabular data

18
New cards
What is SHAP?
Shapley value-based feature attribution for interpretability.
19
New cards
What is feature drift vs concept drift?
Feature drift: change in X distribution; Concept drift: change in P(y|X).
20
New cards
Why calibrate models?
To make predicted probabilities match actual outcome frequencies (important in risk).
21
New cards
How does Random Forest reduce overfitting?
By averaging many decision trees trained on bootstrapped data.
22
New cards
Difference between bagging and boosting?
Bagging trains models in parallel to reduce variance; boosting trains sequentially to reduce bias.
23
New cards
What is t-SNE mainly used for?
Visualizing high-dimensional data in 2D while preserving local structure.
24
New cards
What is transfer learning?
Using pretrained models on new tasks to save time and data.
25
New cards

What problem does batch normalization solve?

Stabilizes layer activations, speeds convergence, and allows for higher learning rates

26
New cards
Name local and global model explanation methods.
Local: LIME; Global: SHAP or feature importance.
27
New cards
What is adversarial training?
Training with perturbed inputs to improve robustness.
28
New cards
What is CatBoost notable for?
Handles categorical variables natively without one-hot encoding.
29
New cards
What is out-of-bag (OOB) error?
Validation metric using unused samples per tree in a Random Forest.
30
New cards
What is model serialization?

Saving trained models for deployment, e.g. using pickle or joblib

31
New cards
What is quantile regression?
Estimates conditional quantiles (like median or 95th percentile).
32
New cards
What is feature hashing?
Hashes categorical features into fixed-length numeric vectors.
33
New cards

What is cross-validation leakage?

Information from validation leaks into training via preprocessing before splitting.

34
New cards
What is bootstrapping used for?
Resampling with replacement to estimate uncertainty and confidence intervals.
35
New cards
When to use precision vs recall?
Precision when false positives are costly; recall when false negatives are costly.
36
New cards
Difference between calibration and discrimination?
Calibration = probability accuracy; discrimination = ranking quality (AUC).
37
New cards

What are Bayesian neural networks used for?

Quantifying uncertainty in predictions, useful for risk-sensitive tasks.

38
New cards
What is knowledge distillation?
Training a smaller model to mimic the outputs of a larger model.
39
New cards

Define concept drift detection best practice.

Monitor feature distributions and predictive performance; retrain when drift detected.

40
New cards
What is active learning?

Model selects most informative samples to query for labels, reducing labelling cost

41
New cards
Formula for sample variance?
s² = (1/(n-1)) sigma (xi - xbar)².
42
New cards
Bayes theorem?
P(A|B) = [P(B|A) * P(A)] / P(B).
43
New cards
Central Limit Theorem (CLT)?
Sample mean of i.i.d. variables approximates a Normal distribution as n approaches infinity
44
New cards
When to use t-test vs z-test?
Use t-test when σ unknown or n small; z-test when σ known or n large.
45
New cards
Use-case for Poisson distribution?

Modeling counts of rare events per time unit (e.g. fraud events)

46
New cards
Law of total probability?
P(B) = sigmai P(B|Ai)P(Ai) for a partition of Ai.
47
New cards
Meaning of p-value?
Probability of observing data at least as extreme assuming H0 is true.
48
New cards
MLE vs Method of Moments?
MLE maximizes likelihood; MoM matches sample and population moments.
49
New cards
When to use Poisson vs Binomial?
Poisson for rare events; Binomial for fixed number of independent trials.
50
New cards
Complexity of BFS?
O(V + E).
51
New cards
Binomial coefficient formula?

C(n, k) = n! / [k!(n-k)!]

52
New cards
Use-case for Dijkstra?
Finding shortest path with non-negative edge weights.
53
New cards
What is dynamic programming?
Breaking problems into overlapping subproblems and caching results.
54
New cards
Master theorem concept?
Compare f(n) vs n^(log_b a) to estimate asymptotic complexity.
55
New cards
Generating function use?
Solve recurrences and count combinatorial objects.
56
New cards
What is modular arithmetic used for?
Cryptography and hashing operations.
57
New cards
Definition of NP-hard?

As hard as hardest problems in NP, not necessarily solvable in polynomial time

58
New cards
Recurrence relation for Fibonacci?

Fn = Fn-1 + Fn-2, F0 = 0, F1 = 1

59
New cards
Example of greedy optimal algorithm?
Kruskals algorithm for Minimum Spanning Tree.