Lecture 9: Decision Trees and Random Forest 2

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/9

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

10 Terms

1
New cards

Q: How do you pick a split in a decision tree?

Choose the variable and cutoff that best separates the data, usually using Gini impurity for classification.

2
New cards

Q: When do you stop growing a decision tree?

When nodes are pure (one label) or have too few data points.

3
New cards

Q: What is bootstrap aggregation (bagging)?

Resampling data with replacement, training models on each sample, and aggregating their results to reduce overfitting.

4
New cards

Q: Why does a fully grown single decision tree overfit?

It perfectly memorizes the training data, losing generalization.

5
New cards

Q: How do random forests improve over single decision trees?

They grow many trees on bootstrapped samples and average or vote across them to reduce variance.

6
New cards

Q: What extra randomness is added in random forests?

Each split considers a random subset of predictor variables instead of all variables.

7
New cards

Q: In Machine Learning, what are the three key steps?

Train on past data, predict on new data, evaluate performance.

8
New cards

Q: Why use ensemble models?

Because no single model is perfect; combining models can improve accuracy.

9
New cards

Q: What is stacking in ensemble modeling?

Using outputs from different models as new features for a final model.

10
New cards

Q: How does linear weighted stacking work?

  1. Split training data into two parts (train1 and train2).
    1. Train several models (e.g., Random Forest, GLM, GBM, SVM) on train1.
    2. Score each model on train2, using the scores as new features.
    3. Combine these new features with original features on train2.
    4. Train a final GLM model (including interaction terms) on this combined data.
    5. Apply the stacked pipeline to test data