Ensemble Learning & Random Forest

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/26

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

27 Terms

1
New cards

Ensemble Learning

a group of predictors and combine them into a voting classifier

2
New cards

Voting classifier

hard voting and soft voting

3
New cards

Hard voting

use the predicted class from each model and return the majority vote

4
New cards

Soft voting

use the estimated probability produced by each classifier and combine probabilities and return the class with the highest probability

5
New cards

Bagging

bootstrap aggregating-create random sets of training data using the original training dataset

6
New cards

Bootstrap

resampling with replacement-randomly select a sample and return them to the original set. Subsequent samples may contain samples that have been picked before

7
New cards

Pasting

creating training sets with random sampling without replacement. There are no duplicates of the sample

8
New cards

Bias

error due to wrong assumptions in the model-high bias leads to underfitting

9
New cards

variance

error due to sensitivity to small fluctuations in the training set-high variance to overfitting

10
New cards

Out-of-bag evaluation

there will be replication of samples in bagging. 37% of the original samples won’t be included in the training of each model

11
New cards

Random patches and subspaces

bagging classifier can also work on random sub-sets of the features in the dataset. Can be useful when you have a large feature set in your data

12
New cards

Random patches

random samples and random features

13
New cards

Random subspace

keep all samples and select randm features

14
New cards

Random Forest

based on decision trees. Brings all the techniques into a single classifier. Uses random patches: bagging & random features

15
New cards

Random Forest Method

has all hyperparamets of decision trees. On each node a random subset of features is used to decide on the best split. With random feature sub-sets it introduces randomness across the generated trees. Greater tree diversity > trade higher bias for lower variance

16
New cards

Extra-Trees

Extremely Randomised Trees. Random feature subset + random threshold on each node. Select the combination with the best performance. Helps with reduced training time

17
New cards

Feature Importance

useful if you need to understand which features contribute the most of the model’s behaviour

18
New cards

Random Forests for Feature Importance

each node uses a feature to reduce impurity. Estimating the weighted average of how each feature contributes across all trained decision trees leads to estimated importance

19
New cards

Boosting

combine multiple weak learners into a strong learner. Train predictors sequentially: use previous predictors to improve the next

20
New cards

AdaBoost

adaptive boosting. Subsequent weak learners are adapted in favour of the samples that were misclassified by previous classifiers. The overall output is the weighted sum of all predictors. After training a weak classifier

21
New cards

Gradient Boosting

sequential training of predictors. Each new predictor is trained on the residual errors of the previous predictor

22
New cards

XGBoost

based on GBDT. a scale and highly accurate implementation of gradient-boosting. Regularisation. Handling sparse data. Weighted quantile sketch. Parallel learning

23
New cards

Regularisation

penalise complex models and prevent overfitting

24
New cards

Handling sparse data

missing values

25
New cards

Weighted quantile sketch

works with weighted data when splitting

26
New cards

Parallel learning

utilise multi-core CPUs/GPU to improve performance

27
New cards

Stacking

training another stronger/high-level predictor based on the outputs(not errors) of weaker/lower-level predictors