Discriminant Analysis

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/33

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

34 Terms

1
New cards

LDA assumptions

normality and same covariance matrix

2
New cards

QDA assumptions

normal, different variance/covariance matrix

3
New cards

Gaussian Naive Bayes assumptions

normal, different variance , no covariance

4
New cards

how to check assumptions graphically

qqplot

side by side boxplot

covariance ellipse

perspective and contour plot

scatter plot

5
New cards

how to check assumptions with tests

boxm

mvn

kolomogrov- smirnov

shapiro-wilk

correlation test

6
New cards

LDA bias

could have high bias when covariance matrix is different

7
New cards

naive bias

reduces variance but could have large bias

8
New cards

naive bias posterior probability

sum of the functions of the input features . GAM generalized additive model you essentially add together the effects of each variable

9
New cards

log odds posterior probability LDA

linear in X

10
New cards

log odds posterior probability QDA

quadratic

11
New cards

log odds posterior probability naive bias

Generalized additive model

12
New cards

log odds posterior probability logistic

linear of x

13
New cards

which two have decision boundaries

LDA and multinomial regression (logistic)

14
New cards

When all covariance matrices are the same (quadratic term is 0) what is LDA a special case of

QDA

15
New cards

Special case of naive bias

LDA and multinomial

16
New cards

which for continuous predictors

LDA and QDA

17
New cards

categorical predictors

Multinomial Regression, Naive Bayes

18
New cards

high value for LD1 indicates

most group seperation happens along that axis

19
New cards

Do you need to specify a k value for Hierarchical clustering

no

20
New cards

why called hierarchal clustering ?

clusters obtained by cutting the dendrogram at given height are nested within the clusters obtained by cutting any higher.

21
New cards

linkage

dissimilarity between clusters with multiple observations

22
New cards

four type of linkage

average, completed, single and centroid

23
New cards

preferred linkage

average and completed

24
New cards

what is hierarchal clustering based on

the distance matrix typically for numerical data

25
New cards

large number of variables

k-methods

26
New cards

structure k-methods vs hierarchal

unstructured hierarchal is more interpretable and informative

27
New cards

easier to determine number of clusters in

hierarchal clustering dendrogram

28
New cards

distinguishes based on prior beliefs

hierarchal clustering may be used to know the number of clusters

29
New cards

Specific number of clusters but the group they belong to is unknown

k-methods

30
New cards

is clustering robust

not robust to perturbations to the data

31
New cards

complete linkage

looks at distance between points in two clusters and pick the largest one.

32
New cards

single linkage

looks at the distance between points in two clusters and picks the smallest one.

33
New cards

average linkage

calculates the average distance between all pairs of points in two cluster

34
New cards

centroid linkage

compares the central points of two clusters, ignoring how spread out the individual points are