1/5
Looks like no tags are added yet.
Name | Mastery | Learn | Test | Matching | Spaced |
---|
No study sessions yet.
T-test Assumptions
random sampling
normal distribution - CLT applies, but heavy tailed data bad!
independent observations - think cluster and serial effect
homogeneity of variance - particularly bad when sample sizes differ
Shapiro-Wilk Normality Test
null hypothesis: data comes from normally distributed population
W-test statistic + p-value to interpret results
note: test is sensitive to sample size - large samples can produce type II error
Testing Homogeneity of Variance
Hartley’s test: largest sample variance/smallest sample variance BUT requires equal sample size, normality, NON ROBUST
Levene’s test: MSbetween/MSwithin, MODERATELY ROBUST
Brown-Forsythe test: uses median, resistant to outliers, MOST ROBUST
Durbin-Watson Test of Independence
tests non-independent observations (autocorrelation)
DW<2 = Positive autocorrelation (residuals have same sign)
DW>2 = Negative autocorrelation (residuals alternate sign)
Outliers
z-score rule: 2SD = 5% are outliers, 3SD = <1% are outliers NON ROBUST
boxplot rule: ROBUST
how outliers effect inference with t:
disturb point estimates
contribute to skew/heavy tails
can affect nonrobust tests of assumptions
T-inferences are inaccurate when…
non-normality is heavy-tailed
data come from nonrandom samples
heterogeneous variances + small sample sizes
nonindependence (serial, cluster, protocol)