PSYC 359: Advanced Research Methods

0.0(0)
studied byStudied by 1 person
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/85

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

86 Terms

1
New cards

statistical inference

specified null that is tested with set significance criterion - based on assumption null is true

2
New cards

4 components of inference

  1. sample size (n)

  2. significance criterion (α)

  3. Statistical power (1 - β)

  4. Effect size (δ or p)

    If you have ¾ you can determine the 4th

3
New cards

type I error (α)

rejecting a null hypothesis when it is actually true - when null is true, for alpha = 0.05 there is a 5% chance of error

4
New cards

type II error (β)

retaining a null when it is actually true - if null is false, β x 100% of the time we make a Type II error

5
New cards

statistical power

probability of correctly rejecting a false null (1 - β)

6
New cards

type S error

rejecting a false null, but getting the direction of the relationship wrong (only for one-tailed tests)

7
New cards

type M error

error in estimation of magnitude of the effect (getting effect size wrong)

8
New cards

type IV error

an incorrect interpretation of a correctly rejected null

9
New cards
<p>statistical power and inferential errors graphed</p>

statistical power and inferential errors graphed

  • type I error = light red region on H0 distribution

  • type II error = blue part of HA distribution in region of retention

  • type s error = dark red region in HA distribution below -1.96

10
New cards

if i retain the null, what is the probability that i have made a Type I error?

0% - can’t make a type I error

11
New cards

if i reject the null, what is the probability that i made a Type I error?

0 or 1 - either didn’t or did make a Type I error

12
New cards

if i reject the null, what is the probability that I made a Type II error?

0% - can’t make a type II error

13
New cards

if the null is false in reality, what proportion of 1000 researchers examining this hypotheses with make a Type I error?

a (usually 5%)

14
New cards

sampling distribution

distribution of a statistic calculated from all possible distinct random samples of of same size n drawn from population

15
New cards

expected value

mean of sampling distribution

16
New cards

standard error

standard deviation of sampling distribution

17
New cards

central limit theorem (CLT)

know mean, standard deviation, form of sample distribution of sample means for the sample distribution of the sample mean when when it is not possible to take all possible samples

18
New cards

mean of sampling distribution of sample means

E(x̄) = μ (equal to population mean)

19
New cards

standard deviation of sample means

standard error of the mean (σ ̄ = σ/√n)

  • n = sample size for mean

  • σ = standard deviation for population

20
New cards

direct route hypothesis testing

compute parameters of interest on entire population using descriptive statistics (never know parameters of entire population)

21
New cards

indirect route hypothesis testing

compute descriptive statistics on random sample from population and infer values of population parameters and draw conclusions

22
New cards

scientific hypothesis

relationship/effect researcher expects to find - restated into null and alt hypotheses for indirect testing (mutually exclusive and exhaustive together)

23
New cards

null hypothesis (H0)

involves a prediction about exact value of population parameter - assumed to be true

  • always contains an equity symbol , , or =

  • H0 :μ = 120, H0: μ 100, H0 : σ = 11, H0 : μ1 μ2 = 0

24
New cards

alternative hypothesis (HA)

accepted if H0 is rejected

  • always contains and inequality symbol: <, >, or ̸=

  • HA :μ < 120, HA : σ ̸= 11

25
New cards

8 steps of hypothesis testing

  1. state problem and nature of data (nature = nominal, ordinal, interval, ratio, censored, truncated)

  2. specify and state hypothesis to be tested (scientific, null, alt)

  3. draw random sample of size n from population (randomly selected and representative)

  4. determine sampling distribution of statistic being calculated (mean, SE, form of null distribution)

  5. specify level of significance

  6. determine regions of retention and rejection

  7. calculate value of statistic of interest

  8. test hypothesis and reject the null if it falls in region of rejection or retain the null if it falls in region of retention

26
New cards

p-value

probability of observed data or more extreme under the null sampling distribution

  • doesn’t tell us the probability of the null (effect size)

27
New cards
<p>test statistic</p>

test statistic

general formula simply number of standard errors from the null sampling distribution

  • based on theoretical sampling distribution of null

  • use common reference distributions (z, t, etc.) to determine probabilities of the data

  • θ is the population parameter of interest

  • estimated by statistic θˆ

  • E(θˆ| H0) is the expected mean of null sampling distribution for the statistic.

  • SEθ | H0 is standard error for the statistic when null is true.

28
New cards

central distributions

distribution of test statistic under the null of no effect

  • λ = 0 - all you require are central distributions

29
New cards
<p>non-centrality parameter (<span style="font-family: CMMI10">λ)</span></p>

non-centrality parameter (λ)

measure of the degree to which a null hypothesis is false

  • forms confidence intervals for effect sizes

  • function of population effect size and sample size

30
New cards

probability function

used to obtain p-value - probability of a specific value and lower or higher

  • given the quantile, calculate the probability

31
New cards

quantile function

determine test statistic that has specified probability of occurring

  • given the probability, calculate the quantile

  • for two-tailed tests (normal and t-distributions), divide alpha by 2

32
New cards

critical values

quantiles on Ho distribution

33
New cards
<p>normal distribution</p>

normal distribution

  • population mean μ

  • variance σ2

34
New cards

Gauss’s 3 assumptions for normal distributions

  1. small errors more likely than large errors

  2. for any error ε, likelihood of ε is same as that of ε

  3. with several measurements containing independent errors, average has highest likelihood of producing observed data (max likelihood estimate of underlying parameter)

35
New cards

probability calculation for normal distribution

pnorm( )

  • calculate 2-tailed tests: 2 * pnorm( )

36
New cards

quantile calculation for normal distribution

qnorm( )

37
New cards
<p><span style="font-family: CMMI12"> </span>chi-square <span style="font-family: CMMI12">χ</span><span style="font-family: CMR8">2</span><span style="font-family: CMR12">(</span><span style="font-family: CMMI12">df</span><span style="font-family: CMR12">)- </span>distribution</p>

chi-square χ2(df)- distribution

sample distribution of sample variation - Function of squared normal distribution

  • each df represents another independent observation from normal distribution

  • sample variance is a function of since variance is a squared value when original observations are normally distributed

  • mean of sampling distribution is its df

38
New cards

probabilities for chi-squared distributions

pchisq(critical value, df, lower.tail)

  • all p-values inherently two-tailed (don’t divide a)

39
New cards

quantiles for chi-squared distributions

qchisq(alpha, df, lower.tail)

  • not divided by 2

40
New cards

degrees of freedom

finite sample size correction - account for variance in sample

41
New cards

assumptions of chi-squared distributions

  • residuals are independent: we have correct df - violating has huge impact on type I error

  • normally distributed: if violated, we don’t know distribution of sample variance

  • estimation of sample variation gets better as sample size increases

42
New cards
<p>t-distribution (student’s t-test)</p>

t-distribution (student’s t-test)

ratio of normal distribution and square root of chi-square distribution

  • statistic in denominator instead of a parameter

  • standard deviation is a function of sqrt(χ2(df))

43
New cards

probabilities for t-distribution

pt(critical value, df, lower.tail)

  • 2x when wanting to find both tails

44
New cards

quantiles for t-distributions

qt(alpha, df, lower.tail)

  • divide alpha by 2

45
New cards
<p>F-distribution</p>

F-distribution

ratio of 2 variance estimates, ratio of χ2 (df ) variables

  • has 2 df associated with it

  • commonly used in context of ANOVA

  • always two-tailed

  • when df1 = 1, it is the square of t-distribution

  • when df2 = infinity, it is the chi-squared distribution/df

46
New cards

probabilities for F-distribution

pf(critical value, df1, df2, lower.tail)

47
New cards

quantiles for F-distribution

qf(a, df1, df2, lower.tail)

  • a not divided by 2

48
New cards

Which statistic has a lower p-value z = 2.0 t(16) = 2.0?

z = 2.0

49
New cards

which test statistics have the same p-values? z = 2.0, t(16) = 2.0, χ2(1) = 4.0 F(1,16) = 4.0?

t and F, z and χ2

50
New cards

why does pnorm(2.00, lower.tail=FALSE) not give us the p-value 4.0 but pchisq(4.0, df=1, lower.tail=FALSE) does?

is inherently two-tailed but normal distributions aren’t, so to get that value, you would have to multiply it by 2

51
New cards

for a = .05 under normal distribution, is critical value of 1.96 (a) a p-value, (b) a quantile, (c ) a parameter, or (d) a test-statistic?

quantile

52
New cards

what do df for a t-test indicate? does this tell us about the amount of info in the means or variance?

Inform us about amount of information (and thus precision) associated with estimate of variance that is being used incomputing the t-test

53
New cards

which statistic’s sampling distribution is the χ2(df)-distribution related to?

sampling distribution of the sample variance0

54
New cards

scatterplot

graph where each individual data is represented by a plot on the graph - use jitter and loess curve to improve its utility

55
New cards

jitter

small amounts of random noise in x or y-axis to avoid overplotting

56
New cards

loess curve

nonparametric smoother (doesn’t presume any relation between the variables)

  • control span (amount of smoothing) and fitting model (relationship you want to use to estimate relationship within each span)

  • simply estimates a relationship based on observed data

  • more properly approximated observed data

57
New cards

pearson product-moment correlation (r )

measure of linear association between two variables after removing units from each variable by standardizing them

  • common effect size measure

  • formula provides correlation for other methods

  • presumes linearity and independence, normality, homogeneity of variance

  • attenuated by restricting range on observed variables (reduces SD and reduces the magnitude)

  • based on t(df) distribution - df = n-2

58
New cards

spearmen correlation (rs)

association between ranks of two variables - replace each observed value for x and y with its rank

  • more robust to assumption of normality, but statistical power is reduced

  • useful for monotonic relationships (association is strictly increasing or decreasing)

59
New cards

point biserial correlation (rpb)

pearson correlation between a dichotomous variable (like treatment/control) and a continuous measure

  • just use Pearson to compute - for distinguishing one of the variable is dichotomous

  • based on t(df) distribution - df = n-2

60
New cards
<p><span style="font-family: CMMI10">φ </span>correlation</p>

φ correlation

measure of association between two dichotomous variables

  • done using Pearson r formula if data are numeric or using cell counts in a table

  • table: define each cell and then calculate - 2×2 table = A, B, C, D

61
New cards

dichotomous

a classification system that divides things into two mutually exclusive categories.

62
New cards
<p><span style="font-family: CMMI10">χ</span><span style="font-family: CMR7">2</span><span style="font-family: CMR10">(</span><span style="font-family: CMMI10">df</span><span style="font-family: CMR10">)</span>-test for independence</p>

χ2(df)-test for independence

tests difference between what we expect to see under null of no association vs observed data

  • significant = rejection of null

  • function of observed count (O) and expected count (E) for cell i as follows where df = (#rows - 1)(#columns - 1)

63
New cards
<p>distance correlation (dr)</p>

distance correlation (dr)

measures dependence between x and y by computing distance of each observation from every other observation

  • addresses issues in Pearson correlation insensitive to deviations from linearity

  • 0 to 1 (0 = complete independence, 1 = perfect dependence)

  • dcor.ttest(x, y) in R

  • probability under the null using a large number of random correlations - dcor.test(x, y, R=10000)

  • unbiased dr is smaller than r

64
New cards

do multiple studies in an article safeguard inferences?

No

  • under standard data-analytic and research practices make significant results easy to obtain (omitting variables of no significance)

65
New cards

familywise type I error rate

probability of at least 1 type I error being. made across set of tests in a manuscript

66
New cards
<p>effect size</p>

effect size

degree to which treatment changes the treated subjects relative to control subjects/degree to which phenomenon is or isn’t present in the population'

  • ratio of systematic variance relative to error or residual variance - signal-to-noise ratio

67
New cards
<p>effect size: Cohen’s d</p>

effect size: Cohen’s d

standardized mean difference - difference between treatment and control mean relative to standard deviation

68
New cards
<p>glass’ <span style="font-family: CMR10">∆</span></p>

glass’

measure of effect size

69
New cards
<p>Hedges’ g</p>

Hedges’ g

unbiased estimate of standardized mean difference based on pooled standard deviation

70
New cards
<p>standardized mean difference for robust t-test (dr)</p>

standardized mean difference for robust t-test (dr)

y1 and y2 = trimmed means from groups 1 and 2

swp = pooled Winsorized standard deviation (.642 adjustment for 20% of each tail)

<p>y1 and y2 = trimmed means from groups 1 and 2</p><p>swp = pooled Winsorized standard deviation (.642 adjustment for 20% of each tail)</p>
71
New cards

statistical power

probability of correctly rejecting a false null

  • function of sample size, alpha, and effect size

72
New cards

determining statistical power for one-sample z-test

  1. Draw null hypothesis theoretical sampling distribution.

  2. Calculate regions of retention and rejection.

  3. Determine and draw alternative hypothesis theoretical sampling distribution on the same graph.

  4. Calculate power as proportion of sample means in alternative hypothesis sampling distribution that is region of rejection.

73
New cards

determining statistical power for independent sample t-tests

same as z-test, except:

  • can’t express hypothesis distributions and regions in terms of specific scores as we don’t know population parameters

  • indicate ncp when calculating pt( ) (treatment mean - control mean / standard deviation

74
New cards

if effect size and alpha remain same, what happens to power when n increases?

it increases

75
New cards

if you decrease effect size, why does power decrease?

mean z-test under HA decreases as a result which reduces power

76
New cards

how are beta and power related

power = 1 - beta

77
New cards

if you increase alpha from .05 to .10, what happens to power?

it increases because increasing alpha also increases the region of rejection

78
New cards

For independent samples t-test, if effect size and α remain the same, and n increases, there are two reasons statistical power increases. What are these two reasons?

  1. noncentrality parameter (mean of the alt distribution on the t-test scale) increases.

  2. critical t-value decreases, which increases region of rejection

79
New cards

confidence interval

index of uncertainty associated with a statistic (usually 95% for alpha = .05)

  • calculated by hand once critical t-values are determined

  • tcrit tells how far estimates are from true population through number of standard errors from 0

  • intervals will cover population mean difference in 95% of random samples

80
New cards

confidence intervals for single means

one sample size

<p>one sample size</p>
81
New cards

standard error of the mean (sy ̄)

standard deviation / sqrt(n)

82
New cards

confidence interval for mean differences

2 sample sizes and 2 means

<p>2 sample sizes and 2 means</p>
83
New cards

sy1 - y2

estimated standard error of difference between 2 means

<p>estimated standard error of difference between 2 means</p>
84
New cards

spooled

pooled standard deviation between 2 means

  • sqrt((var1² + var2²)/2)x

<p>pooled standard deviation between 2 means</p><ul><li><p>sqrt((var1² + var2²)/2)x</p></li></ul>
85
New cards

p-value for difference between two means

differences between sample variances - null hypothesis / SE between means

  • μt μc = null (often = 0)

<p>differences between sample variances - null hypothesis / SE between means</p><ul><li><p><span style="font-family: CMMI10">μ</span><span style="font-family: CMMI7">t </span><span style="font-family: CMSY10">− </span><span style="font-family: CMMI10">μ</span><span style="font-family: CMMI7">c = null (often = 0)</span></p></li></ul>
86
New cards

calculating CIs for standardized effect size estimates

function of population parameter and where it intersects with observed statistic

  • not calculable unless we assume normality and homogeneity for errors, so resampling approximates these CIs

  • values of that have cumulative probabilities between .975 and .025