Statistics - Effect Size & Power

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/51

flashcard set

Earn XP

Description and Tags

Statistics - Advanced Research Methods & Statistics

Statistics

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

52 Terms

1
New cards

Level of significance

<p></p>
2
New cards

⍺ level

Probability of committing a type I error.

3
New cards

Type 1 error

Reject null and accept alt hypothesis.

4
New cards

Significance level

Significance level (i.e., p-value) is the likelihood of observing a trend in the sample (H1) if there is not a trend in the population (H0).

<p>Significance level (i.e., <em>p</em>-value) is the likelihood of observing a trend in the sample (H<sub>1</sub>) if there is not a trend in the population (H<sub>0</sub>).</p>
5
New cards

True or false: Significant results are always meaningful results.

False: With a sufficiently large sample, a statistical test will almost always return a significant result.

6
New cards

Effect size (3 points)

  • A measure of the magnitude/strength of a difference or relationship.

  • It is independent of sample size.

  • Helps quantify practical significance (i.e., importance) – can be compared across studies.

7
New cards

Common effect size measures

  • Cohen’s d for t-tests

  • (Partial) Eta squared for ANOVAs

  • Pearson’s r for correlations

  • Cramer’s V for Chi-square

8
New cards

Cohen’s d

Measure of effect size when comparing two groups (i.e., t-tests): Standardised mean difference.

<p>Measure of effect size when comparing two groups (i.e., t-tests): Standardised mean difference.</p>
9
New cards

Cohen’s d - ≥.20

Small effect

10
New cards

Cohen’s d - ≥.50

Medium effect

11
New cards

Cohen’s d - ≥.80

Large effect

12
New cards

(Partial) Eta Squared

Measure of effect size for ANOVAs: Proportion of variance explained.

13
New cards

Eta Squared - one-way ANOVA formula

knowt flashcard image
14
New cards

Partial eta Squared - two-way ANOVA formula

knowt flashcard image
15
New cards

η22p - ≥.01

Small effect

16
New cards

η22p - ≥.06

Medium effect

17
New cards

η22p - ≥.13

Large effect

18
New cards

Sample size - when should one stop testing?

  • Consider a smaller difference between participant mean scores to reach significance for a small effect.

  • Consider more variance - less participants for a small effect.

  • Consider a more important effect

  • Resource constraints: Limited resources constrain how many participants can be tested.

  • Sample size of previous research: Can expect that the same participant number from significant previous research would give the same result.

  • Central limit theory

19
New cards

Central limit theorem (3 points)

  • Says that the distribution of sample approximates a normal distribution as sample size gets larger.

  • Sufficiently larger sample sizes can predict the characteristics of a population more accurately.

  • Sample size equal to or greater than 30: considered sufficient for theorem to hold.

<ul><li><p>Says that the distribution of sample approximates a normal distribution as sample size gets larger.</p></li><li><p>Sufficiently larger sample sizes can predict the characteristics of a population more accurately.</p></li><li><p>Sample size equal to or greater than 30: considered sufficient for theorem to hold.</p></li></ul><p></p>
20
New cards

Priori power analysis (2 points)

  • Works out how many participants are needed to reliably detect an effect of a certain size or larger.

  • A priori because it is conducted before data collection.

21
New cards

β level

The probability of making a Type II error (usually set at .200).

22
New cards

Graph - hypotheses, levels, power

knowt flashcard image
23
New cards

Power (3 points)

  • The ability to detect a significant effect if H1 is true.

  • Power = 1 - β

  • Power helps us make sense of results that are not significant or results that are statistically but not practically significant.

24
New cards

If we set β at .20, power will be…

.80 (80%)

Power = 1 - β

.80 = 1 - .20

25
New cards

The power of a statistical test depends on 4 parameters…

  • ⍺ (aka significance level)

  • Effect size

  • Sample size

  • Power (1 - β)

26
New cards

Power and ⍺ (2 points)

  • Smaller β = Increased power.

  • More lenient ⍺, the more statistical power a test will have.

<ul><li><p>Smaller β = Increased power.</p></li><li><p>More lenient ⍺, the more statistical power a test will have.</p></li></ul><p></p>
27
New cards

Power and effect size (2 points)

  • The larger the effect, the more likely that the effect will be detected.

  • Small effect = Increased β = Decrease in power

<ul><li><p>The larger the effect, the more likely that the effect will be detected.</p></li><li><p>Small effect = Increased β = Decrease in power</p></li></ul><p></p>
28
New cards

Power and sample size (3 points)

  • The larger the sample size, the more likely that the effect will be detected.

  • Increased β = Decreased power

  • Larger the sample size, higher the power.

<ul><li><p>The larger the sample size, the more likely that the effect will be detected.</p></li><li><p>Increased β = Decreased power</p></li><li><p>Larger the sample size, higher the power.</p></li></ul><p></p>
29
New cards

Ways to increase power of a test (5 points)

  • Set a less stringent significance level (but not usually done in Psychology) or consider a one-tailed test (but this should be based on theory / previous research).

  • Increase effect size by:

    • Increase the predicted difference between population means

    • Decrease the population standard deviation - more controlled environments? Clear procedures?

  • Increase sample size

30
New cards

Given any 3 parameters of power analysis, we can…

Estimate the 4th

31
New cards

Post-hoc power analysis (3 points)

  • Determines the power of a test by looking at significance level, effect size and sample size.

  • We can look at whether a non-significant result is likely to be due to a lack of power.

  • Post-hoc because it is conducted after data collection.

32
New cards

Effect size for priori power analysis (5 points)

  • Based on a previous / pilot study

    • Exact effect size from a previous / pilot study.

    • Assumes the conditions to be similar enough to your study.

  • Based on general benchmarks

    • Effect size benchmarks (i.e., small, medium, and large).

    • Helpful when previous studies do not provide a consistent effect size.

33
New cards

F test in G*Power

ANOVA

34
New cards

How to use G*Power (9 steps)

  1. Select the test you used from test family.

  2. Select the statistical test you used.

  3. Choose the type of power analysis you want to use.

  4. Select the tailedness you want to use.

  5. Input the effect size if you have it - if not, click the determine button. Input the means and standard deviations, and click calculate.

  6. Make sure alpha level is at .05.

  7. Input sample sizes.

  8. Click calculate in the bottom right hand corner.

  9. Power given in power (1-beta err prob).

35
New cards

How to G*Power - t-test post hoc (9 steps)

  1. Select t-test from test family.

  2. Select Means: difference between two independent means (two groups) under statistical tests.

  3. Select post hoc in type of power analysis.

  4. Select tailedness.

  5. If you don’t have the effect size, click the determine button and select n1 = n2. Input the means and standard deviations, and click calculate.

  6. Make sure alpha level is at .05.

  7. Input sample sizes for both groups.

  8. Click calculate in the bottom right hand corner.

  9. Power given in power (1-beta err prob).

36
New cards

Power under .8

The test is non-significant - should not conclude that there was no difference, but that there was not sufficient power to detect a difference.

37
New cards

How to G*Power - paired/t-test priori analysis (9 steps)

  1. Select t-test from test family.

  2. Select Means: difference between two independent means (two groups) under statistical tests / Paired: Means: difference between two dependent means (two groups) under statistical tests.

  3. Select priori in type of power analysis.

  4. Select tailedness.

  5. If you don’t know effect size, click the determine button and select n1 = n2. Input the means and standard deviations, and click calculate.

  6. Make sure alpha level is at .05.

  7. Put power as .8.

  8. Put allocation as 1.

  9. Click calculate in the bottom right hand corner.

38
New cards

Statistical tests in G*Power (2 points)

  • Top 3 mean options - parametric tests.

  • Bottom 3 mean options - non-parametric tests.

39
New cards

Sample sizes through priori power analysis

Tell you how many participants you need to get the same effect size as in the original study.

40
New cards

Effect Size for ANOVAs in G*Power

G*Power does not use η2 as the effect size for ANOVAs. It uses a measure that is named f (not to confuse with F!).

41
New cards

f - ≥.10

Small effect

42
New cards

f - ≥.25

Medium effect

43
New cards

f - ≥.40

Large effect

44
New cards

How to G*Power - one-way ANOVA post hoc (10 steps)

  1. Select F tests from test family.

  2. For a one-way ANOVA, choose 'ANOVA: fixed effects, omnibus, one-way' under statistical tests.

  3. Choose post hoc under type of power analysis.

  4. If you don’t know effect size, click determine, choose 'effect size from variance' under select procedure, click direct.

  5. Input η2p, which is .13??

  6. Click calculate and transfer to main window.

  7. Set alpha level to .05.

  8. Input total sample size.

  9. Input number of groups.

  10. Select calculate in bottom right hand corner. If there's a pop-up, just click okay.

45
New cards

How to G*Power - one-way ANOVA priori analysis (11 steps)

  1. Select F tests from test family.

  2. For a one-way ANOVA, choose 'ANOVA: fixed effects, omnibus, one-way' under statistical tests.

  3. Choose priori under type of power analysis.

  4. If you don’t know effect size, click determine, choose 'effect size from variance' under select procedure, click direct.

  5. Input η2p, which is .13??

  6. Click calculate and transfer to main window.

  7. Set alpha level to .05.

  8. Input total sample size.

  9. Input number of groups.

  10. Put power as .8.

  11. Select calculate in bottom right hand corner. If there's a pop-up, just click okay.

46
New cards

How to G*Power - two-way ANOVA post hoc (11 steps)

  1. Select F tests from test family.

  2. For a two-way ANOVA, choose 'ANOVA: fixed effects, special, main effects and interactions' under statistical tests.

  3. Choose post hoc under type of power analysis.

  4. If you don’t know effect size, click determine, choose 'effect size from variance' under select procedure, click direct.

  5. Input η2p, which is .13??

  6. Click calculate and transfer to main window.

  7. Set alpha level to .05.

  8. Input total sample size.

  9. Input number of groups (experimental conditions).

  10. Input df of experimental component.

  11. Select calculate in bottom right hand corner. If there's a pop-up, just click okay.

47
New cards

How to G*Power - two-way ANOVA priori analysis (11 steps)

  1. Select F tests from test family.

  2. For a two-way ANOVA, choose 'ANOVA: fixed effects, special, main effects and interactions' under statistical tests.

  3. Choose priori under type of power analysis.

  4. If you don’t know effect size, click determine, choose 'effect size from variance' under select procedure, click direct.

  5. Input η2p, which is .13??

  6. Click calculate and transfer to main window.

  7. Set alpha level to .05.

  8. Input total sample size.

  9. Input number of groups.

  10. Put power as .8.

  11. Select calculate in bottom right hand corner. If there's a pop-up, just click okay.

48
New cards

When G*Power priori analysis presents multiple sample sizes…

Choose the largest sample size to ensure that the power for all experimental components is at least .80.

49
New cards

Reporting results - post hoc power analysis (4 steps)

  1. Results of statistical test

  2. Power level (for each effect)

  3. Tailedness (for t-tests only)

  4. Interpretation of power level

50
New cards

Reporting results - 1. Results of statistical test

  • An t-test/ANOVA revealed that [DV] did not differ between [G1] (M = XX.XX, SD = XX.XX) and [G2] (M = XX.XX, SD = XX.XX; t(XX) = XX.XX, p = .XXX, d = XX.XX).

  • "An independent t-test revealed that the scores on the Intolerance of Uncertainty Scale did not differ between adults with anorexia nervosa (M = 12.21, SD = 1.25) and healthy adults (M = 11.52, SD = 1.15; t(40) = 1.86, p = .070, d = 0.58)."

51
New cards

Reporting results - 2. Power level (for each effect)

  • The observed power of the analysis was .XX (one/two-tailed), suggesting that the study was over/underpowered.

  • “The observed power of the analysis was .44 (two-tailed), suggesting that the study was underpowered.”

  • Note: This is technically part of 3. Tailedness (for t-tests only) and 4. Interpretation of power level

52
New cards

Reporting Results – A Priori Power Analysis (7 steps)

  1. Introduction of test used

    • "An a priori power analysis was used to determine the appropriate sample size for the study."

  2. How effect size was chosen

    • "A previous study yielded a medium effect size (d = 0.57)."

  3. Chosen significance level

    • "To detect the same effect at the conventional .050 significance level"

  4. Chosen power level

    • "...with a power of .80"

  5. Chosen statistical test

    • "...using an independent t-test"

  6. Tailedness (for t-tests only)

    • "...(two-tailed)"

  7. Required sample size

    • "...a sample of 98 (49 participants per group) was required."