One-way repeated measures ANOVA

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/29

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

30 Terms

1
New cards

2 t-tests & 2 ANOVAS (fill out table); DV is…

knowt flashcard image
2
New cards

WHEN do we use it?

  • Interval or ratio DV

  • Within-subjects (or matched) IV

  • IV has 3 or more levels

3
New cards

WHY do we use it?

Rather than multiple correlated groups t tests?

  • To avoid inflating Type I error rate (same reason as for one-way between-subjects ANOVA)

  • A single F test keeps alpha @ .05

4
New cards

Logic of Between-Subjects vs. Repeated Measures ANOVA

Between-Subjects ANOVA

  • Total Variability

    • Between groups: Treatment effect & Sampling error (includes individual differences & experimental error); numerator of F

    • Within groups: Sampling error (individual differences & experimental error); denominator of F

    • F = MSB (tells us why people are different across groups) / MSW (tells us why people are different within their group)

Repeated Measures ANOVA

  • Total Variability

    • Between groups: Treatment effect & Sampling error (experimental error only); numerator of F; no individual diffs here, b/c same people in all conditions

    • Within groups: Sampling Error

      • Between persons: individual differences REMOVED

      • Error: Experimental Error (denominator of F)

    • F = MSB/MSE

<p><strong>Between-Subjects ANOVA</strong></p><ul><li><p>Total Variability</p><ul><li><p><u>Between groups</u>: <strong>Treatment effect &amp; Sampling error </strong>(includes <em>individual differences &amp; experimental error</em>); numerator of F</p></li><li><p><u>Within groups</u>: <strong>Sampling error</strong> (<em>individual differences &amp; experimental error</em>); denominator of F</p></li><li><p>F = MSB (<u>tells us why people are different across groups</u>) / MSW (<u>tells us why people are different within their group</u>)</p></li></ul></li></ul><p><strong>Repeated Measures ANOVA</strong></p><ul><li><p>Total Variability</p><ul><li><p><u>Between groups</u>: <strong>Treatment effect &amp; Sampling error </strong>(<em>experimental error only</em>); numerator of F; <strong>no individual diffs here, b/c same people in all conditions</strong></p></li><li><p><u>Within groups</u>: <strong>Sampling Error</strong></p><ul><li><p>Between persons: individual differences REMOVED</p></li><li><p>Error: Experimental Error (denominator of F)</p></li></ul></li><li><p>F = MSB/MSE</p></li></ul></li></ul><p></p>
5
New cards

Part 1: focus on individual differences BETWEEN groups — what is different with rmANOVA:

We will not have individual differences in the between group component, because participants are no longer going to differ ACROSS CONDITIONS

6
New cards

Part 2: focus on individual differences WITHIN groups — what is different with rmANOVA:

Here is the really key piece, and also includes how we are going to do the actual ANOVA test…

There is still variability from person to person WITHIN a group, but because we now have a way to calculate a measure of variability PER PERSON, we can use that knowledge to separate WITHIN group variability into two components: between persons and experimental error.

7
New cards

Comparing the ANOVA Source Tables

knowt flashcard image
8
New cards

Steps in Hypothesis Testing

  1. Establish H0 & H1

    • H0: µp = µa = µn (all population means are the same; there is no variance among the pop. means)

    • H1: The three population means are not all equal; there IS variance among the population means

    • note: n = N in repeated measures

  2. Collect data

  3. Characterize the sampling distribution (F) and determine the critical value from table

  4. Calculate test statistic (F)

  5. Make decision to reject or fail to reject H0

9
New cards

Comparing Definitional Formulas:

SST & SSB are the same; SSW split into:

  • SSE: found via subtraction

  • SSBP

  • The SSW variability that you are familiar with calculating is still important. But, we get to partition SSW into two parts by calculating a measure of variability for EACH PERSON (SSBP). Then, we can exclude that SSBP when we calculate F ratio, which is F = MSB/MSE

<p>SST &amp; SSB are the same; SSW split into:</p><ul><li><p>SSE: found via subtraction</p></li><li><p>SSBP</p></li><li><p>The SSW variability that you are familiar with calculating is still important. But, <em>we get to partition SSW into two parts by calculating a measure of variability for EACH PERSON (SSBP)</em>. Then, we can e<em>xclude that SSBP when we calculate F ratio</em>, which is F = MSB/MSE</p></li></ul><p></p>
10
New cards

Comparing Computational Formulas: SST

Denominator: from N to kN— both of these represent the total number of observations in the data set

<p>Denominator: <strong>from N to kN— both of these represent the total number of observations in the data set</strong></p>
11
New cards

Comparing Computational Formulas: SSB & SSBP

SSB: sum of x columns, square that sum

SSBP: add the squared row columns

Each one is divided by the # of scores (10, 3) it’s based on

<p>SSB: sum of x columns, square that sum</p><p>SSBP: add the squared row columns </p><p>Each one is divided by the # of scores (10, 3) it’s based on </p>
12
New cards

Computational Formulas, denote ∑Xj, ∑Pi², j, i, n, N, k

knowt flashcard image
13
New cards

Degrees of Freedom total, between groups, error, & between persons

kN - 1

k - 1

(k - 1)(N - 1)

N - 1

<p>kN - 1</p><p>k - 1</p><p>(k - 1)(N - 1)</p><p>N - 1</p>
14
New cards

Let’s compute our SS!

PEMDAS

<p>PEMDAS</p>
15
New cards

How to conduct a one-way ANOVA (within groups)

knowt flashcard image
16
New cards

How is P different from the D we calculated for the paired-samples t test?

D removes effect of individual diffs. to create a “pure” treatment effect.

P removes the effect of treatment to get a “pure” measure of independent differences. We compute a SS for independent differences and then remove it.

D— removes individual differences, leaves a clean treatment effect

P— removes treatment effect; isolates person-to-person differences (used to reduce error)

17
New cards

Post Hoc Analyses (Tukey HSD); computing the critical difference

Remember: range = k (number of groups), not dfbetween!

<p>Remember: range = k (number of groups), not dfbetween!</p>
18
New cards

Tukey HSD

knowt flashcard image
19
New cards

Strength of the relationship (partial eta-squared)

PSTT & RMA— eta-squared represents the variability in the DV accounted for by the IV after variability due to individual differences have been removed

<p>PSTT &amp; RMA— eta-squared represents the variability in the DV accounted for by the IV after variability due to individual differences have been removed </p>
20
New cards

Partial eta-squared example

knowt flashcard image
21
New cards

Confidence Intervals!

We are 95% certain that the true population mean differences fall within each of these intervals. If the interval contains 0, then the difference is not statistically significant.

<p>We are 95% certain that the <strong>true population mean differences</strong> fall within each of these intervals. If the interval contains 0, then the difference is not statistically significant.</p>
22
New cards

Assumptions of the F test for repeated measures analyses

  • Sample is independently and randomly selected from the population (i.e., rows are randomly selected).

  • The scores in each population are normally distributed. (robust if N > 30)

  • Populations have homogeneous variances.

  • Sphericity

    • (test is not robust to violations of this assumption; Type I error (false positive) rate will increase if sphericity is violated)

    • Variance of population difference scores should be equal (for all pairs of differences)

    • Effect of the IV should be consistent across all participants (assumes everyone is equally affected by IV).

23
New cards

Explaining Sphercity

Difference scores for each pairwise comparison:

These groups of difference scores should have approximately equal variances (they’re estimating variances in the population).

<p>Difference scores for each pairwise comparison:</p><p>These <strong>groups of difference scores</strong> should have <strong>approximately</strong> <strong>equal variances</strong> (they’re estimating variances in the population).</p>
24
New cards

Visualizing Sphercity

Generally speaking, the effect of condition should be relatively similar for all participants.

Note:

By hand— Assume sphericity has NOT been violated! Just use a Tukey HSD test.

In SPSS:

  • Mauchly test (BUT: The test under-detects sphericity violations in small samples and overdetects them in large samples).

  • THM: Always assume sphericity has been violated and use one of two correction factors:

    • Huynh-Feldt epsilon

    • Greenhouse-Geisser epsilon

<p><strong>Generally speaking, the effect of condition should be relatively similar for all participants. </strong></p><p>Note:</p><p>By hand— <em>Assume sphericity has NOT been violated! Just use a Tukey HSD test.</em></p><p>In <strong>SPSS:</strong></p><ul><li><p><strong>Mauchly test</strong> (BUT: The test <strong>under-detects sphericity violations in small samples</strong> and <strong>overdetects</strong> them in <strong>large</strong> samples).</p></li><li><p><em>THM: Always assume sphericity has been violated and use one of two correction factors:</em></p><ul><li><p><em>Huynh-Feldt epsilon</em></p></li><li><p><em>Greenhouse-Geisser epsilon</em></p></li></ul></li></ul><p></p>
25
New cards

Rules for correction factors in SPSS

  • If Maunchy’s test is significant (p < .05), it means that the sphericity assumption has been violated & Type I error is more likely

  • Depending on what the GG number is, you are going to use either row 2 or row 3 of the source table below.

  • Mauchly’s test isn’t significant, but it under-detects sphericity violations in small samples. Use a correction factor anyway.

    The Greenhouse-Geisser and Huynh-Feldt epsilons (e)

  • When e is close to 1, there is very little violation of sphericity. Lower e = higher violation of sphericity. Note: no leading zero!

  • When the GG is greater than or equal to .75, the correction is too conservative, yet the H-F e tends to overestimate sphericity

SO… we use this general rule:

  • When the G-G e ≥ .75, we use the H-F e correction

  • When the G-G e < .75, we use the G-G e correction

<ul><li><p><strong>If Maunchy’s test is significant (p &lt; .05), it means that the sphericity assumption has been violated &amp; Type I error is more likely</strong></p></li><li><p>Depending on what the GG number is, you are going to use either row 2 or row 3 of the source table below.</p></li><li><p><em>Mauchly’s test isn’t significant, but it under-detects sphericity violations in small samples. Use a correction factor anyway.</em></p><p>The <strong>Greenhouse-Geisser and Huynh-Feldt epsilons</strong> (e)</p></li><li><p>When e is close to 1, there is very little violation of sphericity. Lower e = higher violation of sphericity. Note: no leading zero!</p></li><li><p>When the GG is greater than or equal to .75, the correction is too conservative, yet the H-F e tends to overestimate sphericity </p></li></ul><p>SO… we use this general rule: </p><ul><li><p><strong>When the G-G e </strong><span><strong>≥ .75, we use the H-F e correction</strong></span></p></li><li><p><span><strong>When the G-G e &lt; .75, we use the G-G e correction</strong></span></p></li></ul><p></p>
26
New cards

Rules for Correction Factors continued…

The correction simply adjusts the dfB and dferror by multiplying them by the epsilon. If sphericity hasn’t been violated, e is closer to 1 and won’t affect the df that much.

The epsilons are less than 1, so it makes the df smaller, resulting in a higher F critical value, making the F test more conservative (to correct for the possible Type I error).

You should still report F (2, 18) in your APA-style write-up. (Don’t report fractional df)

Notice that SPSS doesn’t provide dftotal or SStotal.

Be sure to report the partial eta-squared.

<p>The correction simply adjusts the dfB and dferror by multiplying them by the epsilon. If sphericity hasn’t been violated, e is closer to 1 and won’t affect the df that much.</p><p>The epsilons are less than 1, so it makes the df smaller, resulting in a higher F critical value, making the F test more conservative (to correct for the possible Type I error).</p><p>You should still report F (2, 18) in your APA-style write-up. (Don’t report fractional df)</p><p>Notice that SPSS doesn’t provide dftotal or SStotal.</p><p>Be sure to report the partial eta-squared. </p>
27
New cards

Rules for Correction Factors continued *2

Selecting “EM Means” gives you the standard errors for each condition (useful if creating a bar graph).

Ignore these CIs; they are CIs around individual means, not around differences between means.

<p>Selecting “EM Means” gives you the standard errors for each condition (useful if creating a bar graph).</p><p>Ignore these CIs; they are CIs around individual means, not around differences between means.</p>
28
New cards

Understanding the nature of the effect: Pairwise comparisons with a Bonferroni correction

SPSS does not produce a Tukey HSD test for a repeated-measures ANOVA, because sphericity violations (which are common) make using the MSerror in the CD formula problematic.

Instead, pairwise comparisons (paired-samples t tests) are conducted using a Bonferroni procedure to control Type I error rate.

The Bonferroni procedure keeps the experimentwise a at .05 by adjusting the testwise a according to the # of comparisons.

Simply divide a by the # of pairwise comparisons:

  • 3 conditions = 3 pairwise comparisons, so a = .05/3 = .017

  • 4 conditions = 6 pairwise comparisons, so a = .05/6 = .008

These are just paired-samples t tests.

The original p value has been multiplied by the # of pairwise comparisons (in this case, 3). So you can compare the p here to .05. This is analogous to dividing a by 3.

The rest of the output can be ignored:

  • Multivariate tests

  • Tests of within-subjects contrasts

  • Tests of between-subjects effects

<p>SPSS does not produce a Tukey HSD test for a repeated-measures ANOVA, because sphericity violations (which are common) make using the MSerror in the CD formula problematic.</p><p>Instead, pairwise comparisons (paired-samples t tests) are conducted using a Bonferroni procedure to control Type I error rate.</p><p>The Bonferroni procedure keeps the <em>experimentwise</em> a at .05 by adjusting the <em>testwise</em> a according to the # of comparisons.</p><p>Simply divide a by the # of pairwise comparisons:</p><ul><li><p>3 conditions = 3 pairwise comparisons, so a = .05/3 = .017</p></li><li><p>4 conditions = 6 pairwise comparisons, so a = .05/6 = .008</p></li></ul><p>These are just paired-samples t tests.</p><p>The original p value has been multiplied by the # of pairwise comparisons (in this case, 3). So you can compare the p here to .05. This is analogous to dividing a by 3.</p><p>The rest of the output can be ignored:</p><ul><li><p>Multivariate tests</p></li><li><p>Tests of within-subjects contrasts</p></li><li><p>Tests of between-subjects effects</p></li></ul><p></p>
29
New cards

Sample APA-style Results Section (using SPSS)

A one-way repeated measures ANOVA was conducted on pupil dilation (in mm) as a function of type of stimulus viewed (neutral, pleasant, or aversive). Mauchly’s test indicated that the assumption of sphericity was not violated, χ2 (2) = 2.39, p = .303, but because the test is highly sensitive to sample size, the degrees of freedom were still corrected using the Huynh-Feldt estimate of sphericity (e = .94). As hypothesized, pupil dilation was significantly related to type of stimulus, F(2, 18) = 24.93, p < .001, partial η2 = .79. Post-hoc pairwise comparisons using the Bonferroni correction indicated that, as hypothesized, average pupil dilation for neutral stimuli (M = 3.20, SD = 1.03) was significantly smaller than for both pleasant (M = 5.50, SD = 1.27), p = .002; 95% CI [-3.62, -0.98] and aversive stimuli (M = 5.40, SD = 1.17), p < .001; 95% CI [-3.05, -1.35]. Average pupil dilation for pleasant and aversive stimuli did not significantly differ, p = 1.00; 95% CI [-0.92, 1.12].

30
New cards

Visualizing F ratios

reduces variability (noise) by using the same participants in all conditions; each participant acts as their own control, removing between-persons V / increasing sensitivity to real effects

<p><strong>reduces variability </strong>(noise) by using <strong>the same participants</strong> in all conditions; each participant acts as their <strong>own control, removing between-persons V</strong> / increasing sensitivity to real effects</p>