PSY 3001W Exam 3

0.0(0)
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/152

encourage image

There's no tags or description

Looks like no tags are added yet.

153 Terms

1
New cards
Land of Zero Variability:
Every time someone performs a task, they do it exactly the same way
• Everyone behaves like everyone else


Example of land of zero variability:

• Experiment: Will our new "pep pill" make people run faster?
2
New cards
Inferential Statistics:
• We use inferential statistics to try to infer from the sample data what the population might think.
• We use inferential statistics to make judgments of the probability that an observed difference between groups is a dependable one or one that might have happened by chance in this study.
3
New cards
Results:
• Because there is no variability, we can simply compare the sample mean to the population mean
• Now the sample is no longer like the population of undrugged people. There is now a population of drugged people who behave differently than undrugged people.
4
New cards
Do experiment again: First attempt
Pick a guy, give him the pep pill, have him run a mile.
• He runs it in 5:45. What can we conclude about the pep pill?
Nothing.
5
New cards
Do experiment again: Second attempt
Pick a guy, have him run a mile BEFORE he takes the pep pill. It takes him 6:00
• To double-check, have him run it again. Now it takes him only 5:50.
6
New cards
Do experiment again: Third attempt
• Measure how long it takes everyone to run a mile
7
New cards
What is the problem of the third attempt?
We can't measure everyone's running time, so we don't know the distribution of running times for the population (mean and SD)


Problem: the population of undrugged runners might be identical to the population of undrugged runners. (the drug might have no effect)
8
New cards
How can we estimate the population distribution?
By measuring running times for a sample
9
New cards
What is the problem of this?
The population of drugged runners might be identical to the population of undrugged runners (the drug might have no effect)
10
New cards
What is the null hypothesis?
Assume that there is no difference between the populations from which the samples were drawn (drug had no effect)

Ho: Ue \= Uc
11
New cards
Alternative hypothesis or research hypothesis:
There is a difference between the populations (the drug had an effect)

Ha: Ue does not equal Uc
12
New cards
Testing the null hypothesis
Are the sample means "significantly different?"

If the probability of obtaining our result when the null hypothesis is true (p-value) is less than .05, reject the null hypothesis.
Test Ho at the .05 level of significance
13
New cards
T-test:
Tests the significance of the difference between sample means
14
New cards
Compute t based on:
Sample means, standard deviations, sample sizes
15
New cards
Compute SD's on:
Sample sizes (t is very large for large differences)
16
New cards
What is the probability of obtaining that value of t by chance if the samples were drawn from identical populations?
Make a decision based on p-value
17
New cards
For any statistic we compute (t, F, etc.) can we compute its p-value?
Yes
18
New cards
P-value:
Probability of obtaining that value of the statistic or a more extreme value of the statistic if the null hypothesis is true
19
New cards
When doing a t-test in SPSS:
• Select alpha (level of significance)

• Enter your data in SPSS, click on t-test

• SPSS computes t(obs) and tells you the probability of obtaining t(obs) by chance (p-value)

-Determine if the p-value (probability of obtaining Tobs by chance) is less than alpha
If you are doing a one tailed (directional) test, divide p by 2.
20
New cards
Determine if the p-value (probability of obtaining t(obs) by chance) is less than alpha

21
New cards
If you are doing a one-tailed (directional) test: p\=0.12 (divide p by 2)

22
New cards
Test the significance at the .05 level of significance:
Determine p-values
23
New cards
If p < .05, conclude that the sample means are:
significantly different (IV had an effect)
24
New cards
If p \> .05, conclude that the sample means are:
NOT significantly different (IV had no effect)
25
New cards
There is always a chance that:
Our conclusion is wrong
26
New cards
In SPSS, there are different t-tests for independent groups and repeated measures:
Interpret the p-value the same way
27
New cards
Testing the null hypothesis is a...
a decision based on probability
28
New cards
Type 1 error:
H0 is true but you reject the Ho (IV had no effect but you conclude IV had an effect) FLASE ALARM
29
New cards
Type 2 error:
H0 is false but you do not reject H0 (IV had an effect but you conclude that IC had no effect) MISS
30
New cards
Probability of Type 1 error:
Alpha \= level of significance
31
New cards
Probability of Type 2 error:
Beta
32
New cards
Probability of correctly deciding H0 is false:
\= 1- B\=power
33
New cards
Type one and type two error example:
Researcher believes new teaching method can improve math ability in grade school children more than current teaching method can improve math ability in grade school children more than the current teaching method

Ho: new teaching method is no better than old one

H1: new teaching method is better than old one

Type 1 error: researcher concludes that new teaching method is no better than old one
In reality, there is no difference between the old and new teaching methods (false alarm)


Type 2 error: researcher concludes that new teaching method is no better than old one
In reality, the new teaching method is more effective than the old method (miss)
34
New cards
Which type of error is worse?
If you make a type 1 error and conclude that new teaching method is better but its really the same, there is no cost.

If you make a type II error and conclude that new teaching method is no better, you lose out on a beneficial teaching method.

Type II error seems worse than type I error

BUT, what if the new teaching method is inexpensive? A type 1 error is more costly, maybe worse than a type II error.
35
New cards
T-tests:
• Independent-groups t-test
• Related-samples t-test
36
New cards
Independent-groups t-test:
Compares the means of two independent groups in order to determine whether there is statistical evidence that the associated population means are significantly different.
37
New cards
Related-samples t-test:
The Paired Samples t Test compares two means that are from the same individual, object, or related units. The two means typically represent two different times (e.g., pre-test and post-test with an intervention between the two time points) or two different but related conditions or units (e.g., left and right ears, twins). The purpose of the test is to determine whether there is statistical evidence that the mean difference between paired observations on a particular outcome is significantly different from zero.
38
New cards
Type one and type two error example \#2
Test of general feelings of hostility: those who score high have a tendency to exhibit violent behavior in the future.

Ho: student is just like non-aggressive students

H1: student is dangerously aggressive

Type 1 error: conclude that the student is overly aggressive
In reality, student is not aggressive (false alarm)

Type II error: conclude that the student is not aggressive
In reality, student is in danger of becoming violent (miss)

Type II error is worse
39
New cards
Types of ANOVA tests
Independent groups ANOVA (IV manipulated between groups)


related-samples/repeated measures ANOVA (IV manipulated within subjects)
40
New cards
Independent groups ANOVA:
IV manipulated between groups
41
New cards
Related-samples/repeated-measures ANOVA:
IV manipulated within subjects
42
New cards
Chi-square:
Compare counts for nominal data; Measurement scale for DV is nominal; \= chi square
43
New cards
Correlation:
Describe relationship between two continuous variables
44
New cards
Choosing a statistical test:

45
New cards
Degrees of freedom for paired-samples t-test:
df\= N-1
46
New cards
Degrees of freedom for independent samples:
df\= n1 + n2-2
47
New cards
Degrees of freedom (Df)
Related to sample size
48
New cards
Effect size (r)
strength of association
49
New cards
Effect size increases as...
T increases

0 < r < 1
50
New cards
Small effect size:
r\= .15
51
New cards
Medium effect size:
r\= .30
52
New cards
Large effect size:
r\= .40
53
New cards
Power of statistical test:
Probability of correctly detecting an effect
• Power\= 1-B

-power analysis
54
New cards
B\=
Probability of a Type 2 error
55
New cards
Power Analysis:
Given effect size and level of significance, determine N needed to detect effect
56
New cards
Effect size table:

57
New cards
Primary way researcher controls power is to:
Change the sample size


To increase power, increase sample size

To detect smaller effects, increase sample size
58
New cards
To increase power...
Increase the sample size
59
New cards
To detect smaller effects...
Increase the sample size
60
New cards
Interpreting non-significant results:
We "fail to reject H0" rather than "accept H0"
61
New cards
Reasons results may be non-significant even though H0 is false (Type 2 error):
• Level of significance (a, probability of a Type 1 error) is very low...Increases probability of a Type 2 error


• Sample size is too small for effect size

Compute and report effect size for non-significant results
62
New cards
A statistically significant result has little practical significance when:
• The study has poor external validity
• The effect size is very small
• The treatment is too costly to implement
• The effect size is comparable to that for an existing treatment
63
New cards
Multiple tests:
Probability of rolling 3 in at least one of two rolls: 11/36 \=.31

Probability of committing a Type 1 error in AT LEAST ONE OF TWO TESTS, each at the .05 level: .0975

If you perform multiple statistical test, there is a high probability that, by chance, at least one result will be statistically significant
64
New cards
Hypothesis testing:
Probability of committing a type I error (false alarm) when the level of significance is .05: .05
65
New cards
Replication example: dice
Probability of tolling two 3's in two rolls: 1/6 x 1/6 \= 1/36 \= .03

Hypothesis testing
Probability of committing a type 1 error (false alarm) when the level of significance is .05: .05


Probability of committing a Type 1 error in BOTH TESTS when you perform two tests, each at the .05 level: .05 x .05 \= .0025

If you repeatedly obtain significant results (you reject H0) in replications of a study, it is VERY unlikely that you are committing a Type 1 error
66
New cards
Hart and Albarracin (2011)
Examined how verb form affects interpretation of a written passage

Description of one main shooting another after gambling disagreement
"Imperfective" form: smith was firing gun shots
"Perfective" form: smith fired gun shots

"When violent, unlawful actions were described in the imperfective (1) rather than perfective (2) aspect, the perpetrator of the actions was viewed as engaging in them with greater harmful intent. "

Results could have implications for how information is presented to juries
67
New cards
Replication study: Eerland et al (2016)
12 independent research teams performed replications

A meta-analysis of the results did not detect an effect of verb form on "subjects" judgements of criminal intentionality or on the level of detail imagined by subjects while reading the passage.

"Taken together, our studies did not provide evidence that describing actions in imperfect aspect resulted in greater perceived intentionality... or more detailed processing of those actions. This overall pattern of results was consistent across studies."
68
New cards
Quasi-experiments:
Lack some of the features of true experiments
• Involve manipulation of an IV or introduction of a treatment
• Often lack randomization
• One- group posttest-only design
• One-group pretest-posttest design
69
New cards
F tests
Aka Analysis of variance
a more general statistical procedure than the t test
70
New cards
Systematic variance
the deviation of the group means from the grand mean, or the mean score of all individuals in all groups.

between groups
71
New cards
error variance
the deviation of the individual scores in each group from their respective group means.

within groups
72
New cards
Cohens d
Effect size estimate
-expresses effect size in terms of standard deviation units.
73
New cards
Quasi experiments
Lack some of the features of true experiments
Involve manipulation of an IV or introduction of a treatment
Often lack randomization
74
New cards
types of quasi-experiments
one-group posttest-only design
-one-group pretest-posttest design
75
New cards
One-group posttest-only design
called a "one-shot case study" by Campbell and Stanley (1966)

lacks a crucial element of a true experiment: a control or comparison group.


Ex: employees in a company might participate in a 4-hour information session on emergency procedures. At the conclusion of the program, they complete a knowledge test on which their average score is 90%. This result is then used to conclude that the program is successfully educating employees.
76
New cards
One-group pretest-posttest design
way to obtain a comparison is to measure participants before the manipulation (a pretest) and again afterward (a posttest). An index of change from the pretest to the posttest could then be computed

-Include threats to internal validity
77
New cards
History:
Events that occur during participation and effect behavior
78
New cards
Maturation:
Changes due to the passage of time that affect behavior
79
New cards
Testing:
Taking a test can affect subsequent testing
80
New cards
Instrumentation:
Changes in measurement instruments (including observers) over time
81
New cards
Threats to internal validity controlled by true experiments:
1. History
2. maturation
3.testing
4. instrumentation
5. Regression toward the mean
6. Subject attrition (mortality)
7. Selection
82
New cards
Regression toward the mean:
Extreme scores are likely to be followed by more moderate scores
83
New cards
Subject attrition (morality):
Participants selectively drop out of experiment
84
New cards
Selection:
When control and experimental groups are chosen in such a way that they are not equivalent
85
New cards
Nonequivalent control group design:
• Uses an experimental group and control group, but they are not equivalent (e.g. natural groups)
• Groups are "self-selecting"
86
New cards
Nonequivalent control group pretest-posttest design:
Assignment to groups not random;

Ex: Evaluate effects of regular writing exercises on writing quality

• One lab does exercises, other lab does not

• Nonequivalent groups (use pretest to show equivalence; or use pretest to show differential change for the two groups)

• Possible additional problems: Different section leaders (a confound), observer bias, contamination
87
New cards
Selection differences/selection bias
usually occurs when participants who form the two groups in the experiment are chosen from existing natural groups.


The differences become a confounding variable that provides an alternative explanation for the results.
88
New cards
Control series design:
Interrupted time series design with a control group
89
New cards
Interrupted time series design:
Examine a series of observations before and after a treatment and look for a change in behavior

EX; Do daily quizzes increase attendance?

Increase in attendance from one day to the next could be random variability

Instead, record attendance many days before and after introducing quizzes.
90
New cards
Single case experimental designs
NOT case studies
aka single-subject designs

• Traditionally used in studies of reinforcement and behavior modification
• Researcher manipulates an IV (unlike case study)
• Behavior recorded during baseline period-description of behavior as it exists and as it would be in the future without introduction of a treatment
• Behavior is recorded after treatment is introduced
91
New cards
Reversal design (or ABA, or ABAB,...):
A\= baseline period (no treatment)
B\= treatment period


his basic reversal design is called an ABA design; it requires observation of behavior during the baseline control (A) period, again during the treatment (B) period, and also during a second baseline (A) period after the experimental treatment has been removed.
92
New cards
Example of reversal design
For example, the effect of a reinforcement procedure on a child's academic performance could be assessed with an ABA design. The number of correct homework problems could be measured each day during the baseline. A reinforcement treatment procedure would then be introduced in which the child received stars for correct problems; the stars could be accumulated and exchanged for toys or candies. Later, this treatment would be discontinued during the second baseline (A) period.
93
New cards
Horlon (1987)
EX OF ABA (REVERSAL) DESIGN; effects of facial screening on spoon-banging behavior in mentally-impaired 8-year-old girl. Treatment involved placement of soft cloth over the face of the girl; which caused her to spoon banging
94
New cards
Multiple baseline design:
Measure baseline in several situations (e.g. aggressive behavior at home, school, and daycare) \= multiple baseline across situations

Introduce treatment at different times in the different situations

Evidence for treatment effectiveness is that behavior changes only when the treatment is introduced.; To demonstrate the effectiveness of the treatment, such a change must be observed under multiple circumstances to rule out the possibility that other events were responsible.
95
New cards
Types of Multiple Baseline Designs
across subjects, across behaviors, across situations
96
New cards
Multiple baseline across situations
Measure baseline in several situations (e.g. aggressive behavior at home, school, and daycare) \= multiple baseline across situations

Introduce treatment at different times in the different situations
97
New cards
Multiple baseline across subjects:
Measure behavior of several subjects over time; introduce treatment at different times for different subjects
98
New cards
Multiple baseline across behaviors:
Measure different behaviors of a single subject over time; introduce treatment (e.g., rewards) at different times for different behaviors
99
New cards
Problems that even true experiments may not control:
1. Contamination
2. Experimenter expectancy effects/Observer bias
3. Reactivity
100
New cards
experimenter expectancy effects/observer bias
actual change in the behavior of the people or nonhuman animals being observed that is due to the expectations of the observer

tendency of observers to see what they expect to see