Midterm 1

0.0(0)
studied byStudied by 0 people
full-widthCall with Kai
full-widthOpen Podcast
GameKnowt Play
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/139

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

140 Terms

1
New cards

Correlation

  • how two variables covary in relation to each other

    • how they move together/vary together

  • standardized covariance

  • the value of r can range between -1 and +1

    • - = move in opposite directions

    • + = move in the same directions

  • if r = 0 — there is no linear relationship between the two variables

  • the closer r is to ± 1, the stronger the relationship

  • if r = ±1 — there is a perfect linear relationship between the variables

2
New cards

Deviation from the Mean

  • Observation - mean = distance of the observation from the mean

3
New cards

Two ways to define correlation coefficient

  • z scores

  • covariation

4
New cards
<p>z scores</p>

z scores

  • z = 0 — score = mean

  • z > 0 — score ≠ mean

  • z = 1 — score = 1 SD from mean

  • z = 1 — score = 2 SD from mean

5
New cards
<p>Variance Formula</p>

Variance Formula

6
New cards
<p>Covariance Formula</p>

Covariance Formula

7
New cards

Coefficient Determination

  • r2 indicates the percent of the variability in y that is accounted for by the variability in x

8
New cards

Model

  • variability in scores that we can account for

9
New cards

Error

  • variability in scores that we cannot account for

10
New cards
<p>t obtained formula</p>

t obtained formula

11
New cards

Steps in Finding Correlation

  1. create H0 and H1

  2. set parameters (find critical t)

  3. calculate t obtained

  4. compare critical t and t obtained

  5. find p-values in SPSS

12
New cards

Steps in Finding Confidence Intervals

  1. convert Pearson’s r to z score

  2. compute a confidence interval for z score

  3. convert z score back to Pearson’s r

13
New cards

Calculating Confidence Intervals

  • CI = value ± (z critical (SE))

14
New cards
<p>Standardization Formula</p>

Standardization Formula

15
New cards

Reliability

  • in research, the term reliability means “repeatability” or “consistency”

  • a measure is considered reliable if it would give us the same result over and over again (assuming that what we are measuring isn’t changing!)

16
New cards

Internal Consistency Reliability

  • AKA “coefficient alpha”, “Cronbach’s alpha”, “reliability coefficient”

  • judge the reliability of an instrument by estimating how well the items that reflect the same construct yield similar results

  • looks at how consistent the results are for different items for the same construct within the scale

    • questionnaires often have multiple questions dedicated to each factor (some need to be reverse coded)

17
New cards

Purpose of Internal Consistency Reliability

  • used to assess the consistency of results across items within a test

  • how well do the items “hang” together?

  • typically, a reliability analysis is done on items that make up a single scale (items all supposed to measure roughly the same construct) — founded on correlations of items

18
New cards

Reliability Coefficients

  • range from 0-1.00 with higher scores indicating the scales is more internally consistent

  • generally, reliabilities above .80 are considered acceptable for research purposes

  • reliability analyses are carried out on items AFTER recoding (reverse coding)

19
New cards

Validity

  • is the test measuring what it claims to measure

20
New cards

Assumptions of Tests Based on Normal Distribution

  • additivity

  • normality of something or other

  • homogeneity of variance/homoschedasticity

  • independence

21
New cards

Additivity

  • outcome = model + error

    • model used to predict variability

  • the outcome variable (DV) is linearly related to any predictors (IV)

  • if you have several predictors (IVs) then their combined effect is best described by adding their effects together

  • if this assumption is met then your model is valid

22
New cards

Normally Distributed Something or Other

  • the normal distribution is relevant to:

    • parameter estimates

    • confidence intervals around a parameter

    • null hypothesis significance testing

23
New cards

When does the Assumption of Normality Matter?

  • in small samples

    • the central limit theorem allows us to forget about this assumption in larger samples

  • in practical terms, as long as your sample is fairly large, outliers are a much more pressing concern than normality

24
New cards

Spotting Normality

  • we don’t have access to the sampling distributions so we usually test the observed data

  • central limit theorem

  • graphical displays

25
New cards

Spotting Normality with the Central Limit Theorem

  • if N > 30, the sampling distribution is normal anyway

26
New cards

Spotting Normality with graphical displays

  • can use histograms or P-P Plots

27
New cards

P-P Plots

(probability-probability)

  • when it sags: kurtosis is an issue

  • when it forms an “S”: skewness is an issue

28
New cards

Values of Skew/Kurtosis

  • 0 in a normal distribution

  • convert to z (by dividing value by SE)

    • value greater than 1.96 = significantly different from normal

    • should be used for smaller sample sizes, if at all

29
New cards

Kolmogorov-Smirnov Test

  • tests if data differ from a normal distribution

    • significant = non-normal data

    • non-significant = normal data

*for large sample sizes

30
New cards

Shapiro-Wilk Test

  • tests if data differ from a normal distribution

    • significant = non-normal data

    • non-significant = normal data

*for small sample sizes

31
New cards

Homoschedasticity

  • measuring variance of errors

    • variance of outcome variable should be stable across all conditions

32
New cards

Homogeneous

  • uniform error rate across categories

33
New cards

Heterogeneous

  • difference in error rate across categories

    • violation of homoschedasticity

34
New cards

Independence

  • observations are completely independent from each other

    • violation example — 2 participants talk and share notes between the first and second parts of a test so their scores are no longer independent and now have a correlation

35
New cards

Violation of Assumptions of Independence

  • grouped data

36
New cards

Violation of Assumptions of Normality

  • robustness of test

  • transformations

37
New cards

Violation of Assumptions of Homogeneity

  • robustness and unequal sample sizes

  • transformations

    • if you have not normal data, one way to make data normal is to conduct a transformation

38
New cards

Types of Transformations

  • logarithmic transformation

  • square root transformation

  • reciprocal transformation

  • arcsine transformation

  • trimmed samples

  • windsorized sample

39
New cards

Transformations

  • always examine and understand data prior to performing analyses

  • know the requirements of the data analysis technique to be used

  • utilize data transformation with care and never use unless there is a clear reason

40
New cards

Square Root Transformation

  • help decrease skewness and stabilize variances (homoschedasticity)

41
New cards

Reciprocal Transformation

  • reduces influence of extreme values (outliers)

42
New cards

Arcsine Transformation

  • elongates tails (good for leptokurtic distributions)

43
New cards

Trimmed Samples

  • not really a transformation

  • fixed value of extreme values you cut off

44
New cards

Windsorized Samples

  • similar to trimmed samples

  • extreme values replaced by values that occur at 5% in the tails

45
New cards

Steps in Hypothesis Testing

  1. formulate research hypothesis

  2. set up the null hypothesis

  3. obtain the sampling distribution under the null hypothesis (choosing number of participants: past literature, power analysis)

  4. obtain data; calculate statistics

  5. given the sampling distribution, calculate the probability of obtaining a value that is as different as the one you have

  6. on the basis of probability, decide whether to reject or fail to reject the null hypothesis

46
New cards

When do you reject the null hypothesis

  • significance levels

  • conventional levels

  • the score that corresponds to alpha = critical value

    • if p < alpha, reject H0

    • if p > alpha, fail to reject H0

  • the smaller the alpha, the more conservative the test (we are more likely to conserve H0)

47
New cards

Directional hypothesis

  • indicates a direction (ex; time spent in class increases mind wandering)

  • use a one tailed test

48
New cards

Nondirectional hypothesis

  • does not indicate a direction (ex; time spent in class could increase or decrease mind wandering)

  • use a two tailed test

49
New cards

Type I Error

  • your test is significant (p < .05), so you reject the null hypothesis, but the null hypothesis is actually true

50
New cards

Type II Error

  • your test is not significant (p > .05), you don’t reject the null hypothesis but you should have because it’s false

51
New cards

True or False — A significant result means the effect is important

  • False

    • just because it is statistically significant does not mean it is actually significant in the real world

52
New cards

True or False — A non-significant result means that the null hypothesis is true

  • False

    • tells us only that the effect is not big enough to be found with the sample size we had

    • fail to reject the null — doesn’t mean the effect is not there, just means you didn’t find it

53
New cards

True or False — a significant result means that the null hypothesis is false

  • False

    • not a distinct yes or no, probabilistic reasoning — only 95% confident

54
New cards

True or False — The p-value gives you the effect size

  • false

    • we need to do some further calculations to find effect

55
New cards

True or False — The population parameter (μ) will always be within a 95% confidence interval of the sample mean

  • false

    • it’s an estimate

56
New cards

True or False — The sample statistic (M) will always be within a 95% confidence interval of the mean

  • True

    • we calculate the confidence interval based on the sample mean so it is always right in the middle

57
New cards

Trends to Circumvent Problems with NHST

  • effect size calculations — standardized so we can compare

  • confidence intervals

58
New cards

Outcome = __________

(model) + error

  • outcome — dependent variable

  • model — independent variables

59
New cards

B1

Slope

  • increase in y for every increase in x

60
New cards

B0

y-intercept

  • y when x is 0

61
New cards

Parameter Estimates

  • different samples drawn from the same population will likely yield different values for the mean

62
New cards

Sampling Error

  • the difference between my sample and the population value

63
New cards

Sampling error formula

64
New cards

Interval Estimates using σ

  • 95% confidence interval

    • lower limit: M + [z(critical)σM]

    • upper limit: M + [z(critical)σM]

65
New cards

Standard Error of the Sampling Distribution of Means

  • used when we don’t know σ

66
New cards

Standard Error of the Sampling Distribution of Means Formula

67
New cards

Interval Estimates without σ

  • 95% confidence interval

    • lower limit: M + [t(critical)SEM]

    • upper limit: M + [t(critical)SEM]

68
New cards

z scores vs. t scores

  • z scores

    • σ is known

    • large sample sizes

  • t scores

    • σ is not known (we introduce more error with SEM when σ is not known and t corrects for this)

    • small sample sizes

69
New cards

t critical

  • t[critical] = t(n-1)

70
New cards

t critical table

  • degrees of freedom = n-1

    • then find percentage

71
New cards

The mean is sufficient whereas the mode is not (true or false)

  • depends

    • mode is more appropriate for income

72
New cards

One should always plot data (true or false)

  • true

73
New cards

Standard error is a measure of variability (true or false)

  • false

    • estimate of variability across samples (standard deviation is a measure of variability)

74
New cards

You can compare standard deviations across samples (true or false)

  • true

75
New cards

An outlier is a score that falls 3 standard deviations from the mean (true or false)

  • true (usually)

    • this is the most common but it is subjective

76
New cards

Describing and Exploring Data Objectives

  • to reduce data to a more interpretable form using graphical representation and measures of central tendency and dispersion

    • data description and exploration, including plotting, should be the first step in the analysis of any data set

77
New cards

Why Plot Data?

  • gives a quick appreciation of the data set as a whole

  • readily gives information about dispersion/spread and the distributional form/shape

  • clear identification of outliers (if there’s one outlier we might want to go in and find out why)

  • what might cause an outlier?

    • someone with an already high baseline for what you’re measuring (high stress levels make them more/less reactive to something)

78
New cards

What is an outlier?

definitions vary

  • a data point that is far outside the norm for a variable or population and exerts undue influence

  • values that are dubious in the eyes of the researcher

79
New cards

Fringeliers

  • unusual events that occur more than seldom

    • not quite 3 standard deviations away from the mean to be considered outliers but still unusual

80
New cards

Reasons for Outliers

  • data entry or measurement error

  • sampling error

  • natural variations

  • response bias

81
New cards

Bar Chart

  • has spaces between the bars (categorical data)

    • spaces used to distinguish between clear cut categories

82
New cards

Histogram

  • no spaces between the bars (continuous data)

    • scores represent underlying continuity 

      • frequency chart of distribution

83
New cards

Point Estimators / Measures of Central Tendency

  • mean

  • median

  • mode

84
New cards

Properties of Estimators

  • sufficiency

  • unbiasedness

  • efficiency

  • resistance

85
New cards

Sufficiency

  • uses all the data

    • ex; mean

86
New cards

Unbiasedness

  • approximates the population parameter

87
New cards

Efficacy

  • low variability from sample to sample (sample population)

88
New cards

Resistance

  • resistant to outliers

    • ex; median

89
New cards

Mode

  • represents the largest number of people

  • easy to spot

  • unaffected by extreme scores (resistant)

  • can be used for nominal, ordinal, interval, or ratio data

*distributions can have more than one

90
New cards

Median

  • relatively unaffected by extreme scores (resistant to outliers)

  • relatively unaffected by skewed distributions

  • can be used with ordinal, interval, or ratio data

*sometimes not an actual score from the distribution

91
New cards

Mean

most common

  • can be manipulated algebraically

  • unbiased, efficient, and sufficient

  • assumes continuous measurement (can only be used with interval and ratio data)

  • minimizes sum of square deviations from it

    • gives smallest sum of squares of all methods

    • good estimator

  • influenced by outliers

  • influenced by skewness

92
New cards

Four Moments in Statistics

  • Mean

  • Variance

  • Skewness

  • Kurtosis

93
New cards

Mean Formula

94
New cards

Variance Formula

95
New cards

Skewness Formula

96
New cards

Kurtosis Formula

97
New cards

Measures of Dispersion

  • range; interquartile range (only use the middle 50% of data to eliminate outliers)

  • mean absolute deviation

  • variance

  • standard deviation

98
New cards

Mean Absolute Deviation Formula

99
New cards

Symmetrical Distribution

  • when two halves are mirror images

100
New cards

Skewness

  • the bulk of the scores are concentrated on one side of the scale with relatively few scores on the other side