Stats 1

0.0(0)
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/127

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

128 Terms

1
New cards

categorical data

factor, nominal

no particular relationship between the different possibilities

can’t average them/do maths

eg. job, eye colour, drink preference

2
New cards
3
New cards

continuous data

interval

can do maths

eg. weight, height, reaction time, test score

4
New cards

ordinal data

ordered categorical, factor, likert scale

an order to the sequence

can’t do maths

can talk about frequencies but not averages

eg. grades, how religious someone is

5
New cards

descriptive statistics

central tendency: mean, mode, median

variability: range, variance, standard deviation, interquartile range

they describe what’s going on in the data/sample (not the larger population)

6
New cards

mean

average

for continuous data

sum of all values divided by number of values

outliers are a weakness (anomalies)

can be influenced by very high/low scores

7
New cards

mode

value that happens most often

for continuous/categorical data

one = unimodal

two = bimodal

8
New cards

median

the middle number

for continuous data

sort data in ascending order and find number

can only have one

9
New cards

variability

how spread out the data are

how far away from the mean/median the data points are

only suitable for ordinal/continuous data

10
New cards

range

highest value - lowest value

know the boundaries of our data

useful to detect outliers/data input errors

doesn’t tell how common really high'/low numbers are

11
New cards

variance

how far numbers are spread out from the mean

feeds into other statistics that makes it more useful

= (distance of scores from the mean) / (number of scores - 1)

<p>how far numbers are spread out from the mean</p><p>feeds into other statistics that makes it more useful </p><p>= (distance of scores from the mean) / (number of scores - 1)</p>
12
New cards

standard deviation

square root of the variance

describes variability within a single sample - descriptive

measure of average distance of each point from the mean, in the original units

small = data points close to mean

large = data points far away from mean

used to understand the variability in continuous/ordinal data

<p>square root of the variance</p><p>describes variability within a single sample - descriptive </p><p>measure of average distance of each point from the mean, in the original units</p><p>small = data points close to mean</p><p>large = data points far away from mean</p><p>used to understand the variability in continuous/ordinal data</p>
13
New cards

interquartile range

split our data into quarters (each contains 25% of the data points)

  • put data into ascending order

  • find the median (Q2)

  • find the median of the first half (Q1)

  • find the median of the last half (Q3)

  • Q3 - Q1

the range (max-min) is the middle 50% of data

useful as it isn’t affected by outliers

when an odd number of data, find median, then in first half find the median while ignoring the Q2

14
New cards

histograms

use these to graphically visualise the frequency distribution of continuous data (eg. heights, weights, depression, reaction time)

make bin categories and then a frequency table, with frequencies of values in each bin

frequency on Y axis

the mode is the bin with the tallest bar

useful to see the shape/distribution of the data

15
New cards

tukey boxplot/box and whiskers

used to visualise continuous data and provide statistical summaries of data

shows 5 descriptive statistics in one plot

  • minimum bound

  • quartile 1

  • median (Q2)

  • quartile 3

  • maximum bound

  • plus outliers (statistically different to rest of data)

elements:

  • interquartile range = Q3-Q1

  • min and max bounds (not the true minimum and maximum, but the largest data points above and below the thresholds

16
New cards

thresholds in a boxplot

Q1 – (1.5xIQR)

Q3 + (1.5xIQR)

17
New cards

interpreting a boxplot

each section contains 25% of the data

shorter whisker = data will be quite similar (less variability)

longer whisker = data will have more variability

18
New cards

skewness

asymmetry is caused by the presence of extreme values that distort the shape of the data, which is why symmetrical distributions are quite rare

characterised by a long gentle tail on one side of the distribution and a shorter steeper tail on the other

about horizontal shape of distribution

negative skew - less than -2

normal skew - 0 (between -2 and +2)

positive skew - more than 2

<p>asymmetry is caused by the presence of extreme values that distort the shape of the data, which is why symmetrical distributions are quite rare</p><p>characterised by a long gentle tail on one side of the distribution and a shorter steeper tail on the other</p><p>about horizontal shape of distribution </p><p>negative skew - less than -2</p><p>normal skew - 0 (between -2 and +2)</p><p>positive skew - more than 2 </p>
19
New cards

symmetrical distribution

same shape on both sides (bell curve)

mean, mode and median very similar

in a box plot, the median line is central with equal proportions throughout

20
New cards

negative skew

longer tail slopes to the left

mean is lower than the median, which is lower than the mode

in a box plot, median and upper whisker are closer to the right/top

less than -2

<p>longer tail slopes to the left</p><p>mean is lower than the median, which is lower than the mode</p><p>in a box plot, median and upper whisker are closer to the right/top</p><p>less than -2</p>
21
New cards

positive skew

longer tail slopes to the right

mode is lower than the median, which is lower than the mean

in a box plot, median and lower whisker are closer to the left/bottom

more than 2

<p>longer tail slopes to the right</p><p>mode is lower than the median, which is lower than the mean</p><p>in a box plot, median and lower whisker are closer to the left/bottom</p><p>more than 2</p>
22
New cards

kurtosis

used to describe the shape of a distribution

an indicator of the number of extreme values in data

3 categories:

  • mesokurtic - 3, conforms to classic bell curve shape

  • platykurtic - less than 3, flatter profile with shorter tails indicating fewer outliers

  • leptokurtic - more than 3, narrower center and longer tails indicating more outliers

kurtosis between 1 and 5 is within acceptable limits of normality for a given distribution (-2 and +2)

<p>used to describe the shape of a distribution</p><p>an indicator of the number of extreme values in data</p><p>3 categories:</p><ul><li><p>mesokurtic - 3, conforms to classic bell curve shape</p></li><li><p>platykurtic - less than 3, flatter profile with shorter tails indicating fewer outliers</p></li><li><p>leptokurtic - more than 3, narrower center and longer tails indicating more outliers</p></li></ul><p>kurtosis between 1 and 5 is within acceptable limits of normality for a given distribution (-2 and +2)</p>
23
New cards

calculating skew by hand

Pearson’s coefficient of skewness

negative number = negatively skewed

positive number = positively skewed

skew of 0 = symmetrical data

<p>Pearson’s coefficient of skewness</p><p>negative number = negatively skewed</p><p>positive number = positively skewed </p><p>skew of 0 = symmetrical data </p>
24
New cards

bimodality

2 modes

not shown well in a box plot, so making a histogram is useful to show the distribution of the data properly

25
New cards

plotting extreme values

general weirdness

need to consider context of extreme values

26
New cards

robust statistics

the median

less vulnerable to distortion by extreme values

for some data, the median will provide a better estimate of central tendency than the mean

interquartile range less vulnerable too

27
New cards

non-robust statistics

the mean

extreme values can have a noticeable effect on it

SD also vulnerable to distortion by outliers, as it’s calculated using the mean

28
New cards

population

the entire group of individuals or items that a statistical analysis aims to describe or draw conclusions about

all the hypothetical individuals we want to understand something about

isn’t just all physical people currently existing, but a more abstract concept of all infinite individual that have/could exist

29
New cards

sample

randomly chosen individuals from the population who we test/study, which we assume represent the whole population

in reality, many samples are not random when studies are volunteer = bias (random is ideal though)

can calculate descriptive statistics to understand the sample, but these only describe the sample not the wider population

larger the sample size, the more precise the estimates about the population will be

30
New cards

statistical model

a mathematical and simplified representation of observed data/behaviour/reality, used to make predictions or draw inferences about the sample and generalise to the whole population

simplifies reality/complex matters to understand relationships between variables

all models are wrong - they are representations of a thing that doesn’t necessarily capture all the complexity of reality but are still useful to use eg. London underground map

perfect correlation model:

  • data perfectly linear

  • data would all lie on line of best fit

  • can work out one variable if we know the other

can be relatively confident our model is reliable and appropriate for our data by understanding the assumptions it makes

if we reduce the discrepancy between the assumptions and our data, our model allows us to draw useful conclusions about the process that generates our data

they are also machines eg. the Rube-Goldberg machine (intentionally designed to perform a simple task in an overly complicated way

31
New cards

simple model

describing central tendency and variability in a measure

32
New cards

complex model

big network of how different variables connect to each other

33
New cards

normal distribution

a probability distribution as a bell shaped curve

1 peak (mean, median, mode are equal)

can use SD as a unit of measurement (a data points distance from the mean can be measured using a number of SDs)

eg. histograms

defined by its density function (density plots)

follows the 68-95-99.7 rule

properties:

  • symmetric about the mean/centre

  • tail never hits 0

  • characterised by mean and SD

  • X in continuous

    • Y is defined for every value of X

    • what gives the curve its’ smoothness

    • ranges from minus and plus infinity

    • non-continuous/discrete data = binary outcomes, count data, ordinal, psychometric scales

eg. IQ - total score of standardised tests to assess human intelligence (mean = 100, SD = 15)

34
New cards

density plots

shows relative likelihood of x taking on a certain value

normal distributions are defined by its density function

35
New cards

low standard deviation on a normal distribution

tall skinny curve

36
New cards

high standard deviation on a normal distribution (dispersed data)

flat wide curve

37
New cards

proportions under the normal distribution curve

68-95-99.7 rule

68% of data falls within 1 SD of the mean (34% above and 34 below)

95% of data falls within 2 SDs of the mean (47.5% above, 47.5 below)

99.7% of data falls within 3 SDs of the mean (49.85% above, 49.85 below)

the last 0.3% is above/below 3 SDs away from the mean

<p>68-95-99.7 rule </p><p>68% of data falls within 1 SD of the mean (34% above and 34 below)</p><p>95% of data falls within 2 SDs of the mean (47.5% above, 47.5 below)</p><p>99.7% of data falls within 3 SDs of the mean (49.85% above, 49.85 below)</p><p>the last 0.3% is above/below 3 SDs away from the mean </p>
38
New cards

inferential statistics

these make predictions, inferences and conclusions about the wider population based on samples of data

help assess the probability that a certain hypothesis is true

involves estimation of parameters and testing of statistical hypotheses

39
New cards

sample statistics

mean/median/mode

estimates the population mean

can be used to estimate parameters

40
New cards

population parameters

true mean/median/mode

likely the sample mean will be different to the population mean, so the inferences made will always be vulnerable to error

can use sampling distribution to estimate this error

41
New cards

sampling distribution

the SD of the distribution of sample means will be smaller than the SD within the population

means of samples will be less dispersed than they are in the population (same mean but smaller dispersion)

42
New cards

sampling distribution of a statistic

a probability distribution based on a large number of samples from a given population

sample mean = population mean

SD of sample < SD of population

SD here is called standard error (to distinguish it from the SD of one sample population)

43
New cards

standard error (SE)

estimates variability across samples - inferential

the more samples used, the less variable the means will be so SE

estimates whether a sample mean will be bigger/smaller than the population mean

shows how accurately the mean of a sample represents the true mean of the population

the smaller the SE, the more accurate the sample mean is close to the population mean

use it to calculate the confidence interval

44
New cards

difference between SD and SE

SD describes variability within a single sample - descriptive

SE estimates variability across multiple samples - inferential

45
New cards

calculating standard error

divide sample SD by square root of observations/number of samples

<p>divide sample SD by square root of observations/number of samples</p>
46
New cards

size of SE influenced by 3 factors

SD within the sample (larger SD = larger SE)

size of the sample (larger sample = sample mean closer to population mean)

proportion of the population covered by the sample (larger proportion covered in samples = lower variability of the means) - this has less influence on SE

47
New cards

confidence interval (CI)

how accurate an estimation of a population parameter will be

normally indicated as a % where the population mean lies within an upper and lower interval

use this to indicate the range we are pretty sure the population mean lies between 90, 95, 99% CIs

the larger the CI, the less precise the estimation of the population parameter

48
New cards

what would a 95% CI mean

a range of values that you can be 95% confident will contain the true mean of the population

49
New cards

what influences the size of CIs

variation

  • low variation = smaller CI

  • high variation = larger CI

sample size

  • smaller sample = more variation = larger CI

  • larger sample = less variation = smaller CI

50
New cards

calculating confidence intervals

knowt flashcard image
51
New cards

Z value

how many standard deviations a value is from the mean of a distribution

a measure of how many SDs the raw score is above/below the population mean (data that is recorded before going through statistical analysis)

based on a normal distribution

ranges 3 SDs above and below the mean

52
New cards

hypothesis testing steps

process of making inferences based on data from samples in a systematic way, to test ideas about the population

1) defining the null hypothesis

2) defining the alternative hypothesis

3) determining the significance level

4) calculating the p-value

5) reaching a conclusion

53
New cards

hypothesis testing 1) defining the null hypothesis

this assumes there is no difference between groups

nullifies the difference between sample means, by suggesting it is of no statistical significance

tests the possibility that any difference is just the result of sampling variation

54
New cards

hypothesis testing 2) defining the alternative hypothesis

predicts/states there is a relationship between the 2 variables studied

null hypothesis is assumed unless the difference between sample means is too big to believe the samples are the same (reject null hypothesis if differences become too big)

55
New cards

hypothesis testing 3) determining the significance level

typical level = 0.05 / 5%

known as the alpha

defines the probability that the null hypothesis will be rejected

similar outcomes reproduce 95% of potential replications

implies that in 5% of replications we will:

  • observe an effect that isn’t real (type 1 error)

  • fail to find an effect that does exist (type 2 error)

56
New cards

hypothesis testing 4) calculating the p-value

the probability value

how plausible is our data if there was actually no effect (null hypothesis true)

shows probability of an observed effect if the true (unknown) effect is null

between 0 (never) and 1 (always) probability

compare this against the significance level (alpha value)

p<alpha = significant

p>alpha = not significant

a p-value of 0.0125 means if the null hypothesis is true, the probability of obtaining results as/more extreme than those obtained is equal to 0.0125

57
New cards

if the p value is LESS THAN (<) alpha/significance level

reject null hypothesis

in favour of alternate hypothesis

results are statistically significant

we always say ‘reject (or fail to reject) null hypothesis’ when talking about significance, not ‘accept alternative hypothesis’

58
New cards

if the p value is GREATER THAN (>) alpha/significance level

fail to reject the null hypothesis - as this is evidence for the alternate hypothesis

results are NOT statistically significant

data is not inconsistent with null hypothesis

we don’t learn anything as evidence is inconclusive

if p-value is exactly 0.05, we still fail to reject null hypothesis

59
New cards

type 1 error

reject a true null hypothesis

rejecting it when should be accepting it

false positive

probability of making this error is represented by the significance level

can reduce risk by using a lower p-value (0.01 instead of 0.05)

<p>reject a true null hypothesis </p><p>rejecting it when should be accepting it </p><p>false positive </p><p>probability of making this error is represented by the significance level</p><p>can reduce risk by using a lower p-value (0.01 instead of 0.05)</p>
60
New cards

type 2 error

accept a false null hypothesis

accepting it when should be rejecting it

false negative

probability of making this error is know as beta

can reduce risk by making sure the sample size is large enough to detect a difference when one exists

<p>accept a false null hypothesis </p><p>accepting it when should be rejecting it</p><p>false negative </p><p>probability of making this error is know as beta </p><p>can reduce risk by making sure the sample size is large enough to detect a difference when one exists </p>
61
New cards

regions of rejection

determined by the alpha level

known as critical region too

set of values for a test statistic that lead to the rejection of a null hypothesis

the term 1/2 tailed test refers to where these regions are

62
New cards

two-tailed test

non-directional

region where you reject the null hypothesis is on both tails of the curve

eg. alpha level of 0.05 - split between the 2 tails evenly, giving 2.5% to each tail (these are regions you reject null hypothesis)

research question would not specify the direct of effect (eg. difference, impact, affect)

63
New cards

one-tailed upper test

region where you reject the null hypothesis is on the right side/upper end of the curve

eg. alpha level of 0.05 - alpha is all at one end (not split evenly), giving 5% to the upper tail only (region to reject null hypothesis)

upper tail contains upper values in a distribution = higher numbers will appear here so research questions specify the direction of effect (eg. greater/more than, higher)

64
New cards

one-tailed lower test

region where you reject the null hypothesis is on the left side/lower end of the curve

eg. alpha level of 0.05 - alpha is all at one end (not split evenly), giving 5% to the lower tail only (region to reject null hypothesis)

lower tail contains lower values in a distribution = smaller numbers will appear here so research questions specify the direction of effect (eg. lower/less than, decrease)

65
New cards

quantitative research

develop research question and hypotheses using literature review

design study

determine analyses (need to know method, analysis depends on study design)

collect data

analyse data

interpret data

disseminate results

66
New cards

choosing an inferential test

<p></p>
67
New cards

uni-variate data

1 variable

eg. asking 100 people their height and nothing else

summarising - mean and SD
visualising - histograms, box plots

understanding - results, normal distribution/skewness

can only answer simple questions

doesn’t help us understand why people act/think the way they do

68
New cards

bi-variate data

2 variables

eg. asking 100 people their height and weight

can be 2 continuous, 2 categorical or 1 of each

69
New cards

scatterplots

used with 2 continuous variables

a correlation used to measure strength of relationship between 2 variables

shows how much of the variance in 1 is explained by another

each point represents an observation

can see pattern, spread and orientation of the data

bigger sample size is better to reveal if relationship is real, but significant relationships aren’t the same as strong ones

70
New cards

interpreting scatterplots

shows how much of a relationship there is and the type

look at line of best fit, see how far spread out data points are from line, if there’s any anomalies

types of relationships:

  • linear

  • curvilinear

  • exponential

71
New cards

linear relationships

straight line

as one variable changes, the other changes

rate of change remains constant

72
New cards

curvilinear relationships

as one variable changes, the other changes but only up to a certain point

after this point, there’s no relationship or direction of it changes

eg. Yerkes-Dodson

73
New cards

exponential relationships

as one variable changes, another changes exponentially

eg. world population

74
New cards

positive relationship

one variable increases, the other increases

75
New cards

negative relationship

one variable increases, other decreases

76
New cards

no relationship

scatterplot points don’t form a pattern

77
New cards

correlations quantify relationships

strength - how much of a relationship

significance - likelihood of finding the observed relationship in the sample, if there was no relationship in the population

78
New cards

variance explained by our model

line through all data points

line of best fit - smallest distance = bigger effect

models the correlation and relationship between variables

79
New cards

correlation

association between 2 continuous variables

Pearson’s estimates how much of the variance in one variable can be explained by another variable

80
New cards

Pearson’s correlation coefficient (r)

size of the effect = correlation coefficient (r) - between -1 and +1

squaring the correlation coefficient tells how much variance in the data is explained by the model - can convert to %

no relationship - r=0

moderate - r=0.5

perfect relationship - r=+1 or r=-1

81
New cards

linear regression

a modelling technique used to make predictions about an outcome variable, based on 1+ predictor variables (1 = simple linear regression)

establishes link/relationship between outcome and predictor variables

regression is the foundation to other types of regression analysis

tells us:

  • strength of relationship between 2 continuous variables (so does correlations)

  • statistical significance (so does correlations)

  • how much one variable changes as another variable changes

  • the value of one variable if the other variable was 0

  • can predict a person’s score on a variable

includes effect size, slope and intercept

change in outcome variable (DV) for every unit increase in the predictor variable (IV)

  • outcome y can be predicted as linear function of x

82
New cards

regression as a model

predictor variable (x) → outcome variable (y)

arrow points towards variable trying to be predicted

both continuous

still cannot infer causality

83
New cards

effect size - linear regression

shows the strength of relationship between 2 continuous variables

R2 is a proportion of the variance explained by model (x100 to get %)

  • how well does the model (regression line) represent the data

  • strength of relationship (0-1)

modeled by how close our data points are to line of best fit

better the model, closer the points are to line of best fit

84
New cards

slope

beta (β)

shows how much one variable changes as another variable changes

essentially the line of best fit

85
New cards

intercept

a / constant (β0)

where the line of best fit intersects the y-axis

the value of one variable (outcome) if the other variable (predictor) was 0

86
New cards

using regression for predictions

if value of x is known for a given participant, we can predict y

y = bx + a + error

y = (slope x x) + intercept + variance explained by other stuff

the error is taken off for predictions

87
New cards

independent t-test

used to establish whether 2 means collected from independent samples differ significantly (difference between groups)

the null hypothesis is that the population means from the 2 unrelated groups are equal (most the time we reject the null and accept alternative - means are NOT equal)

the difference between the means of 2 independent groups

t = difference between groups / variance within groups

more extreme t value (>2) indicates less overlap between groups

<p>used to establish whether 2 means collected from independent samples differ significantly (difference between groups)</p><p>the null hypothesis is that the population means from the 2 unrelated groups are equal (most the time we reject the null and accept alternative - means are NOT equal)</p><p>the difference between the means of 2 independent groups</p><p>t = difference between groups / variance within groups</p><p>more extreme t value (&gt;2) indicates less overlap between groups </p>
88
New cards

features of independent t-test datasets

1 independent variable with 2 independent groups (between groups design)

1 dependent variable measured using continuous and normally distributed data

89
New cards

what do independent t-tests take into account

the mean difference between samples

variance of scores

sample size

90
New cards

t-statistic

produced from a t-test

the ratio of the mean difference between the 2 groups and the variation in the sampled data

uses the mean difference between samples, sample size and variance of scores

91
New cards

paired samples t-test

repeated measures - compare samples where each group participated in both conditions

data is collected from the same participant for both sets of observations (paired observations)

used to establish whether the mean difference between 2 sets of observations is 0

average size of change for each participant

t = score change / variance of that change

degrees of freedom = number of pairs - 1

92
New cards

features of paired samples t-test datasets

1 independent variable with 2 dependent groups (within groups design)

1 dependent variable measured using continuous and normally distributed data

93
New cards

one-way ANOVA

analysis of variance

tests whether there are statistically significant differences between 3+ samples (between-groups design)

one-way is specifically for samples that are independent groups

tests the null hypothesis that the samples in all groups are drawn from populations with the same mean values (and there is no significant difference between them)

accounts for both the variance between groups and within groups

94
New cards

features of a one-way ANOVA

1 independent variable with 3+ conditions/levels/groups

1 dependent variable measured using continuous and normally distributed data

95
New cards

ANOVA terms

one-way = 1 independent variable

factorial = multiple independent variables

ANCOVA = covariate (control variable)

MANOVA = multiple dependent variables

factors = independent variables

effects = quantitative measure indicating the difference between levels

96
New cards

why is multiple testing/comparisons a problem

if we adopt alpha level of 0.05 and assume the null hypothesis is true, then 5% of the statistical tests would show a significant difference

the more tests we run, the greater likelihood that at least 1 of those tests will be significant by chance (type 1 error - false positive)

97
New cards

when multiple testing issues arise

looking for differences amongst groups on a number of outcome measures

analysing data before data collection has finished, then re-analysing it at end of collection (this violation often used to see whether more data needs to be collected to reach significance)

unplanned analyses (conducting additional ones to try to find something of interest)

98
New cards

how to solve multiple testing issues

avoid over-testing (plan analyses in advance)

use appropriate tests

when multiple tests are run, adjust the alpha threshold

99
New cards

test statistic

a value that is calculated when conducting a statistical test of a hypothesis

shows how closely observed sample data matches the distribution expected under the null hypothesis of that statistical test

used to calculate the p-value of results

eg. t-test produces t statistic, ANOVA produces f statistic

100
New cards

the F statistic

an ANOVA produces this test statistic

a ratio of 2 variances (mean squares)

how much more variability in the data is due to differences between conditions/groups, as opposed to the normal variability

this along with degrees of freedom are used to calculate the p-value

if variants of the between groups is similar to the variants of the within groups, F statistic will be near 1

calculation: variance between groups/variance within groups

<p>an ANOVA produces this test statistic</p><p>a ratio of 2 variances (mean squares)</p><p>how much more variability in the data is due to differences between conditions/groups, as opposed to the normal variability </p><p>this along with degrees of freedom are used to calculate the p-value</p><p>if variants of the between groups is similar to the variants of the within groups, F statistic will be near 1</p><p>calculation: variance between groups/variance within groups</p>