Simple regression & multiple regression

0.0(0)
studied byStudied by 0 people
0.0(0)
full-widthCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/26

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

27 Terms

1
New cards

correlational research

  • allows us to establish whether an association exists, but does not allow us to establish whether the association is causal.

  • can explore things that cannot ethically be looked at in experiments (e.g., drug addiction)

  • can explore things that cannot feasibly be looked at in experiments (due to small effects)

  • can explore things that cannot be put into a classic experimental framework (e.g., age) 

2
New cards

Uses for correlational research

  • Exploring big data

  • Questionnaires/surveys

  • Secondary data analysis

  • Understanding the multivariate world

  • Predictions 

3
New cards

Positive association

as X increases Y increases

4
New cards

Negative association

as X increases Y decreases

5
New cards

strong association

measurements near line of best fit (estimate accurate)

6
New cards

Weak association 

lots of variance around line (not accurate)

7
New cards

Aims of a regression model

  • Whether our model is a ‘good fit’- good at making predictions

  • Whether there are significant relationships between a predictor variable(s) and an outcome variable

  • The direction of these relationships

  • used to make predictions

8
New cards

A regression

identifies and quantifies the relationship between a dependent variable (the outcome) and one or more independent variables (predictors) to make predictions and understand patterns.

9
New cards

line of best fit equation 

Y= bX + a

10
New cards

SSR

Difference from overall mean and predicted value (variance we can explain)

11
New cards

SSE

How far from perfect is our line of best fit (variance we cannot explain)

12
New cards

we want SSR to be__than SSE

bigger 

13
New cards

SST

The total amount of variance/ coefficient of determination

14
New cards

what can R2 range from

0-1

15
New cards

role of Adjusted R2

  • Decreases a lot of variables with little value

  • Should be equal or lower than R2

16
New cards

Linear regression

  • To understand individual predictors

  • To understand the direction of the associations

17
New cards

regression coefficients

The number of units the DV changes for each one unit increase in the IV

usually interpreted alongside standard error (SE should be smaller)

18
New cards

Standardised regression coefficients

  • Beta (β) values commonly reported

  • Explain the association between each IV and DV in terms of standard deviation changes

  • Allows simple comparison of the strength of the association between your IVs and DV

  • The higher the beta the stronger the association is.

19
New cards

Assumptions of simple & multiple regression

  • Normally distributed (ish) data

  • Independent Data

  • Interval/ratio predictors (continuous)

  • Nominal predictors with two categories (dichotomous)

  • No multicollinearity for multiple regression

  • Careful of influencing cases

20
New cards

what is not allowed in a regression and why 

multicategorical predictors-  coding is arbitrary so will give different results depending on how they were coded 

21
New cards

simple regression write up

  1. A simple regression was carried out to investigate the relationship between self-control and BMI.

  2. The regression model was significant/not significant and predicted approx. _% of variance

  3. (adjusted  R2__ (F-stat)=__, p__). ____ was a significant negative/positive predictor of ___ (b=__(se=__), p<__; 95% CI to __).

22
New cards

Multicollinearity and effects

  • occurs when independent variables in a regression model are highly correlated.

    • e.g., height & weight

  • Means they do not provide independent information

  • Can adversely affect regression estimates (b and se)

  • Large amount of variance explained but no significant predictors

23
New cards

Identifying multicollinearity 

  • run a correlation 

  • Look for high correlations between variables in a correlation matrix (rule of thumb r > .80)

  • r = 1.0 is perfect multi-collinearity and likely represents a data issue.

24
New cards

Tolerance statistic

  • Percentage of variance in the IV accounted for by other IVs

  • 1 – R2

  • High tolerance = low multicollinearity

  • Low tolerance = high multicollinearity (a value of < .20 or .10)

25
New cards

Variance Inflation Factor

  • Inverse of tolerance

  • 1/tolerance

  • Indicates of much the standard error will be inflated

  • VIF over 4 suggests multicollinearity with values above 10 suggests there to be substantial multicollinearity.

26
New cards

How to solve multicollinearity issues

  • Increase sample size – this will stabilise the regression coefficients

  • Remove redundant variable(s)

  • If the two or more variables are important, create a variable that takes both of them into account

27
New cards

Multiple regression write up

  1. A multiple regression was conducted to investigate the role of ___, ___, ___ on __. The regression model was significant/not significant and predicted 48% of

  2. variance (adjusted R2: F (_,_)=___,p__). ___ was a significant negative/positive predictor of __ (b= (s.e.= ), p; 95% CI to ) continue with other factors

  3. Variance inflation factors suggest multicollinearity was/was not a concern (do each factor).