Week 5 Lecture Recording
Overview of Regression Analysis
Focused on understanding simple and multiple regression analysis.
Aims to predict outcomes from one or more predictors.
Key concepts: hierarchy and stepwise regression.
Simple Regression
Definition: Analyzing the relationship between two variables to predict an outcome.
Y = a + bX:
Y: Criterion variable (outcome).
a: Intercept (value when X=0).
b: Slope (changes in Y for each unit change in X).
Example: Salary (predictor) and happiness (criterion).
Residuals: Error terms representing the distance of data points from the regression line.
Vertical distance from points to the line indicates error.
Can be assessed using the Durbin-Watson statistic.
Multiple Regression
Definition: Expands simple regression to analyze multiple predictors simultaneously.
Y = a + b1X1 + b2X2 + ... + bnXn:
b1, b2, ... bn: Slopes for each predictor variable.
Outcome Example: Statistics anxiety predicted by personality traits (perfectionism) and effort (revision hours).
Assumptions of Multiple Regression
Continuous Criterion Variable: Must be interval-level.
Normal Distribution of Residuals: Residuals should be normally distributed.
Independence of Errors: Residuals must be independent.
Homogeneity of Variance: Variance should be consistent across levels of predictors.
No Multicollinearity: Predictors should not be highly correlated.
Linear Relationship: A linear connection must exist between predictors and criterion variable.
Regression Models
Hierarchical Regression
Known predictors based on previous research are entered first (theory-driven).
Exploratory predictors can enter in subsequent steps.
Benefits: Establishes unique predictive influence while holding other variables constant.
Forced Entry Regression
All predictors entered simultaneously.
Suitable when all predictors are thought to have equal importance.
Cannot determine which predictor explains more variance beforehand.
Stepwise Regression
Variables entered based on statistical criteria.
Algorithm selects variables iteratively to maximize explained variance.
Useful for exploring significant predictors without prior knowledge.
Not ideal for definitive conclusions due to reliance on mathematical selection process.
Example of Multiple Regression Analysis
Research Question: What enhances exam performance?
Predictors: Revision hours, alcohol consumption, coffee intake.
Model Summary: Multiple correlation coefficient (R) and adjusted R-squared values represent shared variance with the criterion.
ANOVA Table: Determines if the regression model is significant (predicts above chance level).
Coefficients Table: Shows the significance of each predictor variable.
Only revision hours and alcohol consumption were significant predictors; coffee was not.
Coefficients Interpretation
Unstandardized beta coefficients represent the effect of a one-unit change in the predictor on the outcome.
For example, each additional revision hour increases exam scores, while each unit of alcohol decreases scores.
Standardized coefficients allow comparison across predictors.
Reporting Findings
Report coefficients only if the ANOVA is significant.
Example interpretation: "For each additional hour of revision, a student's score increases by 1.53%, while each unit of alcohol consumed drops performance by 2.3%."
Confidence intervals help assess reliability of coefficients (e.g., intervals not crossing zero indicate significance).
Conclusion
Understand and apply regression techniques through practical exercises during computer lab sessions.
Aim for competence in designing, running, and interpreting regression tests in various research contexts.