Regression

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/26

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

27 Terms

1
New cards

Regression Analysis

A statistical technique for finding the best-fitting straight line for a set of data and making predictions based on correlations.

2
New cards

Linear Relationship

A relationship between two variables that allows for the computation of an equation describing the relationship, expressed as Y = bX + a.

3
New cards

Y-intercept

The value of Y when X is equal to 0, represented as 'a' in the equation Y = bX + a.

4
New cards

Slope Constant

Indicates how much Y changes when X increases by 1 point, represented as 'b' in the equation Y = bX + a.

5
New cards

Best Fit Line

The line through the data points that minimizes the distance from each point to the line, providing the best prediction of Y.

6
New cards

Standard Error of Estimate (SEE)

A measure of the accuracy of predictions, indicating the standard distance between the regression line and actual data points.

7
New cards

Coefficient of Determination (r²)

The proportion of variability in the dependent variable (Y) that can be explained by its relationship with the independent variable (X).

8
New cards

Simple Linear Regression

A regression analysis that predicts a dependent variable from one independent variable.

9
New cards

Multiple Linear Regression

A regression analysis that predicts a dependent variable from two or more independent variables.

10
New cards

Homoscedasticity

The assumption that the variance of residuals should remain constant across levels of the dependent variable.

11
New cards

Normal Distribution of Residuals

The assumption that residuals (errors) should be normally distributed in regression analysis.

12
New cards

Multicollinearity

A situation where independent variables are too highly correlated, which can skew results; ideally, r should be less than .80.

13
New cards

Unstandardized Coefficients (B)

Coefficients used in a regression equation to predict Y scores; cannot be compared across predictors unless IVs use the same scale.

14
New cards

Standardized Coefficients (β)

Coefficients that indicate the relative strength of each predictor, allowing comparison across different independent variables.

15
New cards

Durbin-Watson Test

A test used to check the independence of residuals, with values close to 2 being ideal.

16
New cards

Shapiro-Wilk Test

A test used to check for normality in the distribution of residuals.

17
New cards

Adjusted r²

An adjusted version of r² that accounts for the number of predictors in the model, providing a more accurate measure of explanatory power.

18
New cards

Independent Variable (IV)

A variable that is manipulated or categorized to determine its effect on the dependent variable.

19
New cards

Dependent Variable (DV)

The outcome variable being predicted or explained in a regression analysis.

20
New cards

Stepwise Multiple Regression

A regression method where independent variables are entered based on software-driven analysis of data.

21
New cards

Pearson's r

A measure of the linear correlation between two variables, indicating the strength and direction of their relationship.

22
New cards

Assumptions of Regression Analysis

Key assumptions include:
linearity,
independence of residuals,
homoscedasticity (constant variance),
normal distribution of residuals
absence of multicollinearity.

23
New cards

Linearity

Linearity assumes that the relationship between the independent variable (IV) and dependent variable (DV) is a straight line. This means that changes in the IV would lead to proportional changes in the DV.

24
New cards

Independence of Residuals

This assumption states that residuals (the differences between observed and predicted values) should be independent from each other. This means that the value of one residual should not influence the value of another, indicating that all observations should be collected independently.

25
New cards

Homoscedasticity

Homoscedasticity refers to the assumption that the variance of the residuals should remain constant across all levels of the independent variable. If the variance changes (heteroscedasticity), it can lead to inefficiencies in the predictions and estimates.

26
New cards

Normal Distribution of Residuals

This assumption posits that residuals should follow a normal distribution. Normally distributed residuals help ensure the validity of hypothesis tests and confidence intervals, as many statistical tests depend on this condition.

27
New cards

Absence of Multicollinearity

Multicollinearity occurs when independent variables are highly correlated with each other, which can inflate the variance of coefficient estimates and make them unstable. The assumption of absence of multicollinearity suggests that while IVs can be correlated, they should not be so highly correlated that it skews the regression results.