Introduction to Regression – Vocabulary Flashcards

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/43

flashcard set

Earn XP

Description and Tags

Vocabulary flashcards covering fundamental terms and concepts from the lecture on regression algorithms.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

44 Terms

1
New cards

Regression

Task of predicting a continuous target variable by learning an approximation of an unknown function from data.

2
New cards

Predictor Variables (Features)

The input variables x that describe each case and are used to predict the target.

3
New cards

Target Variable

The continuous outcome Y that the model aims to predict.

4
New cards

Regression Model

A function hθ(x) that maps a vector of predictor values x to a real-valued prediction y.

5
New cards

Residual

The observation error εi = yi − hθ(xi) for instance i.

6
New cards

Bias

Systematic error introduced by approximating the true function with a simpler model; error due to wrong assumptions.

7
New cards

Variance

Amount by which a model’s predictions would vary if it were trained on different data sets.

8
New cards

Bias-Variance Trade-off

Relationship where decreasing bias often increases variance and vice versa; total error is their sum.

9
New cards

Loss Function

A function L(y,ŷ) that quantifies the cost of predicting ŷ when the true value is y.

10
New cards

Squared Loss

Loss defined as (y − ŷ)²; heavily penalizes large errors.

11
New cards

Absolute Loss

Loss defined as |y − ŷ|; treats all errors linearly.

12
New cards

Zero-One Loss

Loss equal to 0 if prediction equals true value, 1 otherwise; mainly for classification.

13
New cards

Expected Loss

The average loss E[L(y,ŷ)] over the data distribution; used to assess models.

14
New cards

Mean Squared Error (MSE)

Average of squared prediction errors across N cases; measured in squared units of Y.

15
New cards

Root Mean Squared Error (RMSE)

Square root of MSE; expressed in the same units as Y.

16
New cards

Mean Absolute Error (MAE)

Average of absolute prediction errors; same units as Y.

17
New cards

Relative Error Metrics

Unit-less scores obtained by comparing a model’s error to a baseline model’s error.

18
New cards

Normalized Mean Squared Error (NMSE)

Ratio of model SSE to SSE of constant mean predictor; ranges 0–1 (lower is better).

19
New cards

Normalized Mean Absolute Error (NMAE)

Ratio of model SAE to SAE of constant mean predictor; ranges 0–1 (lower is better).

20
New cards

Correlation Coefficient

Statistic ρŷ,y measuring linear association between predictions and true values; ranges −1 to 1.

21
New cards

Coefficient of Determination (R²)

Proportion of variance in Y explained by the model; ranges 0–1, higher is better.

22
New cards

Simple Linear Regression

Regression with one predictor; model of form y = β0 + β1x + ε.

23
New cards

Multiple Linear Regression

Regression with several predictors; model y = β0 + ΣβjXj + ε.

24
New cards

Sum of Squared Errors (SSE)

Total of squared residuals; minimized to estimate linear regression coefficients.

25
New cards

Multicollinearity

Situation where predictors are highly correlated, leading to unstable and hard-to-interpret coefficients.

26
New cards

Regularization

Technique that adds a penalty term to the loss to keep coefficients small and reduce overfitting.

27
New cards

Ridge Regression

L2-regularized linear regression that adds λΣβj²; shrinks coefficients but none to zero.

28
New cards

Lasso Regression

L1-regularized linear regression that adds λΣ|βj|; can shrink some coefficients to zero, performing feature selection.

29
New cards

Gradient Descent

Iterative optimization algorithm that updates parameters in the negative gradient direction to minimize loss.

30
New cards

Learning Rate

Step-size parameter α controlling how far each gradient descent update moves.

31
New cards

Batch Gradient Descent

Variant that computes gradients using the whole training set before each update.

32
New cards

Stochastic Gradient Descent

Variant that updates parameters after each training example.

33
New cards

Mini-batch Gradient Descent

Variant that updates parameters after processing a small batch of examples.

34
New cards

Regression Tree

Decision tree whose leaves output continuous values; built by recursively partitioning predictor space.

35
New cards

Recursive Partitioning

Process of splitting data into subsets by tests to build a decision or regression tree.

36
New cards

Pre-pruning

Stopping tree growth early using criteria like minimum cases or maximum depth to prevent overfitting.

37
New cards

Post-pruning

Growing a large tree then cutting back branches using error estimates (e.g., cross-validation).

38
New cards

Error-Complexity Pruning

CART method that generates a sequence of sub-trees and chooses the best via cross-validation and the x-SE rule.

39
New cards

Support Vector Regression (SVR)

Support vector machine for regression that fits a function within an ε-insensitive tube and minimizes margin.

40
New cards

Epsilon-insensitive Loss

SVR loss that ignores errors smaller than ε and penalizes only the excess.

41
New cards

LOESS

Locally Estimated Scatterplot Smoothing; non-parametric method that fits locally weighted least-squares models.

42
New cards

MARS

Multiple Additive Regression Splines; non-parametric technique using piecewise linear regressions to model nonlinearities and interactions.

43
New cards

k-Nearest Neighbors Regression

Instance-based method that predicts the average target value of the k closest training instances.

44
New cards

Artificial Neural Network (Regression)

Model composed of layers of neurons with activation functions used to approximate complex continuous functions.