Exam 3 - Daulath

0.0(0)
studied byStudied by 3 people
GameKnowt Play
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/23

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

24 Terms

1
New cards

Gradient Descent

To minimize a loss function by iteratively updating parameters in the direction of the negative gradient.

2
New cards

Key hyperparameters in gradient descent

Initial values, Learning rate (𝜂), Number of epochs, Error/loss function.

3
New cards

Learning rate

The size of the steps taken towards the minimum during each iteration.

4
New cards

Learning rate too small

The algorithm converges very slowly.

5
New cards

Learning rate too large

The algorithm may overshoot the minimum or fail to converge.

6
New cards

Plotting training and validation error

To detect overfitting or underfitting and decide when to stop training.

7
New cards

Identifying overfitting

When training error decreases, but validation error increases.

8
New cards

Identifying underfitting

When errors fail to converge, indicating the model is not learning.

9
New cards

Elbow in an error plot

The point where the validation error levels off, suggesting the optimal number of epochs.

10
New cards

Mean Squared Error (MSE)

The average of squared differences between predicted and actual values: MSE = (1/n) ∑(y^ - y)².

11
New cards

Mean Absolute Error (MAE)

The average of absolute differences between predicted and actual values: MAE = (1/n) ∑|y^ - y|.

12
New cards

MSE vs MAE

MSE penalizes large errors more heavily by squaring them, while MAE treats all errors equally.

13
New cards

Preference for MAE over MSE

MAE is more robust to outliers and easier to interpret.

14
New cards

Purpose of a validation set

To tune hyperparameters and monitor performance during training.

15
New cards

Purpose of a test set

To evaluate the final model's performance on unseen data.

16
New cards

Secant Method

To find the root of a function using an iterative approximation.

17
New cards

Secant method formula

x_{n+1} = x_n - f(x_n) * (x_n - x_{n-1}) / (f(x_n) - f(x_{n-1})).

18
New cards

Initial requirements for the secant method

Two initial guesses (x_0 and x_1) close to the root.

19
New cards

Secant method vs Newton's method

The secant method doesn't require derivatives, while Newton's method does.

20
New cards

Key advantage of the secant method

It works even when derivatives are difficult to compute.

21
New cards

Secant method approximating derivatives

It uses the slope of the secant line instead of the derivative.

22
New cards

Stopping criterion for the secant method

When the error is below a threshold or consecutive iterations are very close.

23
New cards

Convergence rate of the secant method

Superlinear, faster than bisection but slower than Newton's method.

24
New cards

Why the secant method might fail

If initial guesses are too far from the root or the function is discontinuous near the guesses.