ELEC 825 Week 1

studied byStudied by 28 people
5.0(1)
Get a hint
Hint

n-fold cross validation

1 / 26

flashcard set

Earn XP

Description and Tags

basics of ML with a focus on linear classifiers, Naive Bayes, and SVM

27 Terms

1

n-fold cross validation

  • solves problem of noisy estimate of predictive performance when there isn’t enough data for validation

1) split training data into n folds

2) in each run, train on n-1 folds and validate on the remaining 1 fold

3) repeat (2) n times

4) calculate the average performance on the n validation folds

5) select the model that achieves the best average performance

<ul><li><p>solves problem of noisy estimate of predictive performance when there isn’t enough data for validation</p></li></ul><p>1) split training data into n folds</p><p>2) in each run, train on n-1 folds and validate on the remaining 1 fold</p><p>3) repeat (2) n times</p><p>4) calculate the average performance on the n validation folds </p><p>5) select the model that achieves the best average performance</p>
New cards
2

what are the 3 main phases in solving ML problems?

1) training: learning the models.

2) validation: selecting models.

3) testing: evaluating the models you finally obtained above

New cards
3

support vectors

examples closest to the hyperplane

New cards
4

margin

distance between support vectors

<p>distance between support vectors</p>
New cards
5

support vector machine (SVM)

maximum-margin classifier that maps data to a higher dimension space (implicitly using kernel)

<p>maximum-margin classifier that maps data to a higher dimension space (implicitly using kernel)</p>
New cards
6

how do you reduce sensitivity to outliers?

by allowing misclassification during training → bias/variance tradeoff

New cards
7

when is the margin largest?

when the hyperplane is halfway between the 2 samples

New cards
8

how do you classify datasets that aren’t linearly separable?

represent the data into a richer (larger) space which includes some fixed, possibly nonlinear functions of the original inputs/measurements

New cards
9

linear classifier

find the line (or hyperplane) which can “best” (under some criterion/objective) separate two classes:

𝑦(𝑥)=𝑊^𝑇 𝑥+𝑏

New cards
10

what are the differences between regression and classification?

1) regression tries to predict a number while classification predicts a category

2) regression identifies many possible outputs while classification only has a small number of possible outputs

New cards
11

machine learning

computer receives the input and expected output to write a program for solving the problem

New cards
12

when do we need machine learning?

when a) human expertise does not exist, b) models are based on huge amounts of data, c) humans can't explain their expertise, d) models must be customized

New cards
13

supervised learning

given training samples and corresponding desired

outputs (labels), predict outputs on future inputs

New cards
14

what are some examples of supervised learning?

classification, regression, time series prediction

New cards
15

unsupervised learning

  • given training samples without desired outputs, automatically discover representations, features, structures, etc.

  • learn a function f(x) to explore the hidden structure of x

New cards
16

what are some examples of unsupervised learning?

clustering, low-dimensional manifold learning

New cards
17

reinforcement learning

given sequences of states and actions with

(delayed) scalar rewards/punishments, output a policy

New cards
18

what are some examples of what a learning algorithm can do?

recognizing patterns, generating patterns, recognizing anomalies, prediction

New cards
19

regression

given {x_i, y_i} where x_i E R^D (set of all vectors with d real components) and y_i is a real value, learn a function f(x) to predict y given x

<p>given {x_i, y_i} where x_i E R^D (<span style="font-family: -apple-system, system-ui, BlinkMacSystemFont, Segoe UI, Roboto, Oxygen-Sans, Ubuntu, Cantarell, Helvetica Neue, sans-serif">set of all vectors with d real components) and </span>y_i is a real value, learn a function f(x) to predict y given x</p><p></p>
New cards
20

classification

given {x_i, y_i} where x_i E R^D (set of all vectors with d real components) and y_i is categorical, learn a function f(x) to predict y given x

<p>given {x_i, y_i} where x_i E R^D (<span style="font-family: -apple-system, system-ui, BlinkMacSystemFont, Segoe UI, Roboto, Oxygen-Sans, Ubuntu, Cantarell, Helvetica Neue, sans-serif">set of all vectors with d real components) and </span>y_i is categorical, learn a function f(x) to predict y given x</p>
New cards
21

representation learning

  • a form of unsupervised learning that learns a good representation for the inputs, which is then used for subsequent supervised or reinforcement learning

  • compact and low-dimensional representation for inputs

New cards
22

outlier detection

a form of unsupervised learning that detects highly unusual cases from the given data

New cards
23

rule learning

a form of unsupervised learning that learns rules from the data

New cards
24

feature space

multidimensional space where each dimension represents a different feature or characteristic of data

New cards
25

feature normalization

process of transforming numerical features to a standard scale, usually between 0 and 1, to ensure fair comparison and prevent dominance of certain features in machine learning models

New cards
26

what are the steps to Naive Bayes Theorem for discrete features?

1) sort data cases into bins according to Ck

2) compute class prior p(Ck) using frequencies

3) for each class, estimate the distribution of ith variable/feature: 𝑝(𝑥_𝑖│𝐶_𝑘 )

4) compute the posterior

<p>1) sort data cases into bins according to Ck</p><p>2) compute class prior p(Ck) using frequencies</p><p>3) for each class, estimate the distribution of ith variable/feature: 𝑝(𝑥_𝑖│𝐶_𝑘 )</p><p>4) compute the posterior</p><p></p>
New cards
27

how do you deal with continuous features if using Naive Bayes Theorem?

discretization OR fit a known distribution

New cards

Explore top notes

note Note
studied byStudied by 28 people
... ago
5.0(2)
note Note
studied byStudied by 62 people
... ago
4.0(1)
note Note
studied byStudied by 10 people
... ago
5.0(1)
note Note
studied byStudied by 4 people
... ago
5.0(1)
note Note
studied byStudied by 11 people
... ago
5.0(1)
note Note
studied byStudied by 11 people
... ago
4.0(2)
note Note
studied byStudied by 21 people
... ago
5.0(1)
note Note
studied byStudied by 52 people
... ago
5.0(1)

Explore top flashcards

flashcards Flashcard (30)
studied byStudied by 2 people
... ago
4.0(1)
flashcards Flashcard (23)
studied byStudied by 4 people
... ago
5.0(1)
flashcards Flashcard (105)
studied byStudied by 14 people
... ago
5.0(1)
flashcards Flashcard (22)
studied byStudied by 2 people
... ago
5.0(1)
flashcards Flashcard (20)
studied byStudied by 3 people
... ago
5.0(1)
flashcards Flashcard (72)
studied byStudied by 18 people
... ago
5.0(2)
flashcards Flashcard (48)
studied byStudied by 31 people
... ago
5.0(1)
flashcards Flashcard (41)
studied byStudied by 83 people
... ago
5.0(3)
robot