Comprehensive Machine-Learning Vocabulary Review

0.0(0)
studied byStudied by 0 people
GameKnowt Play
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/63

flashcard set

Earn XP

Description and Tags

These vocabulary flashcards cover the principal concepts, algorithms, and statistical foundations presented across the five units of the Machine Learning lecture notes, providing a targeted review for exam preparation.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

64 Terms

1
New cards

Machine Learning

Programming computers to improve performance on a task T with experience E, measured by a performance metric P.

2
New cards

Supervised Learning

Learning a function that maps inputs to outputs from labeled example pairs (x,y); includes classification and regression.

3
New cards

Unsupervised Learning

Discovering patterns or structure in unlabeled data; common tasks are clustering and association.

4
New cards

Reinforcement Learning

Learning to choose actions to maximize cumulative, often discounted, reward through interaction with an environment.

5
New cards

Version Space

The set of all hypotheses in H that are consistent with the observed training examples.

6
New cards

PAC Learning

Probably Approximately Correct framework that bounds the number of samples needed so a learner outputs a near-optimal hypothesis with high probability.

7
New cards

VC Dimension

A measure of the capacity of a hypothesis space; the largest number of points that can be shattered by hypotheses in H.

8
New cards

Decision Tree

A tree-structured model where internal nodes test features and leaves output a prediction.

9
New cards

ID3 Algorithm

Greedy decision-tree inducer that selects splits using highest information gain based on entropy.

10
New cards

CART

Classification And Regression Trees; builds binary trees using Gini or SSE and supports pruning.

11
New cards

Random Forest

Bagging of many decision trees built on bootstrap samples with feature-subsampling to reduce variance.

12
New cards

Bagging

Bootstrap Aggregating; trains learners on different bootstrap samples and averages/votes their outputs.

13
New cards

Boosting

Sequentially builds learners that focus on mistakes of prior learners and combines them into a strong ensemble.

14
New cards

AdaBoost

Adaptive Boosting algorithm that re-weights training data and combines weak classifiers with weighted voting.

15
New cards

Stacking (Stacked Generalization)

Ensemble method that learns a meta-model to combine predictions of several base models.

16
New cards

Linear Regression

Predicts a continuous target as a linear combination of input features plus noise.

17
New cards

Multiple Linear Regression

Linear regression with two or more independent variables.

18
New cards

Logistic Regression

Models the probability of a binary outcome via the logistic (sigmoid) function.

19
New cards

Perceptron

A linear binary classifier that outputs 1 if w·x > 0, else −1; trained with the perceptron learning rule.

20
New cards

Multilayer Perceptron (MLP)

Feed-forward neural network with one or more hidden layers and nonlinear activation functions.

21
New cards

Activation Function

Non-linear function (e.g., sigmoid, tanh, ReLU) applied to neuron inputs to introduce non-linearity.

22
New cards

Support Vector Machine (SVM)

Large-margin classifier that finds the hyperplane maximizing the margin between classes; uses kernels for non-linear data.

23
New cards

Kernel Function

Computes inner products in high-dimensional feature spaces to enable kernelized algorithms.

24
New cards

k-Nearest Neighbors (k-NN)

Instance-based learner that classifies a query by majority vote (or averages) of its k closest training examples.

25
New cards

K-means Clustering

Partitions data into k clusters by iteratively assigning points to closest centroid then updating centroids.

26
New cards

K-Modes Clustering

Extension of k-means for categorical data using modes and dissimilarity measures.

27
New cards

Ensemble Learning

Combining multiple models to obtain better predictive performance than any constituent.

28
New cards

Information Gain

Reduction in entropy achieved by partitioning the data on an attribute.

29
New cards

Entropy (Shannon)

Measure of impurity/uncertainty in a data set; H(S)=−Σp log₂ p.

30
New cards

Gini Index

Impurity measure used by CART; Σ p(1−p) over classes.

31
New cards

Bias–Variance Trade-off

Decomposition of generalization error into bias (under-fit) and variance (over-fit) components.

32
New cards

Sample Error

Fraction of training instances misclassified by a hypothesis.

33
New cards

True Error

Probability that a hypothesis misclassifies a random instance drawn from the distribution.

34
New cards

Confidence Interval

Range that, with specified probability, contains the true parameter (e.g., error rate).

35
New cards

Bootstrap Sample

Sample of size n drawn with replacement from an original dataset of n instances.

36
New cards

Gaussian Mixture Model (GMM)

Probabilistic model assuming data are generated from a mixture of several Gaussian distributions.

37
New cards

Expectation-Maximization (EM)

Iterative algorithm with E-step (compute expectations) and M-step (maximize) to learn latent-variable models like GMMs.

38
New cards

Epanechnikov Kernel

Quadratic kernel function used for kernel smoothing: k(u)=¾(1−u²) for |u|≤1.

39
New cards

KD-Tree

Space-partitioning data structure that accelerates nearest-neighbor search in k-dimensional space.

40
New cards

Q-Learning

Model-free reinforcement learning algorithm that learns state-action value function Q(s,a).

41
New cards

Temporal-Difference Learning

Updates value estimates based on difference between successive predictions, bridging Monte-Carlo and dynamic programming.

42
New cards

Genetic Algorithm (GA)

Evolutionary search method that evolves a population of candidate solutions using selection, crossover and mutation.

43
New cards

Chromosome (in GA)

Encoded representation (often a bit string) of a candidate solution in a genetic algorithm.

44
New cards

Fitness Function

Numerical measure that evaluates how well a candidate solution solves the problem.

45
New cards

Crossover Operator

GA operation that constructs offspring by exchanging substrings between two parent chromosomes.

46
New cards

Mutation Operator

GA operation that randomly alters genes (bits) in a chromosome to maintain diversity.

47
New cards

Genetic Programming (GP)

Evolutionary technique that evolves computer programs, often represented as syntax trees.

48
New cards

Baldwin Effect

Hypothesis that individual learning can indirectly speed evolution by smoothing the fitness landscape.

49
New cards

Lamarckian Evolution (computational)

Evolutionary model where learned traits are explicitly written back into an individual’s genotype.

50
New cards

Distance Measures

Quantitative metrics (e.g., Euclidean, Manhattan) that define similarity between instances.

51
New cards

Manhattan (City-Block) Distance

Sum of absolute differences across dimensions; L₁ norm.

52
New cards

Euclidean Distance

Square-root of summed squared differences across dimensions; L₂ norm.

53
New cards

Kernel Smoother

Non-parametric regression that averages nearby observations weighted by a kernel function.

54
New cards

Learning Rate (α)

Step-size parameter controlling the magnitude of weight updates in many learning algorithms.

55
New cards

Exploration vs. Exploitation

RL dilemma: choosing between trying new actions to gather information or using known rewarding actions.

56
New cards

Discount Factor (γ)

Value in [0,1) that reduces future rewards in reinforcement learning’s cumulative return.

57
New cards

Markov Decision Process (MDP)

Framework for sequential decision making defined by states, actions, transition function and rewards.

58
New cards

Error-Correcting Output Codes

Technique that decomposes multi-class problems into multiple binary classifiers using a code matrix.

59
New cards

Fitness Proportionate Selection

GA selection strategy where the probability of choosing an individual is proportional to its fitness.

60
New cards

Tournament Selection

GA selection method that picks the better of randomly chosen individuals with a preset probability.

61
New cards

Schema (GA)

Template describing a subset of chromosomes with fixed positions and wildcards; used in schema theorem.

62
New cards

Schema Theorem

Statement that short, low-order, above-average fitness schemas receive exponentially increasing trials in GA.

63
New cards

Bias (in Estimators)

Difference between expected estimate and the true parameter value.

64
New cards

Variance (in Estimators)

Expected squared deviation of an estimator from its own mean; measures estimate fluctuation.