ai study notes

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/76

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

77 Terms

1
New cards

Activation Function

A mathematical function that introduces non-linearity into a neural network, determining the output of a neuron.

2
New cards

Backpropagation

An algorithm for calculating gradients of the loss function with respect to network parameters, enabling optimization.

3
New cards

Convolutional Neural Network (CNN)

A neural network architecture designed for processing grid-like data, commonly used in image recognition.

4
New cards

Dense Layer

A fully connected layer in a neural network where each neuron connects to every neuron in the previous layer.

5
New cards

Dropout

A regularization technique where randomly selected neurons are ignored during training, preventing overfitting.

6
New cards

Embedding Layer

A layer that maps categorical variables to low-dimensional continuous vectors, capturing relationships and similarities.

7
New cards

Epoch

One complete pass through the entire training dataset during model training.

8
New cards

GPU (Graphics Processing Unit)

A specialized processor designed for parallel computations, accelerating deep learning tasks.

9
New cards

Loss Function

A function that measures the error between the model's predictions and the actual target values.

10
New cards

Overfitting

When a model learns the training data too well, failing to generalize to unseen data.

11
New cards

Parameter Sharing

Using the same weights and biases for different parts of the input data, common in CNNs.

12
New cards

Recurrent Neural Network (RNN)

A neural network designed for processing sequential data, maintaining a hidden state to capture temporal dependencies.

13
New cards

Regularization

Techniques to prevent overfitting, such as weight decay or dropout.

14
New cards

Softmax

An activation function that outputs a probability distribution over multiple classes.

15
New cards

TensorBoard

A tool for visualizing and monitoring the training process of deep learning models.

16
New cards

Training Set

A subset of the data used to train the model.

17
New cards

Validation Set

A subset of the data used to evaluate the model's performance during training and tune hyperparameters.

18
New cards

Test Set

A subset of the data used to evaluate the final model's performance on unseen data.

19
New cards

Pretrained Model

A model that has been previously trained on a large dataset, providing a starting point for faster and more effective training.

20
New cards

Loss Function

A mathematical function that quantifies the difference between a model's predictions and the actual target values.

21
New cards

Metric

A human-interpretable measure used to evaluate the performance of a trained model, often different from the loss function.

22
New cards

Transfer Learning

The practice of leveraging a pretrained model for a new task, often involving adapting the model's architecture and fine-tuning its weights.

23
New cards

DataBlock

A blueprint for assembling datasets for deep learning in fastai, defining data input/output types, how to access data items, and more.

24
New cards

Convolution

A mathematical operation that forms the basis of convolutional layers in CNNs, extracting features from data.

25
New cards

Pooling Layers

Layers that downsample the feature maps produced by convolutional layers, reducing spatial dimensions while preserving information.

26
New cards

Learning Rate

A hyperparameter that controls the step size at each iteration while moving toward a minimum of a loss function.

27
New cards

Batch Gradient Descent

Refers to using the entire training dataset to compute the gradient and update parameters in a single step.

28
New cards

Mini-batch Gradient Descent

Involves randomly selecting a small subset of the training data to compute the gradient and update parameters.

29
New cards

Fine-tuning

A transfer learning technique where a pretrained model is trained on a new task to adapt it with additional epochs.

30
New cards

Cosine Similarity

A measure of similarity between two vectors, indicating how similar they are in terms of direction.

31
New cards

Latent Factors

Underlying characteristics that influence user preferences in collaborative filtering models.

32
New cards

Positive Feedback Loop

A process where the output of a system reinforces itself, potentially narrowing content recommendations.

33
New cards

Weight Decay

A regularization technique that discourages large weights in the model to prevent overfitting.

34
New cards

Principal Component Analysis (PCA)

A dimensionality reduction technique used to identify significant directions of variation in data.

35
New cards

Max Pooling

A pooling method that selects the maximum value from a defined pooling window, reducing dimensionality.

36
New cards

Average Pooling

A pooling method that calculates the average value within the pooling window.

37
New cards

DeBERTa

A transformer-based language model designed for various natural language processing tasks.

38
New cards

Autoregressive Model

A model that predicts future sequence values based on past observed values.

39
New cards

Masked Language Modeling

A training technique that involves hiding words in the input so that the model learns to predict them.

40
New cards

Permuted Language Modeling

A training technique that randomizes the order of words so the model learns to predict the original sequence.

41
New cards

Sequence Classification

Assigning a category or label to an entire sequence of text.

42
New cards

Stemmer

A tool that reduces words to their base or root form to simplify text data.

43
New cards

Special Tokens

Tokens added to text to provide specific instructions or information for the model.

44
New cards

Tokenization

The process of converting text into individual tokens or words for processing in machine learning.

45
New cards

Vision Transformers

A model that uses self-attention mechanisms for image recognition tasks.

46
New cards

Keras Functional API

A way to build complex neural network models flexibly by defining layers as functions and chaining them.

47
New cards

Sequential API

An API in Keras that allows adding layers to models in a simple, linear sequence.

48
New cards

Deep and Wide Networks

Neural networks that combine deep learning with wide network capabilities for better generalization.

49
New cards

Stochastic Gradient Descent (SGD)

An optimization algorithm that updates model weights incrementally based on small batches of data.

50
New cards

Feedback Loop

The process where the outcomes of a system feed back into that system, potentially reinforcing biases.

51
New cards

Hugging Face Repository

A platform that hosts pre-trained language models and tools for natural language processing tasks.

52
New cards

Multilingual Models

Models trained on multiple languages to handle and understand diverse language inputs.

53
New cards

Normalization

A preprocessing step that adjusts the values in the dataset to fall within a standard range.

54
New cards

Optimization

The process of adjusting the model's parameters to minimize the loss function during training.

55
New cards

Classification

The task of assigning labels to data points based on learned features from the dataset.

56
New cards

Clustering

A method of grouping data points based on similarity, often used in unsupervised learning.

57
New cards

Feature Extraction

The process of selecting and transforming raw data into informative features for model training.

58
New cards

Regularization Techniques

Methods employed to prevent overfitting in machine learning models, such as dropout or weight decay.

59
New cards

Hyperparameters

Parameters set before training a model, influencing the learning process and structure.

60
New cards

Embedding

A representation of high-dimensional data in a lower-dimensional space that captures relationships.

61
New cards

Neural Network

A computational model composed of interconnected nodes that processes data and generates outputs.

62
New cards

Training Process

The phase where the model learns patterns from data by minimizing a loss function over epochs.

63
New cards

Evaluation Metrics

Quantitative measures used to assess the performance of a machine learning model.

64
New cards

AutoML

Automated machine learning processes that facilitate the construction of models without extensive manual intervention.

65
New cards

Data Augmentation

Techniques to artificially expand the size of a training dataset by creating modified versions of existing data.

66
New cards

Image Recognition

The ability of a system to identify objects, people, places, or actions in images.

67
New cards

Natural Language Processing (NLP)

A field of artificial intelligence focused on the interaction between computers and human language.

68
New cards

Sentiment Analysis

The process of determining the emotional tone behind a series of words, used to analyze customer feedback.

69
New cards

Transfer Learning

A machine learning method where a model developed for one task is reused as the starting point for a model on a second task.

70
New cards

Precision

The ratio of true positive predictions to the total positive predictions made by a model.

71
New cards

Recall

The ratio of true positive predictions to the actual positive cases in the dataset.

72
New cards

F1 Score

A measure of a model's accuracy that considers both precision and recall to provide a balance.

73
New cards

Confusion Matrix

A matrix that summarizes the performance of a classification algorithm by showing true vs. predicted classifications.

74
New cards

Anomaly Detection

The identification of rare items, events, or observations that raise suspicions by differing significantly from the majority of the data.

75
New cards

Artificial Neural Network (ANN)

A network structure inspired by biological neural networks that processes inputs and produces outputs based on learned weights.

76
New cards

Feature Scaling

Methods used to normalize the range of independent variables or features of data.

77
New cards

Prototyping

The process of creating an initial model to test and iterate on an idea or product.