ai study notes

studied byStudied by 0 people
0.0(0)
Get a hint
Hint

Activation Function

1 / 76

encourage image

There's no tags or description

Looks like no one added any tags here yet for you.

77 Terms

1

Activation Function

A mathematical function that introduces non-linearity into a neural network, determining the output of a neuron.

New cards
2

Backpropagation

An algorithm for calculating gradients of the loss function with respect to network parameters, enabling optimization.

New cards
3

Convolutional Neural Network (CNN)

A neural network architecture designed for processing grid-like data, commonly used in image recognition.

New cards
4

Dense Layer

A fully connected layer in a neural network where each neuron connects to every neuron in the previous layer.

New cards
5

Dropout

A regularization technique where randomly selected neurons are ignored during training, preventing overfitting.

New cards
6

Embedding Layer

A layer that maps categorical variables to low-dimensional continuous vectors, capturing relationships and similarities.

New cards
7

Epoch

One complete pass through the entire training dataset during model training.

New cards
8

GPU (Graphics Processing Unit)

A specialized processor designed for parallel computations, accelerating deep learning tasks.

New cards
9

Loss Function

A function that measures the error between the model's predictions and the actual target values.

New cards
10

Overfitting

When a model learns the training data too well, failing to generalize to unseen data.

New cards
11

Parameter Sharing

Using the same weights and biases for different parts of the input data, common in CNNs.

New cards
12

Recurrent Neural Network (RNN)

A neural network designed for processing sequential data, maintaining a hidden state to capture temporal dependencies.

New cards
13

Regularization

Techniques to prevent overfitting, such as weight decay or dropout.

New cards
14

Softmax

An activation function that outputs a probability distribution over multiple classes.

New cards
15

TensorBoard

A tool for visualizing and monitoring the training process of deep learning models.

New cards
16

Training Set

A subset of the data used to train the model.

New cards
17

Validation Set

A subset of the data used to evaluate the model's performance during training and tune hyperparameters.

New cards
18

Test Set

A subset of the data used to evaluate the final model's performance on unseen data.

New cards
19

Pretrained Model

A model that has been previously trained on a large dataset, providing a starting point for faster and more effective training.

New cards
20

Loss Function

A mathematical function that quantifies the difference between a model's predictions and the actual target values.

New cards
21

Metric

A human-interpretable measure used to evaluate the performance of a trained model, often different from the loss function.

New cards
22

Transfer Learning

The practice of leveraging a pretrained model for a new task, often involving adapting the model's architecture and fine-tuning its weights.

New cards
23

DataBlock

A blueprint for assembling datasets for deep learning in fastai, defining data input/output types, how to access data items, and more.

New cards
24

Convolution

A mathematical operation that forms the basis of convolutional layers in CNNs, extracting features from data.

New cards
25

Pooling Layers

Layers that downsample the feature maps produced by convolutional layers, reducing spatial dimensions while preserving information.

New cards
26

Learning Rate

A hyperparameter that controls the step size at each iteration while moving toward a minimum of a loss function.

New cards
27

Batch Gradient Descent

Refers to using the entire training dataset to compute the gradient and update parameters in a single step.

New cards
28

Mini-batch Gradient Descent

Involves randomly selecting a small subset of the training data to compute the gradient and update parameters.

New cards
29

Fine-tuning

A transfer learning technique where a pretrained model is trained on a new task to adapt it with additional epochs.

New cards
30

Cosine Similarity

A measure of similarity between two vectors, indicating how similar they are in terms of direction.

New cards
31

Latent Factors

Underlying characteristics that influence user preferences in collaborative filtering models.

New cards
32

Positive Feedback Loop

A process where the output of a system reinforces itself, potentially narrowing content recommendations.

New cards
33

Weight Decay

A regularization technique that discourages large weights in the model to prevent overfitting.

New cards
34

Principal Component Analysis (PCA)

A dimensionality reduction technique used to identify significant directions of variation in data.

New cards
35

Max Pooling

A pooling method that selects the maximum value from a defined pooling window, reducing dimensionality.

New cards
36

Average Pooling

A pooling method that calculates the average value within the pooling window.

New cards
37

DeBERTa

A transformer-based language model designed for various natural language processing tasks.

New cards
38

Autoregressive Model

A model that predicts future sequence values based on past observed values.

New cards
39

Masked Language Modeling

A training technique that involves hiding words in the input so that the model learns to predict them.

New cards
40

Permuted Language Modeling

A training technique that randomizes the order of words so the model learns to predict the original sequence.

New cards
41

Sequence Classification

Assigning a category or label to an entire sequence of text.

New cards
42

Stemmer

A tool that reduces words to their base or root form to simplify text data.

New cards
43

Special Tokens

Tokens added to text to provide specific instructions or information for the model.

New cards
44

Tokenization

The process of converting text into individual tokens or words for processing in machine learning.

New cards
45

Vision Transformers

A model that uses self-attention mechanisms for image recognition tasks.

New cards
46

Keras Functional API

A way to build complex neural network models flexibly by defining layers as functions and chaining them.

New cards
47

Sequential API

An API in Keras that allows adding layers to models in a simple, linear sequence.

New cards
48

Deep and Wide Networks

Neural networks that combine deep learning with wide network capabilities for better generalization.

New cards
49

Stochastic Gradient Descent (SGD)

An optimization algorithm that updates model weights incrementally based on small batches of data.

New cards
50

Feedback Loop

The process where the outcomes of a system feed back into that system, potentially reinforcing biases.

New cards
51

Hugging Face Repository

A platform that hosts pre-trained language models and tools for natural language processing tasks.

New cards
52

Multilingual Models

Models trained on multiple languages to handle and understand diverse language inputs.

New cards
53

Normalization

A preprocessing step that adjusts the values in the dataset to fall within a standard range.

New cards
54

Optimization

The process of adjusting the model's parameters to minimize the loss function during training.

New cards
55

Classification

The task of assigning labels to data points based on learned features from the dataset.

New cards
56

Clustering

A method of grouping data points based on similarity, often used in unsupervised learning.

New cards
57

Feature Extraction

The process of selecting and transforming raw data into informative features for model training.

New cards
58

Regularization Techniques

Methods employed to prevent overfitting in machine learning models, such as dropout or weight decay.

New cards
59

Hyperparameters

Parameters set before training a model, influencing the learning process and structure.

New cards
60

Embedding

A representation of high-dimensional data in a lower-dimensional space that captures relationships.

New cards
61

Neural Network

A computational model composed of interconnected nodes that processes data and generates outputs.

New cards
62

Training Process

The phase where the model learns patterns from data by minimizing a loss function over epochs.

New cards
63

Evaluation Metrics

Quantitative measures used to assess the performance of a machine learning model.

New cards
64

AutoML

Automated machine learning processes that facilitate the construction of models without extensive manual intervention.

New cards
65

Data Augmentation

Techniques to artificially expand the size of a training dataset by creating modified versions of existing data.

New cards
66

Image Recognition

The ability of a system to identify objects, people, places, or actions in images.

New cards
67

Natural Language Processing (NLP)

A field of artificial intelligence focused on the interaction between computers and human language.

New cards
68

Sentiment Analysis

The process of determining the emotional tone behind a series of words, used to analyze customer feedback.

New cards
69

Transfer Learning

A machine learning method where a model developed for one task is reused as the starting point for a model on a second task.

New cards
70

Precision

The ratio of true positive predictions to the total positive predictions made by a model.

New cards
71

Recall

The ratio of true positive predictions to the actual positive cases in the dataset.

New cards
72

F1 Score

A measure of a model's accuracy that considers both precision and recall to provide a balance.

New cards
73

Confusion Matrix

A matrix that summarizes the performance of a classification algorithm by showing true vs. predicted classifications.

New cards
74

Anomaly Detection

The identification of rare items, events, or observations that raise suspicions by differing significantly from the majority of the data.

New cards
75

Artificial Neural Network (ANN)

A network structure inspired by biological neural networks that processes inputs and produces outputs based on learned weights.

New cards
76

Feature Scaling

Methods used to normalize the range of independent variables or features of data.

New cards
77

Prototyping

The process of creating an initial model to test and iterate on an idea or product.

New cards

Explore top notes

note Note
studied byStudied by 41 people
... ago
5.0(1)
note Note
studied byStudied by 56 people
... ago
5.0(1)
note Note
studied byStudied by 1 person
... ago
5.0(1)
note Note
studied byStudied by 18 people
... ago
5.0(1)
note Note
studied byStudied by 16 people
... ago
5.0(1)
note Note
studied byStudied by 19 people
... ago
5.0(1)
note Note
studied byStudied by 37 people
... ago
5.0(1)

Explore top flashcards

flashcards Flashcard (42)
studied byStudied by 2 people
... ago
5.0(1)
flashcards Flashcard (113)
studied byStudied by 27 people
... ago
5.0(1)
flashcards Flashcard (40)
studied byStudied by 14 people
... ago
5.0(1)
flashcards Flashcard (32)
studied byStudied by 47 people
... ago
5.0(1)
flashcards Flashcard (66)
studied byStudied by 1 person
... ago
5.0(1)
flashcards Flashcard (206)
studied byStudied by 27 people
... ago
5.0(1)
flashcards Flashcard (20)
studied byStudied by 3 people
... ago
5.0(1)
flashcards Flashcard (27)
studied byStudied by 116 people
... ago
5.0(1)
robot