1/32
2025 case study key terminology
Name | Mastery | Learn | Test | Matching | Spaced |
---|
No study sessions yet.
Backpropagation Through Time (BPTT)
An algorithm for training Recurrent Neural Networks (RNNs) by unfolding them through time and applying backpropagation at each time step to update weights.
Bag-of-Words
A text representation model that treats a document as a set of individual words, ignoring grammar and order, but keeping word frequency.
Confirmation Bias
A type of AI bias where a model tends to favor data or interpretations that confirm existing beliefs or assumptions.
Historical Bias
A type of AI bias inherited from past data, where the model reflects inequalities or patterns that existed historically.
Labelling Bias
A type of AI bias caused by inconsistently or subjectively labeled data, leading to misleading model learning.
Linguistic Bias
A type of AI bias where language differences like dialects or phrasing affect how the model interprets inputs.
Sampling Bias
A type of AI bias that arises when the data used is not representative of the population it’s meant to model.
Selection Bias
A type of AI bias caused by how data is chosen, potentially excluding important information or skewing the results.
Dataset
A structured collection of data used for training or evaluating machine learning models, often consisting of input-output pairs.
Deep Learning
A subset of machine learning that uses neural networks with multiple layers to learn complex features and patterns.
Graphical Processing Unit (GPU)
A specialized processor optimized for handling parallel operations, crucial for training deep learning models quickly.
Hyperparameter Tuning
The process of adjusting external parameters of a model (like learning rate or batch size) to improve its performance.
Large Language Model (LLM)
A type of deep learning model trained on massive text datasets to perform language tasks such as generation and comprehension.
Latency
The time delay between a user’s input and the chatbot’s response; lower latency improves user experience.
Long Short-Term Memory (LSTM)
A special type of RNN that uses memory cells and gates to retain long-term dependencies in sequential data.
Loss Function
A mathematical function that measures the error between predicted outputs and actual values during training, guiding weight updates.
Memory Cell State
A component of LSTM networks that holds and updates the internal memory of the network over time.
Natural Language Processing (NLP)
A field of artificial intelligence focused on enabling machines to understand, interpret, and generate human language; includes lexical analysis, syntactical analysis, semantic analysis, discourse integration, and pragmatic analysis.
Discourse Integration
An NLP subfield that deals with how the meaning of one sentence depends on previous sentences in a conversation or text.
Lexical Analysis
An NLP subfield focused on breaking down text into individual words or tokens and analyzing them.
Pragmatic Analysis
An NLP subfield that interprets text based on context and intended meaning, such as sarcasm or indirect requests.
Semantic Analysis
An NLP subfield that focuses on determining the meaning of words and sentences, including resolving ambiguities.
Syntactical Analysis (Parsing)
An NLP subfield that examines the grammatical structure of sentences to understand sentence syntax.
Natural Language Understanding (NLU)
A subfield of NLP that focuses on the interpretation and extraction of meaning from human language input.
Pre-processing
The steps taken to clean and prepare raw data for model input, such as tokenization, removing stop words, or lowercasing text.
Recurrent Neural Network (RNN)
A neural network model where connections form loops, allowing it to process sequences and remember prior inputs.
Self-Attention Mechanism
A method used in transformer models to weigh the importance of different words in a sequence, regardless of position.
Synthetic Data
Artificially generated data used to augment training datasets or address issues like bias and data scarcity.
Tensor Processing Unit (TPU)
A hardware chip designed by Google to accelerate deep learning workloads, especially for tensor operations.
Transformer Neural Network (Transformer NN)
A deep learning architecture that uses self-attention to process sequences more efficiently than RNNs or LSTMs.
Vanishing Gradient
A problem during deep network training where gradients become too small to update early layers effectively, hindering learning.
Weights
Parameters in a neural network that determine the strength of connections between neurons and are adjusted during training to minimize loss.
Hyperparameters
Settings or configurations set before training a machine learning model (like learning rate or number of layers) that control how the model learns