A variant of the backpropagation algorithm used for training Recurrent Neural Networks (RNNs), where gradients are propagated backward through time to update the weights.
New cards
2
Bag-of-words
A model used in natural language processing where text is represented as an unordered collection of words, disregarding grammar and word order but keeping track of word frequency.
New cards
3
Biases
Systematic errors in data or algorithms that can lead to unfair or discriminatory outcomes. Biases can occur in various forms, including confirmation, historical, labeling, linguistic, sampling, and selection biases.
New cards
4
Confirmation Bias
A type of bias where data is skewed towards a particular viewpoint or expected outcome, often reinforcing pre-existing beliefs.
New cards
5
Historical Bias
A bias that occurs when training data reflects outdated or historical patterns that may not be relevant to current scenarios, potentially leading to inaccurate predictions.
New cards
6
Labelling Bias
Occurs when the labels applied to training data are subjective, inaccurate, or incomplete, affecting the model's ability to learn correctly.
New cards
7
Linguistic Bias
Bias resulting from training data that favors certain dialects, vocabularies, or linguistic styles, potentially disadvantaging users who use different linguistic forms.
New cards
8
Sampling Bias
Occurs when the training dataset is not representative of the entire population, leading to a model that performs well for certain groups but poorly for others.
New cards
9
Selection Bias
Bias introduced when the training data is not randomly selected but chosen based on specific criteria, potentially missing important variations.
New cards
10
Dataset
A collection of data used to train and evaluate machine learning models. A good dataset is diverse, high-quality, relevant, and up-to-date.
New cards
11
Deep Learning
A subset of machine learning involving neural networks with many layers (deep neural networks) that can learn complex patterns in large datasets.
New cards
12
Graphical Processing Unit (GPU)
Specialised hardware designed to accelerate the processing of large-scale data and complex computations, particularly useful for parallel processing in machine learning tasks.
New cards
13
Hyperparameter Tuning
The process of optimizing the parameters that govern the training of a machine learning model (e.g., learning rate, number of layers) to improve its performance.
New cards
14
Large Language Model (LLM)
Advanced neural networks trained on vast amounts of text data to understand and generate human-like language, such as GPT-3 and Microsoft Co-pilot.
New cards
15
Latency
The delay between a user's query and the chatbot's response. High latency can negatively impact user experience by making the chatbot seem slow.
New cards
16
Long Short-Term Memory (LSTM)
A type of RNN designed to overcome the vanishing gradient problem, using a gating mechanism to retain or forget information over time, enabling the learning of long-term dependencies.
New cards
17
Loss Function
A mathematical function that measures the difference between the predicted output of a model and the actual target output, guiding the optimization process during training.
New cards
18
Memory Cell State
In LSTM networks, the memory cell state represents the information that flows through the network, controlled by input, forget, and output gates to manage long-term dependencies.
New cards
19
Natural Language Processing (NLP)
The field of AI focused on enabling machines to understand, interpret, and generate human language.
New cards
20
Discourse Integration
A stage in NLP where the meaning of a sentence is integrated with the larger context of the conversation to generate coherent and contextually appropriate responses.
New cards
21
Lexical Analysis
The process of breaking down text into individual words and sentences, identifying parts of speech, and preparing it for further processing.
New cards
22
Pragmatic Analysis
Analysing the social, legal, and cultural context of a sentence to understand its intended meaning and implications.
New cards
23
Semantic Analysis
The process of understanding the meaning of words and sentences, going beyond the surface-level structure to interpret the underlying concepts.
New cards
24
Syntactical Analysis (Parsing)
Analysing the grammatical structure of a sentence, identifying the relationships between words and phrases.
New cards
25
Natural Language Understanding (NLU)
A component of NLP focused on understanding the user's input by analysing linguistic features and context.
New cards
26
Pre-processing
The initial step in data preparation, involving cleaning, transforming, and reducing data to improve its quality and make it suitable for training machine learning models.
New cards
27
Recurrent Neural Network (RNN)
A type of neural network designed to handle sequential data, maintaining memory of previous inputs through hidden states to process sequences of data.
New cards
28
Self-Attention Mechanism
A technique in transformer neural networks that captures relationships between different words in a sequence by computing attention weights, enabling better handling of long-term dependencies.
New cards
29
Synthetic Data
Artificially generated data used to supplement real data, covering scenarios that may not be well-represented in the original dataset.
New cards
30
Tensor Processing Unit (TPU)
Custom hardware developed by Google specifically designed to accelerate machine learning workloads, particularly for deep learning models.
New cards
31
Transformer Neural Network (Transformer NN)
A type of neural network that uses a self-attention mechanism for parallel processing of data, particularly effective for natural language processing tasks.
New cards
32
Vanishing Gradient
A problem in training deep neural networks where gradients become very small, making it difficult to update the weights effectively and learn long-term dependencies.
New cards
33
Weights
Parameters in a neural network that are adjusted during training to minimize the loss function and improve the model's predictions.