IB Computer Science HL (Case Study 2025)

studied byStudied by 1 person
0.0(0)
Get a hint
Hint

Backpropagation through time (BPTT)

1 / 32

flashcard set

Earn XP

Description and Tags

All 2025 case study terms for IB Computer Science. Imported from computersciencecafe.com

33 Terms

1

Backpropagation through time (BPTT)

A variant of the backpropagation algorithm used for training Recurrent Neural Networks (RNNs), where gradients are propagated backward through time to update the weights.

New cards
2

Bag-of-words

A model used in natural language processing where text is represented as an unordered collection of words, disregarding grammar and word order but keeping track of word frequency.

New cards
3

Biases

Systematic errors in data or algorithms that can lead to unfair or discriminatory outcomes. Biases can occur in various forms, including confirmation, historical, labeling, linguistic, sampling, and selection biases.

New cards
4

Confirmation Bias

A type of bias where data is skewed towards a particular viewpoint or expected outcome, often reinforcing pre-existing beliefs.

New cards
5

Historical Bias

A bias that occurs when training data reflects outdated or historical patterns that may not be relevant to current scenarios, potentially leading to inaccurate predictions.

New cards
6

Labelling Bias

Occurs when the labels applied to training data are subjective, inaccurate, or incomplete, affecting the model's ability to learn correctly.

New cards
7

Linguistic Bias

Bias resulting from training data that favors certain dialects, vocabularies, or linguistic styles, potentially disadvantaging users who use different linguistic forms.

New cards
8

Sampling Bias

Occurs when the training dataset is not representative of the entire population, leading to a model that performs well for certain groups but poorly for others.

New cards
9

Selection Bias

Bias introduced when the training data is not randomly selected but chosen based on specific criteria, potentially missing important variations.

New cards
10

Dataset

A collection of data used to train and evaluate machine learning models. A good dataset is diverse, high-quality, relevant, and up-to-date.

New cards
11

Deep Learning

A subset of machine learning involving neural networks with many layers (deep neural networks) that can learn complex patterns in large datasets.

New cards
12

Graphical Processing Unit (GPU)

Specialised hardware designed to accelerate the processing of large-scale data and complex computations, particularly useful for parallel processing in machine learning tasks.

New cards
13

Hyperparameter Tuning

The process of optimizing the parameters that govern the training of a machine learning model (e.g., learning rate, number of layers) to improve its performance.

New cards
14

Large Language Model (LLM)

Advanced neural networks trained on vast amounts of text data to understand and generate human-like language, such as GPT-3 and Microsoft Co-pilot.

New cards
15

Latency

The delay between a user's query and the chatbot's response. High latency can negatively impact user experience by making the chatbot seem slow.

New cards
16

Long Short-Term Memory (LSTM)

A type of RNN designed to overcome the vanishing gradient problem, using a gating mechanism to retain or forget information over time, enabling the learning of long-term dependencies.

New cards
17

Loss Function

A mathematical function that measures the difference between the predicted output of a model and the actual target output, guiding the optimization process during training.

New cards
18

Memory Cell State

In LSTM networks, the memory cell state represents the information that flows through the network, controlled by input, forget, and output gates to manage long-term dependencies.

New cards
19

Natural Language Processing (NLP)

The field of AI focused on enabling machines to understand, interpret, and generate human language.

New cards
20

Discourse Integration

A stage in NLP where the meaning of a sentence is integrated with the larger context of the conversation to generate coherent and contextually appropriate responses.

New cards
21

Lexical Analysis

The process of breaking down text into individual words and sentences, identifying parts of speech, and preparing it for further processing.

New cards
22

Pragmatic Analysis

Analysing the social, legal, and cultural context of a sentence to understand its intended meaning and implications.

New cards
23

Semantic Analysis

The process of understanding the meaning of words and sentences, going beyond the surface-level structure to interpret the underlying concepts.

New cards
24

Syntactical Analysis (Parsing)

Analysing the grammatical structure of a sentence, identifying the relationships between words and phrases.

New cards
25

Natural Language Understanding (NLU)

A component of NLP focused on understanding the user's input by analysing linguistic features and context.

New cards
26

Pre-processing

The initial step in data preparation, involving cleaning, transforming, and reducing data to improve its quality and make it suitable for training machine learning models.

New cards
27

Recurrent Neural Network (RNN)

A type of neural network designed to handle sequential data, maintaining memory of previous inputs through hidden states to process sequences of data.

New cards
28

Self-Attention Mechanism

A technique in transformer neural networks that captures relationships between different words in a sequence by computing attention weights, enabling better handling of long-term dependencies.

New cards
29

Synthetic Data

Artificially generated data used to supplement real data, covering scenarios that may not be well-represented in the original dataset.

New cards
30

Tensor Processing Unit (TPU)

Custom hardware developed by Google specifically designed to accelerate machine learning workloads, particularly for deep learning models.

New cards
31

Transformer Neural Network (Transformer NN)

A type of neural network that uses a self-attention mechanism for parallel processing of data, particularly effective for natural language processing tasks.

New cards
32

Vanishing Gradient

A problem in training deep neural networks where gradients become very small, making it difficult to update the weights effectively and learn long-term dependencies.

New cards
33

Weights

Parameters in a neural network that are adjusted during training to minimize the loss function and improve the model's predictions.

New cards

Explore top notes

note Note
studied byStudied by 8 people
... ago
5.0(1)
note Note
studied byStudied by 25 people
... ago
5.0(2)
note Note
studied byStudied by 16 people
... ago
5.0(1)
note Note
studied byStudied by 29 people
... ago
5.0(1)
note Note
studied byStudied by 67 people
... ago
4.7(3)
note Note
studied byStudied by 10 people
... ago
5.0(1)
note Note
studied byStudied by 34 people
... ago
5.0(1)
note Note
studied byStudied by 4699 people
... ago
5.0(6)

Explore top flashcards

flashcards Flashcard (69)
studied byStudied by 6 people
... ago
5.0(2)
flashcards Flashcard (81)
studied byStudied by 24 people
... ago
5.0(1)
flashcards Flashcard (31)
studied byStudied by 6 people
... ago
5.0(1)
flashcards Flashcard (75)
studied byStudied by 43 people
... ago
5.0(2)
flashcards Flashcard (335)
studied byStudied by 5 people
... ago
5.0(1)
flashcards Flashcard (20)
studied byStudied by 11 people
... ago
5.0(1)
flashcards Flashcard (30)
studied byStudied by 5 people
... ago
5.0(1)
flashcards Flashcard (33)
studied byStudied by 28 people
... ago
5.0(1)
robot