AI & Cybersecurity – Core Vocabulary

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/54

flashcard set

Earn XP

Description and Tags

Essential terms and definitions spanning AI foundations, deep-learning architectures, optimization, transformers, tokenisation, anomaly-detection algorithms and key cybersecurity applications.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

55 Terms

1
New cards

Artificial Intelligence (AI)

The field devoted to creating systems able to emulate human reasoning, learning and problem-solving without hard-coded rules.

2
New cards

Machine Learning (ML)

A subset of AI that learns patterns from data to make decisions or predictions.

3
New cards

Neural Network (NN)

A data-driven model composed of interconnected layers of nodes (neurons) that process information via weighted links.

4
New cards

Perceptron

The simplest neural-network unit; computes a weighted sum of inputs, applies an activation function and outputs a value.

5
New cards

Activation Function

A non-linear function (e.g., ReLU, Sigmoid) applied to a neuron’s weighted sum to introduce non-linearity into the model.

6
New cards

ReLU (Rectified Linear Unit)

Activation: f(x)=max(0,x). Fast, non-saturating and widely used in deep nets.

7
New cards

Sigmoid

Activation squashing real numbers into (0,1); useful for probabilities but prone to vanishing gradients.

8
New cards

Deep Learning (DL)

Branch of ML using neural networks with three or more hidden layers to learn hierarchical features.

9
New cards

Shallow Learning

Traditional ML relying on manual feature engineering and simpler models with few layers.

10
New cards

Supervised Learning

ML paradigm using labelled data to learn a mapping from inputs to known outputs.

11
New cards

Unsupervised Learning

ML paradigm discovering hidden patterns in unlabelled data (e.g., clustering, anomaly detection).

12
New cards

Semi-Supervised Learning

Technique combining a small set of labelled data with a large set of unlabelled data to improve learning.

13
New cards

Self-Supervised Learning (SSL)

Learning approach that creates surrogate labels from the data itself (e.g., mask prediction) to learn representations without external labels.

14
New cards

Reinforcement Learning (RL)

Learning paradigm where an agent interacts with an environment, receiving rewards to learn optimal actions over time.

15
New cards

Federated Learning (FL)

Collaborative training of a shared global model from multiple decentralised devices without exchanging raw data.

16
New cards

Transfer Learning (TL)

Technique that adapts a model pre-trained on one task to a related but different task with little new data.

17
New cards

Generative Model

A model that learns the data distribution in order to create synthetic, previously unseen samples.

18
New cards

Adversarial Machine Learning

Field studying how ML models can be fooled (e.g., data poisoning, adversarial examples) and how to defend them.

19
New cards

Gradient Descent (GD)

Optimization algorithm that iteratively adjusts parameters in the direction opposite to the loss gradient.

20
New cards

Stochastic Gradient Descent (SGD)

GD variant updating parameters using randomly selected mini-batches, adding stochasticity and faster convergence.

21
New cards

Momentum

Optimization enhancement that accumulates past gradients to smooth updates and reduce oscillations.

22
New cards

ADAM

Adaptive Moment Estimation; optimizer combining momentum and per-parameter adaptive learning rates.

23
New cards

Batch Size

Number of samples processed before the model’s parameters are updated once.

24
New cards

Underfitting

Modeling error where a model is too simple to capture the underlying pattern, yielding high training loss.

25
New cards

Overfitting

Modeling error where a model fits training data too closely, performing poorly on unseen data.

26
New cards

Feed-Forward Neural Network (FFNN)

Network where information flows in one direction from input to output without cycles.

27
New cards

Recurrent Neural Network (RNN)

Network designed for sequence data, using hidden state to capture temporal dependencies.

28
New cards

Long Short-Term Memory (LSTM)

RNN variant with gated cells (input, forget, output) mitigating vanishing gradients and learning long-term dependencies.

29
New cards

Convolutional Neural Network (CNN)

Architecture using convolution, activation and pooling layers to extract hierarchical features, especially from images.

30
New cards

Graph Neural Network (GNN)

Model that propagates and aggregates information over graph structures to learn node, edge or graph embeddings.

31
New cards

Attention Mechanism

Neural component that weighs the importance of different inputs, enabling models to focus on relevant parts.

32
New cards

Transformer

Encoder-decoder architecture built on self-attention layers, enabling parallel sequence processing and long-range context.

33
New cards

Positional Encoding

Numerical encoding added to transformer inputs to inject information about token positions.

34
New cards

Tokenizer

Module that splits text into tokens, maps tokens to IDs, handles padding/truncation for transformer inputs.

35
New cards

Word2Vec

Technique that learns fixed, non-contextual word embeddings via Skip-Gram or CBOW objectives.

36
New cards

Skip-Gram Model

Word2Vec variant predicting surrounding words given a central word, learning embeddings from co-occurrence.

37
New cards

Masked Language Modeling (MLM)

Pre-training task (e.g., BERT) where random tokens are masked and the model predicts the original words.

38
New cards

Autoencoder (AE)

Neural network trained to reconstruct its input; encodes data to a latent space then decodes back.

39
New cards

Variational Autoencoder (VAE)

Probabilistic AE that learns a latent distribution, enabling generative sampling and regularised representations.

40
New cards

Denoising Autoencoder (DAE)

AE trained to reconstruct clean data from corrupted inputs, learning robust features.

41
New cards

Contrastive Learning

Self-supervised approach that pulls similar (positive) pairs together and pushes dissimilar (negative) pairs apart in embedding space.

42
New cards

Anomaly Detection

Process of identifying data instances that deviate significantly from normal patterns.

43
New cards

One-Class SVM

SVM variant that learns the boundary of normal data and flags points outside as anomalies.

44
New cards

K-Means

Partitional clustering algorithm assigning data to K clusters by minimizing within-cluster variance.

45
New cards

DBSCAN

Density-based clustering algorithm that groups dense regions and labels sparse points as noise/outliers.

46
New cards

Local Outlier Factor (LOF)

Algorithm identifying local density anomalies by comparing a point’s density to that of its neighbors.

47
New cards

Principal Component Analysis (PCA)

Linear dimensionality-reduction method projecting data onto orthogonal components with maximum variance.

48
New cards

t-SNE

Non-linear dimensionality-reduction technique that preserves local similarities for visualising high-dimensional data.

49
New cards

Reconstruction Error

Difference between original input and autoencoder output; high error often signals anomalies.

50
New cards

Latent Space

Compressed representation learned by an encoder, capturing salient features of input data.

51
New cards

Self-Attention

Attention applied within a single sequence, letting tokens weigh relevance of other tokens in the same sequence.

52
New cards

Cross-Attention

Attention where queries come from the decoder and keys/values from the encoder, enabling encoder-decoder interaction.

53
New cards

Causal Masking

Mask used in transformer decoders to prevent a position from attending to future tokens.

54
New cards

Tokenizer Vocabulary

Set of all tokens a tokenizer recognizes; size dictates embedding matrix dimensions.

55
New cards

Perplexity (t-SNE)

Parameter controlling balance between local and global structure when embedding data with t-SNE.