Generative AI – Domain 2 Vocabulary Flashcards

0.0(0)
studied byStudied by 0 people
GameKnowt Play
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/81

flashcard set

Earn XP

Description and Tags

80 vocabulary flashcards covering key Domain 2 Generative AI concepts, AWS services, model architectures, security, RAG, evaluation metrics, and CAF-AI perspectives.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

82 Terms

1
New cards

Generative AI

A subset of deep learning in which models create new, original content (text, images, audio, code) by learning patterns from large datasets.

2
New cards

Foundation Model (FM)

An extremely large, pre-trained neural network with billions of parameters that acts as a base for many downstream tasks.

3
New cards

Parameters

The internal variables a model learns during training; more parameters generally mean greater capacity and capability.

4
New cards

Prompt

User-supplied input (instructions, context, questions, examples) that tells a generative model what to do.

5
New cards

Completion

The output text (or image, etc.) that a generative AI model returns in response to a prompt.

6
New cards

Inference

The run-time process where a trained model uses its knowledge to generate a completion from a prompt.

7
New cards

Prompt Engineering

The skill of designing, structuring, and refining prompts to obtain the desired model output.

8
New cards

In-Context Learning

Technique of providing task examples inside the prompt so the model can mimic them without retraining.

9
New cards

Zero-Shot Learning

Asking the model to perform a task with no examples in the prompt.

10
New cards

One-Shot Learning

Supplying exactly one example of the task inside the prompt.

11
New cards

Few-Shot Learning

Including multiple examples of the task inside the prompt to guide the model.

12
New cards

Transformer Architecture

State-of-the-art neural network design (introduced in “Attention Is All You Need,” 2017) that processes sequences in parallel using self-attention.

13
New cards

Tokenizer

Component that splits human text into tokens and converts them to numeric IDs the model can process.

14
New cards

Token

Basic data unit for an LLM (roughly a word or sub-word) used to measure context window size and pricing.

15
New cards

Vector

Ordered list of numbers representing features of a concept; enables mathematical comparison of similarity.

16
New cards

Embedding

Dense vector representation of a token or item that captures its semantic meaning and context.

17
New cards

Transformer Network

Neural network built from stacked encoder/decoder blocks using self-attention and positional embeddings.

18
New cards

Self-Attention Mechanism

Process that lets a Transformer weigh the importance of every token relative to every other token when generating output.

19
New cards

Positional Embeddings

Extra vectors added to token embeddings to convey each token’s position in the sequence so order is preserved.

20
New cards

Context Window

Maximum number of tokens (prompt + completion) an LLM can handle in a single request.

21
New cards

Encoder (Transformer)

Half of a Transformer that reads the entire input and produces a contextual representation of it.

22
New cards

Decoder (Transformer)

Half of a Transformer that takes encoder context (or previous outputs) and generates output tokens one by one.

23
New cards

Softmax Output Layer

Final function that converts raw model scores into a probability distribution over possible next tokens.

24
New cards

Pre-Training

Compute-intensive initial training phase where the model learns statistical patterns from large, unlabeled data.

25
New cards

Self-Supervised Learning

Training method in which the model generates its own labels (e.g., predicting the next word) instead of using human-labeled data.

26
New cards

Unimodal Model

Generative model that accepts and outputs only one data type (e.g., text-to-text).

27
New cards

Multimodal Model

Model capable of processing and/or generating multiple data types, such as text, images, or audio.

28
New cards

Diffusion Model

Generative model that creates content by reversing a stepwise noising process, refining random noise into coherent output.

29
New cards

Stable Diffusion

Efficient diffusion architecture that performs denoising in a low-dimensional latent space to generate images from text.

30
New cards

Forward Diffusion

Conceptual training process of adding progressive noise to data so the model learns the noise pattern.

31
New cards

Reverse Diffusion

Generative process of starting with noise and iteratively removing it to create new content.

32
New cards

Latent Space

Compressed, abstract feature space where models operate to represent data more efficiently than raw pixels or text.

33
New cards

Retrieval-Augmented Generation (RAG)

Technique that enriches a prompt with retrieved, authoritative data before generation to reduce hallucinations and add freshness.

34
New cards

Knowledge Bases for Amazon Bedrock

Fully managed AWS feature that automates RAG: ingesting data, creating embeddings, storing them, and retrieving context for prompts.

35
New cards

Vector Database

Specialized store that indexes embeddings and returns semantically similar vectors for a query vector.

36
New cards

Ingestion (Knowledge Base)

Process of chunking source documents, generating embeddings, and loading them into a vector database.

37
New cards

Amazon OpenSearch Serverless

Fully managed, pay-per-use AWS vector database option ideal for quick, low-overhead RAG setups.

38
New cards

Pinecone

Purpose-built, high-performance vector database suited for large-scale, low-latency semantic search workloads.

39
New cards

Redis Enterprise Cloud

In-memory database choice for real-time, ultra-low-latency vector search, often used when Redis is already in use.

40
New cards

Amazon Aurora (pgvector)

Relational database (PostgreSQL) with vector search extension, ideal when structured data already resides in Aurora.

41
New cards

MongoDB Atlas

Document database offering vector search; chosen when data is stored in MongoDB JSON documents.

42
New cards

Amazon S3 (RAG Data Source)

Primary storage location where Bedrock Knowledge Bases ingest supported text-centric documents.

43
New cards

Hallucination (LLM)

Model output that is plausible-sounding but factually incorrect or fabricated.

44
New cards

Prompt Injection

Attack where a malicious input causes the model to ignore original instructions and perform unintended actions.

45
New cards

Data Poisoning

Attack that corrupts training data to bias or compromise a model’s behavior.

46
New cards

Model Inversion

Attack attempting to reconstruct private training data by repeatedly querying a model.

47
New cards

ROUGE

Metric that evaluates automatic text summarization quality by comparing model output to reference summaries.

48
New cards

BLEU

Metric that measures machine-translation quality by comparing model output to human reference translations.

49
New cards

Generative Adversarial Network (GAN)

Model consisting of competing generator and discriminator networks that produce high-fidelity synthetic data, especially images.

50
New cards

Variational Autoencoder (VAE)

Encoder–decoder model that learns a latent space to generate new data and allows controlled attribute manipulation.

51
New cards

Reinforcement Learning from Human Feedback (RLHF)

Fine-tuning approach where human-ranked outputs create a reward model to align an LLM with human preferences.

52
New cards

Amazon Bedrock

AWS fully managed service giving API access to multiple foundation models with usage-based pricing.

53
New cards

Amazon SageMaker JumpStart

SageMaker hub offering pre-trained models, notebooks, and 1-click deployments to accelerate ML and generative AI projects.

54
New cards

Amazon Titan

AWS family of foundation models (text and embeddings) available exclusively through Bedrock.

55
New cards

Amazon Q Developer (CodeWhisperer)

Generative AI coding assistant that produces code suggestions directly in an IDE from natural-language comments.

56
New cards

PartyRock

Playground built on Bedrock that lets users experiment with prompt engineering by rapidly creating small AI apps.

57
New cards

AWS Nitro System

Hardware foundation of modern EC2 instances providing isolated, hardware-enforced security for customer workloads.

58
New cards

AWS Trainium

AWS-designed chip optimized for high-performance, cost-efficient training of large ML models.

59
New cards

AWS Inferentia

AWS-designed chip optimized for high-throughput, low-cost inference of ML models.

60
New cards

Transfer Learning

Method of starting with a pre-trained model and fine-tuning it on a smaller, domain-specific dataset.

61
New cards

CAF-AI (Cloud Adoption Framework for AI)

AWS strategic framework guiding organizations across six perspectives to scale AI responsibly and effectively.

62
New cards

CAF-AI Business Perspective

Focuses on aligning AI initiatives with measurable business outcomes and ROI.

63
New cards

CAF-AI People Perspective

Addresses workforce skills, culture, and change management for AI adoption.

64
New cards

CAF-AI Governance Perspective

Ensures responsible, ethical, and compliant AI through policies and risk management.

65
New cards

CAF-AI Platform Perspective

Covers technology architecture, MLOps pipelines, and scalable infrastructure for AI workloads.

66
New cards

CAF-AI Security Perspective

Protects data, models, and intellectual property against threats unique to AI systems.

67
New cards

CAF-AI Operations Perspective

Defines processes for running, monitoring, and continuously improving AI systems in production.

68
New cards

High Availability (HA)

Design goal of minimizing downtime so a system stays accessible (e.g., 99.99 % uptime).

69
New cards

Fault Tolerance (FT)

Capability of a system to keep operating without interruption even when components fail.

70
New cards

AWS Region

Geographically isolated AWS area containing multiple Availability Zones; key to disaster-recovery strategies.

71
New cards

Availability Zone (AZ)

Physically separate data-center cluster within a Region; applications spanning multiple AZs gain HA and FT.

72
New cards

Edge Location

AWS Point of Presence used by CloudFront and Global Accelerator to cache or route traffic closer to users.

73
New cards

Vector Embeddings Model

Model (e.g., Amazon Titan Text Embeddings) that converts text into high-dimensional numeric vectors for similarity search.

74
New cards

Token-Based Pricing

Pay-per-use cost model where charges depend on the number of tokens processed (input + output).

75
New cards

Self-Hosting (LLM)

Running a model on your own EC2/GPU infrastructure, incurring 24/7 compute costs and operational overhead.

76
New cards

Temperature (LLM Parameter)

Inference setting controlling randomness of output; higher values yield more creative but less deterministic text.

77
New cards

Top-p (Nucleus) Sampling

Decoding method where the model samples from the smallest set of top probable tokens whose cumulative probability exceeds p.

78
New cards

Embedding Layer

First neural-network layer that maps discrete token IDs to learned dense vectors.

79
New cards

Statelessness (LLM)

Property that the model does not retain conversational memory between separate calls unless explicitly provided.

80
New cards

Grounding (RAG)

Supplying external, authoritative data to an LLM so it can generate fact-based, context-relevant answers.

81
New cards

Chunking (RAG)

Splitting large documents into smaller text pieces before embedding and storing them for retrieval.

82
New cards

Soft Prompt Tuning

Lightweight fine-tuning approach that learns a small set of prompt tokens instead of updating the entire model.