LLM Architecture: The Basics

0.0(0)
studied byStudied by 0 people
0.0(0)
full-widthCall with Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/15

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No study sessions yet.

16 Terms

1
New cards

What is the term for the maximum amount of input text an LLM can process in a single request?

Context Window.

<p>Context Window.</p>
2
New cards

Context Explosion

Exceeding an LLM's context window leads to _, which causes slow responses, high costs, and poor accuracy.

3
New cards

A performance failure where an LLM's reasoning quality drops as context gets too long, causing it to 'forget' early information.

Term: Attention Decay

4
New cards

Truncation.

What LLM failure mode occurs when data exceeding the context window is cut off, causing the model to miss early parts of a conversation?

5
New cards

Retrieval-Augmented Generation.

What does RAG stand for in the context of LLMs?

6
New cards

The metadata is too complex and high-cardinality to be summarized easily.

Why does traditional RAG often 'break' when applied to massive visual datasets?

7
New cards

A design where an LLM receives a lightweight summary or pointer, while the full dataset remains in system memory.

Term: Pass-by-Reference Architecture

8
New cards

Pass-by-value sends raw data to the LLM, while pass-by-reference sends only a lightweight pointer or summary.

What is the key difference between a pass-by-value and a pass-by-reference architecture for LLMs?

9
New cards

Artifact Pattern

The _ is a mechanism where tools return a summary for the LLM and a detailed 'artifact' for the system to store.

10
New cards

Nodes represent images or objects, while Edges represent relationships like similarity or containment.

In a graph-based knowledge system for visual data, what do 'Nodes' and 'Edges' represent?

11
New cards

An architectural shift that treats data as a network of nodes and edges, allowing an LLM to navigate complex data logically.

Term: Graph-based Knowledge System

12
New cards

To selectively show an LLM only the most relevant metadata based on a user's specific question, reducing the information processed.

What is the goal of Query-Aware Sampling?

13
New cards

Agentic Workflows

Designing AI 'agents' that can perform multi-step reasoning and handle their own failure modes is known as creating _.

14
New cards

Data containing millions of unique values, like labels or GPS coordinates, that cannot be easily chunked or embedded.

Term: High-Cardinality Metadata

15
New cards

Noisy Data.

What kind of data, such as duplicates, blurry images, or inconsistent annotations, is known to 'sabotage' AI projects?

16
New cards

To turn raw, 'invisible' files into machine-readable data layers so engineers can focus on model performance.

What is the purpose of creating a Structured Infrastructure for AI data engineering?