org4

Memory Latency: The time it takes to access memory; DRAM is slower than the CPU.

SRAM: Faster memory than DRAM but takes more space and gets hotter.

Wide Path Memory Access: Reads or writes multiple bytes at once to speed up data transfer.

Memory Interleaving: Splits memory into sections so different parts can be accessed at the same time.

N-way Interleaving: Divides memory into n parts; even and odd addresses can be accessed at the same time in 2-way interleaving.

Cache Memory: Small, fast memory between the CPU and main memory to speed things up.

Cache Line: A small block of memory in the cache (8-64 bytes).

Tags: Labels in the cache that point to data in the main memory.

Cache Hit: When the data the CPU needs is found in the cache.

Cache Miss: When the data the CPU needs isn’t in the cache and must be loaded from main memory.

Hit Ratio: The percentage of times the CPU finds data in the cache.

Least Recently Used (LRU): A method to replace old cache data with new data.

Write-Through: Updates both the cache and main memory immediately when data changes.

Write-Back: Updates the main memory later when the cache data is replaced.

Multi-Level Cache: Multiple cache layers (L1, L2, L3) with different speeds and sizes.

Locality of Reference: Most programs access the same small memory area repeatedly, making caching effective.

Shared Cache: A cache shared by multiple CPU cores.

robot