Unit V – Data Processing Technology

studied byStudied by 37 people
5.0(2)
learn
LearnA personalized and smart learning plan
exam
Practice TestTake a test on your terms and definitions
spaced repetition
Spaced RepetitionScientifically backed study method
heart puzzle
Matching GameHow quick can you match all your cards?
flashcards
FlashcardsStudy terms and definitions

1 / 91

flashcard set

Earn XP

92 Terms

1
1
A. Parallel Processing System
New cards
2
Parallel processing
refers to one or more independent operating systems managing multiple processors and performing multiple tasks.
New cards
3
Parallel processing
is very fast and can share the memory unit.
New cards
4
==Flynn’s taxonomy== and classification by memory structure are leading
examples of parallel processing system classification
New cards
5
2
A. Parallel Processing System

2\. Flynn’s classification of parallel processing systems
New cards
6
2\.1 Single instruction stream - single data stream,

==Single Instruction Stream Single Data Stream (SISD)==
is a single processor system that sequentially processes an instruction and data, one at a time.
New cards
7
2\.1 Single instruction stream - single data stream, Single Instruction Stream Single Data Stream (SISD)
It is the conventional computer architecture that follows von Neumann’s concept.
New cards
8
2\.2 Single instruction stream - multiple data stream,

==Single Instruction Stream Multiple Data Stream (SIMD)==
The structure of processing multiple data with an instruction to simultaneously perform the same operation on multiple data.
New cards
9
2\.2 Single instruction stream - multiple data stream,

Single Instruction Stream Multiple Data Stream (SIMD)
It is also called an array processor, as it enables synchronous parallel processing.
New cards
10
2\.3 Multiple instruction streams -single data stream, Multiple Instruction Stream Single Data Stream (MISD)
\-Each processing unit in the ____ parallel computing architecture ==runs different instructions and processes the same data.== The pipeline architecture is an example. It is not a widely used architecture.
New cards
11
2\.4 Multiple instruction streams - multiple data stream,

^^Multiple Instruction Stream Multiple Data Stream (MIMD)^^
\-In a ____ structure, ^^multiple processors process different programs and different data, and most parallel computers fall into this category^^. It can be classified into a shared memory model and a distributed memory model, depending on how it uses the memory
New cards
12
3\.1 Symmetric multiprocessor (SMP)
is ^^a tightly-coupled system in which al processors use the main memory as the shared memory.^^ It is easy to program since the data transfer can use shared memory
New cards
13
3\.2 Massive parallel processor (MPP)
is ^^a distributed memory type in which each processor has an independent memory.^^ The loosely coupled system exchanges data between processors through a network, such as Ethernet.
New cards
14
3\.3 Non uniform memory access (NUMA)
is ^^a structure the combines the advantages of the SMP which is a shared memory structure^^ that makes it easier to develop programs and the MPP structure, which offers excellent scalability.
New cards
15
4\. Types of parallel processor technology

* ==4.1 Instruction pipelining==
^^The technology improves the CPU performance by dividing an operation into several stages^^ and configuring a hardware unit for processing each stage separately in order to process different instructions simultaneously-
New cards
16
instruction fetching (IF), instruction decoding (ID), operand fetching (OF), and execution (EX).
The stages of the four-stage instruction pipeline are
New cards
17
pipeline hazard
^^refers to the pipeline speed exceptionally slowing down.^^ _____ include the data hazard, the control hazard, and the structural hazard.
New cards
18
Data hazards
^^occur when the next instruction execution has to be delayed until the previous instruction has been completed^^ because of the dependency between instruction operands.
New cards
19
Control hazards
are generated by branch instructions, like branch and jump which change the execution order of the instructions.
New cards
20
Structural hazards
are generated when instructions cannot be processed in parallel in the same clock cycle, due to hardware limitations.
New cards
21
5\. Parallel programming technology

* ==5.1 Compiler technology - OpenMP==
^^is a compiler directive-based parallel programming API^^.
New cards
22
The execution model of OpenMP
^^is the fork/join model.^^
New cards
23
5\. Parallel programming technology

* ==5.2 Message passing parallel programming model, MPI==
^^is a parallel programming model suitable to a distributed memory system structure.^^

\
New cards
24
Parallel programming
tools for message passing include ^^High Performance FORTRAN (HPF^^), ^^Parallel Virtual Machine (PVM),^^ and Message Passing Interface (MPI). MPI has become the standard
New cards
25
5\.3 Load balancing technologies -AMP, SMP, and BMP
^^adequately distributes jobs to the cores in order to increase the multi core performance.^^
New cards
26
AMP , SMP , BMP model
_____ An OS ^^is executed independently^^ in each processor core.

______ An OS ^^manages al processor cores simultaneously^^. Application programs ^^can operate in any core.^^

_______ An OS manages al process cores simultaneously, ^^and an application program can run on a specific core^^
New cards
27
6\. Graphic processing technology

* ==6.1 Graphics processing unit (GPU)==
The hardware specializes in computer graphics calculation and is mainly used for the rendering of 3D graphics.
New cards
28
GPU
dedicated to processing large-capacity image data generates results through parallel jobs using multiple cores.
New cards
29
6\.2 General-purpose GPU (GPGPU)
a ^^GPU shows high computational performance in matrix and vector operations^^ that are mostly used for graphic rendering, the computing system intends to utilize GPUs in the general computing domain as well.
New cards
30

They include CUDA and OpeACC from NVIDIA, OpenCL from Khronos Group, and C++ AMP from Microsoft.

Many models supporting GPGPU programming have appeared.
New cards
31
CUDA
____ ==is a parallel computing platform and a programming model that can significantly improve computing speed== with a large number of GPU cores.
New cards
32
CUDA
It provides intuitive GPI programming, based on the C language, and it enables quick operation using shared memory.
New cards
33
CUDA
is ^^expected to show an excellent performance improvement when applied to performing tasks^^ suitable for parallel processing operations in various fields that require a large amount of computation, such as simulation.
New cards
34
7\. GPU-based parallel programming technology

* ==Open Computing Language (OpenCL)==
^^maintained and managed by Khronos Group^^, is an open, general-purpose parallel computing framework developed by Apple, AMD, IBM, Intel, and NVIDIA.
New cards
35
Open Computing Language (OpenCL)
It is an industry standard programming model for heterogeneous computer systems, consisting of GPUs, CPUs, and other processors. 
New cards
36
C++ Accelerated Massive Parallelism (C++ AMP)
^^was developed by Microsoft in an open programming language for heterogeneous computing, using CPU and GPU, C++^^ AMP, when added to Visual Studio 2021, can increase the execution speed of C++ codes using GPU.
New cards
37
C++ Accelerated Massive Parallelism (C++ AMP)
intends to help developers create general-purpose programs using GPU without a high level of understanding or application capability about DirectX API.
New cards
38
OpenACC
NVIDIA introduced ____, ==a programming model based on compiler directives that abstract CUDA==, _____ is a ^^programming model for higher productivity,^^ since it provides a relatively simple programming environment for developers.
New cards
39
Direct attached storage (DAS)
==The storage== ^^connects a computer system with disks directly through a fiber^^ channel or SCSI cable in order to utilize the storage capacity. It allows the computer system to manage the file system directory.
New cards
40
Network attached storage (NAS)
The storage has a separate file system management server (controller) to manage the storage media such as HDD and SSD.
New cards
41
Storage area network (SAN)
^^was developed to overcome the disadvantages^^. It uses a dedicated fiber channel switch for fast connection, and it enabled the ability to scale up the number of connected servers and storage, with less impact to the connected network load.
New cards
42
3\. IP-SAN
This type of SAN uses the gigabit Ethernet Internet protocol (IP), instead of a fiber channel.
New cards
43
Fiber Channel over IP (FCIP)
is used to connect a remote SAN. It encapsulates data to TCP/IP for interconnection when transferring a frame to a remote location.
New cards
44
Internet fiber channel protocol (iFCP)
provides a TCP/IP connection dedicated to regional SAN, using the iFCP gateway.
New cards
45
Internet SCSI (iSCSI)
encapsulates SCSI commands into IP packets and transfer the I/O block data through TCP/IP. Technologies like \n IPSec ensure reliability.
New cards
46
4\. Storage capacity management

* Thin provisioning
The existing fixed-allocation storage technology uses a thick logical unit number (LUN) wasted data storage space.
New cards
47
4\. Storage capacity management

* Data de-duplication
provides a high efficiency of disk space used by removing any duplicated data when saving the data.
New cards
48
5\. Storage disk scheduling

* Disk scheduling
disk drive that stores data is a device using a rotating magnetic disk.
New cards
49
5\. Storage disk scheduling

* Disk scheduling
is a technique of efficiently processing I/O requests, when multiple users request them, in order to process different tasks.
New cards
50
\-Maximization of I/O requests to service during a unit time

\-Maximization throughput per unit time

\-Minimization of the mean response time

\-Minimization of response time

\-Minimization of the variation of response time
Using disk scheduling has the following purposes:
New cards
51
5\. Storage disk scheduling

* Disk performance measurement indicator
can be compared with the indicators that measure disk performance.
New cards
52
Disk performance measurement indicators
include the access time, seeking time, rotational delay or rotational latency, and data transfer time
New cards
53
seeking time
indicates how long it takes to move the head from the current head position, to the track containing the data.
New cards
54
rotational latency
indicates how long it takes from the moment the head begins rotating to move to the track containing the data, to the moment it reaches the sector that contains the data.
New cards
55
data transfer time
indicates how long it takes to transfer the read data to the main memory. This section describes techniques to minimize the access time by minimizing the seeking time and the rotational latency.
New cards
56
First come first serve (FCFS) disk scheduling
services the requests in the order they are received. The head position moves in the order of the requested tracks in the disk standby queue.
New cards
57
Shortest seeking time first (SSRF) disk scheduling
The scheduling technique first services the request that is closest to the current head position, among the requested services waiting in the queue.
New cards
58
SCAN disk scheduling
The scheduling technique first services the request that has the shortest seeking distance from the current direction of the moving head.
New cards
59
LOOK disk scheduling
The technique is the same as the SCAN disk scheduling, except that the head changes its direction before reaching the outermost or innermost cylinder.
New cards
60
5\. Storage disk scheduling

* Circular SCAN (C-SCAN) disk scheduling
The SCAN technique moves the head by connecting the inner and outer tracks in a circular model.
New cards
61
Circular LOOK (C-LOOK) disk scheduling
it is a LOOK scheduling technique that connects the inner and outer tracks in an annular model in order to make the head move.
New cards
62
3
C. High Availability Storage

1\. Redundant array of independent disks (RAID) technology
New cards
63
Large-capacity storage systems
generally have an error controller and backup function to safely store the massive volume of data.
New cards
64
RAID
==is a storage technology that minimizes the factors that can cause failure==, and it improves access performance by arranging a number of disks, and by creating a separate disk unit by linking them with each other
New cards
65
are improved availability, increased capacity, and increased speed.
main features of RAID
New cards
66
4
C. High Availability Storage

1\. Redundant array of independent disks (RAID) technology
New cards
67
RAID-0 (Striped disk array without fault tolerance)
consists of two or more drives and uses disk striping, which stores data by dividing it into pieces of a specific size and saves it on multiple disks at once.
New cards
68
RAID-1 (Mirroring and Duplexing)
uses a mirroring technique that redundantly stores data on two drives. Since data is stored in redundancy, data can be restored, even if a drive fails.
New cards
69
RAID-4
has a separate parity drive and collects and stores parities for data verification and recovery.
New cards
70
RAID-5
 is an improvement of RAID-4 by distributing the load of the drive that stores the parities.
New cards
71
RAID-6 (Stripe set with dual distributed parity)
is similar to RAID-5, except that while RAID-5 stores one parity, RAID-6 redundantly stores a parity in two drives. The configuration is more durable than RAID-5 and can store data safely.
New cards
72
RAID-10(Striping & Mirroring)
requires at least four drives and is a combination of RAID-0 and RAID-1 to improve I/O speed while providing data stability.
New cards
73
Linear tape-open (LTO)
 is a standard open tape drive technology that supports high-speed data processing and a large capacity.
New cards
74
Virtual tape library (VTL)
is a backup solution that emulates disk storage and makes it into a virtual tape device to compensate for problems such as limited performance, scalability, and the recovery time
New cards
75
5
D. Graphic Compression Technology

* 1. Graphic compression type
New cards
76
1\. Graphic compression type
Video data compression, which accounts for most of the traffic in a multimedia network, ==can be divided into lossless compression (reversible compression)== ^^and lossy compression (irreversible compression).^^
New cards
77
Lossless compression
is also called reversible compression.
New cards
78
Graphic compression type
refers to a method of restoring a compressed image without information loss from the original data while decompressing.
New cards
79
Lossy compression
is also called irreversible compression.
New cards
80
Graphic compression type
refers to a compression method when the compressed data is restored, but it does not match the original data before the compression because some data is lost.
New cards
81
Lossless compression
Since the compression and decompression algorithms are exactly the opposite of each other, the compression method preserves the original data's integrity, and no part of the data is lost during processing.
New cards
82
Lossy compression
compromises some accuracy to increase the compression rate, by allowing the loss of redundant or unnecessary data. There are ==two types of lossy compression methods:== ^^prediction coding and transform coding.^^
New cards
83
The prediction coding method
is used for digitizing the analog signal, Instead of separately quantizing the PCM (Pulse Code Modulation) samples, it quantizes the difference.
New cards
84
The transform coding method
transforms a signal from one domain (mainly a time and space domain) to another domain (mainly a frequency domain), then compresses it.
New cards
85
Multimedia data
includes text, image, video, and audio data. The text has the form of plain text and non-linear hypertext.
New cards
86
Unicode
The basic language is _____ for expressing symbols, and it uses a loss less compression method.
New cards
87
Multimedia data
an image is called a still image and refers to a photo, fax page, or a video frame
New cards
88
Multimedia data
In the transformation process, the JPEG uses DCT (Discrete Cosine Transform) in the first stage of compression, and the decompression uses the inverse DCT method.
New cards
89
Multimedia data
The t==ransformation and inverse transformation apply 8 X 8 blocks.==

The quantization process ^^creates integers from the real number of the DCT^^ transform output and converts some values to zero.
New cards
90
Multimedia data
The coding process arranges data in a zigzag order after quantization and before encoder input, then lossless compression is performed using run-length decoding and arithmetic coding
New cards
91
Video compression standard
The Moving Picture Experts Group (MPEG) is an international standardization organization. The official name of the standard is ISO/IEC JTC1/SC29/WG11.
New cards
92
MPEG
created the following compression formats and additional standards.
New cards

Explore top notes

note Note
studied byStudied by 577 people
645 days ago
5.0(12)
note Note
studied byStudied by 69 people
863 days ago
5.0(1)
note Note
studied byStudied by 16 people
868 days ago
5.0(1)
note Note
studied byStudied by 1 person
703 days ago
5.0(1)
note Note
studied byStudied by 1 person
6 days ago
5.0(1)
note Note
studied byStudied by 7 people
759 days ago
5.0(1)
note Note
studied byStudied by 1 person
61 days ago
5.0(1)
note Note
studied byStudied by 12384 people
697 days ago
4.7(51)

Explore top flashcards

flashcards Flashcard (29)
studied byStudied by 9 people
678 days ago
5.0(1)
flashcards Flashcard (20)
studied byStudied by 5 people
456 days ago
5.0(1)
flashcards Flashcard (40)
studied byStudied by 96 people
789 days ago
5.0(1)
flashcards Flashcard (95)
studied byStudied by 211 people
844 days ago
4.7(3)
flashcards Flashcard (136)
studied byStudied by 5 people
544 days ago
5.0(1)
flashcards Flashcard (45)
studied byStudied by 3 people
213 days ago
5.0(1)
flashcards Flashcard (63)
studied byStudied by 36 people
412 days ago
5.0(1)
flashcards Flashcard (172)
studied byStudied by 25 people
39 minutes ago
5.0(2)
robot