CSE 4310 quiz 1

0.0(0)
studied byStudied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/113

encourage image

There's no tags or description

Looks like no tags are added yet.

Last updated 5:42 PM on 2/2/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

114 Terms

1
New cards

What course/topic is this deck introducing?

Computer Vision (CSE4310) and course overview.

2
New cards

What is computer vision (CV)?

Teaching machines to interpret/understand visual data (images/video).

3
New cards

What’s the basic CV pipeline?

Image/video → sensing (camera/eye) → interpretation (computer/brain) → meaning.

4
New cards

Why is computer vision interdisciplinary?

It combines CS + math + physics + engineering + perception/biology + ML/AI.

5
New cards

What’s the “core gap” CV tries to bridge?

Pixels/numbers → real-world meaning (objects, actions, scenes).

6
New cards

Why is vision considered “important” for humans?

Huge portion of the brain processes visual information.

7
New cards

Why use machines for vision?

Automation, precision, speed, and sensing beyond human vision.

8
New cards

What is “machine vision” in industry?

Automated inspection/quality control using cameras + algorithms.

9
New cards

Example of CV in shopping/retail?

Recognizing products (like produce) for automated checkout.

10
New cards

Face detection vs face analysis—what’s the difference?

Detection finds faces; analysis classifies attributes (smile/age/etc.).

11
New cards

What does a “face makeover” app need to do?

Locate face, align features/landmarks, then modify appearance realistically.

12
New cards

What’s the core idea behind leaf identification apps?

Segment the leaf + extract features + classify against a database.

13
New cards

What does Word Lens-style translation require?

Detect text (OCR) + track the scene + translate + overlay results.

14
New cards

What is the “selling point” of real-time translation overlays?

Translation appears directly on the live camera view (AR feel).

15
New cards

How is the football first-down line possible?

Camera calibration + tracking + field model + consistent overlay.

16
New cards

What does car “night vision” use CV for?

Detecting pedestrians/objects using sensors like infrared.

17
New cards

What is an “around view” parking system?

Stitching multiple cameras into a composite bird’s-eye view.

18
New cards

What’s a key challenge in bird’s-eye camera systems?

Aligning/stitching views correctly (geometry + calibration).

19
New cards

Why did vision in cars accelerate (around 2015)?

Better sensors + compute + deep learning improved detection/understanding.

20
New cards

What is image stitching used for?

Building panoramas or merged views by aligning overlapping images.

21
New cards

What is Photosynth-style reconstruction about?

Using many photos to build navigable/3D-like scene representations.

22
New cards

What does “virtual fitting” require?

Body/pose estimation + realistic overlay/warping of clothing.

23
New cards

Why is CV important for VR interaction?

Tracking hands/objects/body for control in 3D environments.

24
New cards

What’s the big point of “DeepFace”-type systems?

Deep learning learns strong face features for recognition.

25
New cards

What does CVPR attendance growth suggest?

The field is rapidly growing (more researchers/industry interest).

26
New cards

What does CVPR paper growth suggest?

More published research → fast progress + fast-moving topics.

27
New cards

What is “low-level vision”?

Pixel-level operations (filtering, edges, corners, noise reduction).

28
New cards

What is edge detection?

Finding intensity discontinuities often corresponding to boundaries.

29
New cards

What is circle detection an example of?

Detecting geometric shapes/features (often via voting methods like Hough).

30
New cards

What is “mid-level vision”?

Grouping pixels into regions/objects (segmentation, tracking).

31
New cards

What is segmentation?

Partitioning an image into meaningful regions (object vs background, etc.).

32
New cards

What is tracking?

Following an object’s position/motion across video frames.

33
New cards

What is “high-level vision”?

Understanding: recognition, 3D geometry, pose, and scene reasoning.

34
New cards

What is object recognition?

Identifying/classifying objects (labeling what is in the image).

35
New cards

What is scene understanding?

Interpreting the whole scene: objects + relationships + context.

36
New cards

What is the image formation pipeline?

Light → lens → sensor → electrical signal → digitization → image.

37
New cards

What does the pinhole camera model explain?

How 3D points project onto a 2D image plane (basic geometry).

38
New cards

Film vs digital—main difference?

Film chemically records light; digital uses sensor arrays of cells.

39
New cards

How is a grayscale image represented?

2D grid (matrix) of intensity values.

40
New cards

How is a color image represented?

3 channels (R,G,B) → 3 matrices (or one 3D array).

41
New cards

Name common color models.

RGB, CMYK, HSV (and others).

42
New cards

RGB is additive—what does that mean?

Colors add as light; combining channels increases brightness.

43
New cards

CMYK is subtractive—what does that mean?

Ink removes light; used in printing (K improves blacks).

44
New cards

Why use HSV sometimes?

Separates hue/saturation/brightness, often easier for color-based rules.

45
New cards

What is raster graphics?

Pixel grid images; scaling up can pixelate.

46
New cards

What is vector graphics?

Shapes defined mathematically; scale without pixelation.

47
New cards

Lossy vs lossless compression?

Lossy throws away info (smaller); lossless keeps exact data.

48
New cards

What is an alpha channel?

Transparency information (RGBA) used for compositing.

49
New cards

What’s the main idea of image file formats?

Different tradeoffs: size, quality, transparency, compression, purpose.

50
New cards

JPEG—best for what, and what’s the downside?

Great for photos; lossy and no transparency.

51
New cards

GIF—what are its key traits?

Limited colors (palette), supports animation; good for simple graphics.

52
New cards

PNG—why is it popular?

Lossless + supports transparency; good for web graphics/screenshots.

53
New cards

TIFF—why use it?

High-quality, flexible; common in professional imaging/printing.

54
New cards

RAW images—what are they?

Sensor data (“digital negative”), high bit depth, needs processing.

55
New cards

What’s the purpose of the credits slide?

Sources/attribution for slide materials and authors.

56
New cards

What’s the main topic of this deck?

Image filtering: transforming images via math operations.

57
New cards

Conceptually, what is an image in CV?

Data that can be modeled and manipulated mathematically.

58
New cards

How can we represent an image digitally?

As an array/grid of numbers (pixel values).

59
New cards

If values go 0–255, how many bits is that?

8 bits (because 2^8 = 256).

60
New cards

Filtering vs warping?

Filtering changes pixel values; warping changes pixel locations.

61
New cards

Image as a function—what does f(x) mean?

x = [u,v] (pixel coordinate) → intensity value.

62
New cards

What does “range of f(x)” refer to?

The set of possible intensity outputs (e.g., 0–255).

63
New cards

Typical range for an 8-bit grayscale image?

256 levels (0–255).

64
New cards

What does “discrete domain + discrete range” mean?

Pixel coordinates are integer indices; intensities are quantized values.

65
New cards

How do we model a color image as functions?

Separate functions for R, G, B channels.

66
New cards

What is point processing?

Each pixel is transformed independently: y = T(x).

67
New cards

What is neighborhood filtering?

New pixel value depends on nearby pixels (convolution-like).

68
New cards

What defines point processing in one sentence?

Output pixel uses only its own input pixel value (no neighbors).

69
New cards

Two broad categories of image changes shown so far?

Point processing and neighborhood filtering.

70
New cards

What does identity mapping do?

Leaves pixel values unchanged: y = x.

71
New cards

Darkening transform example?

Subtract a constant and clamp to valid range.

72
New cards

How does linear “lower contrast” work?

Scale intensities toward mid/dark: y = x/2 (example).

73
New cards

Why use nonlinear contrast transforms?

They adjust tones differently in shadows/mids/highlights.

74
New cards

Inversion formula?

y = 255 - x (for 8-bit images).

75
New cards

Lightening transform example?

Add a constant (then clamp).

76
New cards

Linear “raise contrast” example?

Multiply intensities (then clamp), e.g., y = 2x.

77
New cards

What does gamma-like mapping do?

Nonlinear remap that changes perceived brightness/contrast.

78
New cards

Examples of stylized point processing?

Sepia, posterization, and other tone remaps.

79
New cards

What is a box filter?

A neighborhood average filter (simple blur).

80
New cards

What’s the 3×3 box filter kernel?

All ones scaled by 1/9 (equal weights).

81
New cards

What is convolution in images (big idea)?

Slide a kernel over the image, compute weighted sums.

82
New cards

Why do early outputs become 0 in the example?

The neighborhood overlaps mostly zeros/background.

83
New cards

What changes as the kernel moves toward a bright region?

Averages increase as more bright pixels enter the window.

84
New cards

Why do we get intermediate values like 10, 20, etc.?

Neighborhood contains a mix of dark and bright pixels.

85
New cards

What does “smoothing an edge” mean numerically?

Sharp step becomes a gradual ramp via averaging.

86
New cards

Why does box filtering blur images?

It replaces each pixel with the mean of its neighbors.

87
New cards

Why do edges become less sharp after blur?

Neighbor mixing reduces high-frequency detail.

88
New cards

What happens deep inside a uniform bright region after box filter?

Output stays close to the original brightness (average ≈ same value).

89
New cards

Why do boundaries look different from interiors?

Boundary neighborhoods include outside/background values.

90
New cards

What is a common border-handling issue in convolution?

Padding choice affects results near edges.

91
New cards

Box filtering is 2D—what does that imply?

It smooths horizontally and vertically (and diagonally via kernel).

92
New cards

One sentence: what does a blur filter do to noise?

Reduces random variation by averaging (but may smear details).

93
New cards

Why is box blur sometimes considered “crude”?

Equal weights can look less natural than tapered weighting.

94
New cards

Big takeaway of the long numeric example sequence?

Convolution = repeated weighted averaging across the grid.

95
New cards

What’s the visual signature of a box filter?

Uniform smoothing; edges soften with a “flat” averaging feel.

96
New cards

What’s a typical use of box blur?

Quick smoothing/noise reduction (cheap computation).

97
New cards

What’s a typical downside of box blur?

Removes fine details and smears edges.

98
New cards

In the example, why does intensity “spread” outward?

Averaging blends bright pixels into neighbors over space.

99
New cards

When the kernel leaves the bright area, what happens?

Output decreases gradually (reverse of entering).

100
New cards

Final box-filter result described in words?

A blurred version where sharp boundaries become smooth gradients.