Lecture 23: Privacy and security 2022

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/17

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

18 Terms

1
New cards

Q: What types of private data does data science often collect?

Medical records, census data, browser history, location data, social media data.

2
New cards

Q: What can be learned from credit card metadata?

Sensitive patterns about a person’s location, identity, and shopping behavior.

3
New cards

Q: What did the study by De Montjoye et al. (2015) find about credit card data?

Four spatiotemporal points can reidentify 90% of individuals.

4
New cards

Q: How does knowing transaction price affect reidentification risk?

Increases reidentification risk by 22% on average.

5
New cards

Q: Are women more or less reidentifiable than men in credit card metadata?

More reidentifiable.

6
New cards

Q: How can medical data combined with voter data breach privacy?

Linking datasets can reveal identities even if medical data is anonymized.

7
New cards

Q: What is one method to anonymize location data?

Coarsening latitude/longitude into larger regions like zipcodes.

8
New cards

Q: What is randomized response used for?

Protecting individual privacy in sensitive surveys.

9
New cards

Q: How does randomized response work?

Individuals answer truthfully or randomly according to a known probability to protect privacy.

10
New cards

Q: What inequality is linked to high-probability accuracy in randomized response?

Chebyshev’s inequality.

11
New cards

Q: What is differential privacy?

A method that ensures the addition or removal of a single data point doesn’t significantly affect outcomes.

12
New cards

Q: What does adversarial machine learning study?

How to secure ML models against attacks like input perturbations.

13
New cards

Q: What famous example shows adversarial ML vulnerabilities?

A panda image slightly perturbed to be misclassified by a neural network.

14
New cards

Q: What are three types of attacks on ML models?

Inversion, extraction, and data poisoning.

15
New cards

Q: What is inversion attack?

Reconstructing sensitive input data from model outputs.

16
New cards

Q: What is extraction attack?

Stealing model parameters or training data.

17
New cards

Q: What is data poisoning?

Maliciously injecting bad data into training to corrupt a model.

18
New cards

Q: What is federated learning?

Training models across decentralized devices without transferring raw data