AI Evaluation and Fairness Flashcards

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/32

flashcard set

Earn XP

Description and Tags

Flashcards covering key vocabulary related to AI evaluation and fairness.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

33 Terms

1
New cards

Fairness

Treating all individuals and groups equitably, without favoring one group over another.

2
New cards

Bias

Producing systematically prejudiced outcomes.

3
New cards

Traceability

Ability to track and document the decision-making processes, data sources, and changes made to an AI model throughout its lifecycle.

4
New cards

Explainability

Ability of an AI system to provide understandable reasons or explanations for its outputs or decisions.

5
New cards

Responsibility

Assigning accountability for the actions and outcomes of AI systems, entails mitigating risks such as harm or misuse.

6
New cards

Liability

Legal responsibility assigned to individuals or organizations for the actions, outcomes, or consequences of an AI system.

7
New cards

Gender Bias

NLP models associate certain professions or behaviors with specific genders.

8
New cards

Racial Bias

NLP models underperform on languages or dialects spoken by minority groups.

9
New cards

Cultural Bias

NLP chatbot fails to understand cultural nuances.

10
New cards

Data Transparency

Knowing the sources of data used to train NLP models.

11
New cards

Versioning

Keeping track of changes in an NLP model.

12
New cards

Reproducibility

Allows developers to reproduce results and check for consistency over time.

13
New cards

Explainability Benefit

Helps users trust NLP systems by explaining outputs.

14
New cards

Content Moderation

NLP models used in social media platforms need to avoid spreading misinformation or hate speech.

15
New cards

Ethical Model Development

Developers must ensure that datasets used for training NLP models are ethically sourced and do not contain biases.

16
New cards

Corrective Mechanisms

Implementing processes to fix or update models when biased or unethical behavior is identified.

17
New cards

Adversarial Examples

Inputs that have been deliberately perturbed to mislead a model.

18
New cards

Synonym Substitution

Changing words to synonyms to confuse a model.

19
New cards

Character-level Attacks

Introducing typos to confuse a model.

20
New cards

Paraphrasing

Rephrasing text to confuse a model.

21
New cards

Clean Data

Remove duplicates, irrelevant content, or highly noisy data.

22
New cards

Cohen's kappa agreement

Take into account the agreement by chance.

23
New cards

Baseline

A simple model or method used as a point of comparison to evaluate the performance of more advanced models.

24
New cards

Random Baseline

Assigns labels randomly.

25
New cards

Majority Class Baseline

Always predicts the most frequent class in the dataset.

26
New cards

BLEU (Bilingual Evaluation Understudy)

Calculates the overlap of n-grams between the machine-generated translation and the reference translation.

27
New cards

ROUGE-N

Measures overlap of n-grams.

28
New cards

ROUGE-L

Measures the longest common subsequence (LCS).

29
New cards

MMLU (Massive Multitask Language Understanding)

Tests knowledge and reasoning across 57 academic subjects.

30
New cards

HellaSwag

Tests commonsense reasoning and narrative plausibility.

31
New cards

TruthfulQA

Tests truthfulness, especially in response to tricky or adversarial questions.

32
New cards

BIG-bench (Beyond the Imitation Game)

A giant collection of 200+ tasks to probe creativity, reasoning, morality.

33
New cards

HLE (Humanity’s Last Exam)

A new gold-standard benchmark to test whether AI systems have reached expert human reasoning.