Lecture 10: Decision trees and random forests 3

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/19

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

20 Terms

1
New cards

Q: What is learning essential for in AI?

Unknown environments where the designer lacks omniscience.

2
New cards

Q: What are the two main uses of learning in systems?

  1. Adapting to unknown environments.
    1. Exposing the agent to reality rather than trying to hand-code every rule.
3
New cards

Q: What does learning modify in an agent?

The decision-making mechanisms to improve performance.

4
New cards

Q: What are the three types of feedback in learning?

• Supervised learning: correct answers given.
• Unsupervised learning: no correct answers given.
• Reinforcement learning: occasional rewards.

5
New cards

Q: What is inductive learning?

Learning a function f from examples (x, f(x)) by finding a hypothesis h such that h ≈ f.

6
New cards

Q: What is Ockham’s razor in learning?

Prefer the simplest hypothesis consistent with the data.

7
New cards

Q: In decision tree learning, what is the goal?

Find a small tree that is consistent with the training examples.

8
New cards

Q: What is an attribute-based representation?

Describing examples using values for different attributes (Boolean, discrete, continuous).

9
New cards

Q: How expressive are decision trees?

They can represent any function of the input attributes.

10
New cards

Q: Why don’t we want giant decision trees that perfectly memorize training data?

They may not generalize to new examples; simpler trees are preferred.

11
New cards

Q: How many distinct decision trees are there with n Boolean attributes?

2^(2ⁿ)
(e.g., with 6 attributes, over 18 quintillion trees!)

12
New cards

Q: How many purely conjunctive hypotheses exist with n attributes?

3^n.

13
New cards

Q: What happens as the hypothesis space gets bigger?

• It’s more likely to contain the target function.
• But it’s also easier to overfit.

14
New cards

Q: What is the main idea when choosing an attribute for a decision tree split?

Pick the attribute that splits the data into subsets that are mostly “pure” (mostly positive or negative).

15
New cards

Q: What formula is used to compute the uncertainty (entropy) of a dataset?

I(P(v₁), ..., P(vₙ)) = -∑ P(vᵢ) log₂ P(vᵢ)

16
New cards

Q: What is Information Gain (IG)?

The reduction in entropy from splitting on an attribute; higher IG is better.

17
New cards

Q: How do you choose the best attribute to split on?

Choose the attribute with the highest information gain.

18
New cards

Q: In the restaurant example, which attribute had the highest Information Gain?

Patrons.

19
New cards

Q: What does a learning curve show?

Prediction accuracy as a function of training set size.

20
New cards

Q: What are two ways to measure if h ≈ f (if your model is good)?

• Try h on a new test set.
• Use computational/statistical learning theory.