1 Decision Trees

studied byStudied by 0 people
0.0(0)
learn
LearnA personalized and smart learning plan
exam
Practice TestTake a test on your terms and definitions
spaced repetition
Spaced RepetitionScientifically backed study method
heart puzzle
Matching GameHow quick can you match all your cards?
flashcards
FlashcardsStudy terms and definitions

1 / 14

encourage image

There's no tags or description

Looks like no one added any tags here yet for you.

15 Terms

1

What is a decision tree, and how is it used in data classification tasks?

A decision tree is a flowchart-like structure used for decision-making and classification tasks. It is used in data classification tasks by dividing the dataset into smaller subsets based on feature values.

New cards
2

Describe the components of a decision tree, including internal nodes, branches, and leaf nodes.

Internal Nodes: Decision points for attributes.

Branches: Outcomes of decisions.

Leaf Nodes: Final predictions or classifications.

New cards
3

Explain the steps involved in building a decision tree.

Steps include: selecting the best attribute (using metrics like entropy/information gain), splitting the dataset, and recursively repeating until all nodes are pure or stopping criteria are met.

New cards
4

Define entropy and information gain in the context of decision trees and explain how they influence the construction of the tree.

Entropy: Measure of impurity or disorder.

Information Gain: Reduction in entropy after a dataset split.

These concepts influence the tree construction process by helping to identify the best feature to split on.

New cards
5

What is the Gini index, and how is it used in building decision trees?

A metric to measure node impurity (probability of incorrectly classifying a randomly chosen data point). It identifies best split.

New cards
6

Describe the ID3 algorithm and its role in decision tree construction.

Builds decision trees by recursively selecting attributes that maximize information gain.

New cards
7

How are attributes selected for splitting at each node in a decision tree?

Attributes are selected based on metrics like information gain or Gini index.

New cards
8

Discuss the techniques used to prevent overfitting in decision trees.

Use pruning, limit tree depth, or ensure sufficient data at nodes.

New cards
9

Explain the difference between pre-pruning and post-pruning in decision tree algorithms.

Pre-Pruning: Halts tree growth early based on criteria.

Post-Pruning: Removes branches after the tree is fully grown.

New cards
10

List and explain the advantages and limitations of using decision trees.

Advantages: Simple, interpretable, handles categorical/continuous data.

Limitations: Prone to overfitting, unstable.

New cards
11

Compare and contrast decision trees with Random Forests.

Random forests use multiple decision trees to improve stability and accuracy through bagging.

New cards
12

How do decision trees handle missing values during the training process?

Methods include ignoring missing values, surrogate splits, assigning missing values to most frequent class, imputing before training.

New cards
13

Provide examples of real-world applications where decision trees are effectively used.

Fraud detection, medical diagnosis, and customer segmentation.

New cards
14

How does a decision tree algorithm handle continuous and categorical variables differently?

Continuous variables are split into ranges; categorical variables are split by category.

New cards
15

What methods are used to evaluate the performance of a decision tree?

Use metrics like accuracy, precision, recall, F1 score, or confusion matrix.

New cards
robot