[go: up one dir, main page]

0% found this document useful (0 votes)
23 views19 pages

UNIT-IV - Decision Tree Induction

Uploaded by

karthiktej890
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
23 views19 pages

UNIT-IV - Decision Tree Induction

Uploaded by

karthiktej890
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 19

Decision Trees for Classification

Decision Trees for Classification

1. Decision Tree is a Supervised learning technique.

2. It can be used for both classification and Regression problems, but


mostly it is preferred for solving Classification problems

3. where the data is continuously split according to a certain


parameter.

4. The tree can be explained by two entities, namely decision nodes


and leaves.

5. The leaves are the decisions or the final outcomes. And the decision
nodes are where the data is split.
Decision Trees for Classification

An example of a decision tree can be explained using binary tree.

The decision nodes here are questions like ‘What’s the age?’, ‘Does he
exercise?’, ‘Does he eat a lot of pizzas’? And the leaves, which are
outcomes like either ‘fit’, or ‘unfit’.
Decision Trees for Classification

A Decision Tree has the following structure:

1. Root Node: The root node is the starting point of a tree. At this point,
the first split is performed.

2. Internal Nodes: Each internal node represents a decision point


(predictor variable) that eventually leads to the prediction of the
outcome.

3. Leaf/ Terminal Nodes: Leaf nodes represent the final class of the
outcome and therefore they’re also called terminating nodes.

4. Branches: Branches are connections between nodes, they’re represented


as arrows. Each branch represents a response such as yes or no.
Decision Trees for Classification

Now let’s try to understand how a Decision Tree is created.

Build A Decision Tree Using ID3 Algorithm

What Is The ID3 Algorithm?

ID3 or the Iterative Dichotomiser 3 algorithm is one of the most


effective algorithms used to build a Decision Tree.

It uses the concept of Entropy and Information Gain to generate a


Decision Tree for a given set of data.
Decision Trees for Classification

Two measures are used to decide the best attribute:(Must be ROOT Node)

Information Gain

Entropy
Decision Trees for Classification

What Is Entropy?

Entropy measures the impurity or


uncertainty present in the data. It is used to
decide how a Decision Tree can split the
data.
Decision Trees for Classification

What Is Information Gain?

1. Information Gain (IG) is the most significant measure used to


build a Decision Tree.

2. Information Gain is important because it used to choose the


variable that best splits the data at each node of a Decision Tree.

3. The variable with the highest IG is used to split the data at the
root node
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification
Decision Trees for Classification

You might also like