Decision Trees 101

Before lookin at those, your going to need to know:

RandomForests says "if one tree is good, why not build 100?"

Example1

Example2:

let us measure diversity using entropy; i.e. -1 ∑ plog2(p)

e.g. 1 orange, 1 apple, 2 bananas, and 4 grapes

Learners like decision trees and random forests use entropy

Example3:

What is the best split for this data?

Here are the options (note that this is four different splits):

Consider the outlook tree. We have three sub-branches so the expected value of the diversity after the split is

The overcast split is easy:

The sunny and rainy split are symmetric

So the expected value after the outlook split is

(BTW, this is an improvement since before the split we have 9 yes, 5 no; ie. entropy was 0.94; i.e. more diversity).

If we repeat this calc over all splits, we get

So we would split on outlook.