decision tree - taoualiw/My-Knowledge-Base GitHub Wiki

-
Entropy controls how a DT splits data.Decision Tree reduces entropy (entropy being between 0 and 1)> Entropy formula :
E = −∑(pi)log2(pi). Entropy definition: measure of impurity in a bunch of samples -
DT algorith maximizes the information Gain Information Gain formula :
IG = entropy(parent)-[weighted average]entropy(children) -
Decision Boundary: axis-parallel rectangles
-
Parameteres:
- min_samples_split
-
In sklearn the default criterion is slightly different from entropy ( something called gini)
See also Random Forest