机器学习(七) — 决策树

model 4 — decision tree

1 decision tree

1. component

usage: classification

  1. root node
  2. decision node

2. choose feature on each node

maximize purity (minimize inpurity)

3. stop splitting

  1. a node is 100% on class
  2. splitting a node will result in the tree exceeding a maximum depth
  3. improvement in purity score are below a threshold
  4. number of examples in a node is below a threshold

2 meature of impurity

use entropy(

H

H

H) as a meature of impurity

H

(

p

)

=

p

l

o

g

2

(

p

)

(

1

p

)

l

o

g

2

(

1

p

)

n

o

t

e

:

0

l

o

g

0

=

0

H(p) = -plog_2(p) – (1-p)log_2(1-p)\\ note: 0log0 = 0

H(p)=−plog2​(p)−(1−p)log2​(1−p)note:0log0=0

在这里插入图片描述

3 information gain

1. definition

i

n

f

o

m

a

t

i

o

n

_

g

a

i

n

=

H

(

p

r

o

o

t

)

(

w

l

e

f

t

H

(

p

l

e

f

t

)

+

w

r

i

g

h

t

H

(

p

r

i

g

h

t

)

)

infomation\_gain = H(p^{root}) – (w^{left}H(p^{left}) + w^{right}H(p^{right}))

infomation_gain=H(proot)−(wleftH(pleft)+wrightH(pright))

2. usage

  1. meature the reduction in entropy
  2. a signal of stopping splitting

3. continuous

find the threshold that has the most infomation gain

在这里插入图片描述

4 random forest

  1. generating a tree sample
given training set of size m
for b = 1 to B:
	use sampling with replacement to create a new training set of size m
	train a decision tree on the training set
  1. randomizing the feature choice: at each node, when choosing a feature to use to split, if n features is available, pick a random subset of k < n(usually

    k

    =

    n

    k = \sqrt{n}

    k=n
    ​) features and alow the algorithm to only choose from that subset of features

本文来自网络,不代表协通编程立场,如若转载,请注明出处:https://net2asp.com/db453b599e.html