toc
Last time:
This lecture:
Announcements:
Section 1
We divide the predictor space – that is, the set of possible values for X1,X2,⋯,Xp — into J distinct and non-overlapping regions, R1,R2,⋯,RJ.
For every observation that falls into the region Rj, we make the same prediction = the mean of the response values for the training observations in Rj.
Find boxes R1,⋯,RJ that minimize
ŷRj = mean response for training observations in jth box
Pick s so that splitting into {X∣Xj /mo> s} and {X∣Xj ≥s} results in largest possible reduction in RSS:
Gini index:
How do you use out of bag error estimation for decision trees?
Section 2
Want to do (but can’t): Build separate models from independent training sets, and average resulting predictions:
Return the average
Boostrap modification:
Return average of predictions (regression)
or majority vote (classification)
Test your understanding: PollEv
Section 3
Goal is to decorrelate the bagged trees:
The random forest fix: