22. Bagging and Random Forests

  Рет қаралды 10,791

Inside Bloomberg

Inside Bloomberg

Күн бұрын

We motivate bagging as follows: Consider the regression case, and suppose we could create a bunch (say B) prediction functions based on independent training samples of size n. If we average together these prediction functions, the expected value of the average is the same as any one of the functions, but the variance would have decreased by a factor of 1/B -- a clear win! Of course, this would require an overall sample of size nB. The idea of bagging is to replace independent samples with bootstrap samples from a single data set of size n. Of course, the bootstrap samples are not independent, so much of our discussion is about when bagging does and does not lead to improved performance. Random forests were invented as a way to create conditions in which bagging works better.
More...Although it's hard to find crisp theoretical results describing when bagging helps, conventional wisdom says that it helps most for models that are "high variance", which in this context means the prediction function may change a lot when you train with a new random sample from the same distribution, and "low bias", which basically means fitting the training data well. Large decision trees have these characteristics and are usually the model of choice for bagging. Random forests are just bagged trees with one additional twist: only a random subset of features are considered when splitting a node of a tree. The hope, very roughly speaking, is that by injecting this randomness, the resulting prediction functions are less dependent, and thus we'll get a larger reduction in variance. In practice, random forests are one of the most effective machine learning models in many domains.
Access the full course at bloom.bg/2ui2T4q

Пікірлер
20. Classification and Regression Trees
1:16:20
Inside Bloomberg
Рет қаралды 14 М.
26.  Gaussian Mixture Models
56:38
Inside Bloomberg
Рет қаралды 18 М.
The Lost World: Living Room Edition
0:46
Daniel LaBelle
Рет қаралды 27 МЛН
Thank you mommy 😊💝 #shorts
0:24
5-Minute Crafts HOUSE
Рет қаралды 33 МЛН
JISOO - ‘꽃(FLOWER)’ M/V
3:05
BLACKPINK
Рет қаралды 137 МЛН
Excel Avoid Offset and use RowAbove
7:18
Access Analytic
Рет қаралды 2 М.
Bagging and Random Forests
15:47
Leslie Myint
Рет қаралды 9 М.
23.  Gradient Boosting
1:24:35
Inside Bloomberg
Рет қаралды 21 М.
5. Excess Risk Decomposition
28:47
Inside Bloomberg
Рет қаралды 11 М.
Machine Learning Tutorial Python - 21: Ensemble Learning - Bagging
23:37
Machine Learning Lecture 31 "Random Forests / Bagging" -Cornell CS4780 SP17
47:25
Regression Trees, Clearly Explained!!!
22:33
StatQuest with Josh Starmer
Рет қаралды 673 М.
12. Feature Extraction
1:14:03
Inside Bloomberg
Рет қаралды 19 М.
StatQuest: Random Forests Part 1 - Building, Using and Evaluating
9:54
StatQuest with Josh Starmer
Рет қаралды 1,2 МЛН
The Lost World: Living Room Edition
0:46
Daniel LaBelle
Рет қаралды 27 МЛН