官术网_书友最值得收藏!

Gradient boosting

Gradient boosted trees are an ensemble of shallow trees (or weak learners). The shallow decision trees could be as small as a tree with just two leaves (also known as decision stump). The boosting methods help in reducing bias mainly but also help reduce variance slightly.

Original papers by Breiman and Friedman who developed the idea of gradient boosting are available at following links:

Intuitively, in the gradient boosting model, the decision trees in the ensemble are trained in several iterations as shown in the following image. A new decision tree is added at each iteration. Every additional decision tree is trained to improve the trained ensemble model in previous iterations. This is different from the random forest model where each decision tree is trained independently from the other decision trees in the ensemble.

The gradient boosting model has lesser number of trees as compared to the random forests model but ends up with a very large number of hyperparameters that need to be tuned to get a decent gradient boosting model.

An interesting explanation of gradient boosting can be found at the following link:  http://blog.kaggle.com/2017/01/23/a-kaggle-master-explains-gradient-boosting/.
主站蜘蛛池模板: 陇南市| 石城县| 修文县| 蒙山县| 云霄县| 黑水县| 张家川| 巴彦淖尔市| 罗山县| 巴马| 宾川县| 冷水江市| 康保县| 宿迁市| 紫云| 濉溪县| 咸阳市| 周至县| 阿克| 中宁县| 南阳市| 邢台市| 分宜县| 蛟河市| 兴义市| 凤台县| 崇文区| 马山县| 乡城县| 永春县| 宕昌县| 通海县| 犍为县| 武强县| 龙门县| 张家口市| 武安市| 刚察县| 于田县| 太和县| 个旧市|