Beta. Content is under active construction and has not been peer-reviewed. Report errors on GitHub.Disclaimer

Gradient Boosting

1 questionsDifficulty 5-5View topic
Intermediate
0 / 1
1 intermediateAdapts to your performance
1 / 1
intermediate (5/10)conceptual
In gradient boosting with squared loss, each new tree is fit to the negative gradient of the loss, which equals the residuals. Why does boosting use the negative gradient rather than directly fitting the residuals for general loss functions?