Menu
XGBoost
XGBoost is an open-source library that offers gradient boosting which is an ensemble method. The algorithm runs on a notebook or scales to support multiple cores, GPUs, threads, and clusters of computing. Supports scikit-learn but has its own interface.ย
Project Background
- Library:ย XGBoost
- Author: Tianqi Chen and DMLC group
- Initial Release: March 2014
- Type: Gradient boosting library
- License: Apache 2.0
- Language: Developed in C++
- GitHub: dmlc/xgboost
- Runs On: Single machine and supports Hadoop, Spark, and Flink
- Twitter:ย XGBoost
Features
- Boosting is an ensemble method
- The algorithm wins lots of Kaggle competitions
- ย Supports regularized boosting which is normalized
- Handles missing values automatically
- Takes advantage of multiple cores, threads, and clusters of compute
- ย Incremental training: stop and save training, then come back to it later
- Tree pruning: deeper and more optimal trees
- 3ย typesย of boosting supported: gradient, stochastic gradient, and regularized boosting