xgboost and chefboost
XGBoost is a production-grade distributed gradient boosting library that would typically be chosen over Chefboost for serious machine learning work, making them direct competitors despite Chefboost's broader coverage of classical decision tree algorithms.
About xgboost
dmlc/xgboost
Scalable, Portable and Distributed Gradient Boosting (GBDT, GBRT or GBM) Library, for Python, R, Java, Scala, C++ and more. Runs on single machine, Hadoop, Spark, Dask, Flink and DataFlow
Implements parallel tree boosting with built-in support for categorical features, missing value handling, and monotonic constraints without preprocessing. Uses a novel column-block structure for cache-aware tree construction and supports GPU acceleration via CUDA for faster training on large datasets. Integrates with ML platforms including scikit-learn, MLflow, and Optuna for hyperparameter optimization, with native support for feature importance analysis and SHAP explainability.
About chefboost
serengil/chefboost
A Lightweight Decision Tree Framework supporting regular algorithms: ID3, C4.5, CART, CHAID and Regression Trees; some advanced techniques: Gradient Boosting, Random Forest and Adaboost w/categorical features support for Python
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work