LightGBM and chefboost
These two tools are competitors, with LightGBM being a more established and widely adopted gradient boosting framework, while ChefBoost offers a lightweight alternative supporting various decision tree algorithms and boosting techniques.
About LightGBM
lightgbm-org/LightGBM
A fast, distributed, high performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.
Implements leaf-wise tree growth with histogram-based learning to reduce memory footprint and accelerate training on CPU and GPU hardware. Provides native bindings for Python, R, and C++, with ecosystem integrations including FLAML for AutoML, Optuna for hyperparameter tuning, and model compilers like Treelite and Hummingbird for production deployment.
About chefboost
serengil/chefboost
A Lightweight Decision Tree Framework supporting regular algorithms: ID3, C4.5, CART, CHAID and Regression Trees; some advanced techniques: Gradient Boosting, Random Forest and Adaboost w/categorical features support for Python
Related comparisons
Scores updated daily from GitHub, PyPI, and npm data. How scores work