A game theoretic approach to explain the output of any machine learning model.
-
Updated
Mar 3, 2026 - Jupyter Notebook
A game theoretic approach to explain the output of any machine learning model.
A fast, distributed, high performance gradient boosting (GBT, GBDT, GBRT, GBM or MART) framework based on decision tree algorithms, used for ranking, classification and many other machine learning tasks.
Machine learning, in numpy
A Python Automated Machine Learning tool that optimizes machine learning pipelines using genetic programming.
A fast, scalable, high performance Gradient Boosting on Decision Trees library, used for ranking, classification, regression and other machine learning tasks for Python, R, Java, C++. Supports computation on CPU and GPU.
Fit interpretable models. Explain blackbox machine learning.
A collection of research papers on decision, classification and regression trees with implementations.
Natural Gradient Boosting for Probabilistic Prediction
A curated list of data mining papers about fraud detection.
[UNMAINTAINED] Automated machine learning for analytics & production
A curated list of gradient boosting research papers with implementations.
A unified ensemble framework for PyTorch to improve the performance and robustness of your deep learning model.
LAMA - automatic model creation framework
python实现GBDT的回归、二分类以及多分类,将算法流程详情进行展示解读并可视化,庖丁解牛地理解GBDT。Gradient Boosting Decision Trees regression, dichotomy and multi-classification are realized based on python, and the details of algorithm flow are displayed, interpreted and visualized to help readers better understand Gradient Boosting Decision Trees
A collection of state-of-the-art algorithms for the training, serving and interpretation of Decision Forest models in Keras.
Building and training Speech Emotion Recognizer that predicts human emotions using Python, Sci-kit learn and Keras
Perpetual is a high-performance gradient boosting machine. It delivers optimal accuracy in a single run without complex tuning through a simple budget parameter. It features out-of-the-box support for causal ML, continual learning, native calibration, and robust drift monitoring, along with Rust core and zero-copy bindings for Python and R
A library to train, evaluate, interpret, and productionize decision forest models such as Random Forest and Gradient Boosted Decision Trees.
Tuning hyperparams fast with Hyperband
Ollama for classical ML models. AOT compiler that turns XGBoost, LightGBM, scikit-learn, CatBoost & ONNX models into native C99 inference code. One command to load, one command to serve. 336x faster than Python inference.
Add a description, image, and links to the gradient-boosting topic page so that developers can more easily learn about it.
To associate your repository with the gradient-boosting topic, visit your repo's landing page and select "manage topics."