A Hands-on Guide To Create Explainable Gradient Boosting Classification models using Bayesian Hyperparameter Optimization.

Boosted decision tree algorithms, such as XGBoost, CatBoost, and LightBoost are popular methods for the classification task. Learn how to split the data, optimize hyperparameters, prevent overtraining, select the best-performing model, and create explainable results.

Erdogan Taskesen
14 min readSep 18, 2022
Photo by Steve Harvey on Unsplash

This blog is written in a series where the first part explains the general concepts of gradient boosting techniques such as XGBoost, CatBoost, and LightBoost, together with the process of tuning hyperparameters, and details about the HGBoost library. In this part 2, I will demonstrate in more detail: 1. how to train a gradient boosting classification model with optimized hyperparameters using Bayesian optimization, 2. how to select the best performing model (and is not overtrained), 3. how to create explainable results by visually explaining the optimized hyperparameter space together with the model performance accuracy.

If you found this article helpful, use my referral link to continue learning without limits and sign up for a Medium membership. Plus, follow me to stay

--

--

Erdogan Taskesen
Erdogan Taskesen

Written by Erdogan Taskesen

Machine Learning | Statistics | D3js visualizations | Data Science | Ph.D | erdogant.github.io