Tag: XGBoost

Binary Class Tabular Model for Kaggle Playground Series Season 3 Episode 3 Using Python and XGBoost

SUMMARY: The project aims to construct a predictive model using various machine learning algorithms and document the end-to-end steps using a template. The Kaggle Playground Series Season 3 Episode 3 dataset is a binary-class modeling situation where we attempt to predict one of two possible outcomes.

INTRODUCTION: Kaggle wants to provide an approachable environment for relatively new people in their data science journey. Since January 2021, they have hosted playground-style competitions to give the Kaggle community a variety of reasonably lightweight challenges that can be used to learn and sharpen skills in different aspects of machine learning and data science. The dataset for this competition was generated from a deep learning model trained on the IBM HR Analytics Employee Attrition & Performance dataset. Feature distributions are close to but different from the original.

ANALYSIS: The performance of the preliminary XGBoost model achieved a ROC/AUC benchmark of 0.8511 after training. When we processed the test dataset with the final model, the model achieved a ROC/AUC score of 0.8969.

CONCLUSION: In this iteration, the XGBoost model appeared to be a suitable algorithm for modeling this dataset.

Dataset Used: Playground Series Season 3, Episode 3

Dataset ML Model: Binary-Class classification with numerical and categorical features

Dataset Reference: https://www.kaggle.com/competitions/playground-series-s3e3

One source of potential performance benchmarks: https://www.kaggle.com/competitions/playground-series-s3e3/leaderboard

The HTML formatted report can be found here on GitHub.

Binary Class Tabular Model for Kaggle Playground Series Season 3 Episode 2 Using Python and XGBoost

SUMMARY: The project aims to construct a predictive model using various machine learning algorithms and document the end-to-end steps using a template. The Kaggle Playground Series Season 3 Episode 2 dataset is a binary-class modeling situation where we attempt to predict one of two possible outcomes.

INTRODUCTION: Kaggle wants to provide an approachable environment for relatively new people in their data science journey. Since January 2021, they have hosted playground-style competitions to give the Kaggle community a variety of reasonably lightweight challenges that can be used to learn and sharpen skills in different aspects of machine learning and data science. The dataset for this competition was generated from a deep learning model trained on the Stroke Prediction Dataset. Feature distributions are close to but different from the original.

ANALYSIS: The performance of the preliminary XGBoost model achieved a ROC/AUC benchmark of 0.8772 after training. When we processed the test dataset with the final model, the model achieved a ROC/AUC score of 0.8730.

CONCLUSION: In this iteration, the XGBoost model appeared to be a suitable algorithm for modeling this dataset.

Dataset Used: Playground Series Season 3, Episode 2

Dataset ML Model: Binary-Class classification with numerical and categorical features

Dataset Reference: https://www.kaggle.com/competitions/playground-series-s3e2

One source of potential performance benchmarks: https://www.kaggle.com/competitions/playground-series-s3e2/leaderboard

The HTML formatted report can be found here on GitHub.

Regression Tabular Model for Kaggle Playground Series Season 3 Episode 1 Using Python and XGBoost

SUMMARY: The project aims to construct a predictive model using various machine learning algorithms and document the end-to-end steps using a template. The Kaggle Playground Series Season 3 Episode 1 Dataset is a regression modeling situation where we are trying to predict the value of a continuous variable.

INTRODUCTION: Kaggle wants to provide an approachable environment for relatively new people in their data science journey. Since January 2021, they have hosted playground-style competitions to give the Kaggle community a variety of reasonably lightweight challenges that can be used to learn and sharpen skills in different aspects of machine learning and data science. The dataset for this competition was generated from a deep learning model trained on the California Housing Dataset. Feature distributions are close to but different from the original.

ANALYSIS: The performance of the preliminary XGBoost model achieved an RMSE benchmark of 0.5232. After a series of tuning trials, the final model processed the test dataset with an RMSE score of 0.5648.

CONCLUSION: In this iteration, the XGBoost model appeared to be a suitable algorithm for modeling this dataset.

Dataset Used: Playground Series Season 3, Episode 1

Dataset ML Model: Regression with numerical features

Dataset Reference: https://www.kaggle.com/competitions/playground-series-s3e1

One source of potential performance benchmarks: https://www.kaggle.com/competitions/playground-series-s3e1/leaderboard

The HTML formatted report can be found here on GitHub.

Multi-Class Tabular Model for Date Fruits Classification Using Python and XGBoost

SUMMARY: The project aims to construct a predictive model using various machine learning algorithms and document the end-to-end steps using a template. The Date Fruits Dataset is a multi-class modeling situation where we attempt to predict one of several (more than two) possible outcomes.

INTRODUCTION: Many factors determine the type of fruit with external appearance features such as color, length, diameter, and shape. However, classifying the variety of fruits simply by looking at their outward appearance requires expertise and a great effort and is also time-consuming. This study aims to classify the types of date fruit using machine-learning methods. The research team obtained 898 images of seven different date fruit types for this project via a computer vision system. The research team was hopeful that machine learning methods could successfully classify date fruit types.

ANALYSIS: The performance of the preliminary XGBoost model achieved an accuracy benchmark of 89.41%. After a series of tuning trials, the final model processed the training dataset with an accuracy score of 90.75%.

CONCLUSION: In this iteration, the XGBoost model appeared to be a suitable algorithm for modeling this dataset.

Dataset Used: Date Fruits Dataset

Dataset ML Model: Multi-Class classification with numerical features

Dataset Reference: https://www.muratkoklu.com/datasets/

One source of potential performance benchmarks: https://doi.org/10.1155/2021/4793293

The HTML formatted report can be found here on GitHub.

Binary-Class Tabular Model for Kaggle Spaceship Titanic Using Python and XGBoost

Template Credit: Adapted from a template made available by Dr. Jason Brownlee of Machine Learning Mastery.

SUMMARY: The project aims to construct a predictive model using various machine learning algorithms and document the end-to-end steps using a template. The Kaggle Spaceship Titanic dataset is a binary-class modeling situation where we attempt to predict one of two possible outcomes.

INTRODUCTION: In 2912, the Spaceship Titanic, an interstellar passenger liner, launched with almost 13,000 passengers on board. The vessel set out on its maiden voyage transporting emigrants from our solar system to three newly habitable exoplanets orbiting nearby stars.

While rounding Alpha Centauri en route to its first destination, the unwary Spaceship Titanic collided with a spacetime anomaly hidden within a dust cloud. Sadly, it met a similar fate as its namesake from 1000 years before. Though the ship stayed intact, almost half of the passengers were transported to an alternate dimension!

This incident presented a challenge to the data scientists of the future. To help rescue crews and retrieve the lost passengers, the data scientists tried to build a model that predicted which passengers were transported by the anomaly using records recovered from the spaceship’s damaged computer system.

ANALYSIS: The performance of the preliminary XGBoost model achieved an accuracy benchmark of 80.69% after training. When we processed the test dataset with the final model, the model achieved an accuracy score of 80.31%.

CONCLUSION: In this iteration, the XGBoost model appeared to be a suitable algorithm for modeling this dataset.

Dataset Used: Spaceship Titanic Dataset

Dataset ML Model: Binary-Class classification with numerical and categorical features

Dataset Reference: https://www.kaggle.com/competitions/spaceship-titanic

One source of potential performance benchmarks: https://www.kaggle.com/competitions/spaceship-titanic/leaderboard

The HTML formatted report can be found here on GitHub.