Skip to content

KKBox churn prediction challenge on Kaggle: dealing with imbalanced data using WRF, autoencoder and xgboost.

Notifications You must be signed in to change notification settings

zihaoxu/KKBox_Churn_Prediction

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 

Repository files navigation

KKBox Churn Prediction Challenge

Group member: Zihao Xu, Xiaotong gui, Minh-Quan Do

Abstract

In this repository, we aim to document our methodologies in approaching the KKBox Churning Prediction Challenge. This challenge is essentially a classification problem, but the response variable is highly imbalanced. In the below sections, we will describe and visually explore the data sets. Then we will talk about several machine learning models we employed that are highly suitable for handling imbalanced data. Our current ranking on Kaggle is 136 out of 535, achieved by the XGBoost model.

Paper & Presentation

Paper:

https://goo.gl/sxwdL9

Presentation:

https://goo.gl/sQMLEa

Notebooks

0. Exploratory Analysis

https://www.kaggle.com/zxql2015/fork-of-0-exploratory-analysis

1. Autoencoder with Keras

https://www.kaggle.com/zxql2015/1-autoencoder-with-keras

2. Weighted Random Forest

https://www.kaggle.com/zxql2015/2-imbalanced-rf/

3. Gxtreme Gradient Boosting

https://www.kaggle.com/zxql2015/3-extreme-gradient-boosting

About

KKBox churn prediction challenge on Kaggle: dealing with imbalanced data using WRF, autoencoder and xgboost.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages