This project is designed to explore machine learning models and methods for the task of classification. We've been tasked with finding a dataset with labeled data with at least 40,000 rows of data and 20 columns. We will generally follow the OSEMN process for a data science project, which includes:
- Obtaining the data
- Scrubing (or cleaning) the data
- Exploring and visualizing the data
- Modeling
- INterpreting the results
Project goal: The data we will be using (more below) is a bank marketing data set. The goal for our project is to develop as model that predicts the success of a bank marketing campaign based on the features we have in out data. This model should therefore help us be better able to identify potential customers and refine our the focus of future campaigns.
The data is related with direct marketing campaigns (phone calls) of a Portuguese banking institution. The classification goal is to predict if the client will subscribe a term deposit (variable y).
This dataset is publicly available for research. The details are described in [Moro et al., 2014]. Please include this citation if you plan to use this database:
[Moro et al., 2014] S. Moro, P. Cortez and P. Rita. A Data-Driven Approach to Predict the Success of Bank Telemarketing. Decision Support Systems, In press, http://dx.doi.org/10.1016/j.dss.2014.03.001
Available at: [pdf] http://dx.doi.org/10.1016/j.dss.2014.03.001 [bib] http://www3.dsi.uminho.pt/pcortez/bib/2014-dss.txt
Link to data: https://archive.ics.uci.edu/ml/datasets/Bank+Marketing
Bank Client Data | ||||
---|---|---|---|---|
Variable | Datatype | Example | ||
age | numeric | 27 | ||
job | categorical | 'admin', 'blue-collar', 'entrepreneur', 'housemaid' | ||
marital | categorical | 'divorced', 'married', 'single', 'unknown' | ||
education | categorical | 'basic.9y', 'high.school', 'illiterate', 'professional.course' | ||
credit_default | categorical | 'yes', 'no', 'unknown' | ||
housing_loan | categorical | 'yes','no','unknown' | ||
personal_loan | categorical | 'yes','no','unknown' |
Campaign Data | ||||
---|---|---|---|---|
Variable | Datatype | Example | ||
contact_type | categorical | 'cellular', 'telephone' | ||
month | categorical | 'jan', 'feb', 'mar' | ||
day_of_week | categorical | 'mon', 'tues', 'wed' | ||
call_duration | numeric | seconds: 240 | ||
campaign | numeric | number of times customer called: 5 | ||
p_days | numeric | number of days that passed by after the client was last contacted from a previous campaign: 6 (999 means client was not previously contacted) | ||
poutcome | categorical | outcome of previous campaign: "failure","success","nonexistant" |
Social and Economic Context | ||||
---|---|---|---|---|
Variable | Datatype | Example | ||
employment_variation_rate | numeric | quarterly indicator of employement: 1.1 | ||
consumer_price_index | numeric | monthly indicator of consumer prices: 93.99 | ||
consumer_confidence_index | numeric | monthly indicator of consumer confidence: -36.4 | ||
euribor_3-month_rate | numeric | daily indicator of EURO interback lending rate: 4.857 | ||
nr.employed | numeric | number of employees: 5191 |
Output variable (desired target):
y - has the client subscribed a term deposit? (binary: 'yes','no')
As this is a project for learning purposes, we're going to try out a lot of models. In this section we will work through:
- Logistic Regression
- K-Nearest Neighbors
- Decision Trees and Random Forest
- XGBoost
- Support Vector Machines (SVM)
We'll also apply a grid search method to each to see if we can work through a range of hyper-parameters to find a what performs best. In short, we'll be doing a lot!
At this time our models do not perform well enough to implement them in the way we have explored
Although on the high end our models may allow us to significantly reduce costs:
Our baseline campaign cost is as much as: $2,890,000 The logreg model could save up to: $1,629,550 The random forest model could save up to $1,758,475 However, the potential drop in revenue is not offset:
Baseline revenue expectations are as high as
Logistic Regression appears to be the method of choice for this problem. Although we do not have it as tuned as we would like it to be, it consistently had the highest recall and therefore lowest false negative rate
Next Steps and Recommendations:
The current method of calling customers relies on a high volume of unproductive calls, and we should continue to work to find a method that helps reduce those We should look into the true cost of a campaign, if true costs are in fact 3-4 times higher, then many of our models may be in place to begin saving money now We should attempt to stack models to gain greater efficiency and potentially improve recall and precision greater data cleaning, feature transformation and feature engineering may help current models perform better on the margins