This article was published as a part of the Data Science Blogathon
Hello, everyone.â
We know the IPL season is going on and we are all eager to know who will win the match beforehand and in the media, there is hype around the winning chances.
What if I say we can make an app that can predict the outcome, Yeah! with the power of Machines and Deep Learning, you can do these types of amazing stuff and this article is all about it.
In particular, here we will be looking at how you can train a model from scratch and embed it in the web app using simple and powerful libraries like sklearn, pandas, and flask. Also, some web development is involved.
The Beneath segment gives an outline and a few rules and you are prescribed to go through it previously.
Let’s get started on building an IPL Score Predictor!
We will dive into each of the sections above, which are common steps a data science team performs. This will also ensure that you learn how to approach the problem with a productive mindset.
NOTE:- It’s better to create a new environment and start working there for each project as it will later help in keeping the essentials separate and increase productivity.
Legends :
H2 – Heading Main Section
H4 – Heading – Consisting of Subsections of Main
bold – Subparts Of Subsections
Before we start our project we need to gather data. This step can be done using the following 3 methods:
For our use case, we are going to use the IPL Scores Dataset (link in reference) which has 76104 observations and 15 features :
total – is our target variable.
We will start by loading some essential libraries needed for the project:
import pandas as pd
import pickle
#Next, we will load our CSV file (ipl.csv) and display the first 2 rows of data
# loading dataset using read_csv()
df = pd.read_csv('ipl.csv')
# displaying data
print(df.head(2))
The first thing to note here is that our dataset has certain columns which are irrelevant and increases resource consumption, so maybe we can remove them in future steps.
“mid,” “venue,” “batsman,” “bowler,” “striker,” and “non-striker.” # irrelevant column
Having looked at the data quickly, let’s dive deeper into the dataset and explore some of the insights. This procedure is very important and will allow us to understand the data and plan our next steps. Luckily pandas provide easy-to-use functions to perform our analysis. So let’s begin.
Even though we know the shape of data (observations and features) the info can be wrong. So it’s better to cross-check our data once. So let’s check it.
We can get the shape of data using the shape attribute of pandas dataframe.
#checking shape of data using df.shape df.shape
Output:
>> (76014, 15)
Note: the output is in the format of (rows, columns).
In real-world cases, the data we find are not of the desired data type, so it will be good if we check it too.
Pandas provide an info method that returns the data type of all columns present in the dataset.
# getting data type df.info()
Returns:
Notice the type- object, it is because panda assumes string as an object datatype.
Having understood the shape and info, we will now check for null values (fields having no data – NAN’s). It is essential for any project as null values can change the whole story depicted by data and can even contribute to making data worse for the use cases. Let’s perform the same
All we have to do is to use the is_null method of the dataframe and then sum the outputs to get the total count for each column.
# check null values and then sum it up df.isnull().sum()
Output:
Image by Author
Seems like we are lucky as there are no null values present(0) in the dataset. and all of them are of type – int64 – takes huge memory
Since there is no nan value doesn’t mean our data is a good representation. To get the gist, we can plot summary measures which generally include:
1. Measurement of Central Tendency – These measures allow us to understand where most of our data lies and mostly include:
2. Measures of Dispersion – These are the measures that allow us to understand how widespread data is and mostly includes:
Performing each check will be cumbersome so pandas pack all these in a single function called describe. Now let’s check the summary stats of our data.
# Check summary stats using df.describe() df.describe()
Image by Author
We see that for all columns the data is almost in a similar range, thus can infer that it is good data to work on.
Note: Description only works on numeric data
Now let’s check how each numerical feature is contributing to our dataset and we can do this by using a panda area plot.
# plot area / contributions in dataset df[['date','venue', 'bat_team', 'bowl_team', 'batsman', 'bowler', 'runs', 'wickets', 'overs', 'runs_last_5', 'wickets_last_5', 'striker', 'non-striker']][:50].plot(kind = 'area', figsize = (10,10), stacked = False)
A Couple Of Things To Note
Returns :
Image by Author
For better illustrations, we have only used 50 observations for plotting.
Further operations can be performed. but this will suffice for now so we can move on to our next step.
After understanding our data, we can now proceed to clean it for our use case. We will start by dropping a few columns so that it becomes easier to work.
A careful inspection of our data states we have many irrelevant columns such as ‘mid’, ‘venue’, ‘batsman’, ‘bowler’, ‘striker’, ‘non-striker’.These do not contribute to data and can be removed to save memory space.
To remove the irrelevant column, all we will do is create a list of the irrelevant columns(cols_to_remove) and pass it to the panda’s drop method.
# Removing unwanted cols - reduce memory size cols_to_remove = ['mid' , 'venue' , 'batsman', 'bowler', 'striker', 'non-striker'] df.drop(labels=cols_to_remove , axis=1 , inplace = True)
tip: here we have used axis = 1 to drop by columns and inplace = True to ensure the operation is performed on the same dataset.
Let’s cross-check our operations by printing out a few rows
#cross check df.head(3)
Output:
Image by Author
We can also check the shape for security purposes.
df.shape
>> (76014, 9)
We see that the columns have been dropped successfully as our column size has been reduced to 9
This operation is irreversible and may raise an error if executed in another cell.
Next, we will filter out teams that are consistently playing. This will allow us to have a basic set of teams that are relevant to IPL’s.
1. Find Unique Teams
For that, we will first find the unique teams in bat_team, create a list out of them and then perform our filtering.
To find unique teams, we will just use the unique method over bat_team
# checking for how many batting teams are there df['bat_team'].unique()
>> array(['Kolkata Knight Riders', 'Chennai Super Kings', 'Rajasthan Royals', 'Mumbai Indians', 'Deccan Chargers', 'Kings XI Punjab', 'Royal Challengers Bangalore', 'Delhi Daredevils', 'Kochi Tuskers Kerala', 'Pune Warriors', 'Sunrisers Hyderabad', 'Rising Pune Supergiants', 'Gujarat Lions', 'Rising Pune Supergiant'], dtype=object)
Note dtype = object here is because pandas assume string data type as object data type
2. Creating List
Now we will carve out a list with consistent teams and store it in consitent_team for reference purposes.
# only keep current team which are present consistent_team = ['Kolkata Knight Riders', 'Chennai Super Kings', 'Rajasthan Royals', 'Mumbai Indians','Kings XI Punjab', 'Royal Challengers Bangalore', 'Delhi Daredevils','Sunrisers Hyderabad']
3. Filtering Consistent Teams
Finally, we can filter our required observations based on the condition
“return only those observations for which a consistent team is present in both bat_team and ball_team”.
# filtering based on consistency df = df[(df['bat_team'].isin(consistent_team)) & (df['bowl_team'].isin(consistent_team))]
The above conditions are achieved by nesting the:
4. Checking Results
Now, let’s check the results by printing the unique values again which should return the same unique values.
# printing out unique team after filtering print(df['bat_team'].unique()) print(df['bowl_team'].unique())
Hurray! As expected, we have now filtered the consistent teams. Off to the next process.
HYPOTHESIS:- We can assume that “In most of the matches the actual game starts after elapsing 5 overs”, so it can serve as a good starting point for our training data.
So following our hypothesis. We will just return all the observations after 5 overs by first accessing the overs column of the dataframe(pdf) and using >= operator.
Returns:
We can see that the over columns have only values > 5.
Using these types of hypotheses is very common and can even yield good results as long as it satisfies in real-world, domain knowledge necessary thoughđ
While examining the info on the features column, we found that the data column is an object data type on which no operations can be performed.
We will use the date-time library to convert the data column into a date-time object.
# converting date cols from string to date time object from datetime import datetime df['date'] = df['date'].apply(lambda x: datetime.strptime(x, '%Y-%m-%d'))
Understanding Code:
Now let’s cross-check:
df['date'].dtype
>> dtype('<M8[ns]')
So it seems like we have successfully converted the date column to date-time data_type
This operation is irreversible and may raise an error if executed again.
There are further steps involved but this will suffice for our use case.
Now, as evident from the info, our data takes a variability of data types including strings, data time, and numbers. But our model requires them all to be in a numeric format, so let’s see if we can perform some operations to make it model-friendly.
On careful inspection, we can find that bat_team and bowl_team are categorical data and can be encoded as numbers(0/1). This is called ONE HOT ENCODING and can be achieved by pandas get_dummy function.
# converting categorical features using 'One Hot Encoding' # for numerical values cat_df = pd.get_dummies(data = df, columns = ['bat_team' , 'bowl_team'])
Understanding Code
Now let’s see what fun has to offer in return:
cat_df.head(2)
Output:
As can be seen, the columns have been increased and one-hot encoded. While it is taking up more space, that’s a trade-off to consider while encoding is up to individuals.
Another method is to use a dictionary(key: value) mapping which can be used with the “apply” method.
It’s now time to split our data into (training and testing set) so that we can fit it into our model.
As a general case, we split our data according to ratio train = 80% and test = 20% but here we will be learning how to adapt for a split for a problem domain.
What we will do is, instead of going for split size based on ratio, we will perform the split based on years.
# split the data into train and test set - based on date column X_train = cat_df.drop(labels = 'total', axis = 1)[cat_df['date'].dt.year <= 2016] X_test = cat_df.drop(labels = 'total', axis = 1) [cat_df['date'].dt.year >= 2017]
So all I did here is define the variables, drop the last column, and filter out the observation based on conditions that return only year values from the date column.
For labels we only use a single column (total) which Panda assumes to be a series so we need to use values over it, else it will return an error sometimes.
# since only one column so cosidered as series y_train = cat_df[cat_df['date'].dt.year <= 2016]['total'].values y_test = cat_df[cat_df['date'].dt.year >= 2017]['total'].values
Theory:
Let’s check the split now for cross-verifying.
>> 37330, 22) (37330,) (2778, 22) (2778,)
The split was successful. Also notice the (2778,) for y_test is because it is a single column with 2778 observations.
One last thing we can do before closing this section is to drop off the data column from our training and test sets, as it is of no use to us now as this will help free up some memory space for further operations.
We will use the same drop method used earlier to perform the dropping.
# since the requirement of our date column is over so we can drop it # dropping date column X_train.drop(labels = 'date', axis = True, inplace = True) X_test.drop(labels = 'date', axis = True, inplace = True)
Now let’s check how splits look using the “display” method – Jupyter specific.
# use display to cross check in single line - only one display("X_train", X_train.head(1)) display("X_test", X_test.head(1))
Output:
Our data looks prettyđ€ and as expected for model feeding (all numeric values).
If you have come this far with me – congrats, you have understood the essence of being a data scientist (yep that’s what a data scientist does!).
Now the only thing left is to choose a model/build it for training, evaluate the results, and finally save it for the latter use case.
After all the hard work we did, it’s now time to choose a model architecture for our use case. This will be a function that will find a way to map our training set to training labels thus allowing prediction of the score and given input data.
Since we have seen that the data mostly have a linear relationship (as evident from the description) we will use a simple linear regression model for our use case from sklearn.
1. Importing Model
Before any further steps, we will first import the Linear Regression model from sklearn’s linear_model class and instantiate a linear regression object as a reg.
The job of the linear regression model is to find the line of best fit, for which error/loss is very low
2. Training Model
Now it’s finally time to train our model. To do so we just use the fit method and pass our training set.
# training model reg.fit(X_train , y_train)
>> LinearRegression(copy_X=True, fit_intercept=True, n_jobs=None, normalize=False)
That’s pretty much it, we have trained our model and can be used to predict our test set.
For more details on Linear Regression, consider checking the official documentation page.
After training comes the evaluation part which tells how our model is performing on data other than the train set. If it performs well, it is good to be deployed in the wild, else you need to reiterate the entire process(or at least a few processes, boils down to how you apply logic).
Let’s test our model too using the prediction method. We will be passing the X_test to the same.
# getting predictions prediction = reg.predict(X_test)
The results returned are predictions and are not user-friendly to understand, so it is better to plot them.
Since we have predictions and actual data(y_test) we can use Seaborn’s distplot which will plot the distribution of prediction vs actual data – in simple words difference between 2(as distance measure).
# plotting our fit import seaborn as sns sns.distplot(y_test-prediction)
Returns:
We can notice that the difference between actual and predicted data (line) is not much and resembles a Gaussian/normal distribution(data symmetric around the mean) as evident from the bell curve.
To further check this, we can also use the evaluation metrics
Plots are a great way to visualize results but it will be much more convenient if we can just summarize the result in a single metric.
Luckily we have a few stats metrics provided by sklearn itself for the task. These include MAE, MSE, and RMSE.
# checking for scores from sklearn import metrics
We just imported metrics from sklearn
# checking for scores from sklearn import metrics import numpy as np # Mean Absolute Error print('MAE: ', metrics.mean_absolute_error(y_test , prediction)) # Mean Squared Error print('MSE: ', metrics.mean_squared_error(y_test, prediction)) # Root Mean Squared Error print('RMSE: ', np.sqrt(metrics.mean_squared_error(y_test, prediction)))
Returns:
Great, we have our MAE to be 12 & RMSE to be 15, we can now proceed further as our score matches the plot. However, if you like you can tweak the data and model to get a score even lower than this (quite possibly!).
NOTE: MSE – Is large because of the square terms in the formula (refer to figure below)
Finally, we can now save our model.
To save our model we will be using the pickle library and dumping our model(refer here to learn more)
Here (dump) means to parse the value to create a serialized object using a pickle.
# creating our model pickel file - saving model file_name = 'ipl_score_predict_model.pkl' pickle.dump(reg , open(file_name,'wb'))
Here ‘wb’ refers to write binary and open is a constructor which creates a file provided in the file_name(File I/O).
This will return the model file in the directory.
Output:
Now let’s see how to deploy our model in the next section.
The next step is to deploy our model but before that let’s choose the deployment architecture. Here is a quick summary image to get started with.
PAAS – Platform As A Service architecture is convenient as we only need to take care of our data part. So we will choose it.
For the platform, we will be using Heroku.
For deployment, one can get lost easily, so we will follow the below steps in sequence
1. Signup
We will just go to Heroku’s official site and signup. Making sure we select the role as a student will allow us to get free dynos.
Finally, we will download the CLI tool provided as it will allow us to see the logs.
2. Creating a Web App Using Flask
Next, we will create a simple web app using HTML, CSS, and JS for the front end & flask and request for the backend.
We will create the following files:
pip freeze > requirements.txt.
This is what our working directory should look like:
Image by Author
Note:
Now let’s head over to git and create a new repository called ipl-demo. Make sure to set it to the public else the deployment will not happen.
Image by Author
Next, we will commit our code and some other files. Folders need to be dragged and dropped manually.
That’s all our work is done here. So let’s move on to the next part.
To let Heroku take care of every step, we need to connect to our GitHub account.
1- You can do so by logging into the created account on the platform and following the instructions below:
1- On the home page, click the new button.
2- Fill in the details. we need to add a unique app name – ipl-score-predictor (in our case).
3- Connect your GitHub account. While searching, we provided the same repository name we created on GitHub (imp) and clicked connect.
To deploy, we can use either automatic deployment or manual deployment. To keep things simple, we will just click the Deploy Branch and then wait for a minute or 3.
OUTPUT:
NOTE: The use of automatic deployment is preferred as it makes changes automatically as we upload new model versions
Looks like we have successfully deployed our model. To access the app, all we need to do is to click on the View button shown in the image above and test our web app.
Here is a glimpse of what we have achieved:
We can now share this with anyone in the world using the site URL.
Here are a few of the conclusions we can derive from this article:
With this, we have come to an end with this daunting article/guide. This has not been an easy one to write, but I hope it will be an easy one to understand and you will apply the knowledge you have gained through this guide.
Feel free to comment on any suggestions in the comments section belowđ . Also, if you want to connect with me, reach me on LinkedIn / Github / Twitter.
Here are a few of the references who are eager to learn more:
Image Ref: Evaluation Metrics
Inspiration: Inspiration For Learning
Code and Related Files: Github Repository