Skip to main content

Machine Learning-Linear Regression

Another post starts with you beautiful people!
In my previous posts we have learnt the Python basics and advanced, statistics techniques for the Data Science track. I suggest you to please read previous post just for 10-15 min. before sleeping daily and then there is no any obstacle to stop you to become a great Data Scientist.
In this post we will start our Machine Learning track with the Linear Regression topic. I Have highlighted the both so please click on the link to know the formal definition of those.

Machine learning- More specifically the field of predictive modeling is primarily concerned with minimizing the error of a model or making the most accurate predictions possible, at the expense of explainability.
In applied machine learning we will borrow, reuse and steal algorithms from many different fields, including statistics and use them towards these ends.

Linear Regression was developed in the field of statistics and is studied as a model for understanding the relationship between input and output numerical variables, but has been borrowed by machine learning. It is both a statistical algorithm and a machine learning algorithm.For more details please see here-tell me more!

Form of linear regression-

  • y=β0+β1x1+β2x2+...+βnxny=β0+β1x1+β2x2+...+βnxn 
  • yy  is the response
  • β0β0  is the intercept
  • β1β1  is the coefficient for  x1x1  (the first feature)
  • βnβn  is the coefficient for  xnxn  (the nth feature)

The  Î²Î²  values are called the model coefficients:




  • These values are estimated (or "learned") during the model fitting process using the least squares criterion.
  • Specifically, we find the line (mathematically) which minimizes the sum of squared residuals (or "sum of squared errors").
  • And once we've learned these coefficients, we can use the model to predict the response.

Agenda of this post-
Introducing the bikeshare dataset

  • Reading in the data
  • Visualizing the data

Linear regression basics

  • Form of linear regression
  • Building a linear regression model
  • Using the model for prediction
  • Does the scale of the features matter?

Working with multiple features

  • Visualizing the data (part 2)
  • Adding more features to the model

Choosing between models
Feature selection

  • Evaluation metrics for regression problems
  • Comparing models with train/test split and RMSE
  • Comparing testing RMSE with null RMSE
  • Creating features

Handling categorical features

  • Feature engineering
  • Advantages/Disadvantages


In this exercise we'll be working with a dataset from Capital Bikeshare that was used in a Kaggle competition.
This dataset contains the hourly and daily count of rental bikes between years 2011 and 2012 in Capital bikeshare system with the corresponding weather and seasonal information.
Bike sharing systems are new generation of traditional bike rentals where whole process from membership, rental and return back has become automatic.
Through these systems, user is able to easily rent a bike from a particular position and return back at another position.

Importing Data set and Basic Data Exploration-


The structure of our dataset will look like below-

We will deal with following questions in our exercise-

  • What does each observation represent?
  • What is the response variable (as defined by Kaggle)?
  • How many features are there?
If you notice the data structure of our dataset, there is a column named as 'count'; which is also a method name. So it's good to rename it first-


Let's visualize the data-

Result-

Now to know how a response variable y changes as an explanatory variable x changes, we will draw the regression line. So you can say in simple word that a regression line is a connection between data points in a set.

Result-

Next, we are going to Build a linear regression model-


Points to think from above code-

Using the above model for prediction-

Does the scale of the features matter?

Let's say that temperature was measured in Fahrenheit, rather than Celsius. How would that affect the model?

Result-

Let's plot again with this Fahrenheit change to see effect-




Conclusion: The scale of the features is irrelevant for linear regression models. When changing the scale, we simply change our interpretation of the coefficients.

Let's explore other features-

We can also use pandas library to built multiple scatter plots-

Are you seeing anything that you did not expect? Yes, the regression line is missing.

Now we will analyze our categorical data (categorical data is data or variables that are separated into different categories that are mutually exclusive from one another) with the help of Cross tabulation tool. tell me more about crosstab!
Let's plot a box plot of rentals wrt to season-


From the plot we notice that :

  • A line can't capture a non-linear relationship.
  • There are more rentals in winter than in spring.
Let's draw a line plot to see the relation between rentals and weather-



Cool! There are more rentals in the winter than the spring, but only because the system is experiencing overall growth and the winter months happen to come after the spring months.

Let's find out the correlation and draw a heat map-



See above plot and think about What relationships do you notice?

Adding more features to the model-


Interpreting the coefficients:
  • Holding all other features fixed, a 1 unit increase in temperature is associated with a rental increase of 7.86 bikes.
  • Holding all other features fixed, a 1 unit increase in season is associated with a rental increase of 22.5 bikes.
  • Holding all other features fixed, a 1 unit increase in weather is associated with a rental increase of 6.67 bikes.
  • Holding all other features fixed, a 1 unit increase in humidity is associated with a rental decrease of 3.12 bikes.
Does anything look incorrect?

Feature selection-
How do we choose which features to include in the model? For this we're going to use train/test split (and eventually cross-validation).

Why not use of p-values or R-squared for feature selection?
  • Linear models rely upon a lot of assumptions (such as the features being independent), and if those assumptions are violated, p-values and R-squared are less reliable. Train/test split relies on fewer assumptions.
  • Features that are unrelated to the response can still have significant p-values.
  • Adding features to your model that are unrelated to the response will always increase the R-squared value, and adjusted R-squared does not sufficiently account for this.
  • p-values and R-squared are proxies for our goal of generalization, whereas train/test split and cross-validation attempt to directly estimate how well the model will generalize to out-of-sample data.
Evaluation metrics for regression problems-
Evaluation metrics for classification problems, such as accuracy, are not useful for regression problems. We need evaluation metrics designed for comparing continuous values.
Here are three common evaluation metrics for regression problems:-

Don't afraid to above formulas, Python's sklearn library make it simple to use these formulas as below-
Comparing these metrics:
  • MAE is the easiest to understand, because it's the average error.
  • MSE is more popular than MAE, because MSE "punishes" larger errors, which tends to be useful in the real world.
  • RMSE is even more popular than MSE, because RMSE is interpretable in the "y" units.
All of these are loss functions, because we want to minimize them.

Comparing models with train/test split and RMSE-
Comparing testing RMSE with null RMSE-
Null RMSE is the RMSE that could be achieved by always predicting the mean response value. It is a benchmark against which you may want to measure your regression model.

Handling categorical features-
Since scikit-learn expects all features to be numeric. So how do we include a categorical feature in our model?
  • Ordered categories: transform them to sensible numeric values (example: small=1, medium=2, large=3)
  • Unordered categories: use dummy encoding (0/1)
What are the categorical features in our dataset?
  • Ordered categories: weather (already encoded with sensible numeric values)
  • Unordered categories: season (needs dummy encoding), holiday (already dummy encoded), workingday (already dummy encoded)
For season, we can't simply leave the encoding as 1 = spring, 2 = summer, 3 = fall, and 4 = winter, because that would imply an ordered relationship
Instead, we create multiple dummy variables:

However, we actually only need three dummy variables (not four), and thus we'll drop the first dummy variable.
Why? Because three dummies captures all of the "information" about the season feature, and implicitly defines spring (season 1) as the baseline level:

In general, if you have a categorical feature with k possible values, you create k-1 dummy variables.
If that's confusing, think about why we only need one dummy variable for holiday, not two dummy variables (holiday_yes and holiday_no).



How do we interpret the season coefficients? 
They are measured against the baseline (spring):
  • Holding all other features fixed, summer is associated with a rental decrease of 3.39 bikes compared to the spring.
  • Holding all other features fixed, fall is associated with a rental decrease of 41.7 bikes compared to the spring.
  • Holding all other features fixed, winter is associated with a rental increase of 64.4 bikes compared to the spring.
Would it matter if we changed which season was defined as the baseline?
No, it would simply change our interpretation of the coefficients.

Important: Dummy encoding is relevant for all machine learning models, not just linear regression models.

Compare original season variable with dummy variables-

That's the end of our model and we come to know following-
Advantages of linear regression:
  • Simple to explain
  • Highly interpretable
  • Model training and prediction are fast
  • No tuning is required (excluding regularization)
  • Features don't need scaling
  • Can perform well with a small number of observations
  • Well-understood
Disadvantages of linear regression:
  • Presumes a linear relationship between the features and the response
  • Performance is (generally) not competitive with the best supervised learning methods due to high bias
  • Can't automatically learn feature interactions
Hope you enjoyed and must learn something today!
Please try above code in your notebook and explore more.











Comments

  1. It is nice blog Thank you porovide importent information and i am searching for same information
    Tableau Online Training

    ReplyDelete
  2. AI Patasala Machine Learning Training in the Hyderabad Program is the best platform for Machine Learning professionals to transform into professionals in their field.
    AI Patasala Machine Learning Course

    ReplyDelete
  3. There are also varieties of options available to choose a career in data science, like you can become a data scientist, developer, data engineer, analyst, database administrator, and many more.
    data science training in patna

    ReplyDelete

Post a Comment

Popular posts from this blog

How to install and compile YOLO v4 with GPU enable settings in Windows 10?

Another post starts with you beautiful people! Last year I had shared a post about  installing and compiling Darknet YOLOv3   in your Windows machine and also how to detect an object using  YOLOv3 with Keras . This year on April' 2020 the fourth generation of YOLO has arrived and since then I was curious to use this as soon as possible. Due to my project (built on YOLOv3 :)) work I could not find a chance to check this latest release. Today I got some relief and successfully able to install and compile YOLOv4 in my machine. In this post I am going to share a single shot way to do the same in your Windows 10 machine. If your machine does not have GPU then you can follow my  previous post  by just replacing YOLOv3 related files with YOLOv4 files. For GPU having Windows machine, follow my steps to avoid any issue while building the Darknet repository. My machine has following configurations: Windows 10 64 bit Intel Core i7 16 GB RAM NVIDIA GeForce G...

How can I make a simple ChatBot?

Another post starts with you beautiful people! It has been a long time of posting a new post. But my friends in this period I was not sitting  where I got a chance to work with chatbot and classification related machine learning problem. So in this post I am going to share all about chatbot- from where I have learned? What I have learned? And how can you build your first bot? Quite interesting right! Chatbot is a program that can conduct an intelligent conversation based on user's input. Since chatbot is a new thing to me also, I first searched- is there any Python library available to start with this? And like always Python has helped me this time also. There is a Python library available with name as  ChatterBot   which is nothing but a machine learning conversational dialog engine. And yes that is all I want to start my learning because I always prefer inbuilt Python library to start my learning journey and once I learn this then only I move ahead for another...

Detecting Credit Card Fraud As a Data Scientist

Another post starts with you beautiful people! Hope you have learnt something from my previous post about  machine learning classification real world problem Today we will continue our machine learning hands on journey and we will work on an interesting Credit Card Fraud Detection problem. The goal of this exercise is to anonymize credit card transactions labeled as fraudulent or genuine. For your own practice you can download the dataset from here-  Download the dataset! About the dataset:  The datasets contains transactions made by credit cards in September 2013 by european cardholders. This dataset presents transactions that occurred in two days, where we have 492 frauds out of 284,807 transactions. The dataset is highly unbalanced, the positive class (frauds) account for 0.172% of all transactions. Let's start our analysis with loading the dataset first:- As per the  official documentation -  features V1, V2, ... V28 are the principal compo...