Linear Regression Complete Derivation With Mathematics Explained!
Last Updated on May 26, 2020 by Editorial Team
Author(s): Pratik Shukla
Machine Learning
Part 3/5 in Linear Regression
Part 1: Linear Regression FromΒ Scratch.
Part 2: Linear Regression Line Through BruteΒ Force.
Part 3: Linear Regression Complete Derivation.
Part 4: Simple Linear Regression Implementation FromΒ Scratch.
Part 5: Simple Linear Regression Implementation Using Scikit-Learn.
In the last article, we saw how we could find the regression line using brute force. But that is not that fruitful for our data, which is usually in millions. So to tackle such datasets, we use python libraries, but such libraries are built on some logical theories, right? So letβs find out the logic behind some creepy looking formulas. Believe me, the math behind it isΒ sexier!
Before we begin, the knowledge of the following topics might beΒ helpful!
- Partial Derivatives
- Summations
Are you excited to find the line of bestΒ fit?
Letβs start by defining a fewΒ things
1) Given n inputs andΒ outputs.
2) We define the line of best fitΒ as:
3) Now we need to minimize the error function we namedΒ S
4) Put the value of equation 2 into equationΒ 3.
To minimize our error function, S, we must find where the first derivative of S is equal to 0 concerning a and b. The closer a and b are to 0, the less total error for each point is. Letβs find the partial derivative of aΒ first.
Finding aΒ :
1 ) Find the derivative of S concerning a.
2 ) Using the chain rule, letβsΒ say
3) Using partial derivative
4) Expanding
5) Simplifying
6) To find extreme values, we put it toΒ zero
7) Dividing the left side withΒ -2
8) Now letβs break the summation in 3Β parts
9) Now the summation of aΒ is
10) Substituting it back in theΒ equation
11) Now we need to solve forΒ a
12) The summation of Y and x divided by n, is simply itβsΒ mean
Weβve minimized the cost function concerning x. Now letβs find the last part which S concerning b.
Finding BΒ :
1 ) Same as we have done withΒ a
2) Finding the partial derivative
3) Expanding it aΒ bit
4) Putting it back in theΒ equation
5) Letβs divide by -2 bothΒ sides
6) Letβs distribute x for ease ofΒ viewing
Now letβs do something fun! Remember, we found the value of earlier in this article? Why donβt we substitute it? Well, letβs see whatΒ happens.
7) Substituting the value ofΒ a
8) Letβs distribute the minus sign andΒ x
Well, you donβt like it? Letβs split up the sum into twoΒ sums
9) Splitting theΒ sum
10) Simplifying
11) Finding B fromΒ it
Great. We did it. We have isolated a and b in the form of x and y. It wasnβt that hard, wasΒ it?
I still have some energy and want to explore it aΒ bit!
12 ) Simplifying theΒ formula
13) Multiplying numerator and denominator by n in equationΒ 11
14) Now if we simplify the value of a using equation 13 weΒ get
SummaryΒ π
If you have a dataset with one independent variable, then you can find the line that best fits by calculating B.
Then substituting B intoΒ a
And finally substituting B and a into the line of bestΒ fit
Moving Onwards,
In the next article, weβll see how we can implement simple linear regression from scratch (without sklearn) inΒ Python.
And please let me know whether you liked this article or not! I bet you likedΒ it.
To find more such detailed explanation, visit my blog: patrickstar0110.blogspot.com
(1) Simple Linear Regression Explained With Its Derivation.
(2)How to Calculate The Accuracy Of A Model In Linear Regression FromΒ Scratch.
(3) Simple Linear Regression Using Sklearn.
You can download the code and some handwritten notes on the derivation on GoogleΒ Drive.
If you have any additional questions, feel free to contact me: [email protected].
Linear Regression Complete Derivation With Mathematics Explained! was originally published in Towards AIβββMultidisciplinary Science Journal on Medium, where people are continuing the conversation by highlighting and responding to this story.
Published via Towards AI