0% found this document useful (0 votes)
8 views

3 - Linear Regression-Least Square Error Fit

Uploaded by

keshavkeshav2608
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
8 views

3 - Linear Regression-Least Square Error Fit

Uploaded by

keshavkeshav2608
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 35

Linear Regression

(Least Square Error Fit)


D r. J A S ME ET S I NGH
A S S I STANT P R OF E SS OR , C S E D
T I E T, PATI ALA
Linear Regression
 In machine learning and statistics, regression attempts to determine the strength
and character of the relationship between one dependent variable (usually
denoted by Y) and a series of other variables (known as independent variables).

 Mathematically, regression analysis uses an algorithm to learn the mapping


function from the input variables to the output variable (Y) i.e. Y = f(x) where Y
is a continuous or real valued variable.

 Regression is said to be linear regression if the output dependent variable is a


linear function of the input variables.
Regression

Regression is a statistical method used to find the relationship between one dependent

variable and a series of other variables.

It attempts to model an output feature between output feature and input features by

fitting an equation to the observed data.

It is supervised machine learning approach.


Regression

Regression is a method of modelling a output feature based on input features.

It finds a relationship between output and input features.

 Simple Regression
 Simple Linear Regression
 Simple Non-Linear Regression

 Multiple Regression
 Multiple Linear Regression
 Multiple Non-Linear Regression
Linear Regression
 Linear regression is a linear model, e.g. a model that assumes a linear relationship between

the input variables (x) and the single output variable (y). More specifically, that y can be

calculated from a linear combination of the input variables (x).

 Examples of Linear Regression are:

 Predicting price of a house on the basis of rooms

 Predicting salary of a person on the basis of his experience


Example of Non Linear Regression
 Nonlinear regression relates input variables (x) and the single output variable (y) in a

nonlinear (curved) relationship.

 Example:

 Prediction of population growth over time. A scatterplot of changing population data

over time shows that there seems to be a relationship between time and population

growth, but that it is a nonlinear relationship, requiring the use of a nonlinear regression

model. A logistic population growth model can provide estimates of the population for

periods that were not measured, and predictions of future population growth.
Linear Regression
 Nonlinear regression and linear regression models are similar in one sense that both seek to

track a particular response(output) from a set of variables(input) graphically.

 Nonlinear models are more complicated than linear models to develop because the function

is created through a series of approximations (iterations) that may stem from trial-and-error.

Mathematicians use several established methods, such as the Gauss-Newton method and the

Levenberg-Marquardt method.
Linear Regression

You have to make sure that a linear relationship exists between the dependent

feature and the independent feature(s).

Figure 1(a) Figure 1(b)


Linear regression
• Given an input feature x,
compute an output feature y
• Output feature is of
continuous type. Y

• For example:
- Predict height from age
- Predict house price from
number of bed rooms
X
X: No of Bed Y: Price(in Lacs)
Rooms
5 101.5
3 79.9
5 99.87
3 56.9
2 66.6
5 105.45
5 126.3
4 89.25
5 99.97
3 87.6
4 112.6
3 85.6
3 78.5
2 74.3
2 74.8
Figure 2(a) Figure 2(b)
Simple Linear Regression (SLR)
 Simple linear regression is a linear regression model with a single explanatory
variable.

 It concerns two-dimensional sample points with one independent variable and


one dependent variable and finds a linear function (a non-vertical straight line)
that, as accurately as possible, predicts the dependent variable values as a
function of the independent variable.

 The adjective simple refers to the fact that the outcome variable is related to a
single predictor.
Simple Linear Regression (SLR) Contd….
 Simple linear regression finds a linear
function (a non-vertical straight line)
that, as accurately as possible, predicts
the dependent variable values as a
function of the independent variable.
 For instance, in the house price
predicting problem (with only one
input variable-plot size), a linear
regressor will fit a straight line with x-
axis representing plot size and y-axis
representing price.
Fitting the Straight Line for SLR
 The linear function that binds the input variable x with the corresponding
predicted value of (yˆ) can be given by the equation of straight line(slope-
intercept form) as:
𝑦ˆ=𝛽0 + 𝛽1 𝑥
 where 𝛽1 is the slope of line (i.e. it measures change in output variable y with
unit change in independent variable x).
 𝛽0 represents y-intercept i.e. the point at which the line touch x-axis
 𝑦ˆ is the predicted value of the output for the particular value of input variable x.
Cost/Error function for SLR
 The major goal of SLR model is to fit the straight line that predicts the output
variable value quite close to the actual value.

 But, in real world scenario, there is always some error (regression residual) in
predicting the values, i.e.
𝑎𝑐𝑡𝑢𝑎𝑙 𝑣𝑎𝑙𝑢𝑒𝑖 = 𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑒𝑑 𝑣𝑎𝑙𝑢𝑒𝑖 + 𝑒𝑟𝑟𝑜𝑟
𝑦𝑖 = 𝑦ˆ𝑖 + 𝜖𝑖
𝑅𝑒𝑠𝑖𝑑𝑢𝑎𝑙 𝐸𝑟𝑟𝑜𝑟 = 𝜖𝑖 = 𝑦𝑖 − 𝑦ˆ𝑖

This error may be positive or negative, as it may predict values greater or lesser
than actual values. So we consider square of each error value.
Cost/Error function for SLR
 The total error for all the n points in the dataset is given by:

𝑇𝑜𝑡𝑎𝑙 𝑆𝑞𝑢𝑎𝑟𝑒 𝐸𝑟𝑟𝑜𝑟 = σ𝑛𝑖=1 𝜖𝑖2 = σ𝑛𝑖=1(𝑦𝑖 − 𝑦ˆ𝑖 )2


𝑛

= ෍ (𝑦𝑖 − 𝛽0 − 𝛽1 𝑥𝑖 )2
𝑖=1

 The mean of square error is called the cost or error function for simple linear function denoted by J(𝛽0 ,
𝛽1 ) and given by:
𝑛
1
J(𝛽0 , 𝛽1 )= ෍(𝑦𝑖 − 𝛽0 − 𝛽1 𝑥𝑖 )2
𝑛
𝑖=1

 There exist many methods to optimize (minimize) this cost/error function to find line of best fit.
Least Square Method for Line of Best Fit
 The least square method aims to find values 𝛽0 ˆand 𝛽1 ˆ for 𝛽0 and 𝛽1 for which the square error
between the actual and the predicted values is minimum i.e. least (So, the name is least square
error fit).
 The values 𝛽0 ˆand 𝛽1 ˆ for 𝛽0 and 𝛽1 for which the square error function (J (𝛽0 , 𝛽1 )) is minimum
are computed using second derivative test as below:
𝜕 J (𝛽0 ,𝛽1 ) 𝜕 J (𝛽0 ,𝛽1 )
1. Compute partial derivatives of J (𝛽0 , 𝛽1 ) w.r.t 𝛽0 and 𝛽1 i.e. 𝜕𝛽0
and 𝜕 𝛽1

𝜕 J (𝛽0 ,𝛽1 )
2. Find values 𝛽0 ˆand 𝛽1 ˆ for which
𝜕𝛽0
=0 and 𝜕 J 𝜕𝛽
(𝛽0 ,𝛽1 )
=0
1
𝜕 2J (𝛽0 ,𝛽1 ) 𝜕2 J (𝛽0 ,𝛽1 )
3. Find second partial derivative and ; and prove it be minimum for 𝛽0 ˆand
𝜕𝛽0 2 𝜕𝛽1 2
𝛽1 ˆ .
Least Square Error Fit- Contd…..
𝑇𝑜𝑡𝑎𝑙 𝑆𝑞𝑎𝑢𝑟𝑒 𝐸𝑟𝑟𝑜𝑟 = 𝐽(𝛽0 , 𝛽1 ) = σ𝑛𝑖=1(𝑦𝑖 − 𝛽0 − 𝛽1 𝑥𝑖 )2
𝜕 J (𝛽0,𝛽1 ) 𝜕 J (𝛽0 ,𝛽1 )
Step 1: Compute partial derivatives of J (𝛽0, 𝛽1 ) w.r.t 𝛽0 and 𝛽1 i.e. 𝜕𝛽0
and 𝜕𝛽1

𝜕 J (𝛽0 ,𝛽1 )
= −2 σ𝑛𝑖=1(𝑦𝑖 − 𝛽0 − 𝛽1 𝑥𝑖 )
𝜕𝛽0
𝜕 J (𝛽0 ,𝛽1 )
= −2 σ𝑛𝑖=1 𝑦𝑖 − 𝛽0 − 𝛽1 𝑥𝑖 𝑥𝑖 = −2 σ𝑛𝑖=1(𝑥𝑖 𝑦𝑖 − 𝛽0 𝑥𝑖 − 𝛽1 𝑥𝑖2)
𝜕𝛽1

𝜕 J (𝛽0,𝛽1 ) 𝜕 J (𝛽0 ,𝛽1 )


Step 2: Find values 𝛽0 ˆand 𝛽1 ˆ for which
𝜕𝛽0
= 0 and 𝜕𝛽1
=0
σ𝑛𝑖=1(𝑦𝑖 − 𝛽0 ˆ − 𝛽1 ˆ𝑥𝑖 ) = 0 (1)
and σ𝑛𝑖=1(𝑥𝑖 𝑦𝑖 − 𝛽0 ˆ𝑥𝑖 − 𝛽1 ˆ𝑥𝑖2) = 0 (𝟐)
Least Square Error Fit- Contd…..
From equation 1:
σ𝑛𝑖=1 𝑦𝑖 − σ𝑛𝑖=1 𝛽0 ˆ − σ𝑛𝑖=1 𝛽1 ˆ𝑥𝑖 = 0
σ𝑛𝑖=1 𝑦𝑖 − 𝑛𝛽0 ˆ − 𝛽1 ˆ σ𝑛𝑖=1 𝑥𝑖 = 0
𝑛𝛽0 ˆ + 𝛽1 ˆ σ𝑛𝑖=1 𝑥𝑖 = σ𝑛𝑖=1 𝑦𝑖 (3)
From equation 2:
σ𝑛𝑖=1 𝑥𝑖 𝑦𝑖 − σ𝑛𝑖=1 𝛽0 ˆ𝑥𝑖 − σ𝑛𝑖=1 𝛽1 ˆ𝑥𝑖2 = 0
σ𝑛𝑖=1 𝑥𝑖 𝑦𝑖 − 𝛽0 ˆ σ𝑛𝑖=1 𝑥𝑖 − 𝛽1 ˆ σ𝑛𝑖=1 𝑥𝑖2 = 0
𝛽0 ˆ σ𝑛𝑖=1 𝑥𝑖 + 𝛽1 ˆ σ𝑛𝑖=1 𝑥𝑖2 = σ𝑛𝑖=1 𝑥𝑖 𝑦𝑖 (4)
Least Square Error Fit- Contd…..
Multiply equation 3 with σ𝑛𝑖=1 𝑥𝑖 and equation 4 by n
𝑛𝛽0ˆ σ𝑛𝑖=1 𝑥𝑖 + 𝛽1 ˆ( σ𝑛𝑖=1 𝑥𝑖 )2 = σ𝑛𝑖=1 𝑥𝑖 σ𝑛𝑖=1 𝑦𝑖 (5)
n𝛽0ˆ σ𝑛𝑖=1 𝑥𝑖 + 𝛽1ˆn σ𝑛𝑖=1 𝑥𝑖2 = 𝑛 σ𝑛𝑖=1 𝑥𝑖 𝑦𝑖 (6)
Subtracting Equation 5 from 6, we get,
𝒏 σ𝒏 𝒏 𝒏
𝒊=𝟏 𝒙𝒊 𝒚𝒊 − σ𝒊=𝟏 𝒙𝒊 σ𝒊=𝟏 𝒚𝒊
𝜷𝟏 ˆ =
n σ𝒏𝒊=𝟏 𝒙𝟐𝒊 −( σ𝒏𝒊=𝟏 𝒙𝒊 )𝟐

From Equation (3),


𝑛𝛽0ˆ + 𝛽1ˆ σ𝑛𝑖=1 𝑥𝑖 = σ𝑛𝑖=1 𝑦𝑖
1 1
𝛽0ˆ = σ𝑛𝑖=1 𝑦𝑖 − 𝛽1ˆ σ𝑛𝑖=1 𝑥𝑖
𝑛 𝑛
𝜷𝟎 ˆ = ഥ
𝒚 − 𝜷𝟏 ˆഥ
𝒙
Least Square Error Fit- Contd…..
𝜕 2J (𝛽0 ,𝛽1 ) 𝜕2 J (𝛽0 ,𝛽1 )
Step 3: Find second partial derivative
𝜕𝛽0 2
and 𝜕𝛽1 2
; and prove it be minimum for
𝛽0 ˆand 𝛽1 ˆ .
𝜕 2J (𝛽0 ,𝛽1 ) 𝜕(−2 σ𝑛
𝑖=1(𝑦𝑖 −𝛽0 −𝛽1 𝑥𝑖 )
= = −2 × −1 = 2
𝜕𝛽0 2 𝜕𝛽0

𝜕 2J (𝛽0 ,𝛽1 ) 𝜕−2 σ𝑛 2


𝑖=1(𝑥𝑖 𝑦𝑖 −𝛽0 𝑥𝑖 −𝛽1 𝑥𝑖 )
2 = = −2 × −𝑥𝑖2 = 2𝑥𝑖2
𝜕𝛽1 𝜕𝛽1

Therefore, both are positive i.e. the cost function continuously increases with respect to 𝛽0
and 𝛽1 beyond 𝛽0 ˆ and 𝛽1 ˆ. But attains it minimum value at 𝛽0 ˆ and 𝛽1 ˆ
Least Square Error Fit- Summary
 The linear function that binds the input variable x with the corresponding predicted value of (yˆ)
can be given by the equation of straight line(slope-intercept form) as:
𝑦ˆ=𝛽0 + 𝛽1 𝑥
 The square error in prediction is minimized when
𝑛 σ𝑛𝑖=1 𝑥𝑖 𝑦𝑖 − σ𝑛𝑖=1 𝑥𝑖 σ𝑛𝑖=1 𝑦𝑖
𝛽1 ˆ =
n σ𝑛𝑖=1 𝑥𝑖2 − ( σ𝑛𝑖=1 𝑥𝑖 )2
𝜎𝑦 σ𝑛𝑖=1(𝑥𝑖 − 𝑥)(𝑦 ҧ 𝑖 − 𝑦) ത
= 𝑟𝑥𝑦 =
𝜎𝑥 σ𝑛𝑖=1(𝑥𝑖 − 𝑥)ҧ 2
and 𝛽0 ˆ = 𝑦ത − 𝛽1 ˆ𝑥ҧ
Least Square Error Fit- Example
Height (m), xi Mass (kg), yi
The data set (shown in table) gives
1.47 52.21
average masses for women as a 1.50 53.12
function of their height in a sample of 1.52 54.48
American women of age 30–39. 1.55 55.84
1.57 57.20
(a) Fit a square line for average mass as 1.60 58.57
function of height using least square 1.63 59.93

error method. 1.65 61.29


1.68 63.11
(b) Predict the average mass of women 1.70 64.47

whose height is 1.40 m 1.73 66.28


1.75 68.10
1.78 69.92
1.80 72.19
1.83 74.46
Least Square Error Fit- Example
i xi yi x i2 x iy i
1 1.47 52.21 2.1609 76.7487
2 1.50 53.12 2.25 79.68
3 1.52 54.48 2.3104 82.8096
4 1.55 55.84 2.4025 86.552
5 1.57 57.20 2.4649 89.804
6 1.60 58.57 2.56 93.712
7 1.63 59.93 2.6569 97.6859
8 1.65 61.29 2.7225 101.1285
9 1.68 63.11 2.8224 106.0248
10 1.70 64.47 2.89 109.599
11 1.73 66.28 2.9929 114.6644
12 1.75 68.10 3.0625 119.175
13 1.78 69.92 3.1684 124.4576
14 1.80 72.19 3.24 129.942
15 1.83 74.46 3.3489 136.2618
Total 24.76 931.17 41.0532 1548.2453
Least Square Error Fit- Example
𝑛 σ𝑛 𝑛 𝑛
𝑖=1 𝑥𝑖 𝑦𝑖 − σ𝑖=1 𝑥𝑖 σ𝑖=1 𝑦𝑖
𝛽1ˆ =
n σ𝑛𝑖=1 𝑥2𝑖 −( σ𝑛𝑖=1 𝑥𝑖 )2

15×1548.2453 −24.76 ×931.17


𝛽1ˆ = 15×41.0532−24.762
= 61.19

𝛽0ˆ = 𝑦ത − 𝛽1ˆ 𝑥ҧ
931.17 24.76
𝛽0 ˆ = − 61.19 × = −38.88
15 15

Therefore the line of best fit is given by: 𝑦ˆ= − 38.88 + 61.19𝑥
Predicted value of y when x is 1.4 is
𝑦ˆ= − 38.88 + 61.19 × 1.4 = 46.78
Multiple Linear Regression (MLR)
 Multiple regression models describe how a single response variable Y depends
linearly on a number of predictor variables.

 Examples:
The selling price of a house can depend on the desirability of the location, the
number of bedrooms, the number of bathrooms, the year the house was built,
the square footage of the lot and a number of other factors.
 The height of a child can depend on the height of the mother, the height of the
father, nutrition, and environmental factors.
Regression Example
 House Value Prediction- The example below shows that the price variable
(output dependent continuous variable) depends upon various input
(independent) variables such as plot size, number of bedrooms, covered area,
granite flooring, distance from city, age, upgraded kitchen, etc.
Multiple Linear Regression Model
 A multiple linear regression model with k independent predictor variables x1,x2...,xk predicts the
output variable as:
𝑦 ˆ = 𝛽 + 𝛽 𝑥 + 𝛽 𝑥 + 𝛽 𝑥 + ⋯ … … … . . +𝛽 𝑥
0 1 1 2 2 3 3 𝑘 𝑘

 There is always some error (regression residual) in predicting the values, i.e.

𝑎𝑐𝑡𝑢𝑎𝑙 𝑣𝑎𝑙𝑢𝑒𝑖 = 𝑝𝑟𝑒𝑑𝑖𝑐𝑡𝑒𝑑 𝑣𝑎𝑙𝑢𝑒𝑖 + 𝑒𝑟𝑟𝑜𝑟


𝑦𝑖 = 𝑦ˆ𝑖 + 𝜖𝑖

𝑦ˆ𝑖 = 𝛽0 + 𝛽1 𝑥𝑖1 + 𝛽2 𝑥𝑖2 + 𝛽3 𝑥𝑖3 + ⋯ … … … . . +𝛽𝑘 𝑥𝑖𝑘 + 𝜖𝑖

The total error can be computed from all the values in dataset i.e. i=1,2,….,n

𝑇𝑜𝑡𝑎𝑙 𝐸𝑟𝑟𝑜𝑟 = σ𝑛𝑖=1 𝜖𝑖 = σ𝑛𝑖=1 𝑦𝑖 − 𝛽0 − σ𝑘𝑗=1 𝛽𝑗 𝑥𝑖𝑗 (7)


Multiple Linear Regression Model
 Equation (7) presented in the previous slide, can be represented in matrix form
as:
𝜖 = 𝑦 − 𝑋𝛽
𝜖1 𝑦1 𝛽0
𝜖2 𝑦2
Where 𝜖 = ⋮ ; y= ⋮ ; β = 𝛽2
⋮ ⋮ ⋮
𝜖𝑛 𝑦𝑛 𝛽𝑘
1 𝑥11 𝑥12 … 𝑥1𝑘
1 𝑥21 𝑥22 … 𝑥2𝑘
and 𝑋 = ⋮ ⋮ ⋮ ⋮

1 𝑥𝑛1 𝑥𝑛2 … 𝑥𝑛𝑘
Least Square Error Fit for MLR
 According to Least Square Error method, we have to find the values of the matrix 𝛽 for which
total square error is minimum.
𝑛
𝑇𝑜𝑡𝑎𝑙 𝑆𝑞𝑢𝑎𝑟𝑒 𝐸𝑟𝑟𝑜𝑟 = 𝐽 𝛽 = ෍ 𝜖𝑖2 = 𝜖 𝑇𝜖
𝑖=1
= (𝑦 − 𝑋𝛽)𝑇(𝑦 − 𝑋𝛽)
= (𝑦 𝑇 − 𝛽 𝑇𝑋 𝑇)(𝑦 − 𝑋𝛽)

𝐽 𝛽 = 𝑦 𝑇 𝑦 − 𝛽 𝑇𝑋 𝑇𝑦 − 𝑦 𝑇𝑋𝛽 + 𝛽 𝑇𝑋 𝑇 𝑋𝛽

𝑱 𝜷 = 𝒚𝑻 𝒚 − 𝟐𝒚𝑻𝑿𝜷 + 𝜷𝑻𝑿𝑻 𝑿𝜷

[Because 𝑦 𝑇𝑋𝛽 and 𝛽 𝑇𝑋 𝑇𝑦 is always equal with only one entry]

 The square error function is minimized using second derivative test.


Least Square Error Fit for MLR
 Step 1: Compute the partial derivate of J(𝜷) w.r.t 𝜷
𝜕J(𝛽) 𝜕(𝑦 𝑇 𝑦 − 2𝑦 𝑇 𝑋𝛽 + 𝛽 𝑇 𝑋 𝑇 𝑋𝛽)
=
𝜕𝛽 𝜕𝛽
𝜕𝑦 𝑇 𝑦 𝜕2𝑦 𝑇 𝑋𝛽 𝜕𝛽 𝑇 𝑋 𝑇 𝑋𝛽
= − +
𝜕𝛽 𝜕𝛽 𝜕𝛽
𝜕𝛽 𝜕𝛽 𝑇 𝑋 𝑇 𝑋𝛽
= 0 − 2𝑋 𝑇 𝑦 +
𝜕𝛽 𝜕𝛽
𝜕𝐴𝑋
[𝐵𝑒𝑐𝑎𝑢𝑠𝑒 = 𝐴𝑇 ]
𝜕𝑋
= −2𝑋 𝑇 𝑦 + 2𝑋 𝑇 𝑋𝛽
𝜕𝑋 𝑇 𝐴𝑋
[𝐵𝑒𝑐𝑎𝑢𝑠𝑒 = 2𝐴𝑋]
𝜕𝑋
Least Square Error Fit for MLR
𝝏J(𝜷)
 Step 2: Compute 𝜷ˆ for 𝜷 for which =𝟎
𝝏𝜷

−2𝑋 𝑇 𝑦 + 2𝑋 𝑇 𝑋𝛽ˆ = 0
𝑋 𝑇 𝑋𝛽ˆ=𝑋 𝑇 𝑦
𝜷ˆ=(𝑿𝑻 𝑿)−𝟏 𝑿𝑻 𝒚

𝝏2 J(𝜷)
 Step 3: Compute and prove it to be minimum for 𝜷ˆ
𝝏𝜷𝟐

𝜕2J(𝛽) 𝜕(−2𝑋 𝑇 𝑦 + 2𝑋 𝑇 𝑋𝛽) 𝑇 = +𝑣𝑒


= = 0 + 2 𝑋𝑋
𝜕𝛽2 𝜕𝛽
Least Square Error Fit for MLR- Example
Example: The Delivery Times Data A soft drink bottler is analyzing the vending machine serving
routes in his distribution system. He is interested in predicting the time required by the distribution
driver to service the vending machines in an outlet. It has been suggested that the two most important
variables influencing delivery time (y in min) are the number of cases of product stocked (x1) and the
distance walked by the driver (x2 in feet). 3 observations on delivery times, cases stocked and walking
times have been recorded.

number of cases of product the distance walked by the Delivery time (in min) y
stocked (x1) driver (x2)
7 560 16.68
3 220 11.50
3 340 12.03
(a) Fit a multiple regression line using least square error fit.
(b) Compute the delivery time when 4 cases are stocked and the distance traveled by driver is 80 feet.
Least Square Error Fit for MLR- Example
Soln
 The multiple linear regression equation is: y = 𝛽1ˆ+𝛽2ˆx1+ 𝛽3ˆx2
𝛽1ˆ
Where 𝛽1ˆ, 𝛽2ˆ, 𝛽3ˆ or 𝛽ˆ= 𝛽2ˆ are regression coefficients for line of best fit.
𝛽3ˆ
We know, 𝜷ˆ=(𝑿𝑻 𝑿)−𝟏 𝑿𝑻 𝒚
1 7 560 1 1 1
𝑋= 1 3 220 𝑎𝑛𝑑 𝑋 𝑇 = 7 3 3
1 3 340 560 220 340

3 13 1120
𝑋 𝑇 𝑋 = 13 67 5600
1120 5600 477600
Least Square Error Fit for MLR- Example
Soln
799/288 79/288 −7/720
(𝑋 𝑇 𝑋)−1= 79/288 223/288 −7/720
−7/720 −7/720 1/7200

7.7696
𝛽ˆ=(𝑋 𝑇 𝑋)−1 𝑋 𝑇 𝑦 = 0.9196
0.0044
The line of best fit is, 𝑦 = 7.7696 + 0.9196𝑥1 + 0.0044𝑥2
When x1= 4, x2= 80
𝑦 = 7.7696 + 0.9196 𝑋 4 + 0.0044 𝑋 80 = 11.80 min

You might also like