Lecture2
Lecture2
regression
with
one
variable
Model
representa6on
Machine
Learning
Andrew
Ng
500
Housing
Prices
400
(Portland,
OR)
300
Price
200
(in
1000s
100
of
dollars)
0
0
500
1000
1500
2000
2500
3000
Size
(feet2)
Supervised
Learning
Regression
Problem
Andrew
Ng
Training
Set
How
do
we
represent
h
?
Learning Algorithm
Andrew
Ng
Linear
regression
with
one
variable
Cost
func6on
Machine Learning
Andrew
Ng
Size
in
feet2
(x)
Price
($)
in
1000's
(y)
Training
Set
2104
460
1416
232
1534
315
852
178
…
…
Hypothesis:
‘s:
Parameters
How
to
choose
‘s
?
Andrew
Ng
3
3
3
2 2 2
1 1 1
0
0
0
0
1
2
3
0
1
2
3
0
1
2
3
Andrew
Ng
y
x
Andrew
Ng
Linear
regression
with
one
variable
Cost
func6on
intui6on
I
Machine
Learning
Andrew
Ng
Simplified
Hypothesis:
Parameters:
Cost Func6on:
Goal:
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameter
)
3 3
2
2
y
1
1
0
0
0
1
2
3
-‐0.5
0
0.5
1
1.5
2
2.5
x
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameter
)
3 3
2
2
y
1
1
0
0
0
1
2
3
-‐0.5
0
0.5
1
1.5
2
2.5
x
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameter
)
3 3
2
2
y
1
1
0
0
0
1
2
3
-‐0.5
0
0.5
1
1.5
2
2.5
x
Andrew
Ng
Linear
regression
with
one
variable
Cost
func6on
intui6on
II
Machine
Learning
Andrew
Ng
Hypothesis:
Parameters:
Cost Func6on:
Goal:
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
500
400
Price
($)
300
in
1000’s
200
100
0
0
1000
2000
3000
Size
in
feet2
(x)
Andrew
Ng
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
Linear
regression
with
one
variable
Gradient
Machine
Learning
descent
Andrew
Ng
Have
some
func6on
Want
Outline:
• Start
with
some
• Keep
changing
to
reduce
un6l
we
hopefully
end
up
at
a
minimum
Andrew
Ng
J(θ0,θ1)
θ1
θ0
Andrew
Ng
J(θ0,θ1)
θ1
θ0
Andrew
Ng
Gradient
descent
algorithm
Andrew
Ng
Linear
regression
with
one
variable
Gradient
descent
intui6on
Machine
Learning
Andrew
Ng
Gradient
descent
algorithm
Andrew
Ng
Andrew
Ng
If
α
is
too
small,
gradient
descent
can
be
slow.
Andrew
Ng
at
local
op6ma
Andrew
Ng
Gradient
descent
can
converge
to
a
local
minimum,
even
with
the
learning
rate
α
fixed.
Andrew
Ng
Gradient
descent
algorithm
Linear
Regression
Model
Andrew
Ng
Andrew
Ng
Gradient
descent
algorithm
update
and
simultaneously
Andrew
Ng
J(θ0,θ1)
θ1
θ0
Andrew
Ng
J(θ0,θ1)
θ1
θ0
Andrew
Ng
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
(for
fixed
,
this
is
a
func6on
of
x)
(func6on
of
the
parameters
)
Andrew
Ng
“Batch”
Gradient
Descent
Andrew Ng