0% found this document useful (0 votes)
26 views

Lectures 17 To 20

The document discusses orthogonal vectors and orthogonal bases. It defines two vectors as orthogonal if their inner product is equal to zero. It then describes how to transform a basis of vectors into an orthonormal basis using the Schmidt orthogonalization process. This involves iteratively making each basis vector orthogonal and normalizing it to have unit length, resulting in an orthonormal basis that spans the same vector space as the original basis.

Uploaded by

Ayushi Raj
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
26 views

Lectures 17 To 20

The document discusses orthogonal vectors and orthogonal bases. It defines two vectors as orthogonal if their inner product is equal to zero. It then describes how to transform a basis of vectors into an orthonormal basis using the Schmidt orthogonalization process. This involves iteratively making each basis vector orthogonal and normalizing it to have unit length, resulting in an orthonormal basis that spans the same vector space as the original basis.

Uploaded by

Ayushi Raj
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 76

Lecture 17

Section A

Pr ofdegree atmost no

(Saus
"*

ISPCqcudi) (Rissdi)

<

"a

roses (cuscai)
powocu fasqasdn a

W(X)> 0

2)"ackg(y)2 ai)"2
#thogonality Inner Product space

V, (s.). We say two vectors are

arthogonal to each other ifinner product is

zero. 17. U.d a are


orthogonal if

(4,x) =

Im Let 92. -
Guy are

non-zero
orthogonal vectors. (i.e. (Mi,4j) =
0

they G41. -

Hub is LI 1Fj)
Proof
④ 2,41 -
+
Cnln 0
=

Aim is to show that21 =2e =


- -
(n 7
=

(21,9,x,7
-

(nMu)

Cu(4).T)
G(u,,H,)
-
((2,,5e)+
+
- 0

j
~.

=>
C.1(1/2 0
=

=>
a 0
=

Similarly 4 =
0. - - (n = c
Consider a Basis
the vectors
Sis. -
Only g.f. are

arthogonal to each other

n = C141 +
-
-
CnUn

r
unique
2
(x,y,) =

2, (M1,41) 4 ((2,1)
=

=> a
ua
=

. .

Sj=
caM;)_.
sn
yre
PAAVectorie s aid
is normalized if

(1211 1.
=

Desirable
-
We banta Basis of vector
-

space stech that all the vectorsare

arthogonal and normalized. (i.e. we cant

authonormal Basis)
⑭Schmidt
orthogonalization

Suppose itis a vectar space with

dim (F)
Un]
Basis n
[U1,
=

Orthonamal Basic
qui-In Sto

authonamal and span GUs-Un]


= spans , -
*n3
V, 41
SpandUr.-Un3

->

((u,))
=
SpansV, 4gc -

Yub

v = Us -

(Ug,v,)U,

Vs:
(V, v.):(Mg,V)
-

(Mg,4)(Y,r,)

(U2,v.)
= -
(U2,4,)

=
0
SpanqUis-UnbESpan9 Vi,We, Ms.
-

Un
(Exercise

⑬3 Us
=
-

(Us,vJVe(Us,v,)W,

Us ((Us,v)Ve f(us,4,)V)
=
-

IV3, Y) (Us,V.)
=
-

(Us,Ve)(UeNX)

=(4s, r.) jyUsiU(Us,YS


=
<(i) (i). () 3
e.g.
R3

Hs 4e Us
Orthonamal Basis using standard dot
produ
u
j(i),v Hz (ue,v,)u,
=
=
-

(i) jj(?) -

v. -

1:(125 (i) () =
-

(a)
5 45
=
-

(43,v,)U, -14s,Vc(Ve

e.g. 12

((i), (i) 3 [is]


A:

x+y
(x,y)
=

(6) ((4.1)
10)(,][I]
:

n =

n.:
ull
Yg Hc
=
- (Ug,Y)V,
2183.
2
SD,c)
exercise: 2, 01,
Pallisdia

arthonamal Basis
Lecture 17

section B.

Orthogonal+y
-
-

orthogonality (4,0)
=> c
=

Si, -7n] arthogonal and nonzero

then 141. -
has are 2.
Consider
otherbasi
a

201.-2nY which is
athoganal
n = C,4, +
-
Cu2n
2
N
unique.

(2,2) (GH1 =
+ -
(nUn,M.)

G.(4,4,)
22(2x,4,)
=
+ -

110
=
2.(2,4,)

2. 1
nnn)
--
6:
215, n=
dired
Ifit a vector is is called normalized if

1121) 1
=

#esirable

Orthogonal Normal =

Orthonormal vecters

Def the set94.-th3 called


is
-

orthe normal ifit is arthogonal and each


nectar is normal ide ((4j)1 tj 1ton
=
=
Schmidt
&
-
Orthogonalization
Given 242. -

Un3 a basi we want

an authonamal Basis 121.- 2n3

SpandUss- Mn3 =span GHis -

Eng

Step 1. 1(41/1 0

v1 =
11 Span24s. -Yn3=span
((411) <H1, Ha, was

us =
U2-(Uc,v,)V, because Hg://U21/02 1
p
(N2, v) (4e -(U0,Y)
=

V,v) (U8gU) = -
(Ueiv,)(v,v)
(H2,V.)
=
-
(H2,V,) 0
=
Span[Ui.-Un]-spanSU,He.-Un3
=
spanSVi,E2, Us. -Gn3
12 G
= (uz,v,)u
+

H1 1(4,71V1
=

Vs:
all
n Us
=
- (Us,v)h, -

(Us,Uc)Vg
u
Us:

(iTs, v, =
0, (M,v2) c
=

[Exercise
5 45
-
-

(MssVisVi
y =
n
114j1)

i 4n = =

(un,vi)vi
W:
null
agus fun, ina
e.g.

u.=
y( m
=

(d)
=

ns 4c (ue,4)u
(j) 2(b) (8)
=
=

- -
=

v
z 1)
=

n=4 -

(Us,veVe

(0)
=
-
(us,v(x
(i) (8)
= -

(B)
n:
vs. is

<(i) (i), 1913,


4
9(0), (,)3
[y3]
A :

(ci,j s (i Az
+

find authonanal Basis


1 1
...

10,112.(110)(42][]
=

=(*) =(4,2][] a
:

(((,3) 5
=

2
=

[1,01(32][]
114,11:(Un,42) =

4
=
i 42
=
-

(42,4)V,

=(i) (>,v.)(1) (i). (k):(*)


-
=

15(%.] (")
(0,v) =
[0,

-[2s]( 2]
2
:

no
("") ve.(i,) = = ( i ) h (i)
=
(IEC112 [-1/c, 1]
[ :?] [ ]
:

[0,43):"):
=
4

112, 11. 4 (Nc/1


=
= 8
Pg - Set ofall quadratic polynomials
u,U2 ②
us

[1,0,6123 Polynomiaarthogan
I Legendre,
f(x)g()dx
I (14,1)
:
f,1.1dx.
V=u1
((U11 2
=

1
=

vg
=
Hz -

e
luivy =
x - (x,
1
&
I

(x,1)
f(.2dx 0
↓v
=

x
=

=
(V2/18: jsdn =

us Us
= -
(43,v)Va
cli2 (usu
=

(x2,a)/
= x2 -
(x,y) -

&

=x2 -

5
1
(7,g) 'w01f()
=

g(1)d3

1():2

Chebyshev polynanial
Lecture 18

Section A
Best approximation
V = PS
x 1
+ W
w*

a
x

# &

(x wt,b) 0
=
-

↓b) W
approximation
LetB -
Let W be a

subspace atv. We whatis


say
the approximation
best ofbeefif

11x-w+1) ((4 -W/)


=
vw W

Im is
It the approximation
best ofGEF
in t ifand only if[ie-wt, w) = 0

XBIV.
80W
1/2-2+1) ((3-01) =

ifand only if
0fWtE
(x w+,w)
=

Assume (29-207,0) = 0 (

8/8 1/2 w* + w*- w/2


112
-
=
-

w*,WX W)
112-w*112 1180 -0*)8 2(n
-
-

+
+
=

2 *(10 1Z w+1)
&
N.
1/2 ((x w
+

81)
-
- = -

((N-LM)e
(18-21
=min
=>
1/2 -w*11A
((2-W15 wth

=>WA is the best approximatic

Conversely Assume

112-0*)) roftht
*
112-811 >

To poswe (4-0*,0) 0
=
f0+N
112-8112 1(4-8+1(8 ((W.W*)(8
=
+

+ 2(2 w
=
*,WEW)

Using (12-0119>,1(2-w*)/2

112
-w *1(2 2(2-W*,W*.0)
+
T0 fwtw
*
u 2 b
ueW
= -

=>

+,u)x0 +4 -5.
114118 2(2
+
-
W

H =

= (0t-WS
(hot (W* -
w)
11 w * -
w1)&
CoI
(AWS Oci)
(ALS)
C
11207 -
81/2

+2 oinofs
WEW) met
2
(200

*,wf W) I
np 110 011
<

2F,wF-cFwtw
-

=>

W)2
-

1127 -

112
+ -W//2

=>
1 (2 -1
*,(0 *-8s /2

7,0

#w((u -
0
=
=>
(2-2 *,20*-5) 0
=

F WCE

=> (U -5 *, W) a
=
f W-F

Tw Bestapproximation of21 in 20

is unique.
Proof Assume We and we are two best
=

approximation of2
W),()) 0
=

(X w,W) 0(X W
10 I
= -
-

I
20-w2
10:
0(X -wz,w,-wq)
=
a

(x W,,w,-Wc)
=

-
w,,w,-we)
(z
= -
= 0

=> 115,
=
-w2112:0 = 1 W, w2 =

Theorem Let tobe


=
a
subspace ofF
and qw,-why is an orthogonal basis ofto

then the best approximation ofleftin t

is
given by
R

1A=5(Honmee
Proofto show (h*,Wil =0
Writton.
(2-I wi) es
was

=
Lecture 18 section B.

v MC
=

2
*

- W* ⑨

F
wt
·

( - W
+,W) 0
=

f W-IN
Approximation: Let z+Fand
Defis
Best
a
subspace ofvi,then notis

called best approximationo f me in two,

if +F.
112-w+1) ((2 -11)
= W

I
hm Itisthe bestapproximation of2

in W ifand only if

f W (W.
(x -

Wt,W) 0
=
Proof V ce
1/2-W+11 HR-W11
-
-

if and only if
F c
(2-204,W) 0
=

Remark IfLot is a fixed nectar in 10 then

4-W written
nectar
BEIN
LANYME
can be as
any
u u
=
- 2*+2F =
Assume (2-2 *, W) =
0 # Wo

To prome (0
f -E.
112 - w) 1/4-5))
<

112-8112:112-W*+W*-8118=1/2-W*112 +1(W *-2814

* (wX W)
+ 8( -

17
=>

112 -811=(/2-20*1187 (/W - WF112 ⑨

=>
1(2 -
8118>, 1(2-5*/12
WCF.
r
((2- 8+1) ((7-81)
=
=>
we Assume

112-w*1) 1(4-601)
< b

To FLOCTN
prove (2-0*,W) 0
= F

112 -
8112 1/2-1*1/8 7((8-8*)/8 +8(4 -W*,w* -W)
=

112-8+1)2 ((W -W+(2+2(21-2*,00*-6)


((2-W118
=

=> -

7,0

110-5*118 +2(7-2*,w/W) x0 fw+

f4+W.
(14118
=> 2(4-2*,4)
+
>0
USE
1IU118 2(2 -w *, 4) T0 F

for waterdefine
any give

(xfc20x-w)
u -
=

1w e

2(2 wf,u)
=

(H,4)
-

Eis

ofc10x2)
-(4-05, 100)
(H,4) =(
=

(1-0),
S
-

2
(107-w//

InPhotl" 1120*-W)
#
lR-NER
I
1/2 Z)))2
X
-
&(x *,u) 2(X 24)
oxcoxios (0x-a
(
- w = - -

WA
g(n*, (n -LX, e
= - w

= -
g(( - wX,2X -
W)/2
xw(12 -

(14118 2(x
+
-

wf,4) =

l
oshoscs,
#WF
I

20)(= 10H.
=>
Is = 0
0

((8 2+1/2
-

2*,W) 0
=

(X
(2-2X,20*.W)
0
-
=

El CIF:
=

=>
# Best
approximation is unique.

Iroof Let we and we are two best

of2 in t
approximations
Fo
C +
(2-W1, ()
=> 0
=

# LCF.
and (U W2,b) 0
=
-

take to -
w0, -w2 Ev

I
=>
(x- W,(, -Ws) 0&(H-W2,W, -We)
= 0
=

(Wg - W,,w,-wa) 0
=

=>- (W,-02112 =

=> w, wg
=

//
Lecture 19 Section B

Best seevin
approximation ofa vector

a
subspace to

11x -w*1) ((4-801)


= wa

loto
=>
(H -

w*,W) 0
=
A
#m Let inbe subspace ofitwith

orthogonal basis Swe...Why, then best

approximation what a in itis given by

CX
sUcUWce
=

Proof V to s
(2 wX,W) 0
=
-
-
-


i=
v 1 to n.
(2-1*,Wi)
= 0
(2-wX, Wil
Co.sncaiusibe
=

=
(x,vi) -

(nmius), wil
(2,wi)
=
-

EncurosislW, a
is

(u,wi)
=
-
vi) toi)
2
will
= (2sWi) -
(22,vi) = 0
Inner Productspace, can
you
approximate rectar
a
using a subspase

-9. V=45, (5):dot product

v (1,2) w
[(1,0) (x (R3 <
=
=

R
·

-
W
W span
=

[(3,0)3 x (1,2)
=

W1: (3,0)

wi)wi
3,0)
w*
(9,0)
=

11 Well
=

·
(1,0)
=

V H3x 21,2,3)
= =

W:
spans (8), (%)]
w1)
log roll a one
of: we

(181/)
2 x4
=

W PC
=

span
[1,0,02-53
=

Sf(n)gx)di
ene
-wAlian)M/s
12
ac+bacle

xir1 +yousatli
et
20* <

ig//
-

=
881 0
+ +
Lecture 19 Section A
BestApproximation
x v
+ and * * InSt.
C

((2-W1/ v i
112-w+)) = ce


(2-W*,W) 0
=
↓ We I

Find a formula wh
Im Letzeerand ofbe the

bestapproximation ofbe in toa subspace)


with an
orthogonal basic
frocs--why
then

wx
wn
=

e 1 1
=

Proof We need to
prove
-
-

wt
t W
( wF,w) 0
=
-


05 =1 ton.
(2.25,Wj)
=
-wicws-cu.cn,avoidsee
(x,vj)
=
=

wr
1 =
1
(wi, wil (Wijwj(
01Fj
=

(2,65)
=
-
(25Wi) i(
-ill2
(2,Wj) (2,5j)
=
=
0
=

//
e.g.E 42 u =
(3) w,

W:soan
[(i)?
x y
=

wis
wX= we

(i)
:E(i):(id
-(3)
43

spans of
e.g. 2.
W:

Basic
codewa lockhere
24 =

(j) E(8) ((Ec I


+ =

1x 1|
*
-
w
setofall
29.3. V =

polynomials
W PC setofall
quadratic polyman
=

jd,(2)
- (
Pa(*) di

Orthogonal Basis at W
G1,x,x 13
=
-

x c0
=

110

-consent issee x
Lecture 20

u x4
=

3 *

112-W*1) ((..W1)
< 10-PC
1 1

min
a,b> <

J &(U - 942 -
bx - 762dx

fpademectan
"
~

f(a,b,x)
=
Discrete feastsquare method.

xiy;
1
1

! ↑

f(x) a ba
=
+

find a & b

②iii. "
m
1

wzi
( E..Absh:21

[0:](b]: [w]
a Sincus b
+

(opci) + (sinces) + dcos(en)

a since bCosini)
+
+(Sincei) +
d(os(eNi)Eg.
/

1]
N

Casinci) bscasoies
11
+

Sin(ai) dCOST21:
=

C
+

my 4 2
-

2..)
ATA
[9]:
T

A [5]
(pi,yi( ... to m

af,(x) b
+
fg(k)
↳ 2

a f,(0i) bfe(si) Yr
+ 1 = 1 tom.

SinE(fiis
2
+ delis -

bi)
j 3.
=

x
[9]
=

A iin j

(941) essuis]
A.
pii i i,#
-
Lecture: Least Square Methods

http://hkkaushik.wordpress.com/courses/mtl107/
Least Square methods

I We discuss how to solve overdetermined linear systems of


equations

Ax ⇡ b, b 2 Rm , x 2 Rn , m > n.

I These systems do in general not have a solution.


Reference
I Ascher-Greif, Chapter 6.
I Burden and Faires: Chapter 8.

http://hkkaushik.wordpress.com/courses/mtl107/
Origins of linear least squares problems Data fitting

http://hkkaushik.wordpress.com/courses/mtl107/
Origins of linear least squares problems(cont.) image
Numerical Methods for Computational Science and Engineering
Introduction

morphing
Origins of linear least squares problems (cont.)
Image morphing

Schaefer, McPhail, Warren: Image deformation using moving least squares


http://faculty.cs.tamu.edu/schaefer/research/mls.pdf

NumCSE, Lecture 8, Oct 14, 2013 4/20

http://hkkaushik.wordpress.com/courses/mtl107/
Linear least squares

Consider overdetermined system on m linear equations in n


unknowns: A [9]
Ax ⇡ b, A 2 Rm⇥n , b 2 Rm , x 2 Rn , m > n.

I We assume the the matrix A has maximal column rank n.


I This means that the columns of A are linearly independent.
In general, this system cannot be solved exactly: b does not lie in
the range space of A.
Reformulate problem:
Find x such that 12 kb Axk22 is minimized.

http://hkkaushik.wordpress.com/courses/mtl107/
Linear least squares(cont.)

We introduce residual

r=b Ax 2 Rm
r = b A x

Minimization problem for a smooth scalar function in n variables

s
1
minx (x), with (x) = krk22
2
Necessary conditions for a minimum:
@
grad (x) = 0 , (x) = 0, k = 1, ..., n.
@xk

http://hkkaushik.wordpress.com/courses/mtl107/
Linear least squares(cont.) f((1,5(e. - S(n)

gradf:of, t, ifneed-

1 1
(x) = rT r = (Ax b)T (Ax b)
2 2
1
= (xT AT Ax xT AT b bT Ax + bT b
2
1 1
= xT AT Ax (AT b)T x + bT b.
2 2
grad (x) = AT Ax AT b = AT (Ax b) = 0
1
(x + x) = kb Ax A xk22
2
1 1
= kb Axk22 (b Ax)T A x + kA xk22
2 2
1
= (x) + kA xk22 (x)
2

http://hkkaushik.wordpress.com/courses/mtl107/
Linear least squares(cont.)

B := AT A, B 2 Rn⇥n
A has maximal rank =) B is SPD with quadratic form

Q(z) = zT Bz = zT AT Az = (Az)T (Az) 0, for all z


Q(z) = 0 , Az = 0. , z = 0

So, in particular,
1
kA xk22 > 0, for all x 6= 0,
2
and

(x + x) > (x), for all x 6= 0.

http://hkkaushik.wordpress.com/courses/mtl107/
Theorem:Least squares

The least squares problem

minx kAx bk2 ,

where A has full column rank, has a unique solution that satisfies
the normal equations

⑦ (AT A)x = AT b

We have x = (AT A) 1 AT b.The matrix multiplying b is called the


pseudo-inverse of A:

A+ = (AT A) 1 T
A 2 Rn⇥m .

http://hkkaushik.wordpress.com/courses/mtl107/
Geometrical interpretation
or Computational Science and Engineering
s
rpretation

From
al interpretation
grad (x) = AT Ax AT b = AT (Ax b) = 0

grad we
(x) AT AAxT r =
see= that AT0.
b = AT (A x b) = 0

hat AT r = 0.
b

lane shows R(A). A x


Ax
ogonal projection of b
. 0

14, 2013 11/20


The blue plane shows R(A). Ax is the orthogonal projection of b
onto R(A).
http://hkkaushik.wordpress.com/courses/mtl107/
Algorithm:Least squares via normal equations

1.B = AT A, y = AT b (normal equations Bx = y)


2.B = LLT (Cholesky factorization)
3.Lz = y, LT x = z (Forward/backward substitution)
4.r = Ax b (Compute residual)
Complexity:
I Step 1:mn2 + O(mn) flops
I Step 2: 13 n3 + O(n2 ) flops
I Steps 3: O(n2 ) flops
I Steps 4: O(mn) flops
Forming B dominates computations cost, in particular if m n

http://hkkaushik.wordpress.com/courses/mtl107/
Example 6.1 from Ascher-Greif
A52T
2 3 2 3
1 0 1 4
6 2 3 57 6 27
6 7 6 7
A=6
6 5 3 27 5⇥3 6
72R , b=6 5772R
5
4 3 5 4 5 4 25
1 6 3 1

For solving the LSP minx kb Axk2 we form B and y:


2 3 2 3
40 30 10 18
T 4
B = A A = 30 79 475 , y = AT b = 4 5 5
10 47 55 21

The final solution is x = (0.3472, 0.3990, 0.7859)T , correct to


the number of digits shown. The residual is
r = (4.4387, 0.0381, 0.495, 1.893, 1.311)T .
http://hkkaushik.wordpress.com/courses/mtl107/
Data fitting

Generally, data fitting problems arise as follows:


I We have observed data b and a model function that for any
candidate model x provides predicted data.
I The task is to find x such that the predicted data match the
observed data to the extent possible.
I We want to minimize the di↵erence of predicted and observed
data in the least squares sense.
I Here, we study the linear case where predicted data are given
by Ax.
(The condition that A has maximal rank means that there is
no redundancy in the representation of the predicted data.)

http://hkkaushik.wordpress.com/courses/mtl107/
Example 6.2 from Ascher-Greif: Linear regression
Consider fitting a given data set of m pairs (ti , bi ) by a straight
line:

v (t) = x1 + x2 t =) v (ti ) ⇡ bi i = 1, ..., m.


2 3
1 t1
6 1 t2 7  Pm
6 7 m Pmi=1 ti2
A = 6 . . 7 B = Pm
4 .. .. 5 i=1 ti i=1 ti
1 tm

i
ti
1
0.0
2
1.0
3
2.0
B:
ATA. A b =
T

bi 0.1 0.9 2.0


 BS=<
0.05
=) x =
0.95

http://hkkaushik.wordpress.com/courses/mtl107/
Polynomial data fitting

Extend linear regression formula to a higher degree polynomial fit:

v (t) = x1 + x2 t + ...xn t n 1
=) v (ti ) ⇡ bi , i = 1, ..., m.

1
The matrix that we now obtain is called a Vandermonde matrix.
2 3
1 t0 · · · t0n 2 t0n 1
6 1 t1 · · · t n 2 t n 17
6 1 1 7
A = 6. . .. .. 7 2 Rm⇥n
4 .. .. . . 5
n 2 tn
1 t m · · · tm 1
m a S1nQLS

MATLAB code lsfit from Ascher-Greif on next slide. b(os(K)


+

x 2x

⑨ e b
+ -
+

http://hkkaushik.wordpress.com/courses/mtl107/
Polynomial data fitting

Extend linear regression formula to a higher degree polynomial fit:

v (t) = x1 + x2 t + ...xn t n 1
=) v (ti ) ⇡ bi , i = 1, ..., m.

The matrix that we now obtain is called a Vandermonde matrix.


2 3
1 t0 · · · t0n 2 t0n 1
6 1 t1 · · · t n 2 t n 17
6 1 1 7
A = 6. . .. .. 7 2 Rm⇥n
4 .. .. . . 5
n 2 tn
1 t m · · · tm 1
m

MATLAB code lsfit from Ascher-Greif on next slide.

http://hkkaushik.wordpress.com/courses/mtl107/

You might also like