Session4: Normsand inner-products
Optimization and Computational Linear Algebra for Data Science Léo MiolaneSession4: Normsand inner-products Optimization and Computational - - PowerPoint PPT Presentation
Session4: Normsand inner-products Optimization and Computational - - PowerPoint PPT Presentation
Session4: Normsand inner-products Optimization and Computational Linear Algebra for Data Science Lo Miolane Contents 1. Norms & inner-products 2. Orthogonality 1 3. Orthogonal projection 4. Proof of the Cauchy-Schwarz inequality
Contents
- 1. Norms & inner-products
- 2. Orthogonality
- 3. Orthogonal projection
- 4. Proof of the Cauchy-Schwarz inequality
1
Normsandinner-products
Questions
Norms and inner-products 2/20 Norms in machinelearning :
- H
- H2
- a. y
ask)
= ftp.Y-yuz-CG-it-3- Use
for
" regularization " minimize Loss( data, a) + thalli with a ER"{
+ MakeQuestions
Norms and inner-products 2/20Questions
Norms and inner-products 2/20Orthogonality
Definition
Orthogonality 4/20 Definition We say that vectors x and y are orthogonal if Èx, yÍ = 0. We write then x ‹ y. We say that a vector x is orthogonal to a set of vectors A if x is- rthogonal to all the vectors in A. We write then x ‹ A.
q
ifeng.se::÷.
5)
Guy> = details> t.itanefa.tk
to
Io
E
=oPythagorean Theorem
Orthogonality 5/20 Theorem (Pythagorean theorem) Let Î · Î be the norm induced by È·, ·Í. For all x, y œ V we have x ‹ y ≈ ∆ Îx + yÎ2 = ÎxÎ2 + ÎyÎ2. Proof. ⇤① €
Katy 112
=( nty , at y >
⇒
= ling tTHE
Application to random variables
Orthogonality 6/20 ✓ = f random variables withfinite
secmoondmeuf) For X, Y , wedefine kx,y>=LfI
the norm induced is 11×11 =T¥ '- Let's
XIY
⇒ ECXYT- o
By Pyth
. them . this isequivalent
to 11×+4112 = 11×112 t 114112- = Efcxt4723
Va¥
=Voix)
+ VaecyjOrthogonal & orthonormal families
Orthogonality 7/20 Definition We say that a family of vectors (v1, . . . , vk) is:- rthogonal if the vectors v1, . . . , vn are pairwise orthogonal,
- rthonormal if it is orthogonal and if all the vi have unit norm:
- =
- Example :
- the
- ( (f) , ft))
- rthogonal
- rthonormal
Coordinates in an orthonormal basis
Orthogonality 8/20 Proposition A vector space of finite dimension admits an orthonormal basis. Proposition Assume that dim(V ) = n and let (v1, . . . , vn) be an orthonormal basis of V . Then the coordinates of a vector x œ V in the basis (v1, . . . , vn) are (Èv1, xÍ, . . . , Èvn, xÍ): x = Èv1, xÍv1 + · · · + Èvn, xÍvn.- ①
Proof
: we have ⇒rvn t .- tan Un
for some
ar -- an ER
Gave t
- tan un , Qi>
HEIL
'Coordinates in an orthonormal basis
Orthogonality 9/20 Let n, y EVERY n = Cue,a) Vz t- tan, a
T ae
→n
y- Cy, y> Vst
- t Lun , ay> Un
- Ps
(my>
= ( airs t .- -tannin ,
- tpnvn)
dept t
- + an pre
- therwise
Proof
Orthogonality 10/20Orthogonalprojection
Picture
Orthogonal projection 12/20 From now, È·, ·Í denotes the Euclidean dot product, and Î · Î the Euclidean norm. What is the vector y of S which is the closest to a ? Iden, s)
t¥s
(sobspace)Orthogonal projection and distance to a subspace
Orthogonal projection 13/20 Definition Let S be a subspace of Rn. The orthogonal projection of a vector x- nto S is defined as the vector PS(x) in S that minimizes the
②
that nphiamjy.es- -
- if
- if
Computing orthogonal projections
Orthogonal projection 14/20 Proposition Let S be a subspace of Rn and let (v1, . . . , vk) be an orthonormal basis of S. Then for all x œ Rn, PS(x) = Èv1, xÍv1 + · · · + Èvk, xÍvk.- Proof :
- taa Va
- y 112
- EYE
for
some h . -- aa ER
- Ily If
- k
- ( NYS
- tamed
Proof
Orthogonal projection 15/20 Ha-YR = Half t o¥dg2-2aiCa- :*÷÷:÷÷:
- i
y
is given byPs(a)
= Ca,# ve t . . - t Lana> re > flail = di- 22, Cami> ¥1
f-
'Gi)- of flail = Zai
- Ka, vis
- →
Consequence
Orthogonal projection 16/20Ps Ca
)
= re t- t
- k
at define
:IEi÷¥;If
'
'il
:÷÷,
Vita = fu
, = (Va ,a)Os t- -
- t Loa , a) Vee
Consequence
Orthogonal projection 17/20 Corollary For all x œ Rn, x ≠ PS(x) is orthogonal to S. ÎPS(x)Î Æ ÎxÎ.=
ProofofCauchy-Schwarz inequality
Cauchy-Schwarz inequality
Proof of Cauchy-Schwarz inequality 19/20 Theorem Let Î · Î be the norm induced by the inner product È·, ·Í on the vector space V . Then for all x, y œ V : |Èx, yÍ| Æ ÎxÎ ÎyÎ. (1) Moreover, there is equality in (1) if and only if x and y are linearly dependent, i.e. x = αy or y = αx for some α œ R.I
=
Proof :
Let
my EV
- If
- From
Proof
Proof of Cauchy-Schwarz inequality 20/20 Wedefine
f
: IR → RE t
Hy
- talk
for TER, fct)
=Ily 112 - It La, y>
+ ¥ Half- Ranch#I :
fettes
- a Tyree
- 2 poltyomialcint)
- Remark #2
fct) 30 for
all t
. Hence the discriminantd
- f fct) is
so
a=a T-4"
to
→ ÷!!Proof
Proof of Cauchy-Schwarz inequality 20/20 There isequality
kn.gs/--HaUllyH
if and only
if
A- O
→ There exists somet
suchthat
fat
- O
- talk -0
y
- ta -0
I;
s
- ta
Questions?
21/20Questions?
21/20 cos O =KI
lyIkIs
Orthogonal matrices
22/20 Definition A matrix A œ Rn×n is called an orthogonal matrix if its columns are an orthonormal family. ==
⇐
ten
- f
go.a.g.ae
Ro
- ( ÷:
Isis:)
matias .A proposition
23/20 Proposition Let A œ Rn×n. The following points are equivalent:- 1. A is orthogonal.
- 2. ATA = Idn.
- 3. AAT = Idn
O
⇒ A invertible and AT = A' G' ⇐ s - A- → µ- y
i÷:t÷÷÷:
Orthogonal matrices & norm
24/20 Proposition Let A œ Rn×n be an orthogonal matrix. Then A preserves the dot product in the sense that for all x, y œ Rn, ÈAx, AyÍ = Èx, yÍ. In particular if we take x = y we see that A preserves the Euclidean norm: ÎAxÎ = ÎxÎ.