Download Least–squares Solution of Homogeneous Equations

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Noether's theorem wikipedia , lookup

Riemannian connection on a surface wikipedia , lookup

Multilateration wikipedia , lookup

Dirac bracket wikipedia , lookup

Transcript
Least–squares Solution of Homogeneous
Equations
supportive text for teaching purposes
Revision: 1.2, dated: December 15, 2005
Tomáš Svoboda
Czech Technical University, Faculty of Electrical Engineering
Center for Machine Perception, Prague, Czech Republic
[email protected]
http://cmp.felk.cvut.cz/~svoboda
Introduction
2/5
We want to find a n × 1 vector h satisfying
Ah = 0 ,
where A is m × n matrix, and 0 is n × 1 zero vector. Assume m ≥ n, and
rank(A) = n. We are obviously not interested in the trivial solution h = 0
hence, we add the constraint
khk = 1 .
Constrained least–squares minimization: Find h that minimizes kAhk
subject to khk = 1.
Derivation I — Lagrange multipliers
3/5
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
Derivation I — Lagrange multipliers
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
3/5
To find an extreme (the sought
h) we must solve
∂
> >
>
h
A
Ah
+
λ(1
−
h
h) = 0 .
∂h
Derivation I — Lagrange multipliers
To find an extreme (the sought
h) we must solve
∂
> >
>
h
A
Ah
+
λ(1
−
h
h) = 0 .
∂h
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
3/5
We derive: 2A>Ah − 2λh = 0.
Derivation I — Lagrange multipliers
To find an extreme (the sought
h) we must solve
∂
> >
>
h
A
Ah
+
λ(1
−
h
h) = 0 .
∂h
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
3/5
We derive: 2A>Ah − 2λh = 0.
After some manipulation we end up with: (A>A − λE)h = 0 which is
the characteristic equation. Hence, we know that h is an eigenvector of
(A>A) and λ is an eigenvalue.
Derivation I — Lagrange multipliers
To find an extreme (the sought
h) we must solve
∂
> >
>
h
A
Ah
+
λ(1
−
h
h) = 0 .
∂h
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
3/5
We derive: 2A>Ah − 2λh = 0.
After some manipulation we end up with: (A>A − λE)h = 0 which is
the characteristic equation. Hence, we know that h is an eigenvector of
(A>A) and λ is an eigenvalue.
The least-squares error is e = h>A>Ah = h>λh.
Derivation I — Lagrange multipliers
To find an extreme (the sought
h) we must solve
∂
> >
>
h
A
Ah
+
λ(1
−
h
h) = 0 .
∂h
h = argminhkAhk subject to khk = 1. We rewrite the constraint as
1 − h>h = 0
3/5
We derive: 2A>Ah − 2λh = 0.
After some manipulation we end up with: (A>A − λE)h = 0 which is
the characteristic equation. Hence, we know that h is an eigenvector of
(A>A) and λ is an eigenvalue.
The least-squares error is e = h>A>Ah = h>λh.
The error will be minimal for λ = mini λi and the sought solution is
then the eigenvector of the matrix (A>A) corresponding to the smallest
eigenvalue.
Derivation II — SVD
4/5
Let A = USV>, where U is m × n orthonormal, S is n × n diagonal with
descending order, and V> is n × n also orthonormal.
Derivation II — SVD
4/5
Let A = USV>, where U is m × n orthonormal, S is n × n diagonal with
descending order, and V> is n × n also orthonormal.
From orthonormality of U, V follows that kUSV>hk = kSV>hk and
kV>hk = khk.
Derivation II — SVD
4/5
Let A = USV>, where U is m × n orthonormal, S is n × n diagonal with
descending order, and V> is n × n also orthonormal.
From orthonormality of U, V follows that kUSV>hk = kSV>hk and
kV>hk = khk.
Substitute y = V>h. Now, we minimize kSyk subject to kyk = 1.
Derivation II — SVD
4/5
Let A = USV>, where U is m × n orthonormal, S is n × n diagonal with
descending order, and V> is n × n also orthonormal.
From orthonormality of U, V follows that kUSV>hk = kSV>hk and
kV>hk = khk.
Substitute y = V>h. Now, we minimize kSyk subject to kyk = 1.
Remember that S is diagonal and the elements are sorted descendently.
Than, it is clear that y = [0, 0, . . . , 1]>.
Derivation II — SVD
Let A = USV>, where U is m × n orthonormal, S is n × n diagonal with
descending order, and V> is n × n also orthonormal.
From orthonormality of U, V follows that kUSV>hk = kSV>hk and
kV>hk = khk.
Substitute y = V>h. Now, we minimize kSyk subject to kyk = 1.
Remember that S is diagonal and the elements are sorted descendently.
Than, it is clear that y = [0, 0, . . . , 1]>.
4/5
From substitution we know that h = Vy from which follows that
sought h is the last column of the matrix V.
Further reading
Richard Hartley and Andrew Zisserman, Multiple View Geometry in
computer vision, Cambridge University Press, 2003 (2nd edition),
[Appendix A5]
Gene H. Golub and Charles F. Van Loan, Matrix Computation, John
Hopkins University Press, 1996 (3rd edition).
Eric W. Weisstein. Lagrange Multiplier. From MathWorld–A Wolfram
Web Resource.
http://mathworld.wolfram.com/LagrangeMultiplier.html
5/5
Eric W. Weisstein. Singular Value Decomposition. From MathWorld–A
Wolfram Web Resource.
http://mathworld.wolfram.com/SingularValueDecomposition.html