Download Optimization Iterative Methods for Smooth Objective Functions

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

System of linear equations wikipedia , lookup

Transcript
Optimization
Iterative Methods for Smooth Objective Functions
 Quadratic Objective Functions
 Stationary Iterative Methods (first/second order)
 Steepest Descent Method
 Landweber/Projected Landweber Methods
 Conjugate Gradient Method
 Non-Quadratic Smooth Objective Functions
 Conjugate Gradient Method
 Newton’s Method
 Trust Region Globalization of Newton’s Method
 BFGS Method
IPIM, IST, José Bioucas, 2015
1
References
[1] O. Axelsson, Iterative Solution Methods. New York: Cambridge
Univ. Press, 1996.
[2] Golub, G.H. and Van Loan, C.F., Matrix Computations, Johns Hopkins
University Press, Baltimore, Maryland, 1983.
[3] C. Byrne, A unified treatment of some iterative algorithms in
signal processing and image reconstruction, Inverse Problems, vol.
20, pp. 103–120, 2004.
IPIM, IST, José Bioucas, 2015
2
Rates of convergence
Suppose that
as
Linear convergence rate: there exits a constant
for which
Superlinear convergence rate: there exits a sequence
of real numbers such that
and
Quadratic convergence rate: there exits a constant
for which
IPIM, IST, José Bioucas, 2015
3
Rates of convergence: example
0
10
linear
superlinear
quadratic
-20
10
-40
10
-60
10
-80
10
-100
10
-120
10
-140
10
0
IPIM, IST, José Bioucas, 2015
5
10
15
20
25
30
35
40
45
50
4
Comparing linear convergence rates
 Many iterative methods for large scale inverse problems
have linear converge rate:
– convergence factor
r - log10 – convergence rate
– number of iterations to reduce the error by a factor of 10
IPIM, IST, José Bioucas, 2015
5
Induced norms and spectral radius
 Given a vector norm
vector norm is
, the matrix norm
induced by the
 When the vector norm is the Euclidian norm, the induced norm is termed
the spectral norm and is given by
 If is Hermitian
spectral radius of A,
IPIM, IST, José Bioucas, 2015
, the matrix norm
is given by the
6
Key results involving the spectral radius
IPIM, IST, José Bioucas, 2015
7
Tikhonov regularization/Gaussian priors
Assume that
is non-singular. Then
The solution is obtained by solving the system
IPIM, IST, José Bioucas, 2015
8
Stationary iterative methods
Consider the system
, where
is non-singular
First Order Stationary Iterative Methods
Let
is nonsingular
be a splitting of
for
must be ease to invert
Jacobi
Gauss-Seidel
IPIM, IST, José Bioucas, 2015
9
Stationary iterative methods
Frequently, we can not access to the elements of A or D, but only
apply these operators. Thus C should depend only on these operators
Example 1: Landweber iterations
Example 2:
Easy to compute when D is
diagonal or a convolution
IPIM, IST, José Bioucas, 2015
10
First order stationary iterative methods: convergence
Consider the system
Let
be a splitting of
is nonsingular
and
Then
iff
IPIM, IST, José Bioucas, 2015
11
First order stationary iterative methods: convergence
Consider the system
Let
be a splitting of
for
Convergence
iff
IPIM, IST, José Bioucas, 2015
12
First order stationary iterative methods (cont.)
Ill-conditioned systems
Number of iterations to attenuate the error norm by 10
 Landweber C = I

Under what conditions?
The eigenvalues of
IPIM, IST, José Bioucas, 2015
tend to be less spread than those of
13
Second order stationary iterative methods: convergence
Consider the system
Let
be a splitting of
Convergence [1]
iff
IPIM, IST, José Bioucas, 2015
14
First/second order stationary iterative methods: comparison
Ill-conditioned systems
First order
Second order
Example
Second order is 100 times faster
IPIM, IST, José Bioucas, 2015
15
Steepest descent method
non-stationary first order iterative
method
Optimal  (line search)
IPIM, IST, José Bioucas, 2015
Convergence
16
Conjugate gradient method
Consider the system
Are conjugate with respect to
if
Equivalently
Let
be a sequence of n mutualy conjugate directions and
Since
Then
and
IPIM, IST, José Bioucas, 2015
17
Conjugate gradient method as an iterative method
Computing the solution of
2- Define
is equivalent to minimize
to as the projection error of
onto the direction
1- minimize along conjugate directions directions
IPIM, IST, José Bioucas, 2015
18
Conjugate gradient and steepest descent paths
steepest descent
conjugate gradient
IPIM, IST, José Bioucas, 2015
19
The resulting algorithm
(
IPIM, IST, José Bioucas, 2015
denotes the negative of the gradient)
20
Some remarks about the CG method
Convergenge [2]
IPIM, IST, José Bioucas, 2015
21
Comparison: CG and First/Second Order Stationary Iterative Methods
4
10
1st order
3
10
2
10
1
10
2nd order
0
10
CG
-1
10
0
50
100
IPIM, IST, José Bioucas, 2015
150
200
250
300
350
400
450
500
22
Preconditioned conjugate gradient (PCG) method
Let
be a s.p.d matrix such that
The eigenvalues of
CG solves the system
system
are more clustered than those of
faster than the
Note: PCG can be written as a small modification of CG: The complexity
of each PCG iteration is that of CG plus the computation of
IPIM, IST, José Bioucas, 2015
23
Constrained Tikhonov regularization/Gaussian priors
where
is a closed convex set
Projection onto a convex set
is non-expansive
IPIM, IST, José Bioucas, 2015
24
Projected iterations
is a closed convex set
Let
be a contraction mapping
Assume that sequence generated by
solution of the unconstrained problem
converges to the
Define the operator:

is a contraction mapping
 for any starting element
, the sequence of sucessive approximations
is convergent and its limit is the unique fixed point of
 the unique fixed point of
optimization problem
IPIM, IST, José Bioucas, 2015
is the solution of the constrained
25