Download SOLVING LINEAR EQUATIONS USING AN OPTIMIZATION

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Quadratic equation wikipedia , lookup

Quartic function wikipedia , lookup

Compressed sensing wikipedia , lookup

Elementary algebra wikipedia , lookup

Linear algebra wikipedia , lookup

Signal-flow graph wikipedia , lookup

System of polynomial equations wikipedia , lookup

History of algebra wikipedia , lookup

Equation wikipedia , lookup

System of linear equations wikipedia , lookup

Transcript
SOLVING LINEAR EQUATIONS USING AN OPTIMIZATION-BASED
ITERATIVE SCHEME
I. Tang
ETS-Rosedale, Mail Stop 58-N, Princeton, NJ 08441-0001
(Submitted September 1996)
A system of linear equations such as
anxx + anx2
a2lxx
+ ••• + alnxn + q = 0
+ ... + a2nx„ + c2 = 0
(1)
+ .-. + annxn + cn = 0
can be solved using either direct methods such as the Gauss-Jordan procedure or iterative
methods such as the Gauss-Seidel procedure. When the equation system is large, and especially
when the coefficients are sparsely distributed, iterative methods are often preferred (see [1], [2])
since iterative methods for these systems are quite rapid and may be more economical in memory
requirements of a computer. Iterative methods usually require a set of starting values as assumed
solution. This article describes a procedure that does not require starting values. The procedure
achieves convergence rapidly and can be applied to dependent systems in which there are fewer
equations than variables.
Consider the case of n simultaneous equations in matrix form:
K1]
A
*\\
\Jn)
*V
\n
(2)
A
n\
vly
We propose to solve the system
:0
(3)
\Jn)
by minimizing a scalar objective function
/„)
(4)
\jnj
The solution of (3) is obtained when the x values are found such that H = 0.
Differentiation of (4) yields
dH_
dt
2(A - / „ )
(5)
dfjdt
But (2) gives
{dfxldt\
dfjdt
a,l i
\an\
*\n
"tin J
(dxxldi
dx„/dt
a,l i
\anl
dxxldt\
*\n
6
dx„ldt
dzldt
(6)
and, together with (2), (5) becomes
248
[JUNE-JULY
SOLVING LINEAR EQUATIONS USING AN OPTIMIZATION-BASED ITERATIVE SCHEME
a,11
dH_
dt
2(/i -
«i„
^
nn
/
/„)
\a„i
A
= 2(Xl -
dxxldt\
dxnl dt
dzldt
nl
<hn
x„ 1)
A
ln
V
dxnldt
dzldt
a„„ 0 ,
\anl
nJ
dxjdt\
^
nn
/
(7)
We now set
dxxl dt
(x.\
A
n\
(8)
dxnldt
dzldt
™\n
\<*nl
0
0
a
C
nn
nJ l
\J
so that dH/dt<0.
By choosing a small value for At (for example, set It equal to the reciprocal of the Euclidean
norm [1] of the matrix), one could approximate the derivatives on the left-hand side of (8) by
finite differences between the (i +1)* and the Ith iterant of each of the x's (with z remaining a
constant = 1 ) , and we write
dxx
x
_ AXj _
dt ~ At ~
dx„
dt ~ At
dz _ Az
dt~ At
i,i+i~xi,i
At
X
• XV
n,i +1
n,i
(9)
At
*M
~h
At
Then (8) becomes
f\ ,
\
fv\
-At
a,
a„-
au
a„.
/+1
aln
cx\
a
C
(\- \
(10)
a
\ nl
nn
2j
vU
with zi+l = zi = l.
Let
[B] =
(I 0
0 1
o^
0
A
n\
AU
a,i i
aln
c^
01)
•At
6 6
nn
\an\
a
nn
C
nJ
and
fv\
(12)
[X] =
then (10) becomes the recursion equation
\h
(13)
1998]
249
SOLVING LINEAR EQUATIONS USING AN OPTIMIZATION-BASED ITERATIVE SCHEME
Starting with an arbitrary set of values [X] - [X]0 with z = 1 as the initial solution, we carry
out the iteration
[X\=[B][X]0
[X]2=[B] [X\
= [Bf[X]Q
(14)
and so on until
[X]k=[Bf[X]0.
Unless the set of equations is an inconsistent system, for sufficiently small At, which serves as an
accelerating factor, [Bf will converge so that
(bn
[*]*->
u
\n
*l,s
X
"un
0
(15)
n,s
0
1
\n
-*V
as£-»oo. It follows from (13) that
hi
[*]* =
u
®nn
0
0
X
n, s
[*1lo-
OS)
1
If the equation system is furthermore not a dependent system, the individual elements by will tend
to zero upon convergence, and
X
l,s
(17)
[X\ =
vly
regardless of the initial starting value, and [X] = [xj are obtained for the solution of (2).
Since the last column of the matrix (15) to which [B] converges contains the solution of (2),
it is clear that to solve a set of linear equations that does not constitute a dependent system, an
assumed starting solution is not required. We only need to multiply [B] of (11) upon itself repeatedly, and if the multiplication is performed by squaring the previous result, convergence is accelerated through the sequence [B],[Bf,[Bf,... . Although the process diverges for an overdetermined system, for a dependent system, one can obtain a solution from (15) according to the
initial starting solution vector.
To illustrate the scheme, consider the following examples.
1. Consider the system: x+y = -2
2x+y = -3
(l 0 0^
(I 2)
[B] = 0 1 0 -At 1 1 1 1 2
2 1 3
1° o lj
250
(o o)
(0.95 -0.03 -0.08"!
-0.03 0.98 -0.05
0
0
1
[JUNE-JULY
SOLVING LINEAR EQUATIONS USING AN OPTIMIZATION-BASED ITERATIVE SCHEME
for which the acceleration factor At = 0.01 is used. Convergence leads to
"0 0 -1
[BT-*\ 0 0 -1
0 0
1
From the last column, the solution (x, y) = (-1, -1) is obtained.
2. Given the dependent system: x+y+z + 2 = 0
2x+y-z+3=0
f 0.95 -.03 0.1 -0.08^
-0.03 0.98
0 -0.05
0.01
0 0.98 0.01
0
0 0 ij
0
0
1 J
for which the acceleration factor At is again 0.01. Iteration leads to
(I
[B] = 0
0
0
0\
(I
2]
1 0 0 -At 1 1
1 -1
0 1 0
0 0
1 1
1 -1
1° °J
'0.2857 -0.428 0.1428 -1.142^
0.6428 -0.214 -0.785
[BT-> -0.428
0.1428 -0.214 0.0714 -0.071
0
1
0
0
v
The solution is found from (15) for any arbitrarily chosen starting value. For instance, if
x0 = j ; 0 =z 0 = 0, then, from the last column, we obtain (x,y,z) = (-1.142, -0.785, -0.071). If
xQ = 2, y0 = 0, and z0 = 1, then (x, y, z) = (-0.714, -1.427,0.143) is obtained.
ACKNOWLEDGMENT
The encouragement of Dr. J. R. Davidson of the Department of Mathematics at Oklahoma
State University (Oklahoma City) is greatly appreciated.
REFERENCES
1.
Curtis F. Gerald & P. O. Wheatley. Applied Numerical Analysis. 4th ed. New York:
Addison Wesley, 1992.
2. M. L. James, G. M. Smith, & J. C. Wolford. Applied Numerical Methods for Digital Computation. New York: Harper Collins, 1993.
AMS Classification Numbers: 15A06, 65F10
1998]
251