Download MATH3033 Information Sheet 1. The Standard Maximum Problem

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Matrix multiplication wikipedia , lookup

Non-negative matrix factorization wikipedia , lookup

Four-vector wikipedia , lookup

Shapley–Folkman lemma wikipedia , lookup

Matrix calculus wikipedia , lookup

Gaussian elimination wikipedia , lookup

Simplex wikipedia , lookup

System of linear equations wikipedia , lookup

Transcript
MATH3033 Information Sheet
1
Linear Programming
1. The Standard Maximum Problem: Find a vector x that maximises
the quantity c1 x1 + . . . + cn xn subject to the constraints
a11 x1 + . . . + a1n xn ≤ b1
..
.
am1 x1 + . . . + amn xn ≤ bm
where x1 , . . . , xn ≥ 0. We often write c = (c1 , . . . , cn )T , b = (b1 , . . . , bm )T


a11 · · · a1n
.. .
..
and A =  ...
.
.
am1 · · · amn
For the rest of the statements in this section, this standard maximum
problem will be referred to as (1).
2. The Standard Minimum Problem: Find a vector y that minimises
the quantity b1 y1 + . . . + bm ym subject to the constraints
a11 y1 + . . . + am1 ym ≥ c1
..
.
a1n y1 + . . . + amn ym ≥ cn
where y1 , . . . , ym ≥ 0.
For the rest of the statements in this section, this standard minimum
problem will be referred to as (2).
3. Theorem 1: If x is feasible for the standard maximum problem (1) and
y is feasible for its dual (2) then
cT x ≤ yT b.
4. Corollary 1: If a standard problem and its dual are both feasible, then
both are bounded feasible.
5. Corollary 2: If there are points x0 and y0 feasible for a standard maximum problem (1) and its dual (2) respectively, such that cT x0 = y0T b
then x0 is optimal for (1) and y0 is optimal for (2).
1
6. The Duality Theorem: If a standard maximum or minimum problem
is bounded feasible, then so is its dual, optimal vectors for both problems
exist, and their objective functions are equal.
7. The Equilibrium Theorem: Let x and y be feasible vectors for a
standard maximum problem (1) and its dual (2) respectively. Then x and
y are optimal if and only if
yi = 0 for every i such that
n
X
aij xj < bi
j=1
and
xj = 0 for every j such that
m
X
yi aij > cj .
i=1
8. The Simplex Method: The simplex tableau for a standard maximum
problem and its dual looks like this:
y1
y2
..
.
x1
a11
a21
..
.
x2
a12
a22
..
.
···
···
···
...
ym am1 am2 · · ·
−c1 −c2 · · ·
xn
a1n
a2n
..
.
b1
b2
..
.
amn bm
−cn 0
If ai0 j0 6= 0 then yi0 and xj0 can exchange places pivoting on the entry ai0 j0 ,
in which case each aij is replaced with âij according to these formulae:
1
ai0 j0
a
âi0 j = aii0jj for j 6= j0 . (Entries in the same row)
0 0
a
âij0 = − aiijj0 for i 6= i0 . (Entries in the same column)
0 0
a a
âij = aij − ijai0 jio j for i 6= i0 and j 6= j0 .
0 0
(i) âi0 j0 =
(ii)
(iii)
(iv)
The simplex method is to follow these rules:
Rule 1: If b ≥ 0, choose any column j0 with the last entry negative. Of the
b
positive aij0 choose i0 such that ai i0j ≤ abiji for all i, and pivot on
0 0
0
ai0 j0 . If there is a tie, any of the tied entries may be used.
Rule 2: If some of the bi are negative, pick the first bk such that bk < 0,
b1 , . . . , bk−1 ≥ 0. In row k, pick any negative entry akj0 < 0. Calculate
bk
and abiji for all i with bi ≥ 0 and aij0 > 0. Pick i0 for which this
akj0
0
ratio is smallest (in the case of a tie, pick any one of the tied entries)
and pivot on ai0 j0 .
2
Alternatively, the simplex method can be applied to the dual problem, in
which case the rules are:
Rule 1*: If −c ≥ 0, choose any row i0 with the last entry negative. Of the
−c
−c
negative ai0 j choose j0 such that ai jj0 ≥ aij j for all j, and pivot on
0 0
0
ai0 j0 . If there is a tie, any of the tied entries may be used.
Rule 2*: If some of the −cj are negative, pick the first −ck such that −ck < 0,
−c1 , . . . , −ck−1 ≥ 0. In column k, pick any positive entry ai0 k > 0.
−c
k
Calculate a−c
and ai jj for all j with −cj ≤ 0 and ai0 j > 0. Pick j0 for
i0 k
0
which this ratio is closest to zero (in the case of a tie, pick any one of
the tied entries) and pivot on ai0 j0 .
9. Sensitivity Analysis
Suppose the standard maximum problem has been solved with dual solution y = (y1 , . . . , ym ). To calculate the new CPF solution under changes
in parameter(s) A → Â, b → b̂ and c → ĉ, follow the following steps:
(a) Construct the matrix [Â : I] and remove the columns corresponding
to nonbasic variables. What remains is B.
(b) Calculate B −1 . (Alternatively, and probably more quickly, find B −1
from the coefficients of the slack variables in the extended solved
tableau.)
(c) Calculate:
b∗ = B −1 b̂,
A∗ = B −1 Â,
z ∗ − ĉ = y − ĉ and
Z ∗ = yb̂.
(d) Construct the simplex matrix
B −1 b∗
A∗
z ∗ − ĉ y Z ∗
(e) Apply Gauss-Jordan elimination so that the columns of this matrix
corresponding to basic variables have zeroes everywhere, except for
the entry in the row of that basic variable, which must be 1.
(f) Removing the columns of the basic variables leaves the new simplex
tableau.
10. Integer Programming
We use the branch and bound procedure. Set Z ∗ = −∞ (or to +∞ for
a standard minimum IP problem).
(a) Divide the feasible region according to values of one decision variable.
3
(b) Solve the LP relaxations of the two new sub-problems, to get bounds
on Z in each sub-region.
(c) A sub-problem is fathomed if any of the following conditions is true:
(i) The solution to the LP relaxation is integer. If the Z-value here
is better than Z ∗ then update the value of Z ∗ to this new value.
(ii) The sub-problem is infeasible.
(iii) The Z-value of the sub-problem is worse than Z ∗ .
A fathomed sub-problem needs no further consideration.
If all sub-problems have been fathomed, stop; the IP solution is Z ∗ .
If there are unfathomed sub-problems remaining, re-start the procedure
on the most recent one. In the case of ties, pick the one with the best
Z-value.
2
Nonlinear Programming
1. A function f (x) is convex (resp. concave) if the line segment joining any
two points of the graph of f lies on or above (resp. below) the graph of
f.
If f is a function of one variable, then f is convex (resp. concave) iff
2
d2 f
≥ 0 (resp. dd xf2 ≤ 0).
d x2
If f is a function of two variables, then f is convex (resp. concave) iff
∆ = fxx fyy − (fxy )2 ≥ 0 and fxx , fyy ≥ 0 (resp. fxx , fyy ≤ 0).
2. A set S is convex if the line segment joining any two elements of S lies
entirely in S.
3. Newton’s Method for functions of one variable: Provided that xi
is sufficiently close to x∗ , where f (x∗ ) = 0, an approximation xi+1 to x∗
can be found using the formula
xi+1 = xi −
f 0 (xi )
.
f 00 (xi )
4. Karush-Kuhn-Tucker conditions: If x∗ minimises f (x) subject to the
conditions that for i = 1, . . . , m and j = 1, . . . , l, gi (x) ≤ 0 and hj (x) = 0,
4
then there exist scalars µi and λj (called KKT multipliers) such that
∇f (x∗ ) +
m
X
µi ∇gi (x∗ ) +
i=1
l
X
λj ∇hj (x∗ ) = 0
j=1
gi (x∗ )
hj (x∗ )
µi gi (x∗ )
µi
≤
=
=
≥
0
0
0
0,
for i = 1, . . . , m and j = 1, . . . , l.
Corollary: If f is convex and the feasible region is also convex, any point
x∗ that satisfies the KKT conditions must be a global minimum.
5. Quadratic Programming: We assume that f is of the form f (x) =
cT x− 21 xT Qx, and that all constraints are linear. To minimise f (x) subject
to the linear constraints Ax ≤ b and x ≥ 0, we do the following:
(i) Write down the KKT conditions for the problem and re-arrange them
in the form
Qx + (−IAT )µ = c
Ax + s = b,
along with a complementarity condition.
(ii) For any ci > 0 and bj < 0 introduce artificial variables z1 , z2 , etc.
(iii) Solve the LP problem of minimising Z = z1 + z2 + · · · (i.e. maximise
−Z = −z1 − z2 − · · · ), with the constraints you found in step (i).
Use the simplex method to do this, rewriting the objective function
in terms of non-z variables and avoiding any pivots that violate the
complementarity conditions.
3
Game Theory
1. Dominance: Player i’s strategy ŝi is strictly dominated by player i’s
strategy si if
ui (si , s−i ) > ui (ŝi , s−i )
for all possible s−i .
2. Weak Dominance: Player i’s strategy ŝi is weakly dominated by player
i’s strategy si if
ui (si , s−i ) ≥ ui (ŝi , s−i )
5
for all possible s−i , and
ui (si , s−i ) > ui (ŝi , s−i )
for some s−i .
3. Best Response: Player i’s strategy ŝi is a best response to other players’
strategies s−i if
ui (ŝi , s−i ) ≥ ui (si , s−i )
for all si .
Equivalently, ŝi solves the problem: choose si to maximise ui (si , s−i ).
4. Nash Equilibrium: A strategy profile (s∗1 , s∗2 , . . . , s∗N ) is a Nash Equilibrium if s∗i is a best response to s∗−i for each i ∈ {1, 2, . . . , N }.
4
Markov Chains
We assume that time increases in discrete units, t = 0, 1, 2, . . . and that the
system in question has a finite number of states, numbered 0 to N .
(n)
1. We define the n-step transition probability pij as the probability that in
n units of time from now, the system will be in state j, given that it is
in state i now. We usually write the one-step transition probabilities in
a transition matrix P = (pij ). The n-step probabilities can be found by
computing P n .
2. For any irreducible ergodic Markov chain, lim pnij = πj exists and is
n→∞
independent of i. The row vector (π0 , . . . , πN ) can be found by solving
(π0 , . . . , πN ) = (π0 , . . . , πN )P.
(n)
3. fij is the probability that the system will transition from state i to state j
N
∞
X
X
(1)
(n)
(n−1)
(n)
in n steps. We have fij = pij and fij =
pik fkj . If
fij = 1
k=0,k6=j
n=1
then the expected first-passage-times µij (the expected time to transition
from state i to state j) satisfy
µij = 1 +
N
X
k=0,k6=j
6
pik µkj .
4. If state k is an absorbing state, fik is the probability that the system will
ever enter state k given that it is in state i now. then the column vector
(f0k , f1k , . . . , fN k )T can be computed from




f0k
f0k
 f 
 f1k 

 .  = P  1k
 ...  ,
 .. 
fN k
fN k
along with the conditions that fkk = 1, and if state l is any other absorbing
state, flk = 0.
7