Download RENEWAL THEORY 1.1. Example: A Dice

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Equation wikipedia , lookup

Fundamental theorem of algebra wikipedia , lookup

Transcript
RENEWAL THEORY
1. RENEWAL PROCESSES
1.1. Example: A Dice-Rolling Problem. In board games like Monopoly, Clue, Parcheesi, and so
on, players take turns rolling a pair of dice and then moving tokens on a board according to rules
that depend on the dice rolls. The rules of these games are all somewhat complicated, but in
most of the games, in most circumstances, the number of spaces a player moves his/her token is
the sum of the numbers showing on the dice.
Consider a simplified board game in which, on each turn, a player rolls a single fair die and
then moves his/her token ahead X spaces, where X is the number showing on the die. Assume
that the board has K spaces arranged in a cycle, so that after a token is moved ahead a total of K
spaces it returns to its original position. Assign the spaces labels 0, 1, 2, . . . , K − 1 in the order they
are arranged around the cycle, with label 0 assigned to the space where the tokens are placed
at the beginning of the game (in Monopoly, this would be the space marked “Go”). Then the
movement of a player’s token may be described as follows: after n turns the token will be on the
space labelled S n mod K , where
(1)
Sn =
n
X
Xi
i =1
and X 1 , X 2 , . . . are the dice rolls obtained on successive turns.
Problem 0: What is the probability that on the tenth time around the board a player’s token lands
on the space labelled K − 1?
Anyone who has played Monopoly will know that this particular space is “Boardwalk”, and
that if another player owns it then it is a bad thing to land on it, especially late in the game.
Of course, the answer to Problem 0 isn’t especially relevant to Monopoly, because the rules of
movement aren’t the simple rules we have imposed above; in fact, the problem is really more
interesting in Monopoly, because there it turns out that the probability of landing on Boardwalk
isn’t quite the same as that of landing on (say) Park Place (space K − 3). But we have to start
somewhere, so Problem 0 is what we’ll consider. Here is a more general problem, formulated in
a more convenient way:
Problem 1: Fix an integer x ≥ 0. What is the probability u x that S n = x for some n?
Observe that when x = 10K − 1, this reduces to Problem 0.
There is a related issue in Monopoly that is of some interest: when you pass Go, there is a
block of spaces (Baltic and Mediterranean Avenues) that you might want to avoid if you don’t
own them. So what is the probability that on the tenth trip around the board you manage to
jump over them altogether (that is, the chance that you don’t land on any of the spaces labelled
1,2,3)? Here is a more general formulation:
1
2
RENEWAL THEORY
Problem 2: Fix t ≥ 0, and define the first-passage and overshoot random variables by
(2)
τ(t ) = min{n ≥ 0 : S n > t }
(3)
R(t ) = S τ(t ) − t
and
What is the probability distribution of R(t )?
1.2. Ordinary Renewal Processes. Problems 1 and 2 above are prototypical problems in renewal
theory. In general, a renewal process (or more precisely, an ordinary renewal process) is the increasing sequence of random nonnegative numbers 0,S 1 ,S 2 , . . . gotten by adding i.i.d. nonnegative random variables X 1 , X 2 , . . . , as in equation (1). The individual terms S n of this sequence are
called renewals, or sometimes occurrences. With each renewal process is associated a renewal
counting process N (t ) that tracks the total number of renewals (not including the initial occurrence) to date: the random variable N (t ) is defined by
(4)
N (t ) = max{n : S n ≤ t } = τ(t ) − 1
where τ(t ) is the first passage time defined by (2) above.
In many instances (but not the dice-rolling example) the renewals have a natural interpretation as the times at which some irregularly occurring event take place. (For example, imagine
that X 1 , X 2 , . . . are the lifetimes of replaceable components in a system, such as light bulbs in a
light socket. If the first component is inserted at time 0, then the renewals S n are the times at
which components must be replaced.) Thus, it is traditional in the subject to refer to the random
variables X i as interoccurrence times, and their common distribution as the interoccurrence time
distribution. The random variables
(5)
A t = A(t ) = t − S N (t ) ,
R t = R(t ) = S τ(t ) − t ,
and
L t = L(t ) = S τ(t ) − S N (t )
are usually called the age, residual lifetime, and total lifetime random variables (although in certain applications it is also common to refer to R(t ) as the overshoot). Note that L(t ) = A(t )+ R(t ).
1.3. Delayed Renewal Processes. It isn’t always reasonable to insist that the first renewal occurs
at time S 0 = 0. For instance, in applications where the occurences are at times when a component of a system must be replaced, one might well be interested in situations where there is
already a working component in place at time 0. For this reason, we define a delayed renewal
process to be a sequence S 0 ,S 1 ,S 2 , . . . where
(6)
Sn = S0 +
n
X
Xj ,
j =1
and (i) the interoccurence times X 1 , X 2 , . . . are positive and i.i.d., as in an ordinary renewal process, and (ii) the initial delay S 0 ≥ 0 is independent of the interoccurence times X i . Notice that
the distribution of the initial delay random variable S 0 is not required to be the same as that of the
interoccurence time random variables X i . The age, residual lifetime, and total lifetime random
variables for a delayed renewal process are defined in exactly the same manner as for ordinary
renewal processes (equations (5) above).
RENEWAL THEORY
3
1.4. There’s a renewal process in my Markov chain. Let {Yn }n≥0 be an irreducible, recurrent
Markov chain on a finite or countable state space Y . Fix a state y ∗ , and let 0 < τ1 < τ2 < · · · be
the times of successive visits to y ∗ . Since the chain is recurrent, all of these times are finite w.p.1.
What’s more, the sequence {τm }m ≥1 constitutes a (delayed) renewal process. This follows from
the (strong) Markov property, because at every visit to y ∗ the Markov chain “starts afresh”, and
so the time until the next return (a) is independent of everything that happened earlier, and (b)
has the same distribution as all other return times (except possibly the first). Let { f k }k ≥1 be the
distribution of the times τm +1 − τm between successive visits, and let µ be its mean. Then by
basic Markov chain theory, the Markov chain Yn is positive recurrent if and only if µ < ∞, and if
µ < ∞ then the stationary probability of state y ∗ is
π(y ∗ ) = 1/µ.
1.5. Poisson and Bernoulli Processes. Two special cases are especially important. When the
interoccurrence time distribution is the exponential distribution with mean 1/λ, then the corresponding renewal process is the Poisson process with intensity λ. When the interoccurrence time
distribution is the geometric distribution with parameter p , that is,
P{X i = k } = p (1 − p )k −1
for k = 1, 2, . . . ,
then the corresponding renewal process is the Bernoulli process with success parameter p . Many
interesting problems that are quite difficult for renewal processes in general have simple solutions for Poisson and Bernoulli processes. For instance, in general the distribution of the count
random variable N (t ) does not have a simple closed form, but for a Poisson process it has the
Poisson distribution with mean λt . Similarly, the age and residual lifetime random variables
A(t ) and R(t ) generally have distributions that do not depend on the interoccurrence time distribution in a simple way, but for the Poisson process they are, for any fixed t , just independent
exponential random variables with mean 1/λ. These properties follow easily from the memoryless property of the exponential distribution. Renewal processes are important because in many
systems the times between successive renewals do not have memoryless distributions.
Exercise: What are the distributions of A t , R t , and L t for a Bernoulli process?
1.6. Renewal Processes: Arithmetic and Nonarithmetic. Renewal theory breaks into two parallel streams, one for discrete time, the other for continuous time. These are called the arithmetic
and nonarithmetic cases, respectively: A renewal process is said to be nonarithmetic if the distribution of the interoccurrence time random variables is not fully supported by an arithmetic
progression h, 2h, 3h, . . . . Otherwise, it is said to be arithmetic, and the maximal value of h > 0
such that the interoccurrence time distribution is fully supported by hZ is called the span. For
simplicity, assume henceforth that in the arithmetic case the span is h = 1, and also, for delayed
arithmetic renewal processes, that the initial delay random variable S 0 is integer-valued.
1.7. Arithmetic Renewal Processes: Associated Markov Chains. If {S n }n ≥0 is an arithmetic renewal process with span 1 then the discrete-time processes A n and R n are Markov chains.
Exercise: Prove this, and prove also that the vector-valued process (A n , R n ) is a Markov chain.
Let’s focus on the residual lifetime process R n . It might seem that the natural state space for
this process is N = {1, 2, ...}. However, if the distribution of the increments f (k ) = f k = P{X 1 = k }
puts no mass on the integers > M then all of these will be inaccessible from state 1. So when
the distribution { f k }k ≥1 has finite support, it’s more natural to let the state space be [m ] :=
4
RENEWAL THEORY
{1, 2, . . . , m } where m is the largest integer such that f m > 0. With this convention, the Markov
chain R n is irreducible. Its transition probabilities are
p (x , x − 1) = 1
if x ≥ 2;
p (1, x ) = f x
for x = 1, 2, ....
This makes it clear (if it wasn’t already) that R n is a recurrent Markov chain, because from any
initial state x ≥ 1 there is P x −probability one that R n will visit state 1. Furthermore, the expected
return time to state 1 is
∞
X
(x − 1)f x = µ − 1,
x =1
where µ is the mean of the increment distribution. Thus, the Markov chain is positive recurrent
if and only if µ < ∞.
Assume then that µ < ∞. Since the Markov chain R n is irreducible and positive recurrent it
has a unique stationary distribution πx . This must satisfy the equations
πx = πx +1 + π1 f x
for all x ≥ 1.
These are easy to solve: the solution is
X
1X
(7)
πx =
f k /π1 =
fk
µ
k ≥x
for all x ≥ 1.
k ≥x
Exercise: Why is 1/µ the correct normalizing factor?
Exercise: What is the stationary distribution of the Markov chain A n ? How about (A n , R n )?
2. THE RENEWAL THEOREM
2.1. Renewal Theorem: Statement. Consider now an arithmetic renewal process {S n }. Let {u m }m ≥0
be the associated renewal measure, that is, u m is the probability that the random walk {S n } ever
visits the point m . Since the random walk makes only positive steps, it can only visit a point m
at most once, so by the law of total probability,
(8)
um =
∞
X
P{S n = m }.
n =0
The cornerstone of renewal theory (in the arithmetic case) is the following theorem, due to
FELLER, ERDÖS, & POLLARD, which asserts that the sequence u m converges to a constant.
Feller-Erdös-Pollard Theorem . Assume that the step distribution of the renewal process is not
fully concentrated on any proper additive subgroup of the integers.1 Then
(9)
lim u m = 1/µ,
m →∞
where µ is the mean of the step distribution.
This gives a partial answer to the dice-rolling problem (Problem 1) of section 1.1: It implies
that if x is large, the chance that S n = x for some n is about 1/µ = 1/(3.5) = 2/7. Later we will see
how to get much more precise numerical approximations.
1Equivalently, there is no integer h ≥ 2 such that the step distribution is concentrated on the set hZ of integer
multiples of h.
RENEWAL THEORY
5
Proof of the Feller-Erdös-Pollard Theorem. Consider the residual lifetime chain R n . Every time
that R n = 1 there will be a renewal at time n + 1; conversely, every renewal (except the one at
m = 0) must be preceded by a visit to 1 by R n . Hence,
u0 = 1
and u m = P{R m −1 = 1} for m = 2, 3, ....
By Kolmogorov’s convergence theorem for positive recurrent Markov chains, in the case µ <
∞ the probability that R m −1 = 1 converges as m → ∞ to the steady-state probability π1 . In
section 1.7 we saw that π1 = 1/µ. Similarly, Kolmogorov’s theorem for null recurrent Markov
chains implies that when µ = ∞ the probability P{R m −1 = 1} converges to 0 = 1/µ.
Feller, Erdös, and Pollard seem not to have realized at the time they wrote their paper that
their theorem was essentially a trivial consequence of Kolmogorov’s theorem, which had been
proved over 15 years earlier (I think). They gave two different analytical proofs2, both harder
than that above. It was pointed out somewhat later that Kolmogorov implies F-E-P, and also that
F-E-P implies Kolmogorov’s theorem (by using the embedded renewal processes exhibited in
section 1.4 above). In section 3 below, I’ll give yet another analytic proof, based on generating
functions, for the special case where the step distribution has finite support.
2.2. The Renewal Equation. The usefulness of the Feller-Erdös-Pollard theorem derives partly
from its connection with another basic theorem called the Key Renewal Theorem (see below)
which describes the asymptotic behavior of solutions to the Renewal Equation. The Renewal
Equation is a convolution equation relating bounded sequences {z (m )}m ≥0 and {b (m )}m ≥0 of
real numbers:
Renewal Equation, First Form:
(10)
z (m ) = b (m ) +
m
X
f (k )z (m − k ).
k =1
Here f (k ) = f k is the interoccurrence time distribution for the renewal process. There is an
equivalent way of writing the Renewal Equation that is more suggestive of how it actually arises
in practice. Set z (m ) = b (m ) = 0 for m < 0; then the upper limit k = m − 1 in the sum in the
Renewal Equation may be changed to m = ∞ without affecting its value. The Renewal Equation
may now be written as follows, with X 1 representing the first interoccurrence time:
Renewal Equation, Second Form:
(11)
z (m ) = b (m ) + E z (m − X 1 ).
As you will come to appreciate, renewal equations crop up all over the place. In many circumstances, the sequence z (m ) is some scalar function of time whose behavior is of some interest;
the renewal equation is gotten by conditioning on the value of the first interoccurrence time. In
carrying out this conditioning, it is crucial to realize that the sequence S ∗1 ,S ∗2 , . . . defined by
S ∗n = S n − X 1 =
n
X
Xj
j =2
is itself a renewal process, independent of X 1 , and with the same interoccurrence time distribution f (x ).
2I know of at least four other proofs of a more probabilistic nature.
6
RENEWAL THEORY
Example 1: The Renewal Measure. The renewal measure is the sequence u (k ) = probability that
there is a renewal at time k . If k = 0, then u (k ) = 1, because in an ordinary renewal process there
is always a renewal at time 0. Suppose that k ≥ 1: then in order that there be a renewal at k , on
the event that X 1 = m , it must be the case that the renewal process S ∗1 ,S ∗2 , . . . has an occurrence
at time k − m . Thus, the renewal measure satisfies the Renewal Equation
u (k ) = δ0 (k ) + E u (k − S 1 ),
(12)
where δ0 (·) is the Kronecker delta function. Here is a formal proof:
u (k ) =
∞
X
P{S n = k }
n =0
= P{S 0 = k } +
= δ0 (k ) +
= δ0 (k ) +
= δ0 (k ) +
= δ0 (k ) +
∞
X
P{S n = k }
n =1
∞ X
∞
X
n =1 x =1
∞ X
∞
X
n =1 x =1
∞
X
P{X 1 = x and S ∗n = k − x }
P{X 1 = x }P{S ∗n = k − x }
f (x )
x =1
∞
X
∞
X
n =1
P{S ∗n = k − x }
f (x )u (k − x )
x =1
= δ0 (k ) + E u (k − X 1 )
The change in the order of summation on the fifth line is justified because all the terms of the
double sum are nonnegative.
Example 2: The Total Lifetime Distribution. Let L(m ) = A(m ) + R(m ) be the total lifetime of the
component in use at time m . (Here A(m ) and R(m ) are the age and residual lifetime random
variables, respectively.) Fix r ≥ 1, and set z (m ) = P{L(m ) = r }. Then z satisfies the Renewal
Equation (11) with
b (m ) = f (r ) for m = 0, 1, 2, . . . , r − 1
(13)
=0
for m ≥ r
EXERCISE: Derive this.
2.3. Solution of the Renewal Equation. Consider the Renewal Equation in its second form z (m ) =
b (m )+ E z (m −X 1 ) where by convention z (m ) = 0 for all negative values of m . Since the function
z (·) appears on the right side as well as on the left, it is possible to resubstitute on the right. This
leads to a sequence of equivalent equations:
z (m ) = b (m ) + E z (m − X 1 )
= b (m ) + E b (m − X 1 ) + E z (m − X 1 − X 2 )
= b (m ) + E b (m − X 1 ) + E b (m − X 1 − X 2 ) + E z (m − X 1 − X 2 − X 3 )
RENEWAL THEORY
7
and so on. After m iterations, there is no further change (because S m +1 > m and z (l ) = 0 for all
negative integers l ), and the right side no longer involves z . Thus, it is possible to solve for z in
terms of the sequences b and p :
z (m ) =
=
=
=
∞
X
E b (m − S n )
n =0
∞ X
∞
X
n =0 x =0
∞ X
∞
X
b (m − x )P{S n = x }
b (m − x )P{S n = x }
x =0 n =0
∞
X
b (m − x )u (x ).
x =0
Note that only finitely many terms in the series are nonzero, so the interchange of summations
is justified. Thus, the solution to the Renewal Equation is the convolution of the sequence b (m )
with the renewal measure:
(14)
z (m ) =
∞
X
b (m − x )u (x )
x =0
2.4. The Key Renewal Theorem. The formula (14) and the Feller-Erdös-Pollard theorem now
combine to give the asymptotic behavior (as m → ∞) of the solution z .
Key Renewal Theorem (Lattice Case) . Let z (m ) be the solution to the Renewal Equation (11). If
the sequence b (m ) is absolutely summable, then
(15)
lim z (m ) = µ−1
m →∞
∞
X
b (k ).
k =0
Proof. The formula (14) may be rewritten as
(16)
z (m ) =
∞
X
b (k )u (m − k )
k =0
For each fixed k , the sequence u (m − k ) → µ−1 as m → ∞, by the Feller-Erdös-Pollard theorem.
Thus, as m → ∞, the k th term of the series (16) converges to b (k )/µ. Moreover, because u (m −
k ) ≤ 1, the k th term is bounded in absolute value by |b (k )|. By hypothesis, this sequence is
summable, so the Dominated Convergence Theorem implies that the series converges as m → ∞
to the right side of (15).
Example 4: Limiting Residual Lifetime Distribution. For each fixed r ≥ 1, the sequence z (m ) =
P{R(m ) = r } satisfies the renewal equation
(17)
z (m ) = P{X 1 = m + r } +
m
X
k =1
z (m − k )P{X 1 = k } = f m +r +
m
X
k =1
z (m − k )f k .
8
RENEWAL THEORY
This reduces to (11) with b (m ) = f (m + r ). The sequence b (m ) is summable, because µ = E X 1 <
∞ (why?). Therefore, the Key Renewal Theorem implies that for each r = 1, 2, 3, . . . ,
(18)
lim P{R(m ) = r } = µ−1
m →∞
∞
X
f (k + r ) = µ−1 P{X 1 ≥ r }.
k =0
This could also be deduced from the convergence theorem for Markov chains, using the fact that
the sequence R m is an irreducible, positive recurrent Markov chain with stationary distribution
(7).
Example 5: Limiting Total Lifetime Distribution. Recall (Example 3 above) that the sequence
z (m ) = P{L(m ) = r } satisfies the Renewal Equation (11) with b (m ) defined by (13). Only finitely
many terms of the sequence b (m ) are
P nonzero, and so the summability hypothesis of the Key
Renewal Theorem is satisfied. Since k ≥0 b (m ) = r f (r ), it follows from (15) that
Corollary 1.
(19)
lim P{L(m ) = r } = r f (r )/µ.
m →∞
Example 6: A renewal equation in disguise. Discrete convolution equations arise in many parts
of probability and applied mathematics, but often with a “kernel” that isn’t a proper probability distribution. It is important to realize that such equations can be converted to (standard)
renewal equations by the device known as exponential tilting. Here is a simple example.
Consider the Fibonacci sequence 1,1,2,3,5,8,. . . . This is the sequence a n defined by the recursion
a n +2 = a n +1 + a n
(20)
and the initial conditions a 0 = a 1 = 1. To convert the recursion to a renewal equation, multiply
a n by a geometric sequence:
z n = θ −n a n
for some value θ > 0. Also, set z n = 0 for n ≤ −1. Then (20) is equivalent to the equation
(21)
z n = z n −1 θ 1 + z n−2 θ 2 + δ0,n
where δi ,j is the Kronecker delta function. (The delta function makes up for the fact that the
original Fibonacci recursion (20) does not by itself specify a 0 and a 1 .) Equation (21) is a renewal
equation for any value of θ > 0 such that θ 1 + θ2 = 1, because then we can set f 1 = θ 1 , f 2 = θ 2 ,
and f m = 0 for m ≥ 3. Is there such a value of θ ? Yes: it’s the golden ratio
p
−1 + 5
.
θ=
2
(The other root is negative, so we can’t use it to produce a renewal equation from (20).) The Key
Renewal Theorem implies that
lim z n = lim θ −n a n = 1/(θ + 2θ 2 ).
n →∞
n→∞
Thus, the Fibonacci sequence grows at an exponential rate, and the rate is the inverse of the
golden ratio.
RENEWAL THEORY
9
3. GENERATING FUNCTIONS AND RENEWAL THEORY
3.1. Generating Functions. The Renewal Theorem tells us how the age and residual lifetime
distributions behave at large times, and similarly the Key Renewal Theorem tells us how the
solution of a renewal equation behaves at infinity. In certain cases, exact calculations can be
done. These are usually done using generating functions.
Definition 1. The generating function of a sequence {a n }n ≥0 of real (or complex) numbers is the
function A(z ) defined by the power series
A(z ) =
(22)
∞
X
anzn.
n =0
Observe that for an arbitrary sequence {a n } the series (22) need not converge for all complex
values of the argumetn z . In fact, for some sequences the series (22) diverges for every z except
z = 0: this is the case, for instance, if a n = n n . But for many sequences of interest, there will
exist a positive number R such that, for all complex numbers z such that |z | < R, the series (22)
converges absolutely. In such cases, the generating function A(z ) is said to have positive radius of
convergence. The generating functions in all of the problems considered in these notes will have
positive radius of convergence. Notice that if the entries of the sequence a n are probabilities,
that is, if 0 ≤ a n ≤ 1 for all n, then the series (22) converges absolutely for all z such that |z | < 1.
If the generating function A(z ) has positive radius of convergence then, at least in principal,
all information about the sequence {a n } is encapsulated in the generating function A(z ). In
particular, each coefficient a n can be recovered from the function A(z ), since n!a n is the nth
derivative of A(z ) at z = 0. Other information may also be recovered from the generating function: for example, if the sequence {a n } is a discrete probability density, then its mean may be
obtained by evaluating A 0 (z ) at z = 1, and all of the higher moments may be recovered from the
higher derivatives of A(z ) at z = 1.
A crucially important property of generating functions is the multiplication law: The generating function of the convolution of two sequences is the product of their generating functions. This is the basis of most uses of generating functions in random walk theory, and all of
the examples considered below. The calculation that establishes this property is spelled out in
section ??. Note that for probability generating functions, this fact is a consequence of the multiplication law for expectations of independent random variables: If X and Y are independent,
nonnegative-integer valued random variables, then
E z X +Y = E z X E z Y .
(23)
3.2. The Renewal Equation.
Let { f k }k ≥1 be a probability distribution on the positive integers
P
with finite mean µ = k f k , and let X 1 , X 2 , . . . be a sequence of independent, identically distributed random variables with common discrete distribution { f k }. Define the renewal sequence
associated to the distribution { f k } to be the sequence
(24)
u m = P{S n = m for some n ≥ 0}
=
∞
X
P{S n = m }
n=0
where S n = X 1 + X 2 + · · · + X n (and S 0 = 0). note that u 0 = 1 A simple linear recursion for the
sequence u m , called the renewal equation, may be obtained by conditioning on the first step
10
RENEWAL THEORY
S1 = X 1:
(25)
um =
m
X
f k u m −k + δ0 (m )
k =1
where δ0 is the Kronecker delta (1 at 0; and 0 elsewhere).
The renewal equation is a particularly simple kind of recursive relation: the right side is just
the convolution of the sequences f k and u m . The appearance of a convolution should always
suggest the use of some kind of generating function or transform (Fourier or Laplace), because
these always convert convolution to multiplication. Let’s try it: define generating functions
∞
X
(26)
U (z ) =
um z m
and
(27)
F (z ) =
m =0
∞
X
fkzk.
k =1
Observe that if you multiply the renewal equation (25) by z m and sum over m then the left side
becomes U (z ), so
∞ X
m
X
(28)
U (z ) = 1 +
f k u m −k z m
=1+
m =0 k =1
∞ X
∞
X
f k z k u m −k z m −k
k =1 m =k
= 1 + F (z )U (z ).
Thus, we have a simple functional equation relating the generating functions U (z ) and F (z ). It
may be solved for U (z ) in terms of F (z ):
(29)
U (z ) =
1
1 − F (z )
3.3. Partial Fraction Decompositions. Formula (29) tells us how the generating function of the
renewal sequence is related to the probability generating function of the steps X j . Extracting useful information from this relation is, in general, a difficult analytical task. However, in the special
case where the probability distribution { f k } has finite support, the method of partial fraction decomposition provides an effective method for recovering the terms u m of the renewal sequence.
Observe that when the probability distribution { f k } has finite support, its generating function
F (z ) is a polynomial, and so in this case the generating function U (z ) is a rational function.3
The strategy behind the method of partial fraction decomposition rests on the fact that a simple pole may be expanded as a geometric series: in particular, for |z | < 1,
∞
X
−1
(30)
(1 − z ) =
zn.
n =0
Differentiating with respect to z repeatedly gives a formula for a pole of order k + 1:
∞ X
n n −k
−k −1
(31)
(1 − z )
=
z
.
k
n =k
3A rational function is the ratio of two polynomials.
RENEWAL THEORY
11
Suppose now that we could write the generating function U (z ) as a sum of poles C /(1−(z /ζ))k +1
(such a sum is called a partial fraction decomposition). Then each of the poles could be expanded
in a series of type (30) or (31), and so the coefficients of U (z ) could be obtained by adding the
corresponding coefficients in the series expansions for the poles.
Example: Consider the probability distribution f 1 = f 2 = 1/2. The generating function F is given
by F (z ) = (z + z 2 )/2. The problem is to obtain a partial fraction decomposition for (1 − F (z ))−1 .
To do this, observe that at every pole z = ζ the function 1 − F (z ) must take the value 0. Thus, we
look for potential poles at the zeros of the polynomial 1 − F (z ). In the case under consideration,
the polynomial is quadratic, with roots ζ1 = 1 and ζ2 = −2. Since each of these is a simple root
both poles should be simple; thus, we should try
C1
C2
1
=
+
.
1 − (z + z 2 )/2 1 − z 1 + (z /2)
The values of C 1 and C 2 can be gotten either by adding the fractions and seeing what works or by
differentiating both sides and seeing what happens at each of the two poles. The upshot is that
C 1 = 2/3 and C 2 = 1/3. Thus,
(32)
U (z ) =
1
2/3
1/3
=
+
.
1 − F (z ) 1 − z 1 + (z /2)
We can now read off the renewal sequence u m by expanding the two poles in geometric series:
(33)
um =
2 1
+ (−2)m .
3 3
There are several things worth noting. First, the renewal sequence u m has limit 2/3. This
equals 1/µ, where µ = 3/2 is the mean of the distribution { f k }. We should be reassured by this,
because it is what the Feller-Erdös-Pollard Renewal Theorem predicts the limit should be. Second, the remainder term (1/3)(−2)−m decays exponentially in m . This is always the case for
distributions { f k } with finite support. It is not always the case for arbitrary distributions { f k },
however.
Problem 1. Recall that the Fibonacci sequence 1, 1, 2, 3, 5, 8, . . . is the sequence a n such that a 1 =
a 2 = 1 and such that
a m +2 = a m + a m +1 .
(A) Find a functional equation for the generating function of the Fibonacci sequence. (B) Use the
method of partial fractions to deduce an exact formula for the terms of the Fibonacci sequence
in terms of the golden ratio.
3.4. Step Distributions with Finite Support. Assume now that the step distribution { f k } has finite support, is nontrivial (that is, does not assign probability 1 to a single point) and is nonlattice
(that is, it does not
P give probability 1 to a proper arithmetic progression). Then the generating
function F (z ) = f k z k is a polynomial of degree at least two. By the Fundamental Theorem of
Algebra, 1 − F (z ) may be written as a product of linear factors:
(34)
1 − F (z ) = C
K
Y
(1 − z /ζ j )
j =1
The coefficients ζ j in this expansion are the (possibly complex) roots of the polynomial equation
F (z ) = 1. Since the coefficients f k of F (z ) are real, the roots of F (z ) = 1 come in conjugate pairs;
thus, it is only necessary to find the roots in the upper half plane (that is, those with nonnegative
12
RENEWAL THEORY
imaginary part). In practice, it is usually necessary to solve for these roots numerically. The
following proposition states that none of the roots is inside the unit circle in the complex plane.
Lemma 2. If the step distribution { f k } is nontrivial, nonlattice, and has finite support, then the
polynomial 1−F (z ) has a simple root at ζ1 = 1, and all other roots ζ j satisfy the inequality |ζ j | > 1.
Proof. It is clear that ζ1 = 1 is a root, since F (z ) is a probability generating function. To see that
ζ1 = 1 is a simple root (that is, occurs only once in the product (34)), note that if it were a multiple
root then it would have to be a root of the derivative F 0 (z ) (since the factor (1 − z ) would occur
at least twice in the product (34)). If this were the case, then F 0 (1) = 0 would be the mean of the
probability distribution { f k }. But since this distribution has support {1, 2, 3 · · · }, its mean is at
least 1.
In order that ζ be a root of 1 − F (z ) it must be the case that F (ζ) = 1. Since F (z ) is a probability
generating function, this can only happen if |ζ| ≥ 1. Thus, to complete the proof we must show
that there are no roots of modulus one other than ζ = 1. Suppose, then, that ζ = e i θ is such that
F (ζ) = 1, equivalently,
X
f k e i θ k = 1.
Then for every k such that f k > 0 it must be that e i θ k = 1. This implies that θ is an integer
multiple of 2π/k , and that this is true for every k such that f k > 0. Since the distribution { f k }
is nonlattice, the greatest common divisor of the integers k such that f k > 0 is 1. Hence, θ is an
integer multiple of 2π, and so ζ = 1.
Corollary 3. If the step distribution { f k } is nontrivial, nonlattice, and has finite support, then
(35)
R
X
Cr
1
1
=
+
1 − F (z ) µ(1 − z ) r =1 (1 − z /ζr )k r
where µ is the mean of the distribution { f k } amd the poles ζr are all of modulus strictly greater
than 1.
Proof. The only thing that remains to be proved is that the simple pole at 1 has residue 1/µ. To
see this, multiply both sides of equation (35) by 1 − z :
X
Cr
1−z
= C + (1 − z )
.
1 − F (z )
(1 − z /ζr )k r
r
Now take the limit of both sides as z → 1−: the limit of the right side is clearly C , and the limit of
the left side is 1/µ, because µ is the derivative of F (z ) at z = 1. Hence, C = 1.
Corollary 4. If the step distribution { f k } is nontrivial, nonlattice, and has finite support, then
(36)
lim u m = 1/µ,
m →∞
and the remainder decays exponentially as m → ∞.
Remark. The last corollary is a special case of the Feller-Erdös-Pollard Renewal Theorem. This
theorem asserts that (9) is true under the weaker hypothesis that the step distribution is nontrivial and nonlattice. Various proofs of the Feller-Erdös-Pollard theorem are known, some of
which exploit the relation (29). Partial fraction decomposition does not work in the general case,
though.
RENEWAL THEORY
13
4. RENEWAL THEORY: NONLATTICE CASE
Renewal theory in the nonlattice case is technically more complicated, but the main results
are very similar to those in the lattice case once the obvious adjustments for continuous time
(e.g., changing sums to integrals) are made. Since many of the arguments are parallel to those in
the lattice case, we shall merely record the main results without proofs.
A nonlattice distribution is one which is not lattice, that is, does not attach all of its probability
to an arithmetic progression hZ. Probability distributions with continuous densities f (x ) for
x > 0 are nonlattice, but so are lots of other distributions that do not have densities, including
some discrete ones. An example that will turn out to be interesting is the discrete distribution
that assigns mass f k to the point 2−k for k = 0, 1, 2, 3, . . . , with f k > 0 for infinitely many values of
k (note that if f k > 0 for only finitely many k then the distribution would be lattice). Throughout
this section we assume that 0,S 1 ,S 2 , . . . is a renewal process with interoccurrence distribution F
satisfying
Assumption 1. The interoccurrence time distribution F is supported by the positive real numbers and is nonlattice.
R
Assumption 2. The interoccurrence time distribution has finite mean µ = x d F (x ).
Renewal Measure: In the nonlattice case, the renewal measure is defined by its cumulative mass
function
X
(37)
U (t ) = 1 + E N (t ) =
P{S n ≤ t }
n =0
Blackwell’s Theorem . For each positive real number h > 0,
(38)
lim (U (t + h) − U (t )) = h/µ.
t →∞
This is the natural analogue of the Feller-Erdös-Pollard Theorem for nonlattice renewal processes. As in the lattice case, the practical importance of the Blackwell Theorem stems from
its connection with a Key Renewal Theorem that describes the asymptotic behavior of solutions
to the Renewal Equation. The Renewal Equation in the nonlattice case is a convollution equation relating two (measurable) functions z (t ) and b (t ) that are defined for all real t , and satisfy
z (t ) = b (t ) = 0 for all t < 0.
Renewal Equation:
t
Z
(39)
z (t ) = b (t ) + E z (t − X 1 ) = b (t ) +
z (t − s ) d F (s ).
0
Technical Note: The notation d F denotes integration with respect to the interoccurrence
disR
tribution. If this distribution has a density f (t ) then d F (t ) = f (t )d t . Otherwise, d F must be
interpreted as a Lebesgue-Stieltjes integral. If you have never heard of these, don’t worry — all
the important examples in this course will either be purely discrete or will have a density.
Just as in the lattice case, the renewal equation has a unique solution z (t ) provided b (t ) is
bounded. There is an explicit formula for this:
14
RENEWAL THEORY
z (t ) =
(40)
∞
X
∞
Z
E b (t − S n ) =
b (t − s ) d U (s ).
s =0
n =0
The Key Renewal Theorem for the nonlattice case differs from the corresponding theorem for
the lattice case in an important technical respect. In the lattice case, it was necessary only to
assume that the sequence b (m ) be absolutely summable. Thus, it is natural to expect that in
the nonlattice case it should be enough to assume that the function b (t ) be L 1 , that is, absolutely
integrable. However, this isn’t so! The right condition turns out to be direct Riemann integrability,
which is defined as follows.
Definition 2. Let b (t ) be a real-valued function of t ∈ R such that b (t ) = 0 for all t < 0. Then b (t )
is said to be directly Riemann integrable if
Z
∞
∞
X
X
(41)
lim
a
inf
b (t ) = lim
a
sup
b (t ) = b (t ) d t .
a →0+
n=1
(n −1)a ≤t ≤na
a →0+
n =1
(n−1)a ≤t ≤na
Notice that the sums would be Riemann sums if not for the infinite limit of summation. If a
function is directly Riemann integrable, then it is Lebesgue integrable (if you don’t know what
this is, don’t worry about it), but the converse is not true. In practice, one rarely needs to deal
with the condition (41), because there are relatively simple sufficient conditons that guarantees
direct Riemann integrability:
Lemma R5. If b (t ) is nonnegative and nonincreasing on [0, ∞), and if the Lebesgue (or Riemann)
∞
integral 0 b (t ) d t < ∞, then b (t ) is directly Riemann integrable.
Lemma 6. If b (t ) has compact support and is Riemann integrable then it is directly Riemann
integrable.
The proofs are omitted but are not extremely difficult.
Key Renewal Theorem (Nonlattice Case) . Assume that the interoccurrence time distribution is
nonlattice and has finite mean µ > 0. If b (t ) is directly Riemann integrable then the solution z (t )
of the renewal equation (39) satisfies
Z∞
(42)
lim z (t ) = µ−1
t →∞
b (t ) d t .
0