Download Use of Compound Poisson Processes in Biological Modeling.

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Statistics wikipedia , lookup

History of statistics wikipedia , lookup

Probability wikipedia , lookup

Randomness wikipedia , lookup

Probability interpretations wikipedia , lookup

Transcript
Use of Compound Poisson Processes in Biological Modeling.
(Lecture notes prepared in GSI, winter 1998)
Ewa Gudowska-Nowak 1;2
1 GSI, Plankstr. 1, D-64291 Darmstadt, Germany
2 Institute of Physics, Jagiellonian University, 30-059 Krakow, Poland.
(30. November 2000)
A brief discussion of the use of Poisson distributions is presented starting
with the denition of random occurrences. Criteria for random occurrences to
be Poisson distributed are presented and followed by a discussion of Poisson
stochastic processes and their generalizations. Examples cover simple applications of the compound Poisson processes in biological modeling.
1
I. BASIC DEFINITIONS
Random events are those which can have more than one possible outcome. One can thus
associate a probability with each outcome. The outcome of a random event is not predicable,
only probabilities of possible outcomes may be specied. With a random event A one can
associate a random variable X which takes dierent possible numerical values x1 ; x2:::xn :::
depending on dierent outcomes of A. The distribution function of a random variable is
determined by assigning corrsponding probabilities P (x1); P (x2):::P (xn):::. Its derivative
P (x + x)
p(x) lim
x!1
x
(1.1)
is called a probability density function. Random variables X; Y are called statistically independent, if they joint probability density p(x; y) is completely factorizable, i.e. p(x; y) =
p(x)p(y).
A stochastic process [2{4] fx(t; !)g is composed by a family of random variables which are
indexed by time, i.e. for each time t, the random variable x(t) takes on value ! = xt with some probability. The most popular example of a stochastic process is a Brownian movement,
discovered by botanician R. Brown in 1827 [5]. Brown has observed under the miscroscope a
strong irregular motion of pollen particles on a surface of water. The trajectories of particles
in a \Brownian process" are irregular and a displacement of a Brownian particle at time t
is a probabilistic, random variable.
Given a random variable X with density p(x), the characteristic function is dened as
Z1
isX
p(x)eisxdx for X continuous
X (s) = E (e ) =
?1
X
X (s) = E (eisX ) = pk eisx
for X discrete
(1.2)
k
k
where E (Y ) stands for the expectation or mean value of a random variable Y . Function X (s)
dtermines completely the probability distribution of the random variable. In particular, if
P (X ) is continuous everywhere and dP (x) = p(x)dx, then
Z1
1
p(x) = 2
X (s)e?ixsds
(1.3)
?1
2
The characteristic function of the sum of independent random variables is the direct product
of the characteristic functions for each one of those variables. The corresponding probability
density function is given by a convolution
Z1
p(z) p(Z = X + Y ) =
(1.4)
pX (z ? s)pY (s)ds
?1
If it exists, the characteristic function X (s) generates the moments of the distribution of
a random variable X . Substituting for Z = eis in eq.(1.2), we get for the discrete random
variable X
G(Z ) = E (Z k ) =
1
X
k=0
pk Z k
(1.5)
where
pk 0;
1
X
k=0
pk = 1
(1.6)
Derivatives of G(Z ) evaluated at Z = 1 are related to the moments:
1
X
0
G (1) = kZ k?1p
G00 (1) =
1
X
k=0
k=0
k jZ =1
= E (k)
k(k ? 1)Z k?2pk jZ =1 = E [k(k ? 1)] = E (k2) ? E (k)
(1.7)
Example 1.
Poisson distribution gives the probability of nding exactly k events in a given length time,
if the events occur independently, at a constant rate. Let us place at random n points in
the time interval (0; T ) with P fk in ti g being the probability that k points will lie in the
interval ti included in (0; T ):
!
n
(1.8)
P fk in tig = k pk qn?k
where p = tT . For n ! 1, T ! 1 in such a way that
converges to a Poisson distribution:
i
P fk in tig = e?t (tki!)
i
3
k
n
T
! , the above distribution
(1.9)
Note, that the interval ti has to be small compared to T . Probability generating function
for the Poisson distribution is
G(Z ) = e(Z ?1)
(1.10)
from which direct calculations of moments gives E (k) = 2(k) = E (k2) ? E 2(k) = .
II. SUMS OF RANDOM NUMBER OF RANDOM VARIABLES
Consider now a sum SN of N independent random variables X
SN =
N
X
i=1
Xi
(2.11)
where N is a random variable with a probability generating function g(s)
g(s) =
1
X
i=0
gisi
(2.12)
Let us assume that each Xi has the same probability generating function f (s) (that means
that Xi0s are sampled from the same probability distribution function):
f (s) =
1
X
j =1
fj s j
(2.13)
By use of the Bayes rule of conditional probabilities
A and B g
ProbfA given B g P (AjB ) = ProbfProb
fB g
(2.14)
the probability that SN takes value j can be then written as
P (SN = j ) hj =
1
X
n=0
P (SN = j jN = n)P (N = n)
(2.15)
For xed value of n, using the statistical independence of Xi's, the sum SN has a probability
generating function being a direct product of f (s):
F (s) = f (s)n =
4
1
X
j =0
Fj sj
(2.16)
from which it follows that P (SN = j jN = n) = Fj . The formula (2.15) can be then rewritten
as
hj =
1
X
n=0
Fj gn
(2.17)
The compound probability generating function of SN is then
h(s) =
=
=
1
X
n=0
1
X
j =0
1 X
1
X
j =0 n=0
hj sj =
Fj gnsj =
gnf (s)n gf (s)
(2.18)
Example 2
A hen lays N eggs, where N has a Poisson distribution with mean . The weight of the nth
egg is Wn 2 f0; 1; 2:::g, where W1; W2 ::: are idependent and identically distributed random
variables with common probability generating function f . By virtue of the above analysis and
use of the formulae (1.10), (2.18), the generating function of the total weight W = PNi=1 Wi
is given by:
gW = exp(?(1 ? f (s)))
(2.19)
W is said to have a compound Poisson distribution. Note that the random observations Wi
can be sampled from any distribution; the crucial point in derivation of (2.18) is that the
number of elements in the their sum N constitutes a Poisson distributed random variable.
If Wi's are distributed according to a Poisson law
f (s) = exp(? + s)
(2.20)
the total weight W is a random variable with a compound Poisson-Poisson (Neyman type
A) distribution:
p(w) =
1 (N)r e?n N e?
X
r!
N!
N =0
5
(2.21)
The compound Poisson distribution (CPD) has a wide application in ecology, nuclear
chain reactions and queing theory. It is sometimes known as the distribution of a \branching process" [3] and as such has been also commonly used to describe radiobiological
eects in cells.
Example 3
In their cluster theory of the eects of ionizing radiation, Tobias et al. [15] have used so
called Neyman [7] distribution (see above, Example 2) which is nothing but a compound
Poisson-binomial (or in a limitting case Poisson-Poisson) distribution. In the derivation the
following reasoning has been used:
when a single heavy ion crosses a cell nucleus, it may produce DNA strand breaks and
chromatin scissions wherever the ionizing track structure overlaps chromatin structure
the multiple yield of such lesions depends on the radial distribution of deposited energy
and on the microdistribution of DNA in the cell nucleus
the number of crossings strongly depends on the geometry of DNA coiling in the cell
nucleus (in a human cell nucleus, the total length of doubled-stranded DNA is more
than one meter and in the nucleus the DNA is packed in coiled strands). For a given
cell line, a \typical" average number n of possible crossings per a particle is assumed.
if p is a probability that a chromatin break occurs at each particle crossing (and q is
the probability that it does not), the distribution of the number of chromatin breaks
in the cluster per one-particle traversal is binomial:
!
n
(2.22)
P (ijn) = i piq(n?i)
with the probability generating function
gs = [sp + (1 ? p)]n
6
(2.23)
the probability that j particles cross the nucleus is given by a Poisson distribution
P = (Fj ! ) e?F
j
(2.24)
with the probability generating function
hs = exp(?F + Fs)
(2.25)
the overall probability that i lesions will be observed after m particles traversed the
nucleus is given by a Neyman distribution
P(ij; F; n) =
1 (nm)!pi q (nm?i) (F )m e?F
X
i!(nm ? i)!m!
m=1
(2.26)
with a compound probability generating function
G(s) = exp(? + [sp + 1 ? p]nm)
(2.27)
where = F .
From the latter, by direct dierentiation one gets expected (mean) value and variance
< i >= np
< i2 > ? < i >2 = n(n ? 1)p2 + np
(2.28)
which explains overdispersion of the probability density function eq.(2.26). By assuming a
repairless cell line, we are able to derive the surviving fraction as a zero class of the initial
distribution, i.e. the proportion of cells with no breaks.
P(0j; F; n) =
1 (nm)!q nm (F )m e?F
X
=
(nm)!m!
m=1
exp[?F (1 ? qn)]
(2.29)
(2.30)
Note the dierence between Neyman and Poisson distribution, for which
P(0j; F; n) = exp[?F ] = exp[? < i >]
7
(2.31)
III. MARKOV PROPERTY
Stochastic processes and Markov processes, in particular, serve as a powerfull tool
to describe and understand biological phenomena at various levels of complexity-from the
molecular to the ecological level. The modeling of biological systems via stochastic processes
allows to incorporate the eects of secondary factors for which the detailed knowledge is
missing. The technique has been widely used to model population growth and extinction
[1,8], population genetics [9,10], chemical kinetics [2,14], ring of neurons [11,12], opening
and closing of biological channels [13] or cell survival after irradiation [15,16]. In what follows. we will briey recall denition of Markovianity.
One says that a real stochastic process is statistically determined if one knows its nth
order or n-point distribution function [2]
P (x1 ; t1; x2 ; t2; x3 ; t3:::xn?1 ; tn?1; xn; tn)
(3.32)
for any n and t, where P (x1; t1 ; x2; t2 ; x3; t3 ; :::; xn?1; tn?1; xn; tn) stands for the probability
that the process fx(t)g is in the state xn (takes the value xn) at time tn and in the state
xn?1 at time tn?1... and in the state x1 at time t1 . These functions are not arbitrary but
they must satisfy certain conditions. A distribution of a given order is determined from a
distribution of lower order by use of the Bayes rule for conditioned probabilities:
P (x1; t1; x2 ; t2; x3 ; t3:::xn?1 ; tn?1; xn; tn) =
P (xn; tnjxn?1; tn?1; :::x1 ; t1):::P (x2 ; t2jx1 ; t1)P (x1; t1)
(3.33)
n?1 ; tn?1 ; xn ; tn )
P (xn; tnjxn?1; tn?1; :::x1 ; t1 ) = P (xP1;(tx1;;xt2 ;; tx2;;xt3 ;; tx3:::x
; t :::x ; t
(3.34)
where
1 1 2 2 3 3
n?1 n?1)
denes the conditioned probability that the process takes on value xn at time tn provided the
sequence of events fxn?1; tn?1; :::x2 ; t2; x1 ; t1g took place at earlier times. A Markov process
8
is a stochastic process fx(t)g which can be fully characterized by a conditioned probability
and a one-point probability functions [2].
The basic denition of Markovianity of the process can be expressed as
P (xn; tnjxn?1; tn?1 ; :::x1; t1 ) = P (xn; tnjxn?1; tn?1)
(3.35)
This criterion is to be complemented by the so-called Smoluchowski-Chapman-Kolmogorov
(SCK) equation:
Z1
P (xn; tnjxk ; tk ) = P (xn; tnjxm ; tm )P (xm; tmjxk ; tk )dxm; n > m > k
(3.36)
1
which follows directly from the denitions eqs.(3.34) and (3.35). Therefore the process which
does not satisfy either the basic denition eq.(3.35) or the SCK equation (3.36)is not Markovian. A non-Markovian process may satisfy one of these relations but both are necessary
conditions of Markovianity (i.e. neither is a sucient one).
9
IV. RANDOM OCCURRENCES
An important example of a discontinuous Markov process is the statistics of the
number of occurrences of an event in time interval (0; t). If we denote that number by x(t),
then fx(t)g is a discontinuous stochastic process taking integer values 0; 1; 2:::n; :::. If the
number of occurrences in an interval (t0; t), assuming x(t0 ) is independent of the past of
x(t), then the process is Markov, according to the above denition. Since x(t) can increase
only, hence
Pij (t; t0 ) P fx(t) = ijx(t0 ) = j g = 0
for j > i
(4.37)
Consider now a small interval t. Neglecting probabilities of order (t)2 and higher, we assume that in the interval we can have only one occurrence and its probability is proportional
to t, namely
P fx(t + t) = ijx(t) = ig 1 ? qi(t)t
P fx(t + t) = i + 1jx(t) = ig = qi(t)t
(4.38)
If qi(t) = (t) is independent of i, we say that the occurrences form random points in
time and the resulting fx(t)g process is Poisson. To see it, let as assume for simplicity 1
that (t) = =constant. Let Pn(t) Pn;0(t) denotes the probability that we have n points
in the interval (0; t), that is, x(t) = n, provided we have started with zero points at time 0,
x(0) = 0. The SCK equation (3.36) can be now written in the form
Pn;0(t + t) = tPn?1;0 (t) + (1 ? t)Pn;0 (t)
(4.39)
from which the following dierential equation results
d P (t) = lim Pn(t + t) ? Pn(t) =
t!0
dt n
t
Pn?1;0(t) ? Pn;0(t)
(4.40)
1 If (t) is not a constant, the process is nonhomogeneous in time but can be transformed to a
R
homogeneous process by a linear time transformation = 0 (s)ds
10
For n = 0
d P = ?P (t)
0
dt 0;0
(4.41)
with the initial condition that P0(0) = 1 and Pn(0) = 0. After using the Laplace transform,
the solution to these equation can be easily found
P~n;0(s) = s + P~n?1;0(s)
P~0(s) = s +1 (4.42)
which leads to
Pn
(t) = e?t (t)
n
(4.43)
n
So far, we have discussed the points random in time, but the same arguments apply to
characterize points randomly distributed on a line. We shall now derive the formula [4] for
the probability distribution function for the distance between n random points on a line.
Example 4
Consider a xed point t0 and its two closest \neigbouring"points to the left and right,
respectively (cf. Figure 1):
z ? y1 ?! x1 !
t?1
t0
t1
Figure 1
Given x1 > 0, the distribution function P (x1) of the random variable X1 representing
the distance between the points t1 and t0 equals the probability that x1 is less than x, and
this is the probability that at least one point has been found in the interval (t0 ; t0 + x):
P (x1) = P fn(t0 ; t0 + x) 1g = 1 ? P fn(t0; t0 + x) = 0g = 1 ? e?x
11
(4.44)
The probability density distribution function for x follows by taking derivative of the above
expression:
p(x) = e?x(x)
(4.45)
where stands for the Heaviside step function:
(x) = 1
for x 0
(x) = 0
otherwise
(4.46)
We can now form a random variable z being the sum of x1 and y1 which represents the
distance from t?1 to t1 . Since both random variables x1 and y1 are independent, the probability density function for their sum equals the convolution of their respective densities
[2{4]:
Z1
Z1
p(z) =
p(z ? y)p(y)dy =
p(z ? x)p(x)dx
(4.47)
?1
?1
Since
p(x) = e?x(x)
p(y) = e?y (y)
and
(4.48)
We obtain for z
Zz
2
p(z) = e?(x?y) e?y dy = 2ze?z
0
(4.49)
In a similar way the density distribution function for n independent variables having exponential distributions with the same parameter can be derived leading to the gamma
density distribution function.
p(z) = N e?z zn?1 n(z)
(4.50)
with a normalisation factor N
N = ?(n)
n
12
(4.51)
The gamma distribution is a basic statistical tool for describing variables bounded at one
side (in our example, x1 ; y1; z 2 [0; 1)).
Let us analyze further a problem of cutting the interval [0; 1) into k fragments by
placing on it randomly n breaking points. The fragments are supposewd to be greater than a
minimum detectable length .Our task is to derive the probability density function p(k; jn).
The problem posed in this way is an adequate model of a chromatin break model discussed
by Schmidt [17].
Example 5.
Seeking the distribution of the number of chromatin fragments detected given the number
of chromatin breaks produced on an arbitrary chromosome. The total number of chromatin
breaks produced in a cell nucleus after irradiation is a random number equal to the sum of
the breaks produced on each chromosome by a random number of particles passing through
the individual chromosomes.
The probability density function for the length xi of the ith fragment is assumed in the
exponential form (note that a mean spacing between the random points have been set up
to unity)
p(xi ) = e?x
i
(4.52)
so that the density of the sum of all n + 1 fragments is
(4.53)
p(z) = ?(n1+ 1) e?z zn
The probability that exactly k of those fragments is greater than some value y is given by
!
n
+
1
p(k; yjn + 1) = k e?ky (1 ? e?y )n+1?k =
Z
p(k; yjn + 1; z)p(z)dz
(4.54)
Let us assume further that z = 1, so that the length y is expressed in units of z, z=y = a
!
Z1
n
+
1
?
ky
?
y
n
+1
?
k
= p(k; a1 jn + 1; ya) ?(n1+ 1) e?ya (ya)nyda (4.55)
p(k; yjn + 1) = k e (1 ? e )
0
13
The binomial coecient on the right hand side of the above equation can be expanded
leading to the identity
!
! X?k
n
+
1
?
k
n + 1 n+1
m
?
(
m
+
k
)
y
=
m
k m=0 (?1) e
Z1
(4.56)
= p(k; 1 jn + 1; ya)??1(n + 1)e?ya (ya)nyda
a
0
By rearranging the terms, the above equation can be rewritten as a Laplace transform of
the function p(k; a1 jn + 1; ya)an:
L[p(k; a1 jn + 1; ya)an] =
!
! n+1
X?k
n
+
1
?
k
n
+
1
?(
n
+
1)
m
?
(
m
+
k
)
y
=
=
m
(y)n+1
k m=0 (?1) e
!
n+1
X?k
n
+
1
?
k
m
L
[(a ? (m + k))n(a ? (m + k))]
(?1)
=
m
m=0
(4.57)
where we have used
L[(a ? (m + k))n(a ? (m + k))] = ?(ynn++1 1) e?(m+k)
From eq.(4.57) the form of the desired probability density function follows:
! n+1
X?k
1
n
+
1
p(k; a jn + 1; z = 1) = k
(?1)m [1 ? a1 (m + k)]n(1 ? m a+ k )
m=0
(4.58)
(4.59)
The density eq.(4.59) refers then to the situation when k random fragments greater than
the minimum \detectable" length a1 are produced by placing randomly n breaks on the line
of a unit length.
14
Probability density
50
40
30
20
10
0
0
200
400
600
800
1000
600
800
1000
x
Probability density
80
60
40
20
0
0
200
400
x
15
REFERENCES
[1] N. Goel and N. Richter-Dyn, Stochastic Processes in Biology, (Academic Press, New York,
1974).
[2] N.G. Van Kampen, Stochastic Processes in Physics and Chemistry, (North Holland, Amsterdam, 1981).
[3] S. Karlin and H. Taylor, First Course in Stochastic Processes, (Academic Press, New York,
1976).
[4] A. Papoulis, Probability, Random Variables and Stochastic Processes, (McGraw-Hill, Tokyo,
1981).
[5] R. Brown, Philos. Mag. 4 (1828) 161.
[6] C.A. Tobias, E. Goodwin and E. Blakely, in Quantitative Mathematical Models in Radiation
Biology, J. Kiefer, ed., Springer Verlag, Berlin 1988, p.135.
[7] J. Neyman, Am. Math. Stat. 10 (1939) 35.
[8] D.R. Nelson and N.M. Shnerb, Phys. Rev. E58 1998 1384.
[9] T. Maruyama, Mathematical Modeling in Genetics, (Springer Verlag, Berlin, 1981).
[10] W. Horsthemke and R. Lefever, Noise-induced transitions. Theory and applications in physics,
chemistry and biology, (Springer Verlag, Berlin, 1984).
[11] J.J. Hopeld, Proc. Natl. Acad. Sci. USA 79 (1982) 2554.
[12] A. Crisanti and H. Sompolinsky, Phys. Rev. A36 (1987) 4922.
[13] B. Hille, Ionic channels of excitable membranes, (Sinauer Inc., Sunderland, MA, 1992).
[14] S. Larsson, Bioch. et Biophys. Acta 1365 (1998) 294.
[15] C.A. Tobias, Radiat. Res. 104 (1985) S77.
[16] N. Albright, Radiat. Res. 118 (1989) 1.
16
[17] J. Schmidt, Heavy ion induced lesions in DNA, Ph.D. thesis, University of California at Berkeley, 1993.
17