Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Lecture Notes 6 Random Processes • Definition and Simple Examples • Important Classes of Random Processes ◦ ◦ ◦ ◦ ◦ IID Random Walk Process Markov Processes Independent Increment Processes Counting processes and Poisson Process • Mean and Autocorrelation Function • Gaussian Random Processes ◦ Gauss–Markov Process EE 278B: Random Processes 6–1 Random Process • A random process (RP) (or stochastic process) is an infinite indexed collection of random variables {X(t) : t ∈ T }, defined over a common probability space • The index parameter t is typically time, but can also be a spatial dimension • Random processes are used to model random experiments that evolve in time: ◦ Received sequence/waveform at the output of a communication channel ◦ Packet arrival times at a node in a communication network ◦ Thermal noise in a resistor ◦ Scores of an NBA team in consecutive games ◦ Daily price of a stock ◦ Winnings or losses of a gambler EE 278B: Random Processes 6–2 Questions Involving Random Processes • Dependencies of the random variables of the process ◦ How do future received values depend on past received values? ◦ How do future prices of a stock depend on its past values? • Long term averages ◦ What is the proportion of time a queue is empty? ◦ What is the average noise power at the output of a circuit? • Extreme or boundary events ◦ What is the probability that a link in a communication network is congested? ◦ What is the probability that the maximum power in a power distribution line is exceeded? ◦ What is the probability that a gambler will lose all his captial? • Estimation/detection of a signal from a noisy waveform EE 278B: Random Processes 6–3 Two Ways to View a Random Process • A random process can be viewed as a function X(t, ω) of two variables, time t ∈ T and the outcome of the underlying random experiment ω ∈ Ω ◦ For fixed t, X(t, ω) is a random variable over Ω ◦ For fixed ω, X(t, ω) is a deterministic function of t, called a sample function X(t, w1 ) t X(t, w2 ) t X(t, w3 ) t1 t2 X(t1 , w) X(t2 , w) EE 278B: Random Processes t 6–4 Discrete Time Random Process • A random process is said to be discrete time if T is a countably infinite set, e.g., ◦ N = {0, 1, 2, . . .} ◦ Z = {. . . , −2, −1, 0, +1, +2, . . .} • In this case the process is denoted by Xn , for n ∈ N , a countably infinite set, and is simply an infinite sequence of random variables • A sample function for a discrete time process is called a sample sequence or sample path • A discrete-time process can comprise discrete, continuous, or mixed r.v.s EE 278B: Random Processes 6–5 Example • Let Z ∼ U[0, 1], and define the discrete time process Xn = Z n for n ≥ 1 • Sample paths: xn Z= 1 2 1 4 1 2 1 8 1 16 n xn Z= 1 4 1 4 1 16 1 64 n xn Z=0 0 1 EE 278B: Random Processes 0 2 0 ... 3 4 5 6 7 ... n 6–6 • First-order pdf of the process: For each n, Xn = Z n is a r.v.; the sequence of pdfs of Xn is called the first-order pdf of the process xn 1 z 0 1 Since Xn is a differentiable function of the continuous r.v. Z , we can find its pdf as 1 1 −1 1 = fXn (x) = xn , 0 ≤ x ≤ 1 nx(n−1)/n n EE 278B: Random Processes 6–7 Continuous Time Random Process • A random process is continuous time if T is a continuous set • Example: Sinusoidal Signal with Random Phase X(t) = α cos(ωt + Θ) , t≥0 where Θ ∼ U[0, 2π] and α and ω are constants • Sample functions: x(t) α Θ=0 π 2ω π ω 3π 2ω 2π ω t x(t) Θ= π 4 t x(t) Θ= EE 278B: Random Processes π 2 t 6–8 • The first-order pdf of the process is the pdf of X(t) = α cos(ωt + Θ). In an earlier homework exercise, we found it to be 1 p , −α < x < +α fX(t)(x) = απ 1 − (x/α)2 The graph of the pdf is shown below fX(t)(x) 1 απ x −α 0 +α Note that the pdf is independent of t. (The process is stationary ) EE 278B: Random Processes 6–9 Specifying a Random Process • In the above examples we specified the random process by describing the set of sample functions (sequences, paths) and explicitly providing a probability measure over the set of events (subsets of sample functions) • This way of specifying a random process has very limited applicability, and is suited only for very simple processes • A random process is typically specified (directly or indirectly) by specifying all its n-th order cdfs (pdfs, pmfs), i.e., the joint cdf (pdf, pmf) of the samples X(t1), X(t2), . . . , X(tn) for every order n and for every set of n points t1, t2, . . . , tn ∈ T • The following examples of important random processes will be specified (directly or indirectly) in this manner EE 278B: Random Processes 6 – 10 Important Classes of Random Processes • IID process: {Xn : n ∈ N } is an IID process if the r.v.s Xn are i.i.d. Examples: ◦ Bernoulli process: X1, X2, . . . , Xn, . . . i.i.d. ∼ Bern(p) ◦ Discrete-time white Gaussian noise (WGN): X1, . . . , Xn, . . . i.i.d. ∼ N (0, N ) • Here we specified the n-th order pmfs (pdfs) of the processes by specifying the first-order pmf (pdf) and stating that the r.v.s are independent • It would be quite difficult to provide the specifications for an IID process by specifying the probability measure over the subsets of the sample space EE 278B: Random Processes 6 – 11 The Random Walk Process • Let Z1, Z2, . . . , Zn, . . . be i.i.d., where ( +1 with probability Zn = −1 with probability 1 2 1 2 • The random walk process is defined by X0 = 0 Xn = n X Zi , n≥1 i=1 • Again this process is specified by (indirectly) specifying all n-th order pmfs • Sample path: The sample path for a random walk is a sequence of integers, e.g., 0, +1, 0, −1, −2, −3, −4, . . . or 0, +1, +2, +3, +4, +3, +4, +3, +4, . . . EE 278B: Random Processes 6 – 12 Example: xn 3 2 1 1 2 3 4 5 1 1 1 −1 1 6 7 8 9 10 n −1 −2 −3 zn : −1 −1 −1 −1 −1 • First-order pmf: The first-order pmf is P{Xn = k} as a function of n. Note that k ∈ {−n, −(n − 2), . . . , −2, 0, +2, . . . , +(n − 2), +n} for n even k ∈ {−n, −(n − 2), . . . , −1, +1, +3, . . . , +(n − 2), +n} for n odd Hence, P{Xn = k} = 0 if n + k is odd, or if k < −n or k > n EE 278B: Random Processes 6 – 13 Now for n + k even, let a be the number of +1’s in n steps, then the number of −1’s is n − a, and we find that n+k k = a − (n − a) = 2a − n ⇒ a = 2 Thus P{Xn = k} = P{ 12 (n + k) heads in n independent coin tosses} n = n+k · 2−n for n + k even and −n ≤ k ≤ n 2 P{Xn = k}, n even . . . −6 −4 −2 EE 278B: Random Processes 0 +2 +4 +6 ... k 6 – 14 Markov Processes • A discrete-time random process Xn is said to be a Markov process if the process future and past are conditionally independent given its present value • Mathematically this can be rephrased in several ways. For example, if the r.v.s {Xn : n ≥ 1} are discrete, then the process is Markov iff pXn+1|Xn (xn+1|xn, xn−1) = pXn+1|Xn (xn+1|xn) for every n • IID processes are Markov • The random walk process is Markov. To see this consider P{Xn+1 = xn+1 | Xn = xn} = P{Xn + Zn+1 = xn+1 | Xn = xn} = P{Xn + Zn+1 = xn+1 | Xn = xn} = P{Xn+1 = xn+1 | Xn = xn} EE 278B: Random Processes 6 – 15 Independent Increment Processes • A discrete-time random process {Xn : n ≥ 0} is said to be independent increment if the increment random variables Xn1 , Xn2 − Xn1 , . . . , Xnk − Xnk−1 are independent for all sequences of indices such that n1 < n2 < · · · < nk • Example: Random walk is an independent increment process because X n1 = n1 X Zi , Xn2 −Xn1 = i=1 n2 X Zi , . . . , Xnk −Xnk−1 = i=n1+1 nk X Zi i=nk−1+1 are independent because they are functions of independent random vectors • The independent increment property makes it easy to find the n-th order pmfs of a random walk process from knowledge only of the first-order pmf EE 278B: Random Processes 6 – 16 • Example: Find P{X5 = 3, X10 = 6, X20 = 10} for random walk process {Xn} Solution: We use the independent increment property as follows P{X5 = 3, X10 = 6, X20 = 10} = P{X5 = 3, X10 − X5 = 3, X20 − X10 = 4} = P{X5 = 3}P{X5 = 3}P{X10 = 4} 5 −5 5 −5 10 −10 = 2 2 2 = 3000 · 2−20 4 4 7 • In general if a process is independent increment, then it is also Markov. To see this let Xn be an independent increment process and define ∆Xn = [X1, X2 − X1, . . . , Xn − Xn−1]T Then pXn+1|Xn (xn+1 | xn) = P{Xn+1 = xn+1 | Xn = xn} = P{Xn+1 − Xn + Xn = xn+1 | ∆Xn = ∆xn, Xn = xn} = P{Xn+1 = xn+1 | Xn = xn} • The converse is not necessarily true, e.g., IID processes are Markov but not independent increment EE 278B: Random Processes 6 – 17 • The independent increment property can be extended to continuous-time processes: A process X(t), t ≥ 0, is said to be independent increment if X(t1), X(t2) − X(t1), . . . , X(tk ) − X(tk−1) are independent for every 0 ≤ t1 < t2 < . . . < tk and every k ≥ 2 • Markovity can also be extended to continuous-time processes: A process X(t) is said to be Markov if X(tk+1) and (X(t1), . . . , X(tk−1)) are conditionally independent given X(tk ) for every 0 ≤ t1 < t2 < . . . < tk < tk+1 and every k ≥ 3 EE 278B: Random Processes 6 – 18 Counting Processes and Poisson Process • A continuous-time random process N (t), t ≥ 0, is said to be a counting process if N (0) = 0 and N (t) = n, n ∈ {0, 1, 2, . . .}, is the number of events from 0 to t (hence N (t2) ≥ N (t1) for every t2 > t1 ≥ 0) Sample path of a counting process: n(t) 5 4 3 2 1 0 t2 t3 t4 t5 0 t1 t1, t2, . . . are the arrival times or the wait times of the events t t1, t2 − t1, . . . are the interarrival times of the events EE 278B: Random Processes 6 – 19 The events may be: ◦ Photon arrivals at an optical detector ◦ Packet arrivals at a router ◦ Student arrivals at a class • The Poisson process is a counting process in which the events are “independent of each other” • More precisely, N (t) is a Poisson process with rate (intensity) λ > 0 if: ◦ N (0) = 0 ◦ N (t) is independent increment ◦ (N (t2) − N (t1)) ∼ Poisson(λ(t2 − t1)) for all t2 > t1 ≥ 0 • To find the kth order pmf, we use the independent increment property P{N (t1) = n1, N (t2) = n2, . . . , N (tk ) = nk } = P{N (t1 ) = n1, N (t2) − N (t1) = n2 − n1, . . . , N (tk ) − N (tk−1) = nk − nk−1} = pN (t1)(n1)pN (t2)−N (t1)(n2 − n1) . . . pN (tk )−N (tk−1)(nk − nk−1) EE 278B: Random Processes 6 – 20 • Example: Packets arrive at a router according to a Poisson process N (t) with rate λ. Assume the service time for each packet T ∼ Exp(β) is independent of N (t) and of each other. What is the probability that k packets arrive during a service time? • Merging : The sum of independent Poisson process is Poisson. This is a consequence of the infinite divisibility of the Poisson r.v. • Branching : Let N (t) be a Poisson process with rate λ. We split N (t) into two counting subprocesses N1(t) and N2(t) such that N (t) = N1(t) + N2(t) as follows: Each event is randomly and independently assigned to process N1(t) with probability p, otherwise it is assigned to N2(t) Then N1(t) is a Poisson process with rate pλ and N2(t) is a Poisson process with rate (1 − p)λ This can be generalized to splitting a Poisson process into more than two processes EE 278B: Random Processes 6 – 21 Related Processes • Arrival time process: Let N (t) be Poisson with rate λ. The arrival time process Tn , n ≥ 0 is a discrete time process such that: ◦ T0 = 0 ◦ Tn is the arrival time of the nth event of N (t) • Interarrival time process: Let N (t) be a Poisson process with rate λ. The interarrival time process is Xn = Tn − Tn−1 for n ≥ 1 • Xn is an IID process with Xn ∼ Exp(λ) Pn • Tn = i=1 Xi is an independent increment process with Tn2 − Tn1 ∼ Gamma(λ, n2 − n1) for n2 > n1 ≥ 1, i.e., fTn2 −Tn1 (t) = λn2−n1 tn2−n1−1 −λt e (n2 − n1 − 1)! • Example: Let N1 (t) and N2(t) be two independent Poisson processes with rates λ1 and λ2 , respectively. What is the probability that N1(t) = 1 before N2(t) = 1? EE 278B: Random Processes 6 – 22 • Random telegraph process: A random telegraph process Y (t), t ≥ 0, assumes values of +1 and −1 with Y (0) = +1 with probability 1/2 and −1 with probability 1/2, and Y (t) changes polarities with each event of a Poisson process with rate λ > 0 Sample path: y(t) +1 0 −1 x1 x2 x3 x4 x5 t EE 278B: Random Processes 6 – 23 Mean and Autocorrelation Functions • For a random vector X the first and second order moments are ◦ mean µ = E(X) ◦ correlation matrix RX = E(XXT ) • For a random process X(t) the first and second order moments are ◦ mean function: µX (t) = E(X(t)) for t ∈ T ◦ autocorrelation function: RX (t1, t2) = E X(t1)X(t2) for t1, t2 ∈ T • The autocovariance function of a random process is defined as CX (t1, t2) = E X(t1) − E(X(t1)) X(t2) − E(X(t2)) The autocovariance function can be expressed using the mean and autocorrelation functions as CX (t1, t2) = RX (t1, t2) − µX (t1)µX (t2) EE 278B: Random Processes 6 – 24 Examples • IID process: µX (n) = E(X1) RX (n1, n2) = E(Xn1 Xn2 ) = ( E(X12) (E(X1))2 n1 = n2 n1 6= n2 • Random phase signal process: Z 2π α cos(ωt + θ) dθ = 0 µX (t) = E(α cos(ωt + Θ)) = 0 2π RX (t1, t2) = E X(t1)X(t2) Z 2π 2 α cos(ωt1 + θ) cos(ωt2 + θ) dθ = 2π 0 = = Z 2π 0 α2 cos(ω(t1 + t2) + 2θ) + cos(ω(t1 − t2)) dθ 4π α2 cos(ω(t1 − t2)) 2 EE 278B: Random Processes 6 – 25 • Random walk: µX (n) = E n X i=1 Zi = RX (n1, n2) = E(Xn1 Xn2 ) n X 0 = 0 i=1 = E [Xn1 (Xn2 − Xn1 + Xn1 )] = E(Xn21 ) = n1 assuming n2 ≥ n1 = min{n1 , n2} in general • Poisson process: µN (t) = λt RN (t1, t2) = E(N (t1)N (t2)) = E [N (t1)(N (t2) − N (t1) + N (t1))] = λt1 × λ(t2 − t1) + λt1 + λ2 t21 = λt1 + λ2t1t2 assuming t2 ≥ t1 = λ min{t1, t2} + λ2t1t2 EE 278B: Random Processes 6 – 26 Gaussian Random Processes • A Gaussian random process (GRP) is a random process X(t) such that [X(t1), X(t2), . . . , X(tn) ]T is a GRV for all t1, t2, . . . , tn ∈ T • Since the joint pdf for a GRV is specified by its mean and covariance matrix, a GRP is specified by its mean µX (t) and autocorrelation RX (t1, t2) functions • Example: The discrete time WGN process is a GRP EE 278B: Random Processes 6 – 27 Gauss-Markov Process • Let Zn , n ≥ 1, be a WGN process, i.e., an IID process with Z1 ∼ N (0, N ) The Gauss-Markov process is a first-order autoregressive process defined by X1 = Z 1 Xn = αXn−1 + Zn , n > 1, where |α| < 1 • This process is a GRP, since X1 = Z1 and Xk = αXk−1 + Zk where Z1, Z2, . . . are i.i.d. N (0, N ), X1 1 0 ··· 0 0 Z1 X2 α Z2 1 · · · 0 0 . .. .. . . . .. X3 = . Z3 . n−2 . α αn−3 · · · 1 0 .. Xn αn−1 αn−2 · · · α 1 Zn is a linear transformation of a GRV and is therefore a GRV • Clearly, the Gauss-Markov process is Markov. It is not, however, an independent increment process EE 278B: Random Processes 6 – 28 • Mean and covariance functions: µX (n) = E(Xn) = E(αXn−1 + Zn) = α E(Xn−1) + E(Zn) = α E(Xn−1) = αn−1 E(Z1) = 0 To find the autocorrelation function, for n2 > n1 we write X n2 = α n2 −n1 X n1 + n2 −n X1−1 αiZn2−i i=0 Thus since Xn1 RX (n1, n2) = E(Xn1 Xn2 ) = αn2−n1 E(Xn21 ) + 0 , and Zn2−i are independent, zero mean for 0 ≤ i ≤ n2 − n1 − 1 Next, to find E(Xn21 ), consider E(X12) = N E(Xn21 ) = E (αXn1−1 + Zn1 )2 = α2 E(Xn21−1 ) + N Thus E(Xn21 ) 1 − α2n1 = N 1 − α2 EE 278B: Random Processes 6 – 29 Finally the autocorrelation function is RX (n1, n2) = α|n2−n1| 1 − α2 min{n1,n2} N 1 − α2 • Estimation of Gauss-Markov process: Suppose we observe a noisy version of the Gauss-Markov process, Yn = X n + W n , where Wn is a WGN process independent of Zn with average power Q We can use the Kalman filter from Lecture Notes 4 to estimate Xi+1 from Y i as follows: Initialization: X̂1|0 = 0 2 σ1|0 =N EE 278B: Random Processes 6 – 30 Update: For i = 2, 3, . . . , X̂i+1|i = αX̂i|i−1 + ki(Yi − X̂i|i−1), 2 2 σi+1|i = α(α − ki)σi|i−1 + N, where ki = 2 ασi|i−1 2 σi|i−1 +Q Substituting from the ki equation into the MSE update equation, we obtain 2 σi+1|i = 2 α2Qσi|i−1 2 σi|i−1 +Q + N, This is a Riccati recursion (a quadratic recursion in the MSE) and has a steady state solution: α2Qσ 2 +N σ2 = 2 σ +Q Solving this quadratic equation, we obtain p N − (1 − α2)Q + 4N Q + (N − (1 − α2)Q)2 2 σ = 2 EE 278B: Random Processes 6 – 31 The Kalman gain ki converges to k= −N − (1 − α2)Q + p 4N Q + (N − (1 − α2)Q)2 2αQ and the steady-state Kalman filter is X̂i+1|i = αX̂i|i−1 + k(Yi − X̂i|i−1) EE 278B: Random Processes 6 – 32