Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
1 Distributed Stochastic Optimization via Correlated Scheduling Observation ω1(t) 1 Observation ω2(t) 2 Fusion Center Michael J. Neely University of Southern California http://www-bcf.usc.edu/~mjneely 2 Distributed sensor reports ω1(t) 1 ω2(t) 2 Fusion Center • ωi(t) = 0/1 if sensor i observes the event on slot t • Pi(t) = 0/1 if sensor i reports on slot t • Utility: U(t) = min[P1(t)ω1(t) + (1/2)P2(t)ω2(t),1] Redundant reports do not increase utility. 3 Distributed sensor reports ω1(t) 1 ω2(t) 2 Fusion Center • ωi(t) = 0/1 if sensor i observes the event on slot t • Pi(t) = 0/1 if sensor i reports on slot t • Utility: U(t) = min[P1(t)ω1(t) + (1/2)P2(t)ω2(t),1] Maximize: U Subject to: P1 ≤ c P2 ≤ c 4 Main ideas for this example • Utility function is non-separable. • Redundant reports do not bring extra utility. • A centralized algorithm would never send redundant reports (it wastes power). • A distributed algorithm faces these challenges: • Sensor 2 does not know if sensor 1 observed an event. • Sensor 2 does not know if sensor 1 reported anything. 5 Assumed structure Agree on plan 0 1 2 3 4 1) Coordinate on a plan before time 0. 2) Distributively implement plan after time 0. t 6 Example “plans” Agree on plan 0 1 2 3 4 Example plan: Sensor 1: • t=even Do not report. • t=odd Report if ω1(t)=1. Sensor 2: • t=even Report with prob p if ω2(t)=1 • t=odd: Do not report. t 7 Common source of randomness Day 1 Day 2 Example: 1 slot = 1 day Each person looks at Boston Globe every day: • If first letter is a “T” Plan 1 • If first letter is an “S” Plan 2 • Etc. 8 Specific example Assume: • Pr[ω1(t)=1] = ¾, Pr[ω2(t)=1] = ½ • ω1(t), ω2(t) independent • Power constraint c = 1/3 Approach 1: Independent reporting • If ω1(t)=1, sensor 1 reports with probability θ1 • If ω2(t)=1, sensor 2 reports with probability θ2 Optimizing θ1, θ2 gives u = 4/9 ≈ 0.44444 9 Approach 2: Correlated reporting Pure strategy 1: • Sensor 1 reports if and only if ω1(t)=1. • Sensor 2 does not report. Pure strategy 2: • Sensor 1 does not report. • Sensor 2 reports if and only if ω2(t)=1. Pure strategy 3: • Sensor 1 reports if and only if ω1(t)=1. • Sensor 2 reports if and only if ω2(t)=1. 10 Approach 2: Correlated reporting X(t) = iid random variable (commonly known): • Pr[X(t)=1] = θ1 • Pr[X(t)=2] = θ2 • Pr[X(t)=3] = θ3 On slot t: • Sensors observe X(t) • If X(t)=k, sensors use pure strategy k. Optimizing θ1, θ2, θ3 gives u = 23/48 ≈ 0.47917 11 Summary of approaches Strategy u Independent reporting 0.44444 Correlated reporting 0.47917 Centralized reporting 0.5 12 Summary of approaches Strategy u Independent reporting 0.44444 Correlated reporting 0.47917 Centralized reporting 0.5 It can be shown that this is optimal over all distributed strategies! 13 General distributed optimization Maximize: U Subject to: Pk ≤ c for k in {1, …, K} ω(t) = (ω1(t), …, ωΝ(t)) π(ω) = Pr[ω(t) = (ω1, …, ωΝ)] α(t) = (α1(t), …, αΝ(t)) U(t) = u(α(t), ω(t)) Pk(t) = pk(α(t), ω(t)) 14 Pure strategies A pure strategy is a deterministic vectorvalued function: g(ω) = (g1(ω1), g2(ω2), …, gΝ (ωΝ)) Let M = # pure strategies: M = |A1||Ω1| x |A2||Ω2| x ... x |AN||ΩN| 15 Optimality Theorem There exist: • K+1 pure strategies g(m)(ω) • Probabilities θ1, θ2, …, θK+1 such that the following distributed algorithm is optimal: X(t) = iid, Pr[X(t)=m] = θm • Each user observes X(t) • If X(t)=m use strategy g(m)(ω). 16 LP and complexity reduction • The probabilities can be found by an LP • Unfortunately, the LP has M variables • If (ω1(t), …, ωΝ(t)) are mutually independent and the utility function satisfies a preferred action property, complexity can be reduced • Example N=2 users, |A1|=|A2|=2 --Old complexity = 2|Ω1|+|Ω2| --New complexity = (|Ω1|+1)(|Ω2|+1) Discussion of Theorem 1 17 Theorem 1 solves the problem for distributed scheduling, but: • Requires an offline LP to be solved before time 0. • Requires full knowledge of π(ω) probabilities. 18 Online Dynamic Approach We want an algorithm that: • Operates online • Does not need π(ω) probabilities. • Can adapt when these probabilities change. Such an algorithm must use feedback: • Assume feedback is a fixed delay D. • Assume D>1. • Such feedback cannot improve average utility beyond the distributed optimum. 19 Lyapunov optimization approach • Define K virtual queues Q1(t), …, QK(t). • Every slot t, observe queues and choose strategy m in {1, …, M} to maximize a weighted sum of queues. • Update queues with delayed feedback: Qk(t+1) = max[Qk(t) + Pk(t-D) - c, 0] 20 Lyapunov optimization approach • Define K virtual queues Q1(t), …, QK(t). • Every slot t, observe queues and choose strategy m in {1, …, M} to maximize a weighted sum of queues. • Update queues with delayed feedback: Qk(t+1) = max[Qk(t) + Pk(t-D) - c, 0] “arrivals” Virtual queue: If stable, then: Time average power ≤ c. “service” 21 Separable problems If the utility and penalty functions are a separable sum of functions of individual variables (αn(t), ωn(t)), then: • There is no optimality gap between centralized and distributed algorithms • Problem complexity reduces from exponential to linear. 22 Simulation (non-separable problem) • • • • • • 3-user problem αn(t) in {0, 1} for n ={1, 2, 3}. ωn(t) in {0, 1, 2, 3, 4, 5, 6, 7, 8, 9} V=1/ε Get O(ε) guarantee to optimality Convergence time depends on 1/ε Utility versus V parameter (V=1/ε) Utility 23 V (recall V = 1/ε) Average power up to time t 24 Average power versus time V=100 V=50 V=10 power constraint 1/3 Time t Adaptation to non-ergodic changes 25 Adaptation to non-ergodic changes 26 Optimal utility for phase 2 Optimal utility for phases 1 and 3 Oscillates about the average constraint c 27 Conclusions • Paper introduces correlated scheduling via common source of randomness. • Common source of randomness is crucial for optimality in a distributed setting. • Optimality gap between distributed and centralized problems (gap=0 for separable problems). • Complexity reduction technique in paper. • Online implementation via Lyapunov optimization. • Online algorithm adapts to a changing environment.