Download T - University of Southern California

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Marginalism wikipedia , lookup

Marginal utility wikipedia , lookup

Choice modelling wikipedia , lookup

Microeconomics wikipedia , lookup

Transcript
1
Distributed Stochastic Optimization
via Correlated Scheduling
Observation ω1(t)
1
Observation ω2(t)
2
Fusion
Center
Michael J. Neely
University of Southern California
http://www-bcf.usc.edu/~mjneely
2
Distributed sensor reports
ω1(t)
1
ω2(t)
2
Fusion
Center
• ωi(t) = 0/1 if sensor i observes the event on slot t
• Pi(t) = 0/1 if sensor i reports on slot t
• Utility: U(t) = min[P1(t)ω1(t) + (1/2)P2(t)ω2(t),1]
Redundant reports do not increase utility.
3
Distributed sensor reports
ω1(t)
1
ω2(t)
2
Fusion
Center
• ωi(t) = 0/1 if sensor i observes the event on slot t
• Pi(t) = 0/1 if sensor i reports on slot t
• Utility: U(t) = min[P1(t)ω1(t) + (1/2)P2(t)ω2(t),1]
Maximize:
U
Subject to:
P1 ≤ c
P2 ≤ c
4
Main ideas for this example
• Utility function is non-separable.
• Redundant reports do not bring extra utility.
• A centralized algorithm would never send
redundant reports (it wastes power).
• A distributed algorithm faces these
challenges:
• Sensor 2 does not know if sensor 1
observed an event.
• Sensor 2 does not know if sensor 1
reported anything.
5
Assumed structure
Agree
on plan 0 1 2 3 4
1) Coordinate on a plan before time 0.
2) Distributively implement plan after time 0.
t
6
Example “plans”
Agree
on plan 0 1 2 3 4
Example plan:
Sensor 1:
• t=even  Do not report.
• t=odd  Report if ω1(t)=1.
Sensor 2:
• t=even  Report with prob p if ω2(t)=1
• t=odd:  Do not report.
t
7
Common source of randomness
Day 1
Day 2
Example: 1 slot = 1 day
Each person looks at Boston Globe every day:
• If first letter is a “T”  Plan 1
• If first letter is an “S”  Plan 2
• Etc.
8
Specific example
Assume:
• Pr[ω1(t)=1] = ¾, Pr[ω2(t)=1] = ½
• ω1(t), ω2(t) independent
• Power constraint c = 1/3
Approach 1: Independent reporting
• If ω1(t)=1, sensor 1 reports with probability θ1
• If ω2(t)=1, sensor 2 reports with probability θ2
Optimizing θ1, θ2 gives u = 4/9 ≈ 0.44444
9
Approach 2: Correlated reporting
Pure strategy 1:
• Sensor 1 reports if and only if ω1(t)=1.
• Sensor 2 does not report.
Pure strategy 2:
• Sensor 1 does not report.
• Sensor 2 reports if and only if ω2(t)=1.
Pure strategy 3:
• Sensor 1 reports if and only if ω1(t)=1.
• Sensor 2 reports if and only if ω2(t)=1.
10
Approach 2: Correlated reporting
X(t) = iid random variable (commonly known):
• Pr[X(t)=1] = θ1
• Pr[X(t)=2] = θ2
• Pr[X(t)=3] = θ3
On slot t:
• Sensors observe X(t)
• If X(t)=k, sensors use pure strategy k.
Optimizing θ1, θ2, θ3 gives u = 23/48 ≈ 0.47917
11
Summary of approaches
Strategy
u
Independent reporting
0.44444
Correlated reporting
0.47917
Centralized reporting
0.5
12
Summary of approaches
Strategy
u
Independent reporting
0.44444
Correlated reporting
0.47917
Centralized reporting
0.5
It can be shown that this is optimal over all
distributed strategies!
13
General distributed optimization
Maximize:
U
Subject to: Pk ≤ c for k in {1, …, K}
ω(t) = (ω1(t), …, ωΝ(t))
π(ω) = Pr[ω(t) = (ω1, …, ωΝ)]
α(t) = (α1(t), …, αΝ(t))
U(t) = u(α(t), ω(t))
Pk(t) = pk(α(t), ω(t))
14
Pure strategies
A pure strategy is a deterministic vectorvalued function:
g(ω) = (g1(ω1), g2(ω2), …, gΝ (ωΝ))
Let M = # pure strategies:
M = |A1||Ω1| x |A2||Ω2| x ... x |AN||ΩN|
15
Optimality Theorem
There exist:
• K+1 pure strategies g(m)(ω)
• Probabilities θ1, θ2, …, θK+1
such that the following distributed
algorithm is optimal:
X(t) = iid, Pr[X(t)=m] = θm
• Each user observes X(t)
• If X(t)=m  use strategy g(m)(ω).
16
LP and complexity reduction
•
The probabilities can be found by an LP
•
Unfortunately, the LP has M variables
•
If (ω1(t), …, ωΝ(t)) are mutually independent
and the utility function satisfies a preferred
action property, complexity can be reduced
•
Example N=2 users, |A1|=|A2|=2
--Old complexity = 2|Ω1|+|Ω2|
--New complexity = (|Ω1|+1)(|Ω2|+1)
Discussion of Theorem 1
17
Theorem 1 solves the problem for distributed
scheduling, but:
• Requires an offline LP to be solved
before time 0.
•
Requires full knowledge of π(ω)
probabilities.
18
Online Dynamic Approach
We want an algorithm that:
• Operates online
• Does not need π(ω) probabilities.
• Can adapt when these probabilities
change.
Such an algorithm must use feedback:
• Assume feedback is a fixed delay D.
• Assume D>1.
• Such feedback cannot improve average
utility beyond the distributed optimum.
19
Lyapunov optimization approach
•
Define K virtual queues Q1(t), …, QK(t).
•
Every slot t, observe queues and choose
strategy m in {1, …, M} to maximize a
weighted sum of queues.
•
Update queues with delayed feedback:
Qk(t+1) = max[Qk(t) + Pk(t-D) - c, 0]
20
Lyapunov optimization approach
•
Define K virtual queues Q1(t), …, QK(t).
•
Every slot t, observe queues and choose
strategy m in {1, …, M} to maximize a
weighted sum of queues.
•
Update queues with delayed feedback:
Qk(t+1) = max[Qk(t) + Pk(t-D) - c, 0]
“arrivals”
Virtual queue: If stable, then:
Time average power ≤ c.
“service”
21
Separable problems
If the utility and penalty functions are a
separable sum of functions of individual
variables (αn(t), ωn(t)), then:
•
There is no optimality gap between
centralized and distributed algorithms
•
Problem complexity reduces from
exponential to linear.
22
Simulation (non-separable problem)
•
•
•
•
•
•
3-user problem
αn(t) in {0, 1} for n ={1, 2, 3}.
ωn(t) in {0, 1, 2, 3, 4, 5, 6, 7, 8, 9}
V=1/ε
Get O(ε) guarantee to optimality
Convergence time depends on 1/ε
Utility versus V parameter (V=1/ε)
Utility
23
V (recall V = 1/ε)
Average power up to time t
24
Average power versus time
V=100
V=50
V=10
power constraint 1/3
Time t
Adaptation to non-ergodic changes
25
Adaptation to non-ergodic changes
26
Optimal utility for phase 2
Optimal utility for phases 1 and 3
Oscillates about the average constraint c
27
Conclusions
• Paper introduces correlated scheduling via
common source of randomness.
• Common source of randomness is crucial for
optimality in a distributed setting.
• Optimality gap between distributed and
centralized problems (gap=0 for separable
problems).
• Complexity reduction technique in paper.
• Online implementation via Lyapunov optimization.
• Online algorithm adapts to a changing
environment.