Download Exploiting Route Redundancy via Structured Peer to Peer

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Multiprotocol Label Switching wikipedia , lookup

Piggybacking (Internet access) wikipedia , lookup

Network tap wikipedia , lookup

Deep packet inspection wikipedia , lookup

IEEE 1355 wikipedia , lookup

Backpressure routing wikipedia , lookup

CAN bus wikipedia , lookup

Net bias wikipedia , lookup

Computer network wikipedia , lookup

Peering wikipedia , lookup

Recursive InterNetwork Architecture (RINA) wikipedia , lookup

IEEE 802.1aq wikipedia , lookup

Airborne Networking wikipedia , lookup

Peer-to-peer wikipedia , lookup

Routing in delay-tolerant networking wikipedia , lookup

Routing wikipedia , lookup

Transcript
Exploiting Route Redundancy via
Structured Peer to Peer Overlays
Ben Y. Zhao, Ling Huang, Jeremy Stribling,
Anthony D. Joseph, and John D. Kubiatowicz
University of California, Berkeley
ICNP 2003
Challenges Facing Network Applications

Network connectivity is not reliable


Disconnections frequent in the wide-area Internet
IP-level repair is slow



Wide-area: BGP  3 mins
Local-area: IS-IS  5 seconds
Next generation network applications




Mostly wide-area
Streaming media, VoIP, B2B transactions
Low tolerance of delay, jitter and faults
Our work: transparent resilient routing infrastructure that
adapts to faults in not seconds, but milliseconds
November 7, 2003
ICNP 2003
[email protected]
Talk Overview






Motivation
Why structured routing
Structured Peer to Peer overlays
Mechanisms and policy
Evaluation
Summary
November 7, 2003
ICNP 2003
[email protected]
Routing in “Mesh-like” Networks


Previous work has shown reasons for long
convergence [Labovitz00, Labovitz01]
MinRouteAdver timer

Necessary to aggregate updates from all neighbors



Contributes to lower bound of BGP convergence time
Internet becoming more mesh-like [Kaat99,labovitz99]


Commonly set to 30 seconds
Worsens BGP convergence behavior
Question

Can convergence be faster in context of structured routing?
November 7, 2003
ICNP 2003
[email protected]
Resilient Overlay Networks (MIT)


Fully connected mesh
Allows each node full
knowledge of network



D
Fast, independent calculation
of routes
Nodes can construct any
path, maximum flexibility
Cost of flexibility


Protocol needs to choose the
“right” route/nodes
Per node O(n) state


Monitors n - 1 paths
O(n2) total path monitoring is
expensive
S
November 7, 2003
ICNP 2003
[email protected]
Leveraging Structured Peer-to-Peer Overlays

0
Key based routing (IPTPS 03)
 Large sparse ID space N
root(k)
(160 bits: 0 – 2160)
 Nodes in overlay network
have nodeIDs  N
 Given some key k  N,
overlay deterministically
maps k to its root node (live
node in the network)
 route message to root (k)
source
k

Distributed Hashtables (DHT) is interface on KBR
 Key is leveraging underlying routing mesh
November 7, 2003
ICNP 2003
[email protected]
Proximity Neighbor Selection

PNS = network aware overlay construction



Important for routing




Within routing constraints, choose neighbors closest in
network distance (latency)
Generally reduces # of IP hops
Reduce latency
Reduce susceptibility to faults
 Less IP links = smaller chance of link/router failure
Reduce overall network bandwidth utilization
We use Tapestry to demonstrate our design


P2P protocol with PNS overlay construction
Topology-unaware P2P protocols will likely perform worse
November 7, 2003
ICNP 2003
[email protected]
System Architecture
v
v
v
v
v
v
OVERLAY
v
v
v
v
v
v
v
Internet



Locate nearby overlay proxy
Establish overlay path to destination host
Overlay traffic routes traffic resiliently
November 7, 2003
ICNP 2003
[email protected]
Traffic Tunneling
A, B are IP addresses
Legacy
Node A
P’(B)
B
Legacy
Node B
register
register
Proxy
P’(A) = A
P’(B) = B
Proxy
get (hash(B)) P’(B)
put (hash(A), P’(A))
put (hash(B), P’(B))
Structured Peer to
Peer Overlay


Store mapping from end host IP to its proxy’s overlay ID
Similar to approach in Internet Indirection Infrastructure (I3)
November 7, 2003
ICNP 2003
[email protected]
Tradeoffs of Tunneling via P2P

Less neighbor paths to monitor per node: O(log(n))
 Large reduction in probing bandwidth: O(n)  O(log(n))



Actively maintain path redundancy





Manageable for “small” # of paths
Redirect traffic immediately when a failure is detected
Eliminate on-the-fly calculation of new routes
Restore redundancy when a path fails
End result


Increase probing frequency
Faster fault detection with low bandwidth consumption
Fast fault detection + precomputed paths = increased
responsiveness to faults
Cons

Overlay imposes routing stretch (more IP hops), generally < 2
November 7, 2003
ICNP 2003
[email protected]
Some Details

Efficient fault detection



Use soft-state to periodically probe log(n) neighbor paths
“Small” number of routes  reduced bandwidth
Exponentially weighted moving average
in link quality estimation




Avoid route flapping due to short term loss artifacts
Loss rate Ln = (1 - )  Ln-1 +   p
p = instantaneous loss rate,  = hysteresis factor
Maintaining backup paths

Each hop has flexible routing constraint



Create and store backup routes at node insertion
Restore redundancy via “intelligent” gossip after failures
Simple policies to choose among redundant paths
November 7, 2003
ICNP 2003
[email protected]
First Reachable Link Selection (FRLS)




Use estimated loss results to
choose shortest “usable” path
Sort next hop paths by latency
Use shortest path with
2299
minimal quality > T
Correlated failures


Reduce with intelligent topology
construction
Key is to leverage redundancy
available
November 7, 2003
ICNP 2003
2225
2274
2046
2286
2281
2530
1111
[email protected]
Evaluation

Metrics for evaluation



How much routing resiliency can we exploit?
How fast can we adapt to faults?
What is the overhead of routing around a failure?



Proportional increase in end to end latency
Proportional increase in end to end bandwidth used
Experimental platforms

Event-based simulations on transit stub topologies


Data collected over different 5000-node topologies
PlanetLab measurements



Microbenchmarks on responsiveness
Bandwidth measurements from 200+ node overlays
Multiple virtual nodes run per physical machine
November 7, 2003
ICNP 2003
[email protected]
% of All Pairs Reachable
Exploiting Route Redundancy (Sim)
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
Instantaneous IP
0
0.05
Tapestry / FRLS
0.1
0.15
0.2
Proportion of IP Links Broken


Simulation of Tapestry, 2 backup paths per routing entry
Transit-stub topology shown, results from TIER and AS graphs similar
November 7, 2003
ICNP 2003
[email protected]
Responsiveness to Faults (PlanetLab)
Time to Switch Routes (ms)
2500
2000
1500
1000
660
alpha=0.2
alpha=0.4
500
0
0
200
300
400
600
800
1000
1200
Link Probe Period (ms)


Response time increases linearly with probe period
Minimum link quality threshold T = 70%, 20 runs per data point
November 7, 2003
ICNP 2003
[email protected]
Link Probing Bandwidth (Planetlab)
Bandwidth Per Node (KB/s)
7
PR=300ms
PR=600ms
6
5
4
3
2
1
0
1
10
100
1000
Size of Overlay


Medium sized routing overlays incur low probing bandwidth
Bandwidth increases logarithmically with overlay size
November 7, 2003
ICNP 2003
[email protected]
Related Work

Redirection overlays





Topology estimation techniques





Detour (IEEE Micro 99)
Resilient Overlay Networks (SOSP 01)
Internet Indirection Infrastructure (SIGCOMM 02)
Secure Overlay Services (SIGCOMM 02)
Adaptive probing (IPTPS 03)
Peer-based shared estimation (Zhuang 03)
Internet tomography (Chen 03)
Routing underlay (SIGCOMM 03)
Structured peer-to-peer overlays

Tapestry, Pastry, Chord, CAN, Kademlia, Skipnet, Viceroy,
Symphony, Koorde, Bamboo, X-Ring…
November 7, 2003
ICNP 2003
[email protected]
Conclusion

Benefits of structure outweigh costs

Structured routing lowers path maintenance costs


Can no longer construct arbitrary paths



Structured routing with low redundancy gets very close to ideal in
connectivity
Incur low routing stretch
Fast enough for highly interactive applications



Allows “caching” of backup paths for quick failover
300ms beacon period  response time < 700ms
On overlay networks of 300 nodes, b/w cost is 7KB/s
Future work


Deploying a public routing and proxy service on PlanetLab
Examine impact of


Network aware topology construction
Loss sensitive probing techniques
November 7, 2003
ICNP 2003
[email protected]
Questions…

Related websites:
Tapestry
http://www.cs.berkeley.edu/~ravenben/tapestry
 Pastry
http://research.microsoft.com/~antr/pastry
 Chord
http://lcs.mit.edu/chord


Acknowledgements

Thanks to Dennis Geels and Sean Rhea for their
work on the BMark benchmark suite
November 7, 2003
ICNP 2003
[email protected]