Exploiting Route Redundancy via Structured Peer to Peer Overlays

Download Report

Transcript Exploiting Route Redundancy via Structured Peer to Peer Overlays

Exploiting Route Redundancy via
Structured Peer to Peer Overlays
Ben Y. Zhao, Ling Huang, Jeremy Stribling,
Anthony D. Joseph, and John D. Kubiatowicz
University of California, Berkeley
ICNP 2003
Challenges Facing Network Applications

Network connectivity is not reliable


Disconnections frequent in the wide-area Internet
IP-level repair is slow



Wide-area: BGP  3 mins
Local-area: IS-IS  5 seconds
Next generation network applications




Mostly wide-area
Streaming media, VoIP, B2B transactions
Low tolerance of delay, jitter and faults
Our work: transparent resilient routing infrastructure that
adapts to faults in not seconds, but milliseconds
November 7, 2003
ICNP 2003
[email protected]
Talk Overview






Motivation
Why structured routing
Structured Peer to Peer overlays
Mechanisms and policy
Evaluation
Summary
November 7, 2003
ICNP 2003
[email protected]
Routing in “Mesh-like” Networks


Previous work has shown reasons for long
convergence [Labovitz00, Labovitz01]
MinRouteAdver timer

Necessary to aggregate updates from all neighbors



Contributes to lower bound of BGP convergence time
Internet becoming more mesh-like [Kaat99,labovitz99]


Commonly set to 30 seconds
Worsens BGP convergence behavior
Question

Can convergence be faster in context of structured routing?
November 7, 2003
ICNP 2003
[email protected]
Resilient Overlay Networks (MIT)


Fully connected mesh
Allows each node full
knowledge of network



D
Fast, independent calculation
of routes
Nodes can construct any
path, maximum flexibility
Cost of flexibility


Protocol needs to choose the
“right” route/nodes
Per node O(n) state


Monitors n - 1 paths
O(n2) total path monitoring is
expensive
S
November 7, 2003
ICNP 2003
[email protected]
Leveraging Structured Peer-to-Peer Overlays

0
Key based routing (IPTPS 03)
 Large sparse ID space N
root(k)
(160 bits: 0 – 2160)
 Nodes in overlay network
have nodeIDs  N
 Given some key k  N,
overlay deterministically
maps k to its root node (live
node in the network)
 route message to root (k)
source
k

Distributed Hashtables (DHT) is interface on KBR
 Key is leveraging underlying routing mesh
November 7, 2003
ICNP 2003
[email protected]
Proximity Neighbor Selection

PNS = network aware overlay construction



Important for routing




Within routing constraints, choose neighbors closest in
network distance (latency)
Generally reduces # of IP hops
Reduce latency
Reduce susceptibility to faults
 Less IP links = smaller chance of link/router failure
Reduce overall network bandwidth utilization
We use Tapestry to demonstrate our design


P2P protocol with PNS overlay construction
Topology-unaware P2P protocols will likely perform worse
November 7, 2003
ICNP 2003
[email protected]
System Architecture
v
v
v
v
v
v
OVERLAY
v
v
v
v
v
v
v
Internet



Locate nearby overlay proxy
Establish overlay path to destination host
Overlay traffic routes traffic resiliently
November 7, 2003
ICNP 2003
[email protected]
Traffic Tunneling
A, B are IP addresses
Legacy
Node A
P’(B)
B
Legacy
Node B
register
register
Proxy
P’(A) = A
P’(B) = B
Proxy
get (hash(B)) P’(B)
put (hash(A), P’(A))
put (hash(B), P’(B))
Structured Peer to
Peer Overlay


Store mapping from end host IP to its proxy’s overlay ID
Similar to approach in Internet Indirection Infrastructure (I3)
November 7, 2003
ICNP 2003
[email protected]
Tradeoffs of Tunneling via P2P

Less neighbor paths to monitor per node: O(log(n))
 Large reduction in probing bandwidth: O(n)  O(log(n))



Actively maintain path redundancy





Manageable for “small” # of paths
Redirect traffic immediately when a failure is detected
Eliminate on-the-fly calculation of new routes
Restore redundancy when a path fails
End result


Increase probing frequency
Faster fault detection with low bandwidth consumption
Fast fault detection + precomputed paths = increased
responsiveness to faults
Cons

Overlay imposes routing stretch (more IP hops), generally < 2
November 7, 2003
ICNP 2003
[email protected]
Some Details

Efficient fault detection



Use soft-state to periodically probe log(n) neighbor paths
“Small” number of routes  reduced bandwidth
Exponentially weighted moving average
in link quality estimation




Avoid route flapping due to short term loss artifacts
Loss rate Ln = (1 - )  Ln-1 +   p
p = instantaneous loss rate,  = hysteresis factor
Maintaining backup paths

Each hop has flexible routing constraint



Create and store backup routes at node insertion
Restore redundancy via “intelligent” gossip after failures
Simple policies to choose among redundant paths
November 7, 2003
ICNP 2003
[email protected]
First Reachable Link Selection (FRLS)




Use estimated loss results to
choose shortest “usable” path
Sort next hop paths by latency
Use shortest path with
2299
minimal quality > T
Correlated failures


Reduce with intelligent topology
construction
Key is to leverage redundancy
available
November 7, 2003
ICNP 2003
2225
2274
2046
2286
2281
2530
1111
[email protected]
Evaluation

Metrics for evaluation



How much routing resiliency can we exploit?
How fast can we adapt to faults?
What is the overhead of routing around a failure?



Proportional increase in end to end latency
Proportional increase in end to end bandwidth used
Experimental platforms

Event-based simulations on transit stub topologies


Data collected over different 5000-node topologies
PlanetLab measurements



Microbenchmarks on responsiveness
Bandwidth measurements from 200+ node overlays
Multiple virtual nodes run per physical machine
November 7, 2003
ICNP 2003
[email protected]
% of All Pairs Reachable
Exploiting Route Redundancy (Sim)
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
Instantaneous IP
0
0.05
Tapestry / FRLS
0.1
0.15
0.2
Proportion of IP Links Broken


Simulation of Tapestry, 2 backup paths per routing entry
Transit-stub topology shown, results from TIER and AS graphs similar
November 7, 2003
ICNP 2003
[email protected]
Responsiveness to Faults (PlanetLab)
Time to Switch Routes (ms)
2500
2000
1500
1000
660
alpha=0.2
alpha=0.4
500
0
0
200
300
400
600
800
1000
1200
Link Probe Period (ms)


Response time increases linearly with probe period
Minimum link quality threshold T = 70%, 20 runs per data point
November 7, 2003
ICNP 2003
[email protected]
Link Probing Bandwidth (Planetlab)
Bandwidth Per Node (KB/s)
7
PR=300ms
PR=600ms
6
5
4
3
2
1
0
1
10
100
1000
Size of Overlay


Medium sized routing overlays incur low probing bandwidth
Bandwidth increases logarithmically with overlay size
November 7, 2003
ICNP 2003
[email protected]
Related Work

Redirection overlays





Topology estimation techniques





Detour (IEEE Micro 99)
Resilient Overlay Networks (SOSP 01)
Internet Indirection Infrastructure (SIGCOMM 02)
Secure Overlay Services (SIGCOMM 02)
Adaptive probing (IPTPS 03)
Peer-based shared estimation (Zhuang 03)
Internet tomography (Chen 03)
Routing underlay (SIGCOMM 03)
Structured peer-to-peer overlays

Tapestry, Pastry, Chord, CAN, Kademlia, Skipnet, Viceroy,
Symphony, Koorde, Bamboo, X-Ring…
November 7, 2003
ICNP 2003
[email protected]
Conclusion

Benefits of structure outweigh costs

Structured routing lowers path maintenance costs


Can no longer construct arbitrary paths



Structured routing with low redundancy gets very close to ideal in
connectivity
Incur low routing stretch
Fast enough for highly interactive applications



Allows “caching” of backup paths for quick failover
300ms beacon period  response time < 700ms
On overlay networks of 300 nodes, b/w cost is 7KB/s
Future work


Deploying a public routing and proxy service on PlanetLab
Examine impact of


Network aware topology construction
Loss sensitive probing techniques
November 7, 2003
ICNP 2003
[email protected]
Questions…

Related websites:
Tapestry
http://www.cs.berkeley.edu/~ravenben/tapestry
 Pastry
http://research.microsoft.com/~antr/pastry
 Chord
http://lcs.mit.edu/chord


Acknowledgements

Thanks to Dennis Geels and Sean Rhea for their
work on the BMark benchmark suite
November 7, 2003
ICNP 2003
[email protected]
Backup Slides
November 7, 2003
ICNP 2003
[email protected]
Another Perspective on Reachability
Proportion of All Paths
1
0.9
0.8
Portion of all pairwise paths where
no failure-free
paths remain
0.7
A path exists, but
neither IP nor
FRLS can locate
the path
0.6
0.5
0.4
0.3
0.2
0.1
Portion of0 all paths
where IP0 and
FRLS both route
successfully
November 7, 2003
0.05
0.1
0.15
Proportion of IP Links Broken
ICNP 2003
FRLS finds
path, 0.2
where
short-term IP
routing fails
[email protected]
Constrained Multicast




Used only when all paths are below
quality threshold
Send duplicate messages on multiple
paths
Leverage route convergence
2299
2274
 Assign unique message
IDs
 Mark duplicates
 Keep moving window of IDs
2046
2281
 Recognize and drop duplicates
? ? ?
Limitations
 Assumes loss not from congestion
1111
 Ideal for local area routing
November 7, 2003
ICNP 2003
2225
2286
2530
[email protected]
Latency Overhead of Misrouting
Proportional Latency Overhead
Hop 0
Hop 1
Hop 4
Hop 3
Hop 2
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
20
40
60
80
100
120
160
140
180
Latency from source to destination (ms)
November 7, 2003
ICNP 2003
[email protected]
Bandwidth Cost of Constrained Multicast
Proportional Bandwidth Overhead
Hop 0
Hop 1
Hop 2
Hop 3
Hop 4
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
20
40
60
80
100
120
140
160
Latency from source to destination (ms)
November 7, 2003
ICNP 2003
[email protected]
Challenge 2: Tunneling Application Traffic

Basic idea



Desired properties




Tunnel “legacy” traffic via overlay proxies
Should be protocol independent
Traffic redirection transparent to end hosts
Stable mapping between a proxy and its legacy nodes
Incremental deployment
Details

Making a connection to D




Determine if D is reachable from overlay
If so, retrieve D’s overlay ID from storage in overlay
Redirect traffic through overlay to D’s overlay ID
Assign legacy nodes IDs in the overlay

Routing to IDs automatically reach respective proxies
November 7, 2003
ICNP 2003
[email protected]
Assigning Overlay Identifiers




0
A proxy close by in network
assigns a legacy node a
non-random overlay ID
Assign closest possible
ID to proxy’s ID

proxy
proxy
source
node ID
such that: root (ID) = proxy
Chord/Tapestry: ID = proxyID – 1
Stable ID mapping


Probability of a new node
“hijacking” legacy node very low
1 million nodes,
1000 nodes / proxy,
160 bit names, Ph < 10-40
November 7, 2003
ICNP 2003
destination
node ID
[email protected]
Unused Slides
November 7, 2003
ICNP 2003
[email protected]
Summary

Highly adaptive to link failures

Fast adaptation time w/ low monitoring BW cost



300ms beacon period, 7KB/s (300 overlay nodes),
response time < 700ms
Low latency and BW overhead for misrouting
Limitation: less useful for congested IP links



Constrained multicast can exacerbate congestion
Can utilize more physically-aware overlay construction
Issue: using loss as congestion indicator

Can fix with more intelligent protocols such as XCP
November 7, 2003
ICNP 2003
[email protected]
Possible Approaches

Modify basic IP protocols



“Helper” infrastructures to assist protocols


BGP, OSPF, IS-IS
Millions of deployed routers  Significant
deployment challenge
Manipulate input to improve performance without
modification
Deploy intelligent protocols above IP layer


Overlay approach taken by many
Relies on basic point to point IP routing
November 7, 2003
ICNP 2003
[email protected]