TC90: Presentation Title

Download Report

Transcript TC90: Presentation Title

Energy in Networks & Data Center
Networks
Yanjun Yao
Department of EECS
University of Tennessee, Knoxville
1
Network Architecture
Internet
End
Host
Router
Router
Switch
Switch
Switch
End
Host
End
Host
Switch
End
Host
End
Host
Switch
End
Host
End
Host
2
A Feasibility Study for Power
Management in LAN Switches
Maruti Gupta, Satyajit Grover and Suresh Singh
Computer Science Department
Portland State University
3
Motivation and Goals
 Motivation
 Few dynamic power management schemes for internet
devices
 Goal
 Power management scheme for LAN switches
 Why switches?
 Switches comprise bulk of network devices in LAN
 Consumes largest percentage of energy in internet devices
Device
Approximate Number Deployed
Total AEC TW-h
Hubs
93.5 million
1.6 TW-h
LAN Switches
95,000
3.2 TW-h
WAN Switches
50,000
0.15 TW-h
router
3,257
1.1 TW-h
4
Related Works
 Estimate power consumption in switch fabrics:
 Developing statistical traffic models [Wassal et al. 2001]
 Various analytical models [G. Essakimuthu et al. 2002, D.
Langen et al. 2000, C. Patel et al. 1997, Hang et al. 2002,
Ye et al. 2002]
 Power management schemes for interconnection
network fabrics:
 Using DVS with links [Li et al. 2003]
 Using on/off links [L. Peh et al. 2003]
 Router power throttling [Li et al. 2003]
5
Feasibility
 What to do?
 Put LAN switch components, interfaces or entire switches in
sleep.
 Are there enough idle periods to justify sleeping?
Individual Switch Interface
Activity at Switch
Low activity time)
60% of time has interactivity time
Greater than 20 seconds)
Interactive time (seconds)
Percentage of 2 hours
Percentage of 2 hours
High activity time
Low activity time)
High activity time
Interactive time (seconds)
6
Models for Sleeping
 Basic sleep components:
 No sleep model for switches
 Each port has a line card
 Each line card with a processor and
buffers
Ingress Buffer
Network
Processor
Egress Buffer
 Sleep model for a line card is obtained
from the sleep model of its constituent
parts
 Develop sleep model based on the
functionality of the line card
7
Models for Sleeping
 Interface state is preserved
Wake
 HABS (Hardware Assisted Buffered Sleep):
HABS
 Incoming packet wakes up the interface and is buffered
 Power on input buffer, input circuits for receiving
 HAS (Hardware Assisted Sleep):
HAS
 Incoming packet wakes up switch interface and is lost
 Power on receiver circuits
 Simple Sleep:
Simple
Wake
 Set a sleep timer
 Only wakes up when timer expires
 Assumption:
 Transmitting from a deeper sleep to lighter sleep takes time
and results in a spike in energy consumption
8
Implication of Sleeping
 Simple Sleep:
 All packets are lost
 Poor throughput, energy saving will be offset by
retransmission
 To use this state, we need:
 Interface connected to end host: ACPI (Advanced Configuration
and Power Interface) to inform the switch that it is going to sleep
 Interface connecting switches: guarantee no packets will be
sent to a sleeping interface
 HAS:
 The packets wake up the interface get lost
 To use it, we need:
 Send a dummy packet ahead of the packets to be sent to the
sleeping interface
9
Implication of Sleeping
 HABS:
 Lower energy saving
 Further simplify the model:
 Simple sleep:
 Switch interface connected to end
hosts with extended ACPI
 HABS:
 Switch to switch
 Switch to router
 Switch interface connected to
hosts without extended ACPI
10
Algorithms for Sleeping
 Questions:
 When can interface go to sleep?
 Length of sleep interval ts?
 Length of wake interval between consecutive sleeps t I ?
 Wake and Simple Sleep:
 Switch interface sleep when the end host goes to sleep
 Wakes up periodically to check if host has woken up:
 End hosts wakes up and send packets to switch interface with
period   t I
 Remains awake if end host awake until end hosts sleep
again
11
Algorithms for Sleeping
 Wake and HABS:
 Make decision after processing the last packet in the buffer:
 If ( x   )es   ew  xeI , then sleep time ts  x  
 Otherwise, stays awake
 Two simple practical algorithm:
 Estimated algorithm:
 Use an estimator for x , sleep if ( x   )es   ew  xeI ,
where x   xt 1  (1  ) xt
 Sleeps until woken up by an incoming packet
 Estimated and Periodic Algorithm:
 For periodic traffics
 Get time to next periodic packet y, determine x
 Interface sleeps if (min( x , y)   )es   ew  min( x , y)eI
12
Estimated Energy Savings
 Determine energy saving:

E
Es
Energy with no Sleeping/Energy when Sleeping
Individual Switch Interface
es = 0.1
Low activity period
High activity period
es = 0.5
Low activity period
High activity period
Time to wake up (seconds)
13
Performance of Three Algorithms
Host M to Switch Interface
Light
Optimal, Estimated and
Estimated & Period
Heavy
Energy with no Sleeping/Energy when Sleeping
Energy with no Sleeping/Energy when Sleeping
Host Y to Switch Interface
Light & Heavy
All Algorithms
Time to wake up (seconds)
Switch to Switch Interface
Switch to Switch Interface
Optimal, Estimated and
Estimated & Period
Light
Light
Heavy
Heavy
Time to wake up (seconds)
Energy with no Sleeping/Energy when Sleeping
Energy with no Sleeping/Energy when Sleeping
Three algorithms have
very similar performance
Time to wake up (seconds)
Optimal, Estimated and
Estimated & Period
Light
Light
Heavy
Heavy
14
Time to wake up (seconds)
Simulation Results
 Topology:
 Six switches
 Each host runs STP protocol
in addition to different data
streams
 Data for simulations is
generated using Markov
Modulated Poisson Process
 Simulation on Opnet
 Evaluate Interfaces:
 Sw0 to sw4
 Sw2 to mmpp22
15
Simulation Result
 Switch to switch saves more
energy
Energy with no Sleeping/Energy when Sleeping
Switch Interfaces, HABS Simulation
Time to wake up (seconds)
Switch Interfaces, Simple Sleep Simulation
Percentage of Packets Lost
Energy with no Sleeping/Energy when Sleeping
Switch Interfaces, Simple Sleep Simulation
Time to wake up (seconds)
Time to wake up (seconds)
16
Impact of Sleeping On protocols and
Topology Design
 Simple Sleep’s impact on protocol design:
 For periodic messages, the sleep time must be fine tuned.
 Wake up all interfaces for broadcasting.
 Impact of network topology and VLANs on sleeping:
 For redundant paths:
 Aggregate traffic loads to some of the paths and put the rest to
sleep.
 However, the STP generated a spanning tree
17
Conclusion
 Sleeping in order to save energy is a feasible option
in the LAN.
 Three sleeping models are proposed.
 Two types of algorithms for transmitting from wake
state and sleeping state are shown.
 Simulations are done to evaluate the performance of
HABS and Simple Sleep.
18
Critique
 Three sleeping models are proposed but only two of
them are evaluated. HAS is eliminated without a
good reason.
 Modifications on hardware are needed to support the
three sleep models.
 For the first simulation, it is said that the HABS are
used for both experiments, but different transision
energies are used.
 Did not evaluate packet delay
19
VL2: A Scalable and Flexible Data
Center Network
Albert Greenberg. James R. Hamilton. Navendu
Jain. Srikanth Kandula. Changhoon Kim, et al
Microsoft Reseach
20
Architecture of Data Center Networks (DCN)
21
Conventional DCN Problems
CR
CR
AR
AR
AR
AR
S
S
S
S
1:240
S
IS wantS1:80
more
S
1:5
…
…
 Static network assignment
 Fragmentation of resource
...
I have spare ones,
S
S
S
S
but…
…
…
 Poor server to server connectivity
 Traffics affects each other
 Poor reliability and utilization
22
Objectives:
 Uniform high capacity:
 Maximum rate of server to server traffic flow should be limited only
by capacity on network cards
 Assigning servers to service should be independent of network
topology
 Performance isolation:
 Traffic of one service should not be affected by traffic of other
services
 Layer-2 semantics:
 Easily assign any server to any service
 Configure server with whatever IP address the service expects
 VM keeps the same IP address even after migration
23
Measurements and Implications of DCN
 Data-Center traffic analysis:
 Traffic volume between servers to entering/leaving data
center is 4:1
 Demand for bandwidth between servers growing faster
 Network is the bottleneck of computation
 Flow distribution analysis:
 Majority of flows are small, biggest flow size is 100MB
 The distribution of internal flows is simpler and more uniform
 50% times of 10 concurrent flows, 5% greater than 80
concurrent flows
24
Measurements and Implications of DCN
 Traffic matrix analysis:
 Poor summarizing of traffic patterns
 Instability of traffic patterns
 Failure characteristics:
 Pattern of networking equipment failures: 95% < 1min, 98%
< 1hr, 99.6% < 1 day, 0.09% > 10 days
 No obvious way to eliminate all failures from the top of the
hierarchy
25
Virtual Layer Two Networking (VL2)
 Design principle:
 Randomizing to cope with volatility:
 Using Valiant Load Balancing (VLB) to do destination
independent traffic spreading across multiple intermediate
nodes
 Building on proven networking technology:
 Using IP routing and forwarding technologies available in
commodity switches
 Separating names from locators:
 Using directory system to maintain the mapping between names
and locations
 Embracing end systems:
 A VL2 agent at each server
26
VL2 Addressing and Routing
Switches run link-state routing
and
maintain only switch-level
topology
LAs
ToR1 . . . ToR2
ToR3 y payload
ToR34 z payload
AAs
x
...
ToR3
y,yz
. . . ToR4
z
Directory
Service
…
x  ToR2
y  ToR3
z  ToR34
…
Lookup &
Response
Servers use flat
names
27
Random Traffic Spreading over Multiple Paths
IANY
IANY
IANY
Links used
for up paths
Links used
for down paths
T1
IANY T53
T2
T3
x
y
T4
T5
T6
yz payload
z
28
VL2 Directory System
RSM
RSM
Servers
3. Replicate
RSM
RSM
4. Ack
(6. Disseminate)
2. Set
...
DS
DS
2. Reply
...
DS
2. Reply
1. Lookup
...
Directory
Servers
5. Ack
1. Update
Agent
Agent
“Lookup”
“Update”
29
Evaluation
 Uniform high capacity:
 All-to-all data shuffle stress test:
 75 servers, deliver 500MB
 Maximal achievable goodput is 62.3
 VL2 network efficiency as 58.8/62.3 = 94%
30
Evaluation
 Fairness:
Fairness Index
 75 nodes
 Real data center workload
 Plot Jain’s fairness index for traffics to intermediate switches
1.00
0.98
0.96
Aggr1
0.94
0
100
200
300
Time (s)
Aggr2
400
Aggr3
500
31
Evaluation
 Performance isolation:
 Two types of services:
 Service one: 18 servers do single TCP transfer all the time
 Service two: 19 servers starts a 8GB transfer over TCP every 2
seconds
 Service two: 19 servers burst short TCP connections
32
Evaluation
 Convergence after link failures
 75 servers
 All-to-all data shuffle
 Disconnect links between intermediate and aggregation
switches
33
Conclusion
 Studied the traffic pattern in a production data center
and find the traffic patterns
 Design, build and deploy every component of VL2 in
an 80 server testbed
 Apply VLB to randomly spreading traffics over
multiple flows
 Using flat address to split IP addresses and server
names
34
Critique
 The extra servers are needed to support the VL2
directory system,:
 Brings more cost on devices
 Hard to be implemented for data centers with tens of
thousands of servers.
 All links and switches are working all the times, not
power efficient
 No evaluation of real time performance.
35
Comparison
LAN Switch
VL2
Target
Save power on LAN
switches
Achieve agility on DCN
Networks
LAN
DCN
Traffic Pattern Light for most time
Highly unpredictable
Object
Switches
Whole network
Experiment
Simulation on Opnet
Real testbed
36
 Q&A
37