FairCloud: Sharing the Network in Cloud Computing

advertisement
FairCloud: Sharing the Network
in Cloud Computing
Computer Communication Review(2012)
Arthur : Lucian Popa
Arvind Krishnamurthy
Sylvia Ratnasamy
Ion Stoica
Presenter : 段雲鵬
Outline
•
•
•
•
•
Introduction
challenges sharing networks
Properties for network sharing
Mechanism
Conclusion
2/20
Some concepts
• Bisection bandwidth
– Each node has a unit weight
– Each link has a unit weight
• Flow def
– standard five-tuple in packet headers
• B denotes bandwidth
• T denotes traffic
• W denotes the weight of a VM
3/20
Background
• Resource in cloud computing
– Network , CPU , memory
• Network allocation
– More difficult
• Source, destination and cross traffic
– Tradeoff
• payment proportionality VS bandwidth guarantees
4/20
Introduction
• Network allocation
– Unkown to users , bad predictability
• Fairness issues
– Flows, source-destination pairs, or sources alone ,
destination alone
• Difference with other resource
– Interdependent Users
– Interdependent Resources
5/20
Assumption
• From a per-VM viewpoint
• Be agnostic to VM placement and routing
algorithms
• In a single datacenter
• Be largely orthogonal to work on network
topologies to improve bisection bandwidth
6/20
Traditional Mechanism
• Per flow fairness
– Unfair when simply instantiating more flow
• Per source-destination pair
– Unfair when one VM communicates with more
VMs
• Per source
– Unfair to destinations
• Asymmetric
– Only be fair for source or destination only
7/20
Examples
• Per source-destination pair
If there is little traffic on the A-F and B-E ,
B(A)=B(B) =B(E) =B(F) =2*B(C) =2*B(D)
=B(G) =B(H)
Per source
B(E) =B(F) =0.25*B(D) , In the opposite
direction, B(A) =B(B) =0.25*B(C)
8/20
Properties for network sharing(1)
• Strategy proofness
– Can’t increase bandwidth by modifying behavior
at application level
• Pareto Efficiency A
– X and Y is bottlenecked , when B(X-Y) increases,
B(A-B) must decrease ,otherwise congestion will
be worse
B
10 M
1M
10 M
9/20
Properties for network sharing(2)
• Non-zero Flow Allocation
– A strictly +B() between each pairs are expected
• Independence
L1
– When T2 increase , B1 should not be
affected
• Symmetry
– If all flows’ direction are swiched,
the allocation should be the same
L2
10/20
Network weight and user’s payment.
• Weight Fidelity(provide incentive)
– Strict Monotonicity (Monotonicity)
• If W(VM) increases ,then all its traffic must increase
(not decrease) .
– Proportionality
• Guaranteed Bandwidth
Subset
P(2/3)
– Admission control
• They are conflicting, tradeoff
Subset
Q(1/3)
No communication between P and Q
11/20
Per Endpoint Sharing (PES)
• Can explicitly trade between weight fidelity
and guaranteed bandwidth
– NA denote the number of VMs A is communicating
with
– WS-D=f(WS,WD) , WA-B=WB-A
– Normalized by L1 normalization
• Drawback : Static Method (out of discussion)
12/20
Example
• WA-D=WA/NA+WD/ND
=1/2+1/2=1
• WA-C=WB-D=1/2+1/1=1.5
• Total Weight=4(4 VMs)
• So WA-D=1/4=0.25 WA-C=WB-D=1.5/4=0.325
13/20
Comparison
14/20
PES
• For one host , B ∝ (closer VMs) instead of
(remote VMs)
• Higher guarantees for the worst case
• WA−B = WB−A =α*WA/NA+ β*WB/ NB
– α and β can be designed to weight between
bandwidth guarantees and weight fidelity
15/20
One Sided PES (OSPES)
• Designed for tree-based topology
• WA−B = WB−A =α*WA/NA+ β*WB/ NB
• When closer to A, α = 1 and β = 0
• When closer to B, α = 0 and β = 1
16/20
OSPES
• fair sharing for the traffic towards or from
the tree root
– Resource allocation are depended on the root
When
– Non-strict monotonicity
W(A) = W(B) , If the
access
link is 1 Gbs, then
each VM is
guaranteed 500
Mbps
WA-VM1=1/1 WB-VMi=1/10(i=2,3……,11)
17/20
Max-Min Fairness
• The minimum data rate that a dataflow
achieves is maximized
– The bottleneck is fully utilized
• Can be applied
18/20
Conclusion
• Problem : sharing the network within a cloud
computing datacenter
• Tradeoff between payment proportionality
and bandwidth guarantees
• A mechanism to make tradeoff between
conflicting requirements
19/20
Reference
•
•
•
•
•
•
•
•
•
[1] Amazon web services.
http://aws.amazon.com.
[2] M. Al-Fares, A. Loukissas, and A. Vahdat.
A scalable, commodity data center
network architecture. In SIGCOMM. ACM,
2008.
[3] M. Al-Fares, S. Radhakrishnan, B.
Raghavan, N. Huang, and A. Vahdat.
Hedera: Dynamic Flow Scheduling for Data
Center Networks. In NSDI, 2010.
[4] H. Ballani, P. Costa, T. Karagiannis, and A.
Rowstron. Towards Predictable
Datacenter Networks. In ACM SIGCOMM,
2011.
[5] D. P. Bertsekas and R. Gallager. Data
networks (2. ed.). Prentice Hall, 1992.
[6] B. Briscoe. Flow rate fairness: Dismantling
a religion. ACM SIGCOMM
•
•
•
•
•
•
•
•
Computer Communication Review, 2007.
[7] N. G. Duffield, P. Goyal, A. G. Greenberg,
P. P. Mishra, K. K. Ramakrishnan,
and J. E. van der Merwe. A flexible model for
resource management in virtual
private networks. In SIGCOMM, 1999.
[8] A. Ghodsi, M. Zaharia, B. Hindman, et al.
Dominant resource fairness: fair
allocation of multiple resource types. In
USENIX NSDI, 2011.
[9] A. Greenberg, J. R. Hamilton, N. Jain, S.
Kandula, C. Kim, P. Lahiri, D. A.
Maltz, P. Patel, and S. Sengupta. VL2: A
Scalable and Flexible Data Center
Thanks !!
21/20
Download