the only constant is change
play

The Only Constant is Change: Incorporating Time-Varying Bandwidth - PowerPoint PPT Presentation

The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1 Cloud Computing is Hot Private Cluster 2 Key Factors for Cloud Viability Cost


  1. The Only Constant is Change: Incorporating Time-Varying Bandwidth Reservations in Data Centers Di Xie, Ning Ding, Y. Charlie Hu, Ramana Kompella 1

  2. Cloud Computing is Hot Private Cluster 2

  3. Key Factors for Cloud Viability • Cost • Performance 3

  4. Performance Variability in Cloud • BW variation in cloud Bandwidth (Mbps) 1000 due to contention 900 [Schad’10 VLDB] 800 700 600 500 • Causing unpredictable 400 performance 300 200 100 0 Local Cluster Amazon EC2 4

  5. Reserving BW in Data Centers • SecondNet [Guo’10] – Per VM-pair, per VM access bandwidth reservation • Oktopus [Ballani’11] – Virtual Cluster (VC) – Virtual Oversubscribed Cluster (VOC) 5

  6. How BW Reservation Works Only fixed-BW reservation Request <N, B> Bandwidth B Time Virtual Switch 0 T . . . N VMs Virtual Cluster Model 2. Allocate and enforce the model 1. Determine the model 6

  7. Network Usage for MapReduce Jobs Hadoop Sort, 4GB per VM 7

  8. Network Usage for MapReduce Jobs Hadoop Sort, 4GB per VM Hadoop Word Count, 2GB per VM 8

  9. Network Usage for MapReduce Jobs Hadoop Sort, 4GB per VM Hadoop Word Count, 2GB per VM Hive Join, 6GB per VM 9

  10. Network Usage for MapReduce Jobs Hadoop Sort, 4GB per VM Hadoop Word Count, 2GB per VM Hive Join, 6GB per VM Hive Aggregation, 2GB per VM 10

  11. Network Usage for MapReduce Jobs Time-varying network usage Hadoop Sort, 4GB per VM Hadoop Word Count, 2GB per VM Hive Join, 6GB per VM Hive Aggregation, 2GB per VM 11

  12. Motivating Example • 4 machines, 2 VMs/machine, non-oversubscribed Not enough 1Gbps BW network 500Mbps 500Mbps • Hadoop Sort – N: 4 VMs – B: 500Mbps/VM 12

  13. Motivating Example • 4 machines, 2 VMs/machine, non-oversubscribed 1Gbps network 500Mbps • Hadoop Sort – N: 4 VMs – B: 500Mbps/VM 13

  14. Under Fixed-BW Reservation Model 1Gbps Bandwidth 500 500Mbps Job1 Job2 Job3 Time 0 5 10 15 20 25 30 Virtual Cluster Model 14

  15. 15

  16. 16

  17. 17

  18. 18

  19. 19

  20. Temporally-Interleaved Virtual Cluster (TIVC) • Key idea: Time-Varying BW Reservations • Compared to fixed-BW reservation – Improves utilization of data center • Better network utilization • Better VM utilization – Increases cloud provider’s revenue – Reduces cloud user’s cost – Without sacrificing job performance 20

  21. Challenges in Realizing TIVC Q1: What are right model functions? Q2: How to automatically derive the models? Bandwidth Bandwidth B B Time Time Virtual Switch 0 T 0 T Request Request <N, B> <N, B(t)> . . . N VMs Virtual Cluster Model 21

  22. Challenges in Realizing TIVC Q3: How to efficiently allocate TIVC? Q4: How to enforce TIVC? 22

  23. Challenges in Realizing TIVC • What are the right model functions? • How to automatically derive the models? • How to efficiently allocate TIVC? • How to enforce TIVC? 23

  24. Challenges in Realizing TIVC • What are the right model functions? • How to automatically derive the models? • How to efficiently allocate TIVC? • How to enforce TIVC? 24

  25. How to Model Time-Varying BW? Hadoop Hive Join 25

  26. TIVC Models B B Bandwidth Bandwidth B b B b 0 0 T 1 T 2 T T 11 T 12 T 21 T 22 T 31 T 32 T Time Time B Bandwidth Virtual Cluster B b 0 T 32 T 11 T 11 T 12 T 21 T 22 T 31 T 32 T Time 26

  27. Hadoop Sort 27

  28. Hadoop Word Count v 28

  29. Hadoop Hive Join 29

  30. Hadoop Hive Aggregation 30

  31. Challenges in Realizing TIVC  What are the right model functions? • How to automatically derive the models? • How to efficiently allocate TIVC? • How to enforce TIVC? 31

  32. Possible Approach • “White - box” approach – Given source code and data of cloud application, analyze quantitative networking requirement – Very difficult in practice • Observation: Many jobs are repeated many times – E.g., 40% jobs are recurring in Bing’s production data center [Agarwal’12] – Of course, data itself may change across runs, but size remains about the same 32

  33. Our Approach • Solution: “Black - box” profiling based approach 1. Collect traffic trace from profiling run 2. Derive TIVC model from traffic trace • Profiling: Same configuration as production runs – Same number of VMs How much BW – Same input data size per VM should we give to – Same job/VM configuration the application? 33

  34. Impact of BW Capping 34

  35. Impact of BW Capping No-elongation BW threshold 35

  36. Choosing BW Cap • Tradeoff between performance and cost – Cap > threshold: same performance, costs more – Cap < threshold: lower performance, may cost less • Our Approach: Expose tradeoff to user 1. Profile under different BW caps Only below 2. Expose run times and cost to user threshold ones 3. User picks the appropriate BW cap 36

  37. From Profiling to Model Generation • Collect traffic trace from each VM – Instantaneous throughput of 10ms bin • Generate models for individual VMs • Combine to obtain overall job’s TIVC model – Simplify allocation by working with one model – Does not lose efficiency since per-VM models are roughly similar for MapReduce-like applications 37

  38. Generate Model for Individual VM 1. Choose B b 2. Periods where B > B b , set to B cap B cap BW B b Time 38

  39. Maximal Efficiency Model Applicatio n Traffic Volume Efficiency  • Reserved Bandwdith Volume • Enumerate B b to find the maximal efficiency model B cap BW B b Time 39

  40. Challenges in Realizing TIVC  What are the right model functions?  How to automatically derive the models? • How to efficiently allocate TIVC? • How to enforce TIVC? 40

  41. TIVC Allocation Algorithm • Spatio-temporal allocation algorithm – Extends VC allocation algorithm to time dimension – Employs dynamic programming • Properties – Locality aware – Efficient and scalable • 99 th percentile 28ms on a 64,000-VM data center in scheduling 5,000 jobs 41

  42. Challenges in Realizing TIVC  What are the right model functions?  How to automatically derive the models?  How to efficiently allocate TIVC? • How to enforce TIVC? 42

  43. Enforcing TIVC Reservation • Possible to enforce completely in hypervisor – Does not have control over upper level links – Requires online rate monitoring and feedback – Increases hypervisor overhead and complexity • Observation: Few jobs share a link simultaneously – Most small jobs will fit into a rack – Only a few large jobs cross the core – In our simulations, < 26 jobs share a link in 64,000-VM data center 43

  44. Enforcing TIVC Reservation • Enforcing BW reservation in switches – Avoid complexity in hypervisors – Can be implemented on commodity switches • Cisco Nexus 7000 supports 16k policers 44

  45. Challenges in Realizing TIVC  What are the right model functions?  How to automatically derive the models?  How to efficiently allocate TIVC?  How to enforce TIVC? 45

  46. Proteus: Implementing TIVC Models 1. Determine the model 2. Allocate and enforce the model 46

  47. Evaluation • Large-scale simulation – Performance – Cost – Allocation algorithm • Prototype implementation – Small-scale testbed 47

  48. Simulation Setup • 3-level tree topology – 16,000 Hosts x 4 VMs 50Gbps – 4:1 oversubscription 20 Aggr Switch … 10Gbps • Workload 20 ToR Switch … … – N: exponential distribution 1Gbps 40 Hosts around mean 49 … … … … – B(t): derive from real Hadoop apps 48

  49. Batched Jobs • Scenario: 5,000 time-insensitive jobs 1/3 of each type All rest results are for mixed Completion 42% 21% 23% 35% time reduction 49

  50. Varying Oversubscription and Job Size 25.8% reduction for non-oversubscribed network 50

  51. Dynamically Arriving Jobs • Scenario: Accommodate users’ requests in shared data center – 5,000 jobs, Poisson arrival, varying load Rejected: VC: 9.5% TIVC: 3.4% 51

  52. Analysis: Higher Concurrency • Under 80% load 28% higher 7% higher job 28% higher Rejected jobs Charge VM utilization VMs are large concurrency revenue VM 52

  53. Tenant Cost and Provider Revenue • Charging model – VM time T and reserved BW volume B – Cost = N ( k v T + k b B ) Amazon target – k v = 0.004$/hr, k b = 0.00016$/GB utilization 12% less cost for tenants Providers make more money 53

  54. Testbed Experiment • Setup – 18 machines – Tc and NetFPGA rate limiter • Real MapReduce jobs • Procedure – Offline profiling – Online reservation 54

  55. Testbed Result Baseline suffers elongation, TIVC finishes job faster than VC, TIVC achieves similar Baseline finishes the fastest performance as VC 55

  56. Conclusion • Network reservations in cloud are important – Previous work proposed fixed-BW reservations – However, cloud apps exhibit time-varying BW usage • We propose TIVC abstraction – Provides time-varying network reservations – Uses simple pulse functions – Automatically generates model – Efficiently allocates and enforces reservations • Proteus shows TIVC benefits both cloud provider and users significantly 56

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend