nested qos providing flexible performance in
play

Nested QoS: Providing Flexible Performance in Shared IO Environment - PowerPoint PPT Presentation

Nested QoS: Providing Flexible Performance in Shared IO Environment Hui Wang Peter Varman Rice University Houston, TX 1 Outline l Introduction l System model l Analysis l Evaluation l Conclusions and future work 2 Resource


  1. Nested QoS: Providing Flexible Performance in Shared IO Environment Hui Wang Peter Varman Rice University Houston, TX 1

  2. Outline l Introduction l System model l Analysis l Evaluation l Conclusions and future work 2

  3. Resource consolidation in data centers l Centralized storage l Economies of scale l Easier management VM l High reliability Scheduling VMs l VM-based server consolidation Storage Server Virtualized Host 3

  4. Issues in resource sharing l Challenges l Performance guarantees § QoS models l Resource management l Capacity provisioning l Difficulties: sharing of multiple clients § bursty nature of storage workloads § 4

  5. System model for shared I/O Client queues Client 1 Client 2 Client 3 Scheduler Storage array Client n Sharing : The server has to properly allocated resource to concurrent clients to guarantee their performance. 5

  6. Providing QoS for Bursty Workloads l Requests have response time QoS l Storage workloads are bursty Large capacity needed to • meet response time during bursts Low average server utilization • l Providing QoS for bursty workloads which have response time QoS requirement Eg. Open Mail trace, with 100ms window size • Average rate:~700 IOPS • Peak rate: 4500 IOPS 6

  7. Related Work l Proportional Resource Sharing l Algorithms: l Fair Queuing, WFQ, WF2Q, Start Time Fair Queuing , Self-Clocking l Allocate active clients bandwidth (IOPS) in proportion to their weight w i l Limitations: l Response time is not independently controlled Low throughput transactions requiring short response time • High throughput file transfer insensitive to response time • l No provisioning for bursts 7

  8. Related work (cont’d) l Providing response time guarantees l Algorithms: l SCED, p Clock l Client traffic must be within a specified traffic envelope then client requests are guaranteed a maximum response time of δ i l Limitations: l No isolation of non-compliant part of workload Loss of QoS guarantee over extended (unbounded) portions § l Only a single response time guarantee is supported Lack of flexibility & high capacity requirement § 8

  9. Performance QoS l QoS often specified as a percentage of workload meeting the response time bound l Absolute percentage guarantees are hard to support l Can provide response time guarantees if entire workload is bounded by a traffic envelope l Requires high capacity l Guarantee any fixed percentage (say 90%) of the workload l Unrestricted traffic above the traffic envelope can decrease the guaranteed percentage arbitrarily 9

  10. Nested QoS l We propose: l Multiple traffic envelops (classes) to describe one bursty workload l Performance guarantees based on portion of traffic that satisfies traffic envelope (not percentage) l Different performance guarantees for different classes 10

  11. Outline l Introduction l System Model l Analysis l Evaluation l Conclusions and future work 11

  12. Traffic envelopes l Abstract model Class 2 ( ! 2, " 2, # 2 � ) Class 3 l Each class i has ( ! 3, " 3, # 3) l Traffic envelope (Token bucket) ( σ i, ρ i) Class 1 l Response time δ i z ( ! 1, " 1, # 1) l Eg: 3-class Nested QoS model l (30, 120 IOPS, 500ms) l (20, 110 IOPS, 50ms) l (10, 100 IOPS, 5ms) 12

  13. Token Bucket Regulation l Traffic Envelope Tokens arrive at rate ρ Arrival Curve Limit l ( σ , ρ ) Token Bucket Model • Bucket of capacity is σ tokens; • Arriving request takes a token from the bucket and enters system • Tokens replenished at a constant rate of ρ tokens/sec • Maximum number of tokens in bucket is capped at σ • A request that arrives when there are no tokens is a violation of traffic envelope (constraints) l Service Level Agreement (SLA): • Client traffic limited by the Traffic Envelope • Response time is guaranteed on requests 13

  14. Bounding the arrival curve with traffic envelope (token bucket) � Token-bucket regulator: ρ : token-generation rate %&#&'()"*$ +,,"*(' 233$, 4.&/5 σ : maximum tokens / -".'()"./ instantaneous burst size � � Maximum # requests arriving in any time interval t: ≤ σ + ρ *t +,,"*(' %&,*$ � 1 ( 0 !"#$ If the arrival curve lies below the Upper Bound then all requests will meet their deadlines 14

  15. Architecture in VM environment • Request Classification • Multiple token buckets VM 1 VM 2 VM n . . . . . . Request Request Request Scheduler in Hypervisor Classifier Classifier Classifier • Request Scheduling • Two levels: EDF within VM queues and FQ Q1 Q1 Q1 . . . . . . across VMs Q3 Q3 Q3 Q1 Q2 Q1 Q2 Q1 Q2 • Alternative: 1-level EDF Request Scheduler • Pros: Capacity & Simplicity Storage • Cons: Low robustness to Server capacity variation 15

  16. Request Classification • Queues Classifier Classifier Classifier ( ! 3, " 3) ( ! 2, " 2) ( ! 1, " 1) Requests Arrival • Token Buckets Q1, # 1 Q2, # 2 Q3, # 3 16

  17. Outline l Introduction l System Model l Analysis l Evaluation l Conclusions and future work 17

  18. Analysis • Proof see paper. 18

  19. Outline l Introduction l System Model l Analysis l Evaluation l Conclusions and future work 19

  20. Evaluation l Determine the parameters empirically l Number of classes & traffic envelope l Tradeoff between capacity required (cost) and performance. l Workloads l Block-level workloads from trace repository 20

  21. Nested QoS for a single workload • Workloads '&!!!" • WebSearch1: (3, 650IOPS, 5ms) '%!!!" • WebSearch2: (3, 650IOPS, 5ms) '$!!!" • FinTrans: (4, 400 IOPS, 5ms) 7%8%'-9:*;<13$=* '#!!!" • OLTP: (3, 650IOPS, 5ms) • Exchange: (33, 6600IOPS, 5ms) '!!!!" &!!!" ()*+),-./0" %!!!" • Goal $!!!" 01234)-5)6)4"./0" • 90% requests in class 1 (5ms) #!!!" • 95% requests in class 2 !" !"#$"%&'()* !"#$"%&'(+* 12/3* 45'(%.6"* ,-./&%.0* (50ms) • 100% requests in class 3 (500ms) Capacity Requirement • Singe level QoS • 100% requests in 5 ms 21

  22. Nested Nested QoS for a single workload • Goal *#'"##$% • 90% requests in class 1 (5ms) 06"&%77*8"&'"-9%5"*5:%&%-9"";* *##"##$% • 95% requests in class 2 &!"##$% (50ms) • 100% requests in class 3 &)"##$% (500ms) +%,%*% &("##$% +%,%'% &'"##$% • Singe level QoS +%,%-% &#"##$% • 100% requests in 5 ms !!"##$% !"#$"%&'()* !"#$"%&'()* +,-.&%-/* 01.2* 34'(%-5"* Performance for Nested QoS 22

  23. Nested QoS for Concurrent Workloads • Two workloads • W1: Web Search; ~350 IOPS • W2: Financial Transaction; ~170 IOPS • Total capacity 528 IOPS • Response times: • 50ms for class 1; 500ms for class 2 and 5000ms for class 3 $" $" !#," !#," !#+" !#+" !#*" !#*" !#)" !#)" !"#$%&'( !"#$%&'( !#(" !#(" 234564785"9:;" 234564785"9:;" !#'" !#'" <=>:?@" !#&" !#&" <=>:?@" AB%9" !#%" !#%" !#$" !#$" AB%9" !" !" -(!./" (!0$!!./" $!!0%!!./" %!!0(!!./" 1(!!./" " " " " " / / / / / . . . . . ! ! ! ! ! ( ! ! ! ! - $ % ( ( 0 0 0 1 ! ! ! ( ! ! $ % )*+,&'+*(-./*( )*+,&'+*(-./*( WebSearch performance FinTrans performance 23

  24. Nested QoS for Concurrent Workloads • Two workloads • W1: Web Search; ~350 IOPS • W2: Financial Transaction; ~170 IOPS • Total capacity 528 IOPS • Response times: • 50ms for class 1; 500ms for class 2 and 5000ms for class 3 $" $" !#," !#," !#+" !#+" !"#$$%&$'()*%+)($,-.($ !"#$%&$'()*%+)($,-.($ !#*" !#*" !#)" !#)" !#(" !#(" 234564785"9:;" 234564785"9:;" !#'" !#'" <=>:?@" <=>:?@" !#&" !#&" AB%9" AB%9" !#%" !#%" !#$" !#$" !" !" -(!./" (!0$!!./" $!!0%!!./" %!!0(!!./" 1(!!./" " " " " " / / / / / . . . . . ! ! ! ! ! ( ! ! ! ! - $ % ( ( 0 0 0 1 ! ! ! ( ! ! $ % '()*%+)($,-.($ '()*%+)($,-.($ WebSearch: CDF of Response time FinTrans: CDF of Response time 24

  25. Outline l Introduction l System Model l Analysis l Evaluation l Conclusions and future work 25

  26. Conclusions and future work l Conclusions Large reduction in server capacity without significant performance loss l Analytical estimation of the server capacity l Providing flexible SLOs to clients with different performance/cost tradeoffs l Providing a conceptual structure of SLOs in workload decomposition l l Future work Workload characteristics for nested model parameters l 26

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend