Nested QoS: Providing Flexible Performance in Shared IO Environment - - PowerPoint PPT Presentation

nested qos providing flexible performance in
SMART_READER_LITE
LIVE PREVIEW

Nested QoS: Providing Flexible Performance in Shared IO Environment - - PowerPoint PPT Presentation

Nested QoS: Providing Flexible Performance in Shared IO Environment Hui Wang Peter Varman Rice University Houston, TX 1 Outline l Introduction l System model l Analysis l Evaluation l Conclusions and future work 2 Resource


slide-1
SLIDE 1

Nested QoS: Providing Flexible Performance in Shared IO Environment

Hui Wang Peter Varman

Rice University Houston, TX

1

slide-2
SLIDE 2

Outline

l Introduction l System model l Analysis l Evaluation l Conclusions and future work

2

slide-3
SLIDE 3

Resource consolidation in data centers

l Centralized storage

l Economies of scale l Easier management l High reliability l VM-based server

consolidation

Storage Server Virtualized Host

VMs

VM Scheduling

3

slide-4
SLIDE 4

Issues in resource sharing

l Challenges

l Performance guarantees

§ QoS models

l Resource management l Capacity provisioning l Difficulties:

§

sharing of multiple clients

§

bursty nature of storage workloads

4

slide-5
SLIDE 5

System model for shared I/O

Storage array

Scheduler

Client queues Client 1 Client 2 Client 3 Client n Sharing: The server has to properly allocated resource to concurrent clients to guarantee their performance.

5

slide-6
SLIDE 6

Providing QoS for Bursty Workloads

l Requests have response

time QoS

l Storage workloads are

bursty

  • Large capacity needed to

meet response time during bursts

  • Low average server utilization

l Providing QoS for bursty

workloads which have response time QoS requirement

  • Eg. Open Mail trace, with 100ms window size
  • Average rate:~700 IOPS
  • Peak rate: 4500 IOPS

6

slide-7
SLIDE 7

Related Work

l Proportional Resource Sharing

l Algorithms:

l Fair Queuing, WFQ, WF2Q, Start Time Fair Queuing , Self-Clocking

l Allocate active clients bandwidth (IOPS) in proportion to their

weight wi

l Limitations:

l Response time is not independently controlled

  • Low throughput transactions requiring short response time
  • High throughput file transfer insensitive to response time

l No provisioning for bursts

7

slide-8
SLIDE 8

Related work (cont’d)

l Providing response time guarantees

l Algorithms:

l SCED, pClock

l Client traffic must be within a specified traffic envelope then client

requests are guaranteed a maximum response time of δi

l Limitations:

l No isolation of non-compliant part of workload

§

Loss of QoS guarantee over extended (unbounded) portions

l Only a single response time guarantee is supported

§

Lack of flexibility & high capacity requirement

8

slide-9
SLIDE 9

Performance QoS

l QoS often specified as a percentage of workload

meeting the response time bound

l Absolute percentage guarantees are hard to support

l Can provide response time guarantees if entire workload is

bounded by a traffic envelope

l Requires high capacity

l Guarantee any fixed percentage (say 90%) of the workload

l Unrestricted traffic above the traffic envelope can decrease the

guaranteed percentage arbitrarily

9

slide-10
SLIDE 10

Nested QoS

l We propose:

l Multiple traffic envelops (classes) to describe one bursty

workload

l Performance guarantees based on portion of traffic that satisfies

traffic envelope (not percentage)

l Different performance guarantees for different classes

10

slide-11
SLIDE 11

Outline

l Introduction l System Model l Analysis l Evaluation l Conclusions and future work

11

slide-12
SLIDE 12

Traffic envelopes

z Class 1 (!1, "1, #1) Class 2 (!2, "2, #2) Class 3 (!3, "3, #3)

l Abstract model l Each class i has

l Traffic envelope (Token bucket)

(σi, ρi)

l Response time δi

l Eg: 3-class Nested QoS model

l (30, 120 IOPS, 500ms) l (20, 110 IOPS, 50ms) l (10, 100 IOPS, 5ms)

12

slide-13
SLIDE 13

Token Bucket Regulation

l Traffic Envelope

Arrival Curve Limit

l (σ, ρ) Token Bucket Model

Tokens arrive at rate ρ

  • Bucket of capacity is σ tokens;
  • Arriving request takes a token from the bucket and enters

system

  • Tokens replenished at a constant rate of ρ tokens/sec
  • Maximum number of tokens in bucket is capped at σ
  • A request that arrives when there are no tokens is a

violation of traffic envelope (constraints)

l Service Level Agreement (SLA):

  • Client traffic limited by the Traffic Envelope
  • Response time is guaranteed on requests

13

slide-14
SLIDE 14

Bounding the arrival curve with traffic envelope (token bucket)

!"#$

  • %&#&'()"*$ +,,"*('
  • ".'()"./
  • (

1 +,,"*(' %&,*$ 233$, 4.&/5 Token-bucket regulator: ρ: token-generation rate σ: maximum tokens / instantaneous burst size Maximum # requests arriving in any time interval t: ≤ σ + ρ*t

If the arrival curve lies below the Upper Bound then all requests will meet their deadlines

14

slide-15
SLIDE 15

Architecture in VM environment

Storage Server VM n VM 1

Request Scheduler Request Classifier

Q1 Q1 Q3 Q2

VM 2

. . . . . . . . . . . .

Scheduler in Hypervisor

Q1 Q1 Q3 Q2 Q1 Q1 Q3 Q2

Request Classifier Request Classifier

  • Request Classification
  • Multiple token buckets
  • Request Scheduling
  • Two levels: EDF within

VM queues and FQ across VMs

  • Alternative: 1-level EDF
  • Pros: Capacity &

Simplicity

  • Cons: Low robustness to

capacity variation

15

slide-16
SLIDE 16

Request Classification

Requests Arrival Classifier (!1, "1) Classifier (!2, "2) Classifier (!3, "3) Q1, #1 Q3, #3 Q2, #2

  • Queues
  • Token Buckets

16

slide-17
SLIDE 17

Outline

l Introduction l System Model l Analysis l Evaluation l Conclusions and future work

17

slide-18
SLIDE 18

Analysis

  • Proof see paper.

18

slide-19
SLIDE 19

Outline

l Introduction l System Model l Analysis l Evaluation l Conclusions and future work

19

slide-20
SLIDE 20

Evaluation

l Determine the parameters empirically

l Number of classes & traffic envelope l Tradeoff between capacity required (cost) and performance.

l Workloads

l Block-level workloads from trace repository

20

slide-21
SLIDE 21

Nested QoS for a single workload

!" #!!!" $!!!" %!!!" &!!!" '!!!!" '#!!!" '$!!!" '%!!!" '&!!!"

!"#$"%&'()* !"#$"%&'(+* ,-./&%.0* 12/3* 45'(%.6"* 7%8%'-9:*;<13$=* ()*+),-./0" 01234)-5)6)4"./0"

  • Workloads
  • WebSearch1: (3, 650IOPS, 5ms)
  • WebSearch2: (3, 650IOPS, 5ms)
  • FinTrans: (4, 400 IOPS, 5ms)
  • OLTP: (3, 650IOPS, 5ms)
  • Exchange: (33, 6600IOPS, 5ms)
  • Goal
  • 90% requests in class 1 (5ms)
  • 95% requests in class 2

(50ms)

  • 100% requests in class 3

(500ms)

  • Singe level QoS
  • 100% requests in 5 ms

Capacity Requirement

21

slide-22
SLIDE 22

Nested Nested QoS for a single workload

!!"##$% &#"##$% &'"##$% &("##$% &)"##$% &!"##$% *##"##$% *#'"##$%

!"#$"%&'()* !"#$"%&'()* +,-.&%-/* 01.2* 34'(%-5"* 06"&%77*8"&'"-9%5"*5:%&%-9"";* +%,%*% +%,%'% +%,%-%

Performance for Nested QoS

  • Goal
  • 90% requests in class 1 (5ms)
  • 95% requests in class 2

(50ms)

  • 100% requests in class 3

(500ms)

  • Singe level QoS
  • 100% requests in 5 ms

22

slide-23
SLIDE 23

Nested QoS for Concurrent Workloads

!" !#$" !#%" !#&" !#'" !#(" !#)" !#*" !#+" !#," $"

  • (!./"

(!0$!!./" $!!0%!!./" %!!0(!!./" 1(!!./" !"#$%&'( )*+,&'+*(-./*( 234564785"9:;" <=>:?@" AB%9" !" !#$" !#%" !#&" !#'" !#(" !#)" !#*" !#+" !#," $"

  • (

! . / " ( ! $ ! ! . / " $ ! ! % ! ! . / " % ! ! ( ! ! . / " 1 ( ! ! . / " !"#$%&'( )*+,&'+*(-./*( 234564785"9:;" <=>:?@" AB%9"

  • Two workloads
  • W1:

Web Search; ~350 IOPS

  • W2:

Financial Transaction; ~170 IOPS

  • Total capacity 528 IOPS
  • Response times:
  • 50ms for class 1; 500ms for class 2 and 5000ms for class 3

FinTrans performance WebSearch performance

23

slide-24
SLIDE 24

Nested QoS for Concurrent Workloads

!" !#$" !#%" !#&" !#'" !#(" !#)" !#*" !#+" !#," $"

  • (

! . / " ( ! $ ! ! . / " $ ! ! % ! ! . / " % ! ! ( ! ! . / " 1 ( ! ! . / " !"#$%&$'()*%+)($,-.($ '()*%+)($,-.($ 234564785"9:;" <=>:?@" AB%9" !" !#$" !#%" !#&" !#'" !#(" !#)" !#*" !#+" !#," $"

  • (!./"

(!0$!!./" $!!0%!!./" %!!0(!!./" 1(!!./" !"#$$%&$'()*%+)($,-.($ '()*%+)($,-.($ 234564785"9:;" <=>:?@" AB%9"

  • Two workloads
  • W1:

Web Search; ~350 IOPS

  • W2:

Financial Transaction; ~170 IOPS

  • Total capacity 528 IOPS
  • Response times:
  • 50ms for class 1; 500ms for class 2 and 5000ms for class 3

FinTrans: CDF of Response time WebSearch: CDF of Response time

24

slide-25
SLIDE 25

Outline

l Introduction l System Model l Analysis l Evaluation l Conclusions and future work

25

slide-26
SLIDE 26

Conclusions and future work

l Conclusions

l

Large reduction in server capacity without significant performance loss

l

Analytical estimation of the server capacity

l

Providing flexible SLOs to clients with different performance/cost tradeoffs

l

Providing a conceptual structure of SLOs in workload decomposition

l Future work

l

Workload characteristics for nested model parameters

26