Adap%ve policies for balancing performance and life%me of mixed SSD - - PowerPoint PPT Presentation

adap ve policies for balancing performance and life me of
SMART_READER_LITE
LIVE PREVIEW

Adap%ve policies for balancing performance and life%me of mixed SSD - - PowerPoint PPT Presentation

Adap%ve policies for balancing performance and life%me of mixed SSD arrays through workload sampling Sangwhan Moon A. L. Narasimha Reddy Texas A&M University 2 / 16 Outline Introduc%on Mixed SSD Arrays Workload distribu%on of


slide-1
SLIDE 1

Adap%ve policies for balancing performance and life%me of mixed SSD arrays through workload sampling

Sangwhan Moon

  • A. L. Narasimha Reddy

Texas A&M University

slide-2
SLIDE 2

2 / 16

Outline

  • Introduc%on

– Mixed SSD Arrays – Workload distribu%on of mixed SSD array

  • Problem Statement
  • Selec%ve caching policies
  • Our approach

– Online sampling – Adap%ve workload distribu%on

  • Evalua%on
  • Conclusion
slide-3
SLIDE 3

3 / 16

Different classes of SSDs

0.1 1 10 100 0.1 1 10 100 Cost ($/GB) Device Writes Per Day (DWPD, higher is be>er)

High-­‑end SSDs Low-­‑end SSDs

slide-4
SLIDE 4

4 / 16

Mixed SSD array

  • High-­‑end SSDs cache

– Faster: PCIe interface – Reliable: SLC eMLC (write endurance = 100K) – Expensive per gigabyte

  • Low-­‑end SSDs main storage

– Slower: Serial ATA interface – Less reliable: MLC TLC (write endurance < 30K) – Cheap per gigabyte

slide-5
SLIDE 5

5 / 16

Workload distribu%on of mixed SSD array

  • LRU Caching Policy

High-­‑end SSDs

Low-­‑end SSDs

3.mrr 1.r 2.w 5.(mrr + mww)⋅d 4.mrr read miss dirty entry evic%on read write wC = mrr + w NC ⋅CC wS = mww NS ⋅CS

min lC wC , lS wS ! " # $ % &

Lifetime =

mr,mw NC, NS CC,CS lC,lS wC,wS Writes per flash cell

Cache read/write miss rate The number of SSDs The capacity of SSD Write endurance of cache/storage

r,w

Read/write workload

slide-6
SLIDE 6

6 / 16

Workload distribu%on of mixed SSD array

  • 1 high-­‑end SSD cache for 3 low-­‑end SSDs

High-­‑end SSDs

Low-­‑end SSDs

3.mrr 1.r 2.w 4.mrr read miss dirty entry evic%on read write

Item DescripKon SpecificaKon High-­‑end SSD (SLC) Capacity 100 GB Write Endurance 100 K Low-­‑end SSD (MLC) Capacity 200 GB Write Endurance 10 K Workload Read/write (MB/s) 100 / 250 Read/write cache hit rate 50% / 15% Read / write length 4KB / 64KB

wC = 0.5⋅100MB / s+ 250MB / s 1⋅100GB wC = 0.85⋅250MB / s 1⋅100GB

min 1.47years, 6.34years ! " # $ % &

Lifetime =

high-­‑end low-­‑end 5.(mrr + mww)⋅d

slide-7
SLIDE 7

7 / 16

Problem statement

  • High-­‑end SSDs cache can wear out faster than low-­‑end

SSDs main storage

– Caching less results in poor performance – Caching more results in poor reliability

  • Sta%c workload classifiers can be less efficient
  • The characteris%cs of workload can change over %me
  • Objec%ves

– Balance the performance and life%me of cache and storage at the same %me

metric : Latency over Life0me (less is be5er)

slide-8
SLIDE 8

8 / 16

  • Request Size based Caching Policy
  • Hotness based Caching Policy

Selec%ve caching policies

I/O requests whose sizes are 4KB are domina%ng 90% of workload is reference

  • nce and never accessed

Sta0c workload classifiers cannot distribute workload across cache and storage precisely

slide-9
SLIDE 9

9 / 16

Selec%ve caching policies

  • Control trade-­‑offs between performance and life%me

Frontend Cache

Backend Storage

2.mrr 1.hrr 6.mwwp 5.(1− p)w 3.mrpr 4.pw read write read miss bypassed writes dirty entry evic%on ProbabilisKc Caching Policy bypassed read miss 7.mr(1− p)r

p (threshold): the probability of caching data

p is more: cache wears out faster, performance enhances p is less: cache wears slower, performance degrades

slide-10
SLIDE 10

10 / 16

LRU Main Storage LRU

Selec%ve Cache

90 % LRU LRU LRU . . . 0.1 0.2 0.3 0.9 1.0

Sampling Cache Sampling Cache Sampling Cache Sampling Cache Sampling Cache

. . . 1% 1% 1% 1% 1% LRU p 1.0 – p . . . Sampling Rate: 10%

Es%mate latency over life%me for each sampling cache Employ best value of p, the proximity of caching

Online sampling

slide-11
SLIDE 11

11 / 16

Simula%on environment

  • Trace-­‑driven simulator
  • Microsog Research Cambridge I/O Block Trace

– 13 enterprise applica%ons trace for a week

  • Cache provisioning = 5%

– Cache size / Storage size

  • Unique data size of workload / Storage Size = 0.5
  • Caching policies

– LRU, size-­‑based (+ sampling), hotness-­‑based (+ sampling), probabilis%c (+ sampling)

slide-12
SLIDE 12

12 / 16 Threshold

0.1 1 10 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1

Threshold latency life%me metric

Threshold

0.1 1 10 0.9 0.8 0.7 0.6 0.5 0.4 0.3 0.2 0.1

Threshold latency life%me metric

StaKc threshold based analysis Sampling based analysis Cache less Cache more Cache less Cache more Cache less Cache more Cache less Cache more

Adap%ve threshold

Hardware monitoring Web server

slide-13
SLIDE 13

13 / 16

Different workload traces

  • Overall, reduced latency over life%me by 60%.

– Very effec%ve on some traces (mds, stg, web, prn, usr, proj, src1, src2) – Less effec%ve on very skewed workload (wdev, rsrch, ts, hm, prxy)

slide-14
SLIDE 14

14 / 16

Different sampling rates

  • Higher sampling rate results in more accurate

es%ma%on (beneficial) and less space for adap%ve cache (harmful)

slide-15
SLIDE 15

15 / 16

Conclusion

  • We showed that high-­‑end SSD cache can wear
  • ut faster than low-­‑end SSD main storage.
  • We proposed sampling based selec%ve

caching to balance the performance and life%me of cache and storage.

  • Trace-­‑based simula%on showed that the

proposed caching policy is effec%ve.

slide-16
SLIDE 16

16 / 16

Q & A