On Average Latency for File Access in Distributed Coded Storage - - PowerPoint PPT Presentation

on average latency for file access in distributed coded
SMART_READER_LITE
LIVE PREVIEW

On Average Latency for File Access in Distributed Coded Storage - - PowerPoint PPT Presentation

On Average Latency for File Access in Distributed Coded Storage Parimal Parag Archana Bura Jean-Fran cois Chamberland Electrical Communication Engineering Indian Institute of Science Electrical and Computer Engineering Texas A&M


slide-1
SLIDE 1

1/ 19

On Average Latency for File Access in Distributed Coded Storage

Parimal Parag Archana Bura Jean-Fran¸ cois Chamberland

Electrical Communication Engineering Indian Institute of Science Electrical and Computer Engineering Texas A&M University

IEEE INFOCOM May 2, 2017

slide-2
SLIDE 2

2/ 19

Dominant traffic on Internet

Peak Period Traffic Composition (North America)

Upstream Downstream Aggregate 20 40 60 80 100 Real-Time Entertainment Web Browsing Marketplaces Filesharing Tunneling Social Networking Storage Communications Gaming Outside Top 5

◮ Real-Time Entertainment: 64.54% for downstream and 36.56

% for mobile access1

1https://www.sandvine.com/downloads/general/global-internet-phenomena/2015/global-internet- phenomena-report-latin-america-and-north-america.pdf

slide-3
SLIDE 3

3/ 19

Established Solutions – Content Delivery Network

Vault File 1 File 2 File 3 File 4 File 5 File 6 File 1 File 3 File 5 File 1 File 4 File 6 File 2 File 3 File 6 File 2 File 4 File 5 Routed Requests

Congestion Prevention and Outage Protection

◮ Mirroring content with local servers ◮ Media file on multiple servers

slide-4
SLIDE 4

4/ 19

Question: Duplication versus MDS Coding

B B A A D C B A

Reduction of access time

◮ How many fragments for a single message? ◮ How to encode and store at the distributed storage nodes?

slide-5
SLIDE 5

5/ 19

Pertinent References (very incomplete)

  • N. B. Shah, K. Lee, and K. Ramchandran, “When do redundant requests reduce latency?” IEEE Trans.

Commun., 2016.

  • G. Joshi, Y. Liu, and E. Soljanin, “On the delay-storage trade-off in content download from coded

distributed storage systems” IEEE Journ. Spec. Areas. Commun., 2014. Dimakis, Godfrey, Wu, Wainwright, and Ramchandran, “Network Coding for Distributed Storage Systems ” IEEE Trans. Info. Theory, 2010.

  • A. Eryilmaz, A. Ozdaglar, M. M´

edard, and E. Ahmed, “On the delay and throughput gains of coding in unreliable networks,” IEEE Trans. Info. Theory, 2008.

  • D. Wang, D. Silva, F. R. Kschischang, “Robust Network Coding in the Presence of Untrusted Nodes”, IEEE
  • Trans. Info. Theory, 2010.
  • A. Dimakis, K. Ramchandran, Y. Wu, C. Suh, “A Survey on Network Codes for Distributed Storage”,

Proceedings of IEEE, 2011. Karp, Luby, Meyer auf der Heide, “Efficient PRAM simulation on a distributed memory machine”, ACM symposium on Theory of computing, 1992. Adler, Chakrabarti, Mitzenmacher, Rasmussen, “Parallel randomized load balancing”, ACM symposium on Theory of computing, 1995. Gardner, Zbarsky, Velednitsky, Harchol-Balter, Scheller-Wolf, “Understanding Response Time in the Redundancy-d System”, SIGMETRICS, 2016.

  • B. Li, A. Ramamoorthy, R. Srikant, “Mean-field-analysis of coding versus replication in cloud storage

systems”, INFOCOM, 2016.

slide-6
SLIDE 6

6/ 19

System Model

File storage

◮ Each media file divided into k pieces ◮ Pieces encoded and stored on n servers

Arrival of requests

◮ Each request wants entire media file ◮ Poisson arrival of requests with rate λ

Time in the system

◮ Till the reception of whole file

Service at each server

◮ IID exponential service time with rate k/n

slide-7
SLIDE 7

7/ 19

Parallel Processing of Requests

D C B A

◮ Service rate available to each request is proportional to

number of servers processing the requests in parallel

slide-8
SLIDE 8

8/ 19

State Space Structure

Keeping Track of Partially Fulfilled Requests

◮ Element of state vector YS(t) is number of users with given

subset S of pieces

Continuous-Time Markov Chain

◮ Y(t) = {YS(t) : S ⊂ [n], |S| < k} is a Markov process

slide-9
SLIDE 9

9/ 19

Priority Scheduling

D C B A D C B A

Mean shortest remaining time processing

◮ Priority to jobs with higher number of pieces ◮ FIFO scheduling among jobs with equal number of pieces

slide-10
SLIDE 10

10/ 19

State Space Collapse

Theorem

For duplication and coding schemes under priority scheduling and parallel processing model, collection S(t) = {S ⊂ [n] : YS(t) > 0, |S| < k}

  • f information subsets is totally ordered in terms of set inclusion

Corollary

Let Yi(t) be number of requests with i information symbols at time t, then Y(t) = (Y0(t), Y1(t), . . . , Yk−1(t)) is Markov process

slide-11
SLIDE 11

11/ 19

State Transitions of Collapsed System

Arrival of Requests

◮ Unit increase in Y0(t) = Y0(t−) + 1 with rate λ

Getting Additional Symbol

◮ Unit increase in Yi(t) = Yi(t−) + 1 ◮ Unit decrease in Yi−1(t) = Yi−1(t−) − 1

Getting Last Missing Symbol

◮ Unit decrease in Yk−1(t) = Yk−1(t−) − 1

slide-12
SLIDE 12

12/ 19

Tandem Queue Interpretation

γ1 Y1(t) γ0 Y0(t) λ

Tandem Queue with Pooled Resources

◮ Servers with empty buffers help upstream ◮ Aggregate service at level i becomes li(t)−1

  • j=i

γj where li(t) = k ∧ {l > i : Yl(t) > 0}

◮ No explicit description of stationary distribution for

multi-dimensional Markov process

slide-13
SLIDE 13

13/ 19

Bounding and Separating

µ1 µ0 λ

Theorem†

When λ < min µi, tandem queue has product form distribution π(y) =

k−1

  • i=0

λ µi

  • 1 − λ

µi yi

Uniform Bounds on Service Rate

Transition rates are uniformly bounded by γi ≤

li(y)−1

  • j=i

γj ≤

k−1

  • j=i

γj Γi

†F. P. Kelly, Reversibility and Stochastic Networks. New York, NY, USA: Cambridge University Press, 2011.

slide-14
SLIDE 14

14/ 19

Bounds on Tandem Queue

γ1 Y1(t) γ0 Y0(t) λ Γ1 Y1(t) Γ0 Y0(t) λ γ1 Y1(t) γ0 Y0(t) λ

Lower Bound

Higher values for service rates yield lower bound on queue distribution π(y) =

k−1

  • i=0

λ Γi

  • 1 − λ

Γi yi

Upper Bound

Lower values for service rate yield upper bound on queue distribution π(y) =

k−1

  • i=0

λ γi

  • 1 − λ

γi yi

slide-15
SLIDE 15

15/ 19

Approximating Pooled Tandem Queue

γ1 Y1(t) γ0 Y0(t) λ ˆ µ1 ˆ Y1(t) ˆ µ0 ˆ Y0(t) λ

Independence Approximation with Statistical Averaging

Service rate is equal to base service rate γi plus cascade effect, averaged over time ˆ µk−1 = γk−1 ˆ µi = γi + ˆ µi+1ˆ πi+1(0) ˆ π(y) =

k−1

  • i=0

λ ˆ µi

  • 1 − λ

ˆ µi yi

slide-16
SLIDE 16

16/ 19

Mean Sojourn Time

0.1 0.2 0.4 0.6 0.8 0.95 5 10 15 Arrival Rate Replication Coding Upper Bound Simulation Approximation Lower Bound 0.1 0.2 0.4 0.6 0.8 0.95 5 10 15 Arrival Rate (4, 2) MDS Code Upper Bound Simulation Approximation Lower Bound ◮ MDS coding significantly outperforms replication ◮ Bounding techniques are only meaningful under light loads ◮ Approximation is accurate over range of loads

slide-17
SLIDE 17

17/ 19

Comparing Replication versus MDS Coding

2 4 8 12 16 20 1 2 3 4 5 Number of Servers Mean Sojourn Time Repetition Simulation Repetition Approximation MDS Simulation MDS Approximation

Arrival rate 0.3 units and coding rate n/k = 2

slide-18
SLIDE 18

18/ 19

1 4 8 12 16 20 24 2 4 6 8 10 Message length k Mean Sojourn Time W Mean Sojourn Time versus Message Length Repetition Coding Simulation Repetition Coding Approximation MDS Coding Simulation MDS Coding Approximation

Figure: For rate λ = 0.45 and n = 24 servers.

slide-19
SLIDE 19

19/ 19

Summary and Discussion

Main Contributions

◮ Analytical framework for study of distributed computation and

storage systems

◮ Upper and lower bounds to analyze replication and MDS codes ◮ A tight closed-form approximation to study distributed storage

codes

◮ MDS codes are better suited for large distributed systems ◮ Mean access time is better for MDS codes for all code-rates