Hakim W Hakim Weather eatherspoon spoon Joint with Lakshmi - - PowerPoint PPT Presentation

hakim w hakim weather eatherspoon spoon joint with
SMART_READER_LITE
LIVE PREVIEW

Hakim W Hakim Weather eatherspoon spoon Joint with Lakshmi - - PowerPoint PPT Presentation

Hakim W Hakim Weather eatherspoon spoon Joint with Lakshmi Ganesh, Tudor Marian, Mahesh Balakrishnan, and Ken Birman File and Storage Technologies (FAST) San Francisco, California February 26 th , 2009 U.S. Department of Treasury Study


slide-1
SLIDE 1

Hakim W Hakim Weather eatherspoon spoon Joint with Lakshmi Ganesh, Tudor Marian, Mahesh Balakrishnan, and Ken Birman File and Storage Technologies (FAST) San Francisco, California February 26th, 2009

slide-2
SLIDE 2

 U.S. Department of Treasury Study

  • Financial Sector vulnerable to significant data loss in disaster
  • Need new technical options

 Risks are real, technology available, Why is problem not solved?

slide-3
SLIDE 3

 Want asynchronous performance to local data center  And want synchronous guarantee

Primary site Remote mirror async sync

Conundrum: there is no middle ground

slide-4
SLIDE 4

 Want asynchronous performance to local data center  And want synchronous guarantee

Primary site Remote mirror sync

Conundrum: there is no middle ground

Local-sync Remote-sync

slide-5
SLIDE 5

 How can we increase reliability of local-sync protocols?

  • Given many enterprises use local-sync mirroring anyways

 Different levels of local-sync reliability

  • Send update to mirror immediately
  • Delay sending update to mirror – deduplication reduces BW
slide-6
SLIDE 6

 Introduction  Enterprise Continuity

  • How data loss occurs
  • How we prevent it
  • A possible solution

 Evaluation  Discussion and Future Work  Conclusion

slide-7
SLIDE 7

Primary site Remote mirror

 Rather, where do failures occur?  Rolling disasters Packet loss Partition Site Failure Power Outage

slide-8
SLIDE 8

Local-sync Network-sync Remote-sync Wide-area network Primary site Remote mirror

slide-9
SLIDE 9

 Use network level redundancy and exposure

  • reduces probability data lost due to network failure

Primary site Remote mirror

Data Packet Repair Packet Network-level Ack Storage-level Ack

slide-10
SLIDE 10

 Network-sync increases data reliability

  • reduces data loss failure modes, can prevent data loss if
  • At the same time primary site fail network drops packet
  • And ensure data not lost in send buffers and local queues

 Data loss can still occur

  • Split second(s) before/after primary site fails…
  • Network partitions
  • Disk controller fails at mirror
  • Power outage at mirror

 Existing mirroring solutions can use network-sync

slide-11
SLIDE 11

 A file system constructed over network-sync

  • Transparently mirrors files over wide-area
  • Embraces concept:

file is in transit (in the WAN link) but with enough recovery data to ensure that loss rates are as low as for the remote disk case!

  • Group mirroring consistency
slide-12
SLIDE 12

B2 B1

append(B1,B2)

V1 R1 I2 B4 B3 I1

append(V1..)

V1 R1 I2 I1 B2 B1 B3 B4

slide-13
SLIDE 13

 Introduction  Enterprise Continuity  Evaluation  Conclusion

slide-14
SLIDE 14

 Demonstrate SMFS performance over Maelstrom

  • In the event of disaster, how much data is lost?
  • What is system and app throughput as link loss increases?
  • How much are the primary and mirror sites allowed to diverge?

 Emulab setup

  • 1 Gbps, 25ms to 100ms link connects two data centers
  • Eight primary and eight mirror storage nodes
  • 64 testers submit 512kB appends to separate logs
  • Each tester submits only one append at a time
slide-15
SLIDE 15

 Local-sync unable to recover data dropped by network  Local-sync+FEC lost data not in transit  Network-sync did not lose any data

  • Represents a new tradeoff in design space

Primary site Remote mirror

  • 50 ms one-way latency
  • FEC(r,c) = (8,3)

Local- sync Network- sync Remote- sync

slide-16
SLIDE 16

 c = 0, No recovery packets: data loss due to packet loss  c = 1, not sufficient to mask packet loss either  c > 2, can mask most packet loss

 Network-sync can prevent loss in local buffers

Primary site Remote mirror

  • 50 ms one-way latency
  • FEC(r,c) = (8,varies)
  • 1% link loss

Local- sync Network- sync Remote- sync

0.1 1 10 100 1000 10000 100000 1 2 3 # Messages Value of C Local-sync+FEC total msgs sent Network-sync total msgs sent Unrecoverable lost msgs

slide-17
SLIDE 17
slide-18
SLIDE 18

 App throughput measures application perceived performance  Network and Local-sync+FEC tput significantly greater than

Remote-sync(+FEC)

slide-19
SLIDE 19
slide-20
SLIDE 20
slide-21
SLIDE 21
slide-22
SLIDE 22

 Introduction  Enterprise Continuity  Evaluation  Discussion and Future Work  Conclusion

slide-23
SLIDE 23

 Do (semi-)private lambda networks drop packets?

  • E.g. Teragrid

 Cornell National Lambda Rail (NLR) Rings testbed

  • Up to 0.5% loss

 Scale network-sync solution to 10Gbps and beyond

  • Commodity (multi-core) hardware
slide-24
SLIDE 24
slide-25
SLIDE 25
slide-26
SLIDE 26
slide-27
SLIDE 27

 Do (semi-)private lambda networks drop packets?

  • E.g. Teragrid

 Cornell National Lambda Rail (NLR) Rings testbed

  • Up to 0.5% loss

 Scale network-sync solution to 10Gbps and beyond

  • Commodity (multi-core) hardware
slide-28
SLIDE 28

 Introduction  Enterprise Continuity  Evaluation  Discussion and Future Work  Conclusion

slide-29
SLIDE 29

 Technology response to critical infrastructure needs  When does the filesystem return to the application?

  • Fast — return after sending to mirror
  • Safe — return after ACK from mirror

 SMFS — return to user after sending enough FEC  Network-sync:

Lossy NetworkLossless NetworkDisk!

 Result: Fast, Safe Mirroring independent of link length!

slide-30
SLIDE 30

 Questions?

Email: hweather@cs.cornell.edu Network-sync code available: http://fireless.cs.cornell.edu/~tudorm/maelstrom Cornell National Lambda Rail (NLR) Rings testbesb http://www.cs.cornell.edu/~hweather/nlr