Capability testing of data transfer tools on a high latency 100 - - PowerPoint PPT Presentation

capability testing of data transfer tools on a high
SMART_READER_LITE
LIVE PREVIEW

Capability testing of data transfer tools on a high latency 100 - - PowerPoint PPT Presentation

Capability testing of data transfer tools on a high latency 100 Gbit/s light path Kees de Jong University of Amsterdam MSc System and Network Engineering Research Project 1 Presentation Supervisor: dhr. dr. ing. Leon Gommans (KLM) February 6,


slide-1
SLIDE 1

Capability testing of data transfer tools on a high latency 100 Gbit/s light path

Kees de Jong

University of Amsterdam MSc System and Network Engineering Research Project 1 Presentation Supervisor: dhr. dr. ing. Leon Gommans (KLM)

February 6, 2018

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 1 / 33

slide-2
SLIDE 2

Background

  • Airplanes not only transport people and cargo, but also data
  • Sensor readings
  • Engine data
  • And more. . .
  • Accumulating to several TB’s of data per flight
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 2 / 33

slide-3
SLIDE 3

Background (continued)

  • Critical to transport data fast, to shorten and improve maintenance
  • KLM challenges for the future
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 3 / 33

slide-4
SLIDE 4

Background (continued)

  • Internet is not private nor fast enough
  • 100 Gbit/s path from Amsterdam to Chicago (95 ms RTT)
  • Compare capabilities of high performance GridFTP data transfer tools
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 4 / 33

slide-5
SLIDE 5

Globus GridFTP

  • Globus GridFTP
  • Concurrency (concurrent FTP connections for multiple transfers)
  • Pipelining (latency transparency)
  • Parallelism (divide blocks over multiple transport streams)
  • Third party data transfer
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 5 / 33

slide-6
SLIDE 6

mdtmFTP features

  • Build on top of the Globus GridFTP module
  • Multicore-Aware Data Transfer Middleware
  • Application level scheduler (mostly independent from OS scheduling)
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 6 / 33

slide-7
SLIDE 7

mdtmFTP features (continued)

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 7 / 33

slide-8
SLIDE 8

mdtmFTP features (continued)

  • NUMA: Dedicated NIC and I/O threads + buffers pinned
  • Large virtual file mechanism (LOSF)
  • Direct I/O (disk –> memory)
  • Splice (storage –> NIC)
  • Pipelining
  • Parallelism
  • Third party data transfer
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 8 / 33

slide-9
SLIDE 9

Related work

  • mdtmFTP and Globus GridFTP evaluated by L. Zhang et al.
  • Simulated shared network loop between Chicago and Oakland
  • RTT 95 ms, 100 Gbit/s
  • Concluded that mdtmFTP was on average 20% to 30% faster
  • Globus GridFTP over TCP compared to UDT by John Bresnahan et al.
  • Application level improvement for Globus GridFTP: UDT
  • Tested network with highest latency was 204 ms RTT (ANL to

Auckland)

  • "Best of their knowledge" 1 Gbit/s
  • In most cases UDT outperformed TCP (Reno), often by a factor of 3
  • r 4 in throughput
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 9 / 33

slide-10
SLIDE 10

UDT feature excerpt

  • Application level protocol build on top of UDP
  • Globus XIO module (substitution of transport protocols)
  • Adapts faster to available bandwidth and more features
  • Because this is done in the application layer, it consumes more RAM
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 10 / 33

slide-11
SLIDE 11

Research question

Main research question: "What are the capabilities of mdtmFTP compared to Globus GridFTP on a 100 Gbit/s light path between Amsterdam and Chicago?"

1 Which features and/or design allows optimum throughput? 2 How do these data transfer tools behave with various sets of different file sizes and

quantity?

3 Is the conclusion still valid that Globus GridFTP over UDT outperforms TCP on a

high latency network? And is it enough to beat mdtmFTP?

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 11 / 33

slide-12
SLIDE 12

Methodology

  • Map bottlenecks in the test setup
  • Pinpoint the limitations of the data transfer tools
  • Single and concurrent transfer of a large contiguous file
  • Handling LOSF
  • Transfer of KLM flight data
  • Measure performance/behavior of throughput
  • Throughput on network level
  • TTC on application level
  • Script experiments
  • Drop buffers/caches
  • Repeat tests multiple times (10x)
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 12 / 33

slide-13
SLIDE 13

Network overview

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 13 / 33

slide-14
SLIDE 14

Disk performance

KLM DTN Chicago DTN 1 Chicago DTN 2 Max read speed ∼1500 MB/s ∼1000 MB/s ∼1200 MB/s Max write speed ∼800 MB/s ∼700 MB/s ∼700 MB/s # disks 2 6 6

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 14 / 33

slide-15
SLIDE 15

Results: 100GB (node-to-node + 3rd party)

  • All experiments were done with 4 parallel data streams
  • Anything above 16 parallel streams is regarded wasteful
  • Initial experimentation verified this
  • Globus GridFTP
  • Parallelism
  • mdtmFTP
  • Parallelism
  • Direct I/O (disk –> memory)
  • Splice (storage –> NIC)
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 15 / 33

slide-16
SLIDE 16

Results: 100GB (node-to-node)

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 16 / 33

slide-17
SLIDE 17

Results: 3rd party 100GB (6*100 GB)

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 17 / 33

slide-18
SLIDE 18

Results: 3rd party, GridFTP TCP, TTC=195 sec.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 18 / 33

slide-19
SLIDE 19

Results: 3rd party, GridFTP UDT, TTC=161 sec., 40% diff.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 19 / 33

slide-20
SLIDE 20

Results: 3rd party, mdtmFTP, TTC=520 sec.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 20 / 33

slide-21
SLIDE 21

Results: 3rd party, mdtmFTP, TTC=215 sec., 80% diff.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 21 / 33

slide-22
SLIDE 22

Results: LOSF + KLM

  • Node-to-node
  • 3rd party folder transfer only available for mdtmFTP (crashed)
  • Globus GridFTP
  • Concurrency
  • Pipelining
  • Parallelism
  • mdtmFTP
  • Parallelism
  • Pipelining
  • Virtual file mechanism for LOSF
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 22 / 33

slide-23
SLIDE 23

Results: LOSF GridFTP, concurrency 2

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 23 / 33

slide-24
SLIDE 24

Results: LOSF GridFTP, concurrency 4, 50% diff.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 24 / 33

slide-25
SLIDE 25

Results: LOSF mdtmFTP, with Direct I/O a 30% diff.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 25 / 33

slide-26
SLIDE 26

Results: KLM mdtmFTP, with Direct I/O a 65% diff.

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 26 / 33

slide-27
SLIDE 27

Results: KLM GridFTP, without pipelining

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 27 / 33

slide-28
SLIDE 28

Results: KLM GridFTP, with pipelining

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 28 / 33

slide-29
SLIDE 29

Discussion

  • mdtmFTP still in development
  • Unclear error messages
  • Limited documentation available
  • Large file performance slow
  • High CPU (90%) usage observed, even when idle
  • Limited testing done in a controlled test environment?
  • Large files
  • Globus GridFTP with UDT performed best, 75% faster than

mdtmFTP

  • Did not observe more RAM usage
  • LOSF/KLM data
  • mdtmFTP’s virtual file system greatly benefits performance
  • Globus GridFTP over UDT with concurrency of 2 and pipelining

performs equally with KLM data

  • Network may not have been fully reserved/stable during testing
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 29 / 33

slide-30
SLIDE 30

Conclusion

  • mdtmFTP is a very promising project
  • Needs more testing and improvements
  • Design is capable of more
  • Performed excellent with LOSF
  • Globus GridFTP is here to stay, for now
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 30 / 33

slide-31
SLIDE 31

Future work

  • Test Splice and 3rd party folder transfer
  • Future testing fo mdtmFTP when it matures
  • compare UDT with TCP BBR
  • If implemented, test UDT with mdtmFTP
  • Redo experiment with a hard network reservation
  • K. de Jong (UvA)

RP1: #57 February 6, 2018 31 / 33

slide-32
SLIDE 32

Questions

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 32 / 33

slide-33
SLIDE 33

Network performance baseline (iPerf)

  • K. de Jong (UvA)

RP1: #57 February 6, 2018 33 / 33