BOTTLENECKS Past, Present, Future IN THE BEGINNING Circa 1980: - - PowerPoint PPT Presentation

bottlenecks
SMART_READER_LITE
LIVE PREVIEW

BOTTLENECKS Past, Present, Future IN THE BEGINNING Circa 1980: - - PowerPoint PPT Presentation

BOTTLENECKS Past, Present, Future IN THE BEGINNING Circa 1980: 10Mb/s local network was a fair alternative to a cheap local disk with 5Mb/s transfer and 50-100ms average seek. Multidrop network installations often had problems.


slide-1
SLIDE 1

BOTTLENECKS

Past, Present, Future

slide-2
SLIDE 2

IN THE “BEGINNING”

Circa 1980: 10Mb/s local network was a fair alternative to a cheap local disk with 5Mb/s transfer and 50-100ms average seek. Multidrop network installations often had problems.

slide-3
SLIDE 3

INEFFICIENCIES

File fragmentation and placement limit operational throughput to around 5% of the devices’ maximum 2MB/s. Affordable disk capacities O(1GB) outstripped affordable tape backup (6250 bpi 9-track tape). Poorly performing Ethernet cards were common. Routers could buffer tens of packets. System memory: 20sec. File system: 10min. In theory.

slide-4
SLIDE 4

UNIX AND NSFNET

BSD Unix brought the “Fast File System” with fewer seeks per file (or directory). TCP over NSFnet WAN (56kb – 45Mb/s) was so fragile, files were generally broken into 50–100kB chunks for transfer.

slide-5
SLIDE 5

MID-80’S

Distributed processing becomes feasible on the LAN: uux, Berknet, rsh, SUNRPC, Condor. Small datasets move over the wide area in real time...on dedicated links. Large data sets move

  • ver lunch, or overnight.
slide-6
SLIDE 6

CURRENT PLATFORMS

Servers’ 1–10Gb/s network speeds are well matched to disk or disk array speeds. High CPU power  stealing memory for compute jobs. System memory: 1min. File system: 90min. In theory.

150MB/s 750MB/s 1250MB/s

SERVER DISKS

slide-7
SLIDE 7

THE LHC ERA

HEP groups brought file catalogs, subscriptions, transfer managers: fewer user commands per file. The entire production and analysis chain is so heterogenous (and in some places, fragile), data sets are broken up into 1–5GB files for transfer and processing.

slide-8
SLIDE 8

CURRENT BOTTLENECKS

File servers: Concurrent analysis access is ok: client jobs act compute bound from server’s point of view. Concurrent whole-file (typically WAN) transfer slows all peers.

150MB/s O(10ms) seek ~ 1.5MB

  • f time. Not much ...

... unless interleaved with transfers ≲ 15MB.

slide-9
SLIDE 9

MITIGATING SEEKS FOR SOME READERS

Disk-based file system Silicon overlay file system Fast readers Writers, Slow readers Full copy on

  • pen for read
slide-10
SLIDE 10

SECURITY BOTTLENECKS

GSI requires each endpoint to check several RSA signatures, an O(N2) operation each (for N bit keys). Each endpoint must also generate one signature, an O(N3) operation. To receive a delegation*, the server must do an O(N4)

  • peration.

The work is concentrated on the server. Kerberos does less work, and almost all on the client.

slide-11
SLIDE 11

KERNEL BOTTLENECKS

No load Background load 10, CPU and disk only Throughput ~ 15% Preempted network receiving→delayed TCP processing.

slide-12
SLIDE 12

HARDWARE BOTTLENECKS

Advanced NIC functions RSS: multiple queues, distributed interrupts. Flow Director: packets

  • f a flow pinned to
  • ne core.

! !

!"#$%&%'()*+,%-+)+"."/#%'01)+22%3",4%-55%

Linux sometimes does TCP in interrupt context, sometimes in process: large reordering happens. None of these pins the flow to the Application!

slide-13
SLIDE 13

PUSHING THE BOTTLENECKS DOWN

Network quality File system inefficiency Protocol implementation Network capacity and reliability Fragile middleware Kernel scheduling vs. protocol Multicore, multi-cache, interrupts

O l d p r

  • b

l e m s New problems

slide-14
SLIDE 14