it s time for low latency
play

Its Time for Low Latency Steve Rumble , Diego Ongaro, Ryan Stutsman, - PowerPoint PPT Presentation

Its Time for Low Latency Steve Rumble , Diego Ongaro, Ryan Stutsman, Mendel Rosenblum, John Ousterhout Stanford University 1 Future Web Applications Need Low Latency They will access more bytes of data Bandwidth problem


  1. It’s Time for Low Latency Steve Rumble , Diego Ongaro, Ryan Stutsman, Mendel Rosenblum, John Ousterhout Stanford University 1

  2. Future Web Applications Need Low Latency ‣ They will access more bytes of data ‣ Bandwidth problem ‣ Commodity net bandwidth has increased > 3,000x in 30 years ‣ But also more pieces of inter-dependent data ‣ Latency problem ‣ Commodity net latency has decreased only ~30x in 30 years ‣ Facebook is a glimpse into future applications ‣ Huge datasets, DRAM-based storage, small requests, random dependent data accesses, low locality ‣ Dependent on network latency: Can only afford 100-150 dependent accesses per page request 2

  3. Datacenter Latency Is Too High Simple RPCs take 300-500us in current datacenters NIC: 2.5-32us NIC: 2.5-32us App Server Kernel Kernel Switches: 100+ us Switches: 100+ us NIC NIC Kernel: 15us Kernel: 15us Server: 1us Component Delay Round-Trip Switch 10-30us/hop 100-300us NIC 2.5-32us 10-128us OS Net Stack 15us 60us Server Code 1us 1us Speed of Light 5ns/m < 2us Not limited by server execution or propagation delay! 3

  4. On The Cusp Of Low Latency ‣ Low latency available in the HPC space (Infiniband) ‣ 100ns switches ‣ < 1us NIC latencies ‣ OS Bypass (U-Net style) ‣ But , won’t displace Ethernet ‣ Some migration into commodity Ethernet space ‣ Fulcrum Microsystems, Mellanox: Sub-500ns switches ‣ RDMA on commodity NICs (e.g. iWarp) ‣ Now we need to pull in the rest of the ideas ‣ Let’s get the OS community involved and do it right ‣ Goal: 5-10us RTTs in the short term 4

  5. An Opportunity To Define The Right Structure OS OS OS Net Stack App NIC App NIC App NIC Net Stack Net Stack Ethernet Infiniband U-Net ‣ Re-think APIs: Apps need speed and simplicity ‣ Infiniband verbs too complex, RDMA too low-level ‣ Developers used to sockets, but can we make them fast? ‣ Network Protocols ‣ Can we live with TCP? (Needs in-order delivery, Slow stacks) ‣ How do we scale low-latency to 100,000+ nodes? ‣ Closed datacenter ecosystem makes new protocols feasible 5

  6. Getting The Lowest Possible Latency The NIC will become the bottleneck under 10us ‣ 500ns round-trip propagation in 50m diameter ‣ 1us round-trip switching latency (10 x 100ns hops) ‣ Even fast NICs take nearly 2us on each end! Today: 5-10 Years: CPU Cache Cache NIC MEM CPU MEM PCIe NIC PCIe accesses & memory accesses too slow Transmit/Receive directly from/to cache One microsecond RTTs possible in 5-10 years 6

  7. Low Latency Is Up To Us ‣ Low latency is the future of web applications ‣ If we don’t take action to make it happen, we risk: ‣ Not getting it at all, or ‣ Missing the opportunity to re-architect (and getting something that sucks) 7

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend