Reconfigurable hardware for big ig data Gustavo Alonso Systems - - PowerPoint PPT Presentation

reconfigurable hardware for big ig data
SMART_READER_LITE
LIVE PREVIEW

Reconfigurable hardware for big ig data Gustavo Alonso Systems - - PowerPoint PPT Presentation

Reconfigurable hardware for big ig data Gustavo Alonso Systems Group Department of Computer Science ETH Zurich, Switzerland www.systems.ethz.ch Systems Group 7 faculty ~40 PhD ~8 postdocs Researching all aspects of system


slide-1
SLIDE 1

Reconfigurable hardware for big ig data

Gustavo Alonso Systems Group Department of Computer Science ETH Zurich, Switzerland

slide-2
SLIDE 2

www.systems.ethz.ch

Systems Group

  • 7 faculty
  • ~40 PhD
  • ~8 postdocs

Researching all aspects of system architecture, sw and hw

slide-3
SLIDE 3

The team behind the work:

David Sidler Zsolt Istvan Kaan Kara Muhsen Owaida

slide-4
SLIDE 4

Data processing today: Appliances Data Centers (Cloud)

slide-5
SLIDE 5

What is a database engine?

  • As complex or more complex than an operating system
  • Full software stack including
  • Parsers, Compilers, Optimizers
  • Own resource management (memory, storage, network)
  • Plugins for application logic
  • Infrastructure for distribution, replication, notifications, recovery
  • Extract, Transform, and Load infrastructure
  • Large legacy, backward compatibility, standards
  • Hugely optimized
slide-6
SLIDE 6

Databases are blindly fast at what they do well

slide-7
SLIDE 7

From Oracle documentation

Databases = think big

ORACLE EXADATA

slide-8
SLIDE 8

Database engine trends: Appliances

Oracle: T7, SQL in Hardware, RAPID SAP: OLTP+OLAP on main memory Hana on SGI supercomputer

SAP Hana on SGI UV 300H SGI documentation

slide-9
SLIDE 9

The challenge of hardware acceleration

slide-10
SLIDE 10

If it sounds too good to be true ..

slide-11
SLIDE 11

Usual unspoken caveats in HW acceleration

  • Where is the data to start with?
  • Where does the data has to be at the end?
  • What happens with irregular workloads?
  • What happens with large intermediate states?
  • What is the architecture?
  • Is the design preventing the system from doing something else?
  • Can the accelerator be multithreaded?
  • Is the gain big enough to justify the additional complexity?
  • Can the gains be characterized?
slide-12
SLIDE 12

Do not replace, enhance Help the CPU to do what it does not do well

slide-13
SLIDE 13

Text search in databases

INTEL HARP: This is an experimental system provided by Intel any results presented are generated using pre- production hardware and software, and may not reflect the performance

  • f production or future systems.

Istvan et al, FCCM’16

slide-14
SLIDE 14

100% processing on FPGA

slide-15
SLIDE 15

Hybrid Processing CPU/FPGA

slide-16
SLIDE 16

Inside a real database …

Sidler et al., SIGMOD’17

slide-17
SLIDE 17

Accelerating real engines

slide-18
SLIDE 18

Accelerators to come

From Oracle M7 documentation

slide-19
SLIDE 19

If the data moves, do it efficiently

Bumps in the wire(s)

slide-20
SLIDE 20

(Woods, VLDB’14)

IBEX

slide-21
SLIDE 21

Sounds good?

The goal is to be able to do this at all levels:

Smart storage On the network switch (SDN like) On the network card (smart NIC) On the PCI express bus On the memory bus (active memory)

Every element in the system (a node, a computer rack, a cluster) will be a processing component

slide-22
SLIDE 22

Disaggregated data center

Near Data Computation

slide-23
SLIDE 23

18-Nov-16 23

Consensus in a Box (Istvan et al, NSDI’16)

Xilinx VC709 Evaluation Board SFP+ SFP+ SFP+ SFP+

DRAM (8GB)

FPGA

Networking Atomic Broadcast Replicated key-value store

Reads Writes SW Clients / Other nodes Other nodes Other nodes TCP Direct Direct

slide-24
SLIDE 24
  • Drop-in replacement for memcached with Zookeeper’s

replication

  • Standard tools for benchmarking (libmemcached)
  • Simulating 100s of clients

24

The system

X 12

10Gbps Switch 3 FPGA cluster Clients

  • Comm. over TCP/IP
  • Comm. over direct

connections

+ Leader election + Recovery

slide-25
SLIDE 25

25

Latency of puts in a KVS

Consensus 15-35μs ~10μs Memaslap (ixgbe) TCP / 10Gbps Ethernet ~3μs Direct connections

slide-26
SLIDE 26

1000 10000 100000 1000000 10000000 1 10 100 1000 Througput (consensus rounds/s) Consensus latency (us) FPGA (Direct) FPGA (TCP) DARE* (Infiniband) Libpaxos (TCP) Etcd (TCP) Zookeeper (TCP)

Specialized solutions

26

The benefit of specialization…

General purpose solutions

[1] Dragojevic et al. FaRM: Fast Remote Memory. In NSDI’14. [2] Poke et al. DARE: High-Performance State Machine Replication on RDMA Networks. In HPDC’15. *=We extrapolated from the 5 node setup for a 3 node setup.

10-100x

slide-27
SLIDE 27

This is the end …

There is a killer application (data science/big data) There is a very fast evolution of the infrastructure for data processing (appliances, data centers) Conventional processors and architectures are not good enough FPGAs great tools to: Explore parallelism Explore new architectures Explore Software Defined X/Y/Z Prototype accelerators