SLIDE 7 Why using FPGA for accelerating neural networks?
13
- The inference process for neural networks can be performed using
different computing devices; ○ These alternatives can be classified on different dimensions: ■ Performances, energy efficiency, ease of programming.
CPU GPU ASIC (TPU) FPGA
implementing an ISA and having a programmable datapath;
towards data-parallel (SIMD/SIMT) processing.
fabric implementing digital logic;
frequency;
logic having the highest efficiency.
reconfigured when the needs change. Easy to program / less efficient More efficient / less flexible
Why using FPGA for accelerating neural networks?
14
- The main advantage of FPGA is their flexibility:
- Potentially, it is possible to implement the inference process using a
hardware accelerator custom-designed for the specific NN. ○ Efficient datapath tailored for the specific inference process. ○ If the network change the accelerator can be redesigned. ■ Same chip fit different types of machine learning models;
- As a result, FPGA-accelerated inference of NN is typically more
energy-efficient compared to CPUs and GPUs; ○ High throughput without ramping up the clock frequency.
Ease of development Performance per Watt CPU GPU FPGA HDL FPGA HLS