understanding and tackling the hidden memory
play

Understanding and Tackling the Hidden Memory Latency for Edge-based - PowerPoint PPT Presentation

DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) Understanding and Tackling the Hidden Memory Latency for Edge-based Heterogeneous Platform Zhendong Wang, Zhen Wang, Cong Liu, and Yang Hu


  1. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) Understanding and Tackling the Hidden Memory Latency for Edge-based Heterogeneous Platform Zhendong Wang, Zhen Wang, Cong Liu, and Yang Hu Presented by Zhendong Wang HotEdge 2020 Jun. 25, 2020 Pervasive and Emerging Architecture Research Lab, UT Dallas

  2. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 1. Background 2. Motivation and Challenges Edge intelligence Integrated GPU Latency 3. Proposed design 4. Evaluation and Conclusion CPU GPU GPU data data Computation allocation initialization kernel 2

  3. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 1. Background ML/DNN enables a series of edge applications Widely deployed in integrated CPU/GPU(iGPU) platform Size Integrated backed by GPU Weight Power Deployment in iGPUs are stymied Rigorous requirements of (1) Limited memory space memory footprint and e.g., TX2: 8GPU, AGX:16GB processing latency for iGPU platform (2) Application stringent latency requirements e.g., driving automation is safety-critical and latency-sensitive 3

  4. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 1. Background Unified Memory (UM) Management model has relieved the situation (1) Ease memory management (2) Save memory footprint CUDA: cudaMallocManaged() Is current Unified Memory (UM) model good enough? 4

  5. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 2. Motivation Limits of current Unified Memory (UM) model – hidden latency t Init Alloc CPU Def Data processing flow under Def. and UM memory model GPU Execution Init Alloc Def: copy-then-execute memory model CPU UM: unified memory model UM GPU Execution Kernel time Autonomous driving workloads – large matrix operation scale (M.O.S.) --- Matric Addition and Matric Multiplication DNN YOLO2 YOL03 SSD DAVE-2 49K 81K 10K 250K M.O.S 5

  6. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 2. Motivation Limits of current Unified Memory (UM) model – hidden latency t Init Alloc 1. Def: copy-then-execute memory model CPU Others = H2D copy + D2H copy + kernel time Def Init.: data initialization GPU Execution Init Alloc 2. UM: unified memory model CPU Others = kernel time (No copy) UM Init.: data initialization GPU Execution UM still spends excessive time on initialization Def: init ~50% latency UM: init ~90% latency matrix multiplication 6 matrix addition

  7. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 2. Motivation Limits of current Unified Memory (UM) model – hidden latency UM also slows down the computation kernel 1. Def kernel = kernel execution 2. UM = kernel execution + mapping latency matrix multiplication matrix addition t Init Alloc CPU Def GPU Execution Observations: Unnecessary initialize data in Kernel time Init Alloc CPU side CPU UM (1) Save initialization latency GPU (2) Benefit kernel /overall application response Execution performance 7 Kernel time

  8. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 3. Proposed design Enhanced Unified Memory Management (eUMM) (1) Initializing data in GPU side Existing mechanism of legacy Unified Management model GPU-side data initialization in eUMM 8

  9. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 3. Proposed design Enhanced Unified Memory Management (eUMM) (2) Prefetch-enhanced GPU-Init performance 9

  10. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 4. Evaluation Platform: Jetson TX2, Xavier AGX Benchmark: matrix addition, matrix multiplication, Needleman-Wunsch (NW), random access (RA) Faster data initialization Computation kernel is not longer slowed down 10

  11. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) 5. Conclusion Characterization of legacy unified memory management ◆ Initialization latency ◆ Kernel launch latency An enhanced data initialization model based on Unified Memory management (eUMM) ◆ Initializing data in GPU side ◆ Overlapping page mapping with data initialization to further reduce latency 11

  12. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) Prospect & Future work Extend eUMM to a broad spectrum of workloads ◆ Autonomous driving workloads (object detection, object tracking) Reduce the inherent overhead of GPU-side data initialization ◆ GPU-side data initialization does not outperform when data size is small GPUDirect ◆ Bypass CPU to accelerate the communication between GPU and peripheral storage 12

  13. DEPARTMENT OF ELECTRICAL AND COMPUTER ENGINEERING Pervasive and Emerging Architecture Research Lab (PEARL) Thank You If you have any questions, please contact zhendong.wang@utdallas.edu 13

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend