TVM TVM f for ed or edge c e com omputin ting p g pla latf - - PowerPoint PPT Presentation

tvm tvm f for ed or edge c e com omputin ting p g pla
SMART_READER_LITE
LIVE PREVIEW

TVM TVM f for ed or edge c e com omputin ting p g pla latf - - PowerPoint PPT Presentation

TVM TVM f for ed or edge c e com omputin ting p g pla latf tform orm NTT Software Inno nnovation n Ce Center Ka Kazutaka Mo Morita In Inference in 5G era Edge Devices Offload MEC (Mobile edge computing) server Offload


slide-1
SLIDE 1

TVM TVM f for ed

  • r edge c

e com

  • mputin

ting p g pla latf tform

  • rm

NTT Software Inno nnovation n Ce Center

Ka Kazutaka Mo Morita

slide-2
SLIDE 2

Edge Devices

  • In

Inference in 5G era

Base station MEC (Mobile edge computing) server

Cloud Internet

Offload Offload ~10 10 ms ms la latency

slide-3
SLIDE 3
  • Be

Benefits of offloading inference

Device Edge

GPU

Cloud Device Edge Device

AI chip CPU 5G data Big data

Computing resource Inference with data

Edge is one of the targets of AI accelerators High-end server- spec accelerators are available AI chip is unavailable for low-end devices Real-time inference with big data

5G

Interaction with other devices

slide-4
SLIDE 4
  • Ex

Exam ample le – Aug Augmented R d Reality

Object segmentation inference Plane detection

  • bject will not collide

Object detection inference can also provide collider from moving real world objects bouncing object

Occlusion Point cloud

Ma Many Inferenc nce tasks Inferenc nce with h bi big da data in n the he cloud ud

HYPER-REALITY: https://vimeo.com/166807261

Cloud

Point cloud data Captured images

slide-5
SLIDE 5

Internet

  • Edge computing platform with TVM

VM

Developing framework for edge computing

Device SDK Device

TVM VM

Developer Edge Cloud

  • ffload
  • ffload
  • ffload

Offload inference if necessary, based on device and communication status Distribute runtimes to device, edge, and cloud

data data

slide-6
SLIDE 6
  • What are required for TVM

VM?

He Heter erogen eneo eous ru runti time with th

  • f
  • ffl

floa

  • ading

g suppor

  • rt

Dy Dyna namic r run untime Sma Smart NI NIC support

Auto tuning support would be also nice

Switch based on device and communication status Execute on edge via RPC

Edge Edge

Smart NIC CPU CPU GPU NIC FPGA

No overhead of PCIe communication or host memory access

Device Device

On device On edge On device On edge

Scheduler

slide-7
SLIDE 7

Th Thank You!

  • Email: kazutaka.morita.fp@hco.ntt.co.jp