SLIDE 10 5 10 15 20
Training epoch
59 60 61 62 63 64 65
Test Error (%) (a) DNN (BN50) (Speech)
Training with Single Precision (FP32) Training with DLFloat (FP16)
50 100 150 200
Training epoch
10 20 30 40 50 60
Test Error (%) (b) ResNet32 (CIFAR10) (Image)
Training with Single Precision (FP32) Training with DLFloat (FP16)
20 40 60 80
Training epoch
20 40 60 80 100
Test Error (%) (c) ResNet50 (Imagenet) (Image)
Training with Single Precision (FP32) Training with DLFloat (FP16)
10 20 30 40 50
Training epoch
40 50 60 70 80
Test Error (%) (d) AlexNet (Imagenet) (Image)
Training with Single Precision (FP32) Training with DLFloat (FP16)
Results – comparison with Baseline (IEEE-32)
indistinguishable from baseline
did not need to adjust network hyper-parameters to obtain good convergence
development to be decoupled from compute precision in hardware