on mathematical theories of deep learning
play

On Mathematical Theories of Deep Learning 1 Yuan YAO HKUST - PowerPoint PPT Presentation

On Mathematical Theories of Deep Learning 1 Yuan YAO HKUST Acknowledgement A following-up course at HKUST: https://deeplearning-math.github.io/ Outline Why mathematical theories of Deep Learning? The tsunami of deep learning in


  1. On Mathematical Theories of Deep Learning 1 Yuan YAO HKUST

  2. Acknowledgement A following-up course at HKUST: https://deeplearning-math.github.io/

  3. Outline ´ Why mathematical theories of Deep Learning? ´ The tsunami of deep learning in recent years… ´ What Theories Do We Have or Need? ´ Harmonic Analysis: what are optimal representation of functions? ´ Approximation Theory: when deep networks are better than shallow ones? ´ Optimization: what are the landscapes of risk and how to efficiently find a good optimum? ´ Statistics: how deep net models can generalize well?

  4. Reaching Human Performance Level 1997 2004 AlphaGo “LEE” 2016 Deep Blue in 1997 AlphaGo ”ZERO” ¡D ¡Silver ¡ et ¡al. ¡Nature 550, ¡ 354–359 ¡(2017) ¡doi:10.1038/nature24270

  5. ImageNet Dataset 14,197,122 labeled images 21,841 classes Labeling required more than a year of human effort via Amazon Mechanical Turk

  6. ImageNet Top 5 classification error ImageNet (subset): 1.2 million training images 100,000 test images 1000 classes ImageNet large-scale visual recognition Challenge source: https://www.linkedin.com/pulse/must-read-path-breaking-papers-image-classification-muktabh-mayank 13 /

  7. Crowdcomputing: researchers raising the competition record

  8. Depth as function of year [He et al., 2016]

  9. Growth of Deep Learning

  10. New Moore’s Laws CS231n attendance NIPS registrations

  11. "We’re at the beginning of a new day… This is the beginning of the AI revolution.” — Jensen Huang, GTC Taiwan 2017

  12. Some Cold Water: Tesla Autopilot Misclassifies Truck as Billboard Problem: Why? How can you trust a blackbox?

  13. Deep Learning may be fragile in generalization against noise! [Goodfellow et al., 2014] Small but malicious perturbations can result in severe misclassification Malicious examples generalize across different architectures What is source of instability? Can we robustify network?

  14. Kaggle survey: Top Data Science Methods https://www.kaggle.com/surveys/2017 Academic Industry

  15. What type of data is used at work? https://www.kaggle.com/surveys/2017 Academic Industry

  16. What’s wrong with deep learning? Ali Rahimi NIPS’17: Machine (deep) Learning has become alchemy . https://www.youtube.com/watch?v=ORHFOnaEzPc Yann LeCun CVPR’15, invited talk: What’s wrong with deep learning? One important piece: missing some theory ! http://techtalks.tv/talks/whats-wrong-with-deep-learning/61639/

  17. Perceptron: single-layer Invented by Frank Rosenblatt (1957) b x 1 w 1 x 2 w 2 · f ( z ) z = − → w · − → x + b · · w d x d

  18. Locality or Sparsity of Computation Minsky and Papert, 1969 Perceptron can’t do XOR classification Perceptron needs infinite global locality of computation? information to compute connectivity Locality or Sparsity is important: Locality in time? Locality in space?

  19. Multilayer Perceptrons (MLP) and Back-Propagation (BP) Algorithms Rumelhart, Hinton, Williams (1986) Learning representations by back-propagating errors, Nature, 323(9): 533-536 BP algorithms as stochastic gradient descent algorithms ( Robbins–Monro 1950; Kiefer- Wolfowitz 1951 ) with Chain rules of Gradient maps MLP classifies XOR, but the global hurdle on topology (connectivity) computation still exists

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend