What Is There to Be Skeptical About?
About the Fantastic Success of Deep Learning & The Imperative of Incorporating Priors Into Learning
Oliver Brock Robotics and Biology Laboratory
What Is There to Be Skeptical About? About the Fantastic Success of - - PowerPoint PPT Presentation
What Is There to Be Skeptical About? About the Fantastic Success of Deep Learning & The Imperative of Incorporating Priors Into Learning Oliver Brock Robotics and Biology Laboratory The Fantastic Success of Deep Learning Lots of blowout
What Is There to Be Skeptical About?
About the Fantastic Success of Deep Learning & The Imperative of Incorporating Priors Into Learning
Oliver Brock Robotics and Biology Laboratory
The Fantastic Success of Deep Learning
► Lots of blowout successes in CV and beyond ► End-to-end ► Pixels to Torque ► By using a general function, a gradient, and lots of data
DNNs
Is it Science? Engineering? Protoscience? Art?
Hennig Brand (1630–1692, depicted: 1669)
The Alchymist, in Search
Stone, Discovers Phosphorus, and prays for the successful Conclusion of his
custom of the Ancient Chymical Astrologers (1771, 1795) Joseph Wright of Derby (1734–1797)
David Wolpert, William Macready and many others…
1.36 × 1050 bits/(s kg)
Bremermann’s Limit
1075ops/s
10103ops/s
10200 À 10103 600 ¿ 300.000.000
The Imperative of Incorporating Priors Into Learning
► Assumption: NFL captures a property of the problem
space
► Hypothesis 1: The problems we want to solve are
actually really simple and the NFL does not apply
► Hypothesis 2: Neural nets capture exactly the right prior
for the type of problems we want to solve
► Hypothesis 3: Hypothesis 1 & 2 together ► Hypothesis 4: We must incorporate task-specific priors
into learning 4
I am the deep learning hammer!
Towards Combined Tools
Merging Two Ends of the Spectrum
Rico Jonschkowski
Incorporate Priors Into Learning
► Representation Learning with Newton’s Laws ► Learning with Side Information
Rico Jonschkowski and Oliver Brock. Learning State Representations with Robotic Priors. Autonomous Robots. Springer US 39(3):407-428, 2015.
Direct Pattern Multi-Task Pattern
Rico Jonschkowski, Sebastian Höfer, and Oliver Brock. Patterns for Learning with Side Information. arXiv:1511.06429. 2016.
Rico Jonschkowski Sebastian Höfer Rico Jonschkowski
Learning With Side Information: A Simple Example
f: 3 → 14 f: 5 → 30 f: 2 → 9 What is f?
Rico Jonschkowski, Sebastian Höfer, and Oliver Brock. Patterns for Learning with Side Information. arXiv:1511.06429. 2016.
Learning With Side Information: A Simple Example
f: 3 → 9 → 14 f: 5 → 25 → 30 f: 2 → 4 → 9 What is f?
Rico Jonschkowski, Sebastian Höfer, and Oliver Brock. Patterns for Learning with Side Information. arXiv:1511.06429. 2016.
Problem and illustrations adapted from [Thrun, Burgard, Fox: Probabilistic Robotics. MIT Press,2006.]
Toy problem: 1D hallway navigation with identical doors
Case Study: Recursive State Estimation
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Neural Histogram Filter
Measurement Model Motion Model Motion Update Measurement Update
Learned Models (from 1600 training steps)
Measurement Model Motion Model Motion Update Measurement Update
Learned Models (from 1600 training steps)
Measurement Model Motion Model Motion Update Measurement Update
Learned Models (from 1600 training steps)
NHF predictions for 32 test steps (from 1600 training steps)
NHF predictions for 32 test steps (from 1600 training steps)
NHF
Qualitative Comparison (models from 1600 training steps)
NHF RNN 2x RNN LSTM 2x LSTM
Qualitative Comparison (models from 1600 training steps)
First Results
Conclusion
I am the Gaussian process saw! I am the SVM screw driver! We are all important!