Introduction to Deep Learning
M S Ram
- Dept. of Computer Science & Engg.
Indian Institute of Technology Kanpur
Reading of Chap. 1 from “Learning Deep Architectures for AI”; Yoshua Bengio; FTML Vol. 2, No. 1 (2009) 1–127
1 Date: 12 Nov, 2015
Introduction to Deep Learning M S Ram Dept. of Computer Science - - PowerPoint PPT Presentation
Introduction to Deep Learning M S Ram Dept. of Computer Science & Engg. Indian Institute of Technology Kanpur Reading of Chap. 1 from Learning Deep Architectures for AI; Yoshua Bengio; FTML Vol. 2, No. 1 (2009) 1 127 Date: 12 Nov,
M S Ram
Indian Institute of Technology Kanpur
Reading of Chap. 1 from “Learning Deep Architectures for AI”; Yoshua Bengio; FTML Vol. 2, No. 1 (2009) 1–127
1 Date: 12 Nov, 2015
them in natural language
machines to interact with humans using these concepts
Interactions Scene Description
abstractions with little supervision
Courtesy: Yoshua Bengio, Learning Deep Architectures for AI
2
, Fei-Fei; CVPR 20 2015 15)
"boy is doing backflip
“two young girls are playing with lego toy.” "man in black shirt is playing guitar." "construction worker in
working on road." http://cs.stanford.edu/people/karpathy/deepimagesent/
3
4
5
6
to learn a higher level abstraction
7
input vector
hidden layers
Back-propagate error signal to get derivatives for learning
Compare outputs with correct answer to get error signal
Source: Hinton’s 2009 tutorial on Deep Belief Networks
8
minima
=> 784 * 500 + 500 * 10 ≈ 0.4 million weights
initialization to arrive at the global optimum
Pre-trained N/W Weights Fast unsupervised pre-training Good Solution Slow Fine-tuning (Using Back-propagation) Very slow Back-propagation (Often gets stuck at poor local minima) Random Initial position Very high-dimensional parameter space
series of generative models
for the traditional back-propagation
Intel QuadCore 2.83GHz, 4GB RAM MLP: Python :: DBN: Matlab
underlying the data
multiple features of the previous layer
are not necessarily mutually exclusive
15
Localized Representation Distributed Representation
16
17
Source: https://en.wikipedia.org/wiki/Multi-task_learning
18
Primary
Yoshua Bengio, Learning Deep Architectures for AI, Foundations and Trends in Machine Learning Vol. 2, No. 1 (2009) 1–127 Hinton, G. E., Osindero, S. and Teh, Y. A fast learning algorithm for deep belief nets. Neural Computation 18 (2006), pp 1527-1554 Rumelhart, David E., Geoffrey E. Hinton, and R. J. Williams. Learning Internal Representations by Error Propagation. David E. Rumelhart, James L. McClelland, and the PDP research group. (editors), Parallel distributed processing: Explorations in the microstructure of cognition, Volume 1: Foundations. MIT Press, 1986.
Secondary
Hinton, G. E., Learning Multiple Layers of Representation, Trends in Cognitive Sciences, Vol. 11, (2007) pp 428-434. Hinton G.E., Tutorial on Deep Belief Networks, Machine Learning Summer School, Cambridge, 2009 Andrej Karpathy, Li Fei-Fei. Deep Visual-Semantic Alignments for Generating Image