E9 205: Machine Learning for Signal Processing Introduction to - - PowerPoint PPT Presentation

e9 205 machine learning for signal processing
SMART_READER_LITE
LIVE PREVIEW

E9 205: Machine Learning for Signal Processing Introduction to - - PowerPoint PPT Presentation

E9 205: Machine Learning for Signal Processing Introduction to 16-10-2019 Neural Network Models Perceptron Algorithm Perceptron Model [McCulloch, 1943, Rosenblatt, 1957] Targets are binary classes [-1,1] What if the data is not linearly


slide-1
SLIDE 1

E9 205: Machine Learning for Signal Processing

16-10-2019

Introduction to Neural Network Models

slide-2
SLIDE 2

Perceptron Algorithm

What if the data is not linearly separable Perceptron Model [McCulloch, 1943, Rosenblatt, 1957] Targets are binary classes [-1,1]

slide-3
SLIDE 3

Multi-layer Perceptron

Multi-layer Perceptron [Hopfield, 1982] thresholding function non-linear function (tanh,sigmoid)

slide-4
SLIDE 4

Neural Networks

Multi-layer Perceptron [Hopfield, 1982] thresholding function non-linear function (tanh,sigmoid)

  • Useful for classifying non-linear data boundaries -

non-linear class separation can be realized given enough data.

slide-5
SLIDE 5

Neural Networks

tanh sigmoid ReLu Cost-Function are the desired outputs Mean Square Error Cross Entropy Types of Non-linearities

slide-6
SLIDE 6

Learning Posterior Probabilities with NNs

Choice of target function

  • Softmax function for classification
  • Softmax produces positive values that sum to 1
  • Allows the interpretation of outputs as posterior

probabilities

slide-7
SLIDE 7

Parameter Learning

Error function for entire data Typical Error Surface as a function of parameters (weights and biases)

slide-8
SLIDE 8

Parameter Learning

Non-linear nature of error function

  • Move in the reverse

direction of the gradient Error back propagation Error surface close to a local optima

slide-9
SLIDE 9

Parameter Learning

  • Solving a non-convex
  • ptimization.
  • Iterative solution.
  • Depends on the initialization.
  • Convergence to a local
  • ptima.
  • Judicious choice of learning

rate