neural network backpropagation
play

Neural Network Backpropagation 3-2-16 Recall from Monday... - PowerPoint PPT Presentation

Neural Network Backpropagation 3-2-16 Recall from Monday... Perceptrons can only classify linearly separable data. Multi-layer networks Can represent any boolean function. We dont want to build them by hand, so we need a way to


  1. Neural Network Backpropagation 3-2-16

  2. Recall from Monday... Perceptrons can only classify linearly separable data.

  3. Multi-layer networks ● Can represent any boolean function. ● We don’t want to build them by hand, so we need a way to train them. ● Algorithm: backpropagation. ○ You’ve already seen this in action in yesterday’s lab.

  4. Backpropagation networks sigmoid ● Backpropagation can be applied to activation functions any directed acyclic neural network. -0.5 ● Activation functions must be 0.2 differentiable. 2.7 0.8 ● Activation functions should be non- 3.0 -0.3 -1.9 linear. -1.2 -1.6 OK not OK 1.5 2.2 0.1 ● Layered networks allow training to be parallelized within each layer.

  5. Sigmoid activation functions ● We want something like a threshold. ○ Neuron is inactive below the threshold; active above it. ● We need something differentiable. ○ Required for gradient descent.

  6. Gradient descent ● Define the squared error at each output node as: ● Update weights to reduce error. ○ Take a step in the direction of steepest descent: w 0 derivative of w 1 learning rate error w.r.t. weight

  7. Computing the error gradient … algebra ensues ...

  8. Gradient descent step for output nodes 2 1.04 2 1 -.97 1.2 -1 1.2

  9. Backpropagation Key idea: at hidden units, use the next-layer change instead of the error function. ● Determine the node’s contribution to its successors. w 0 w 1 ● Update incoming weights using this “error”

  10. Backpropagation algorithm for 1:training runs for example in training_data: run example through network compute error for each output node for each layer (starting from output): for each node in layer: gradient descent update on incoming weights

  11. Exercise: run a backprop step on this network 2 -0.5 0.2 2.7 t = 0.1 0.8 -0.3 3.0 0 -1.2 -1.6 1.5 t = 0.8 0.1 -1

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend