cse neubeh 528 modeling synapses and networks
play

CSE/NEUBEH 528 Modeling Synapses and Networks (Chapter 7) Image - PDF document

CSE/NEUBEH 528 Modeling Synapses and Networks (Chapter 7) Image from Wikimedia Commons R. Rao, 528: Lecture 9 1 Lecture figures are from Dayan & Abbotts book Course Summary (thus far) F Neural Encoding What makes a neuron fire? (STA,


  1. CSE/NEUBEH 528 Modeling Synapses and Networks (Chapter 7) Image from Wikimedia Commons R. Rao, 528: Lecture 9 1 Lecture figures are from Dayan & Abbott’s book Course Summary (thus far) F Neural Encoding What makes a neuron fire? (STA, covariance analysis) Poisson model of spiking F Neural Decoding Spike-train based decoding of stimulus Stimulus Discrimination based on firing rate Population decoding (Bayesian estimation) F Single Neuron Models RC circuit model of membrane Integrate-and-fire model Conductance-based Models R. Rao, 528: Lecture 9 2

  2. Today’s Agenda F Computation in Networks of Neurons Modeling synaptic inputs From spiking to firing-rate based networks Feedforward Networks Linear Recurrent Networks R. Rao, 528: Lecture 9 3 SYNAPSES ! R. Rao, 528: Lecture 9 4 Image Credit: Kennedy lab, Caltech. http://www.its.caltech.edu/~mbkla

  3. What do synapses do? Spike Increase or decrease postsynaptic membrane potential R. Rao, 528: Lecture 9 5 Image Source: Wikimedia Commons An Excitatory Synapse Input spike  Spike Neurotransmitter release (e.g., Glutamate)  Binds to receptors  Ion channels open  positive ions (e.g. Na+) enter cell  Depolarization due to EPSP (excitatory postsynaptic potential) R. Rao, 528: Lecture 9 6 Image Source: Wikimedia Commons

  4. An Inhibitory Synapse Input spike  Spike Neurotransmitter release (e.g., GABA)  Binds to receptors  Ion channels open  positive ions (e.g., K+) leave cell  Hyperpolarization due to IPSP (inhibitory postsynaptic potential) R. Rao, 528: Lecture 9 7 Image Source: Wikimedia Commons Flashback Membrane Model V  dV ( V E ) I    L e c , or equivalent ly m  m = r m c m = R m C m dt r A m membrane time dV      ( V E ) I R constant m L e m dt R. Rao, 528: Lecture 9 8

  5. Flashback! The Integrate-and-Fire Model V Models a passive leaky membrane dV      ( V E ) I R   E 70 mV m L e m L dt (resting potential) If V > V threshold  Spike   V 50 mV threshold Then reset: V = V reset  V E reset L R. Rao, 528: Lecture 9 9 Flashback! Hodgkin-Huxley Model L K Na dV I    e c i m m dt A       4 3 ( ) ( ) ( ) i g V E g n V E g m h V E m L , max L K , max K Na , max Na E L = -54 mV, E K = -77 mV, E Na = +50 mV R. Rao, 528: Lecture 9 10

  6. How do we model the effects of a synapse on the membrane potential V ? Synapse ? R. Rao, 528: Lecture 9 11 Modeling Synaptic Inputs Synaptic V conductance Synapse dV        ( V E ) r g ( V E ) I R m L m s s e m dt  g g P P Probability of postsynaptic channel opening s s , max rel s (= fraction of channels opened) Probability of transmitter release given an input spike R. Rao, 528: Lecture 9 12

  7. Basic Synapse Model F Assume P rel = 1 fraction of channels F Model the effect of a single spike input on P s opened F Kinetic Model of postsynaptic channels:    s Open Closed    s dP      s ( 1 P ) P s s s s dt Closing rate Opening rate Fraction of channels open Fraction of channels closed R. Rao, 528: Lecture 9 13 What does P s look like over time? t   1  K ( t ) e s  s Exponential function K ( t ) gives reasonable fit to biological data (other options: difference of exponentials, “alpha” function) R. Rao, 528: Lecture 9 14

  8. Linear Filter Model of Synaptic Input to a Neuron Synaptic weight Input Spike w b Train  b (t)  b ( t ) =  i δ ( t-t i ) ( t i are the input spike times) t   1  Filter for synapse b : K ( t ) e s  s    I ( t ) w K ( t t ) Synaptic current for b : b b i  t t i t        w K ( t ) ( ) d b b   R. Rao, 528: Lecture 9 15 Modeling Networks of Neurons F Option 1: Use spiking neurons Advantages : Model computation and learning based on: Spike Timing Spike Correlations/Synchrony between neurons Disadvantages : Computationally expensive F Option 2: Use neurons with firing-rate outputs (real valued outputs) Advantages : Greater efficiency, scales well to large networks Disadvantages : Ignores spike timing issues F Question: How are these two approaches related? R. Rao, 528: Lecture 9 16

  9. From Spiking to Firing Rate Models w 1 w N Spikes Spikes  1 (t)  N (t)   I ( t ) I ( t ) Total synaptic current s b b t   Spike train  b (t)       I ( t ) w K ( t ) ( ) d s b b   b t        Firing rate u b (t) w K ( t ) u ( ) d b b   b R. Rao, 528: Lecture 9 17 Synaptic Current Dynamics in Firing Rate Model t   1  K ( t ) e F Suppose synaptic kernel K is exponential: s  s t        Differentiating w.r.t. time t , I ( t ) w K ( t ) u ( ) d s b b   b dI      s we get I w u s s b b dt b    w  I u s R. Rao, 528: Lecture 9 18

  10. Output Firing-Rate Dynamics F How is the output firing rate v related to synaptic inputs? dv dI         w  s v F ( I ( t )) I u s s r s dt dt F Looks very much like membrane equation: dV      ( V E ) I R m L e m dt F On-board derivations of special cases obtained from comparing the relative magnitudes of  r and  s … (see also pages 234-236 in the text) R. Rao, 528: Lecture 9 19 How good are Firing Rate Models? Input I(t) = I 0 + I 1 cos(  t) Firing rate model v(t) = F(I(t)) describes this well but not this case R. Rao, 528: Lecture 9 20

  11. Feedforward versus Recurrent Networks d v      v F ( W u M v ) dt Output Decay Input Feedback For feedforward networks, matrix M = 0 R. Rao, 528: Lecture 9 21 Example: Linear Feedforward Network d v     Dynamics: v W u dt Steady State  v W u (set d v / dt to 0): ss    1 0 0 0 1     1    1 1 0 0 0   2      0 1 1 0 0     What is v ss ?   W u 2     0 0 1 1 0  2      0 0 0 1 1       1     1 0 0 0 1  R. Rao, 528: Lecture 9 22

  12. Linear Feedforward Network      1 0 0 0 1 0       1  1 1 0 0 0   1     2        0 1 1 0 0 0          v ss W u 2     0 0 1 1 0   0   2        0 0 0 1 1 1         1          1 0 0 0 1 0 What is the network doing? R. Rao, 528: Lecture 9 23 Linear Filtering for Edge Detection     Filter 0 1 1 0 0 (and shifted versions in W)   0   1     1 Input   2     0       Input 2 Output   0   2      1       1   0 Output R. Rao, 528: Lecture 9 24

  13. Example of Edge Detection in a 2D Image http://www.alexandria.nu/ai/blog/entry.asp?E=51 R. Rao, 528: Lecture 9 25 Edge detectors in the visual system Examples of receptive fields in primary visual cortex (V1) V1 (From Nicholls et al., 1992) R. Rao, 528: Lecture 9 26

  14. Filtering network is computing derivatives!     df f ( x h ) f ( x )   0 1 1 0 0 lim  dx h h 0    Discrete approximat ion f ( x 1 ) f ( x )       2 d f f ( x h ) f ( x )   0 1 2 1 0 lim 2  dx h h 0           Disc. approx. f ( x 1 ) f ( x ) f ( x ) f ( x 1 )      f ( x 1 ) 2 f ( x ) f ( x 1 ) R. Rao, 528: Lecture 9 27 Feedforward Networks: Example 2 Coordinate Transformation Output: Premotor Cortex Neuron with Body-Based Tuning Curves Input: Area 7a Neurons with Gaze-Dependent Tuning Curves (From Section 7.3 in Dayan & Abbott) R. Rao, 528: Lecture 9 28

  15. Output of Coordinate Transformation Network Head fixed; gaze shifted to g 1 g 2 g 3 Same tuning curve regardless of gaze angle Premotor cortex neuron responds to stimulus location relative to body , not retinal image location R. Rao, 528: Lecture 9 (See section 7.3 in Dayan & Abbott for details) 29 Linear Recurrent Networks d v     W  v u M v dt Output Decay Input Feedback R. Rao, 528: Lecture 9 30

  16. Next Class: Recurrent Networks F To Do: Homework 2 Find a final project topic and partner(s) R. Rao, 528: Lecture 9 31

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend