encoding and decoding neural information
play

Encoding and decoding neural information Encoding : building - PDF document

CSE/NB 528 Final Lecture: All Good Things Must CSE/NB 528: Final Lecture 1 Course Summary Where have we been? Course Highlights Where do we go from here? Challenges and Open Problems Further Reading CSE/NB 528: Final


  1. CSE/NB 528 Final Lecture: All Good Things Must… CSE/NB 528: Final Lecture 1 Course Summary • Where have we been? • Course Highlights • Where do we go from here? • Challenges and Open Problems • Further Reading CSE/NB 528: Final Lecture 2

  2. What is the neural code? What is the nature of the code? Representing the spiking output: single cells vs populations rates vs spike times vs intervals What features of the stimulus does the neural system represent? CSE/NB 528: Final Lecture 3 Encoding and decoding neural information Encoding : building functional models of neurons/neural systems and predicting the spiking output given the stimulus Decoding : what can we say about the stimulus given what we observe from the neuron or neural population? CSE/NB 528: Final Lecture 4

  3. Key concepts: Poisson & Gaussian Spike trains are variable Models are probabilistic Deviations are close to independent CSE/NB 528: Final Lecture 5 Highlights: Neural Encoding spike-triggering stimulus features f 1 multidimensional x 1 decision function spiking output r (t) stimulus X(t) f 2 x 2 f 3 x 3 CSE/NB 528: Final Lecture 6

  4. Highlights: Finding the feature space of a neural system Gaussian prior stimulus distribution covariance STA Spike-conditional distribution CSE/NB 528: Final Lecture 7 Highlights: Finding an interesting tuning curve P(s) P(s|spike) P(s) P(s|spike) s s CSE/NB 528: Final Lecture 8

  5. Decoding: Signal detection theory z p(r|-) p(r|+) <r> + <r> - Decoding corresponds to comparing test to threshold. a (z) = P[ r ≥ z| -] false alarm rate, “size” b (z) = P[ r ≥ z|+] hit rate, “power” CSE/NB 528: Final Lecture 9 Highlights: Neurometric curves CSE/NB 528: Final Lecture 10

  6. Decoding from a population e.g. cosine tuning curves RMS error in estimate Theunissen & Miller, 1991 CSE/NB 528: Final Lecture 11 More general approaches: MAP and ML MAP: s* which maximizes p[s|r] ML: s* which maximizes p[r|s] Difference is the role of the prior: differ by factor p[s]/p[r] For cercal data: CSE/NB 528: Final Lecture 12

  7. Highlights: Information maximization as a design principle of the nervous system CSE/NB 528: Final Lecture 13 The biophysical basis of neural computation CSE/NB 528: Final Lecture 14

  8. Excitability is due to the properties of ion channels • Voltage dependent • transmitter dependent (synaptic) • Ca dependent CSE/NB 528: Final Lecture 15 Highlights: The neural equivalent circuit Ohm’s law: and Kirchhoff’s law - Capacitive Ionic currents Externally current applied current CSE/NB 528: Final Lecture 16

  9. Simplified neural models A sequence of neural models of increasing complexity that approach the behavior of real neurons Integrate and fire neuron: subthreshold, like a passive membrane spiking is due to an imposed threshold at V T Spike response model: subthreshold, arbitrary kernel spiking is due to an imposed threshold at V T postspike, incorporates afterhyperpolarization Simple model: complete 2D dynamical system spiking threshold is intrinsic have to include a reset potential CSE/NB 528: Final Lecture 17 Simplified models: integrate-and-fire V Integrate-and- Fire Model dV      ( ) V E I R m L e m dt If V > V threshold  Spike Then reset: V = V reset CSE/NB 528: Final Lecture 18

  10. Simplified models: spike response model Gerstner; Keat et al. 2001 CSE/NB 528: Final Lecture 19 Highlights: Dendritic computation Filtering Shunting Delay lines Information segregation Synaptic scaling Direction selectivity CSE/NB 528: Final Lecture 20

  11. Highlights: Compartmental models Neuronal structure can be modeled using electrically coupled compartments Coupling conductances CSE/NB 528: Final Lecture 21 Connecting neurons: Synapses Spike Presynaptic spikes cause neurotransmitters to cross the cleft and bind to postsynaptic receptors, allowing ions to flow in and change postsynaptic potential CSE/NB 528: Final Lecture 22

  12. EPSPs and IPSPs Size of PSP is a measure of synaptic strength Can vary on the short term due to input history Long term due to synaptic plasticity (LTP/LTD) CSE/NB 528: Final Lecture 23 Networks CSE/NB 528: Final Lecture 24

  13. Modeling Networks of Neurons d v      ( W M ) v F u v dt Output Decay Input Feedback CSE/NB 528: Final Lecture 25 Highlights: Unsupervised Learning •   T T w u u w For linear neuron: v w  • d Basic Hebb Rule:  u v w dt • Average effect over many inputs: Hebb rule performs principal component analysis (PCA) d w    u v Q w w w dt • Q is the input correlation matrix:  T Q uu CSE/NB 528: Final Lecture 26

  14. Highlights: Generative Models Mathematical Droning lecture Lack of sleep derivations CSE/NB 528: Final Lecture 27 Highlights: Generative Models and the Connection to Statistics Unsupervised learning = learning the hidden causes of input data G = (  v,  v ) Causes v [ | ; ] p v [ ; G ] p v u G (posterior) (prior) Causes of clustered Generative Use EM data algorithm for model learning the parameters G “Causes” of natural images [ | ; ] p u v G Data u (data likelihood) CSE/NB 528: Final Lecture 28

  15. Highlights: Supervised Learning: Neurons as Classifiers Perceptron: Weighted Sum Threshold Inputs u j Output v i (-1 or +1) (-1 or +1) u 2 Separating hyperplane: u 1 CSE/NB 528: Final Lecture 29 Highlights: Supervised Learning: Regression Backpropagation for Multilayered Networks    m m ( ( )) v g W g w u i ij jk k j k Finds W and w that minimize errors:   1  m m 2 ( , ) ( ) E W w d v ij jk i i 2 , m i Desired output Example: Truck backer upper m u k CSE/NB 528: Final Lecture 30

  16. Highlights: Reinforcement Learning • Learning to predict delayed rewards (TD learning): (http://employees.csbsju.edu/tcreed/pb/pdoganim.html)           ( ) ( ) [ ( ) ( 1 ) ( )] ( ) w w r t v t v t u t • Actor-Critic Learning: 2.5 1 • Critic learns value of each state using TD learning • Actor learns best actions based on value of next state (using the TD error) CSE/NB 528: Final Lecture 31 The Future: Challenges and Open Problems • How do neurons encode information? • Topics : Synchrony, Spike-timing based learning, Dynamic synapses • How does a neuron’s structure confer computational advantages? • Topics : Role of channel dynamics, dendrites, plasticity in channels and their density • How do networks implement computational principles such as efficient coding and Bayesian inference? • How do networks learn “optimal” representations of their environment and engage in purposeful behavior? • Topics : Unsupervised/reinforcement/imitation learning CSE/NB 528: Final Lecture 32

  17. Further Reading (for the summer and beyond) • Spikes: Exploring the Neural Code, F. Rieke et al., MIT Press, 1997 • The Biophysics of Computation, C. Koch, Oxford University Press, 1999 • Large-Scale Neuronal Theories of the Brain , C. Koch and J. L. Davis, MIT Press, 1994 • Probabilistic Models of the Brain , R. Rao et al., MIT Press, 2002 • Bayesian Brain , K. Doya et al., MIT Press, 2007 • Reinforcement Learning: An Introduction , R. Sutton and A. Barto, MIT Press, 1998 CSE/NB 528: Final Lecture 33 Next meeting: Project presentations! • Project presentations will be on Monday, June 10, 10:30am-12:20pm in the same classroom • Keep your presentation short: ~8 slides, 8 mins/group • Slides: • Bring your slides on a USB stick to use the class laptop OR • Bring your own laptop if you have videos etc. • Projects reports (10-15 pages total) due by midnight Tuesday, June 11 (by email to both Adrienne and Raj) CSE/NB 528: Final Lecture 34

  18. Have a great summer! Au revoir! CSE/NB 528: Final Lecture 35

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend