Chapter 13. Neurodynamics
Neural Networks and Learning Machines (Haykin)
Lecture Notes on
Self-learning Neural Algorithms
Byoung-Tak Zhang School of Computer Science and Engineering Seoul National University
Version 20171109
Chapter 13. Neurodynamics Neural Networks and Learning Machines - - PowerPoint PPT Presentation
Chapter 13. Neurodynamics Neural Networks and Learning Machines (Haykin) Lecture Notes on Self-learning Neural Algorithms Byoung-Tak Zhang School of Computer Science and Engineering Seoul National University Version 20171109 Contents 13 .1
Version 20171109
13.1 Introduction ……………………………………………………..…………………………….... 3 13.2 Dynamic Systems ………………………………………………….…….………………..…. 5 13.3 Stability of Equilibrium States ……………………………....………………….…….... 8 13.4 Attractors ……….………..………………….……………………………………..………..... 14 13.5 Neurodynamic Models ……………………………..…………………………...…...…. 15 13.6 Attractors and Recurrent Networks ….……..……..……….…….……………….. 19 13.7 Hopfield Model ….………………………………………..……..…….……….………….. 20 13.8 Cohen-Grossberg Theorem ……………….…………..………….………….....……. 27 13.9 Brain-State-In-A-Box Model ………………………….……..…......................…. 29 Summary and Discussion …………….…………….………………………….……………... 33
(c) 2017 Biointelligence Lab, SNU 2
(c) 2017 Biointelligence Lab, SNU 3
itself in the learning process:
1. A static neural network (NN) is made into a dynamic mapper by stimulating it via a memory structure, short term or long term. 2. Time is built into the operation of a neural network through the use of feedback.
1. Local feedback, which is applied to a single neuron inside the network; 2. Global feedback, which encompasses one or more layers of hidden neurons—or better still, the whole network.
improperly, it can produce harmful effects. In particular, the application of feedback can cause a system that is originally stable to become unstable. Our primary interest in this chapter is in the stability of recurrent networks.
particular emphasis on the stability problem, is referred to as neurodynamics.
(c) 2017 Biointelligence Lab, SNU 4
– The presence of stability always implies some form of coordination between the individual parts of the system.
– Deterministic neurodynamics, in which the neural network model has a deterministic behavior. Described by a set of nonlinear differential equations This chapter. – Statistical neurodynamics, in which the neural network model is perturbed by the presence of noise. Described b y stochastic nonlinear differential equations, thereby expressing the solution in probabilistic terms.
(c) 2017 Biointelligence Lab, SNU 5
j j j
d x t F x t j N dt
1 2
[ , , ... , ]T
N
t x t x t x t x
d t t dt x F x
Figure 13.1 A two-dimensional trajectory (orbit) of a dynamic system.
(c) 2017 Biointelligence Lab, SNU 6
Figure 13.2 A two-dimensional state (phase) portrait of a dynamic system. Figure 13.3 A two-dimensional vector field of a dynamic system.
(c) 2017 Biointelligence Lab, SNU 7
( ) ( ) K F x F u x u
M x x. x u Lipschitz condition Let denote the norm, or Euclidean length, of the vector Let and be a pair of vectors in an open set in a noraml vector (state) space. Then, acco K M x u rding to the Lipschitz condition, there exists a constant for all and in .
S V
If the divergence ( ) (which is a scalar) is zero, the system is conservative and if it is negative the system is dissipative , , . F x
Divergence Theorem
(c) 2017 Biointelligence Lab, SNU 8
Table 13.1
F x ( ) = + ( ) t t x x x ( ) + ( ) t F x x A x = ( )|
x x
A F x x ( ) ( ) d t t dt x A x
(c) 2017 Biointelligence Lab, SNU 9
Figure 13.4 (a) Stable node. (b) Stable focus. (c) Unstable node. (d) Unstable
(c) 2017 Biointelligence Lab, SNU 10
The equilibrium state is stable if, in a small neighborhood of there exists a positive - definite function such that its derivative with respect to time is negative semidefinite in t V Theorem1. x x, (x) hat region. The equilibrium state is asymptotically stable if, in a small neighborhood of , there exists a positive - definite function such that its derivative with respect to time is negative de V Theorem2. x x (x) finite in that region.
(c) 2017 Biointelligence Lab, SNU 11
Figure 13.5 Illustration of the notion of uniform stability of a state vector.
(c) 2017 Biointelligence Lab, SNU 12
0 for x d V u dt x x According to Theorem 1,
x d V u dt x x According to Theorem 2,
( ) ( ) . ( ) = 0 V V V x x x x Requirement : The Lyapunov function to be a positive - definite function
has continous partial derivatives with respect to the elements
2. . ( ) > 0
V u u x x x x 3. if where is a small neighborhood around
(c) 2017 Biointelligence Lab, SNU 13
1 2 3
c < < x
c c c
Figure 13.6 Lyapunov surfaces for decreasing value of constant , with . The equilibribum state is denoted by the point .
(c) 2017 Biointelligence Lab, SNU 14
which is defined by the set of equations
which regions of initial conditions of a nonzero state-space volume converge as time t increases.
1 2
1, 2, ... , , , ... , 0,
j N
j k M x x x k N
Figure 13.7 Illustration of the notion of a basin of attraction and the idea of a separatrix.
(c) 2017 Biointelligence Lab, SNU 15
1. A large number of degrees of freedom
The human cortex is a highly parallel, distributed system that is estimated to possess about 10 billion neurons, with each neuron modeled by one or more state variables. The system is characterized by a very large number of coupling constants represented by the strengths (efficacies) of the individual synaptic junctions.
2. Nonlinearity
A neurodynamic system is inherently nonlinear. In fact, nonlinearity is essential for creating a universal computing machine.
3. Dissipation
A neurodynamic system is dissipative. It is therefore characterized by the convergence
4. Noise
Finally, noise is an intrinsic characteristic of neurodynamic systems. In real-life neurons, membrane noise is generated at synaptic junctions
(c) 2017 Biointelligence Lab, SNU 16
Figure 13.8 Additive model of a neuron, labeled j.
(c) 2017 Biointelligence Lab, SNU 17
1 N j j ji i j i j j
j j
, 1
= 1, 2, ...,
N j j j ji i j i j
dv t v t C w x t I j N dt R
j j
Additive model
(c) 2017 Biointelligence Lab, SNU 18
= φ , = 1, 2, ...,
j j ji i j i
dv t v t w v t I j N dt
j
dx t = x t φ x t , = 1, 2, ..., dt
j ji i j i
w K j N
Related model
=
k kj j j
v t w x t
=
k kj j j
I w K
(c) 2017 Biointelligence Lab, SNU 19
(c) 2017 Biointelligence Lab, SNU 20
set of neurons and a corresponding set of unit-time delays, forming a multiple-loop feedback system.
network, we use the neurodynamic model which is based on the additive model of a neuron.
Figure 13.9 Architectural graph of a Hopfield network consisting of N = 4 neurons.
1 1
N N ji i j i j i j
(c) 2017 Biointelligence Lab, SNU 21
The Discrete Hopfield Model as a Content-Addressable Memory
1 N j ji i j i
where bj is a fixed bias applied externally to neuron j. Hence, neuron j modifies its state xj according to the deterministic rule
1 if for 1 if for
j j j
v x v
j j
(c) 2017 Biointelligence Lab, SNU 22
The Discrete Hopfield Model as a Content-Addressable Memory Storage of samples (= learning) – According to the outer-product rule of storage—that is, the generalization of Hebb’s postulate of learning—the synaptic weight from neuron i to neuron j is defined by
, , 1
1
M ji j i
w N
0 for all
ii
w i Figure 13.13 Illustration of the encoding–decoding performed by a recurrent network.
(c) 2017 Biointelligence Lab, SNU 23
The Discrete Hopfield Model as a Content-Addressable Memory 1. Storage Phase
– The output of each neuron in the network is fed back to all other neurons. – There is no self-feedback in the network (i.e.,wii = 0). – The weight matrix of the network is symmetric, as shown by the following relation
2. Retrieval Phase
1
1
M
M N
W Ι
W W
, j = 1, 2, …, N
1
sgn
N j ji i j i
y w y b
sgn( ) y Wy b
(c) 2017 Biointelligence Lab, SNU 24
(c) 2017 Biointelligence Lab, SNU 25 Figure 13.14 (a) Architectural graph of Hopfield network for N = 3 neurons. (b) Diagram depicting the two stable states and flow
, j = 1, 2, …, N
j j j
(c) 2017 Biointelligence Lab, SNU 26
Spurious States
viewpoint of the discrete Hopfield network used as a content-addressable memory
1. The discrete Hopfield network acts as a vector projector in the sense that it projects a probe onto a subspace spanned by the fundamental memory vectors. 2. The underlying dynamics of the network drive the resulting projected vector to one of the corners of a unit hypercube where the energy function is minimized.
1. The need to preserve the fundamental memory vectors as fixed points in the state space. 2. The desire to have few spurious states.
(c) 2017 Biointelligence Lab, SNU 27
, j = 1, 2, …, N
1
N j j j j j ji i i i
1 1 1
j
N N N u ji i i j j j j i j j
ij ji
j j
j j j j j
(c) 2017 Biointelligence Lab, SNU 28
The Hopfield Model as a Special Case of the Cohen–Grossberg Theorem
1 1 1
j
N N N v j ji i i j j j j i j j j
(c) 2017 Biointelligence Lab, SNU 29
Figure 13.15 (a) Block diagram of the brain-state-in-a-box (BSB) model. (b) Signal-flow graph of the linear associator represented by the weight matrix W.
( ) ( ) ( ) n n n y x Wx
( 1) ( ( )) n n x y
1 if ( ) 1 ( 1) ( ( )) ( ) if 1 ( ) 1 1 if ( ) 1
j j j i i i
y n x n y n y n y n y n
(c) 2017 Biointelligence Lab, SNU 30
BSB Model
Figure 13.16 Piecewise-linear activation function used in the BSB model.
1
( 1) ( ) , 1,2, ,
N j ji i i
x n c x n j N
ji ji ji
c w
1
( ) ( ) ( ) , 1,2, ,
N j j ji i i
d x t x t c x t j N dt
1
( ) ( )
N j ji i i
x t c v t
1
( ) ( )
N j ji i i
v t c x t
1
( ) ( ) ( ( )), 1,2, ,
N j j ji i i
d v t v t c v t j N dt
Continuous form Equivalent form
To transform this into the additive model, we introduce
31
Lyapunov (Energy) Function of BSB
1 1
2 2
N N ji j i i j
E w x x
x Wx
1 1 1
1 ( ) ( ) ( ) 2
j
N N N v ji i j i j j
E c v v v v dv
Dynamics of the BSB Model
that minimizes the energy function E.
unit huypercube and its orgin.
W should be diagonal dominant, which means that
to be a fixed point attractor, the weight matrix W should be strongly diagonal dominant:
w for 1,2, ,
jj ij i j
w j N
w + for 1,2, ,
jj ij i j
w j N
(c) 2017 Biointelligence Lab, SNU 32
Clustering
Figure 13.17 Illustrative example of a two-neuron BSB model, operating under four different initial conditions:
figure represent the model’s basins of attraction;
trajectories of the model are plotted in blue;
trajectories terminate, are printed as black dots.
(c) 2017 Biointelligence Lab, SNU 33