MIT Department of Brain and Cognitive Sciences Instructor: Professor Sebastian Seung

advertisement
MIT Department of Brain and Cognitive Sciences
9.641J, Spring 2005 - Introduction to Neural Networks
Instructor: Professor Sebastian Seung
Excitatory-inhibitory networks
Sebastian Seung
Two neural populations
• “excitatory” and “inhibitory”
• interactions
– within populations: symmetric
– between populations: antisymmetric
The two populations of an
excitatory-inhibitory network
behave as if they have
opposing goals.
Minimax
• An excitatory-inhibitory network is a
method of solving a minimax problem.
min max S ( x, y )
x
!
y
Multiple goals
• Analogy to game theory
– zero-sum game
• Equilibrium
• Oscillations
• Complex non-periodic behavior
Synaptic interactions
excitatory
population
inhibitory
population
T
B
A
y
x
!C
!B
• A and C symmetric
• excitatory-inhibitory interpretation
– A, B, C nonnegative matrices
Matrix-vector notation
" x x˙ + x = f ( u + Ax # By )
T
˙
" y y + y = g(v + B x # Cy )
!
Saddle function
• Excitatory neurons try to minimize
• Inhibitory neurons try to maximize
T
1
2
T
T
1
2
T
S = "u x " x Ax + v y " y Cy
T
T
T
T
+1 F ( x ) + y B x "1 G ( y )
• Platt & Barr (1987)
• Mjolness & Garrett (1990)
!
Saddle function gradients
#S
"1
"
= u + Ax " By " f ( x )
#x
"1
"1
˙
= f ($ x x + x ) " f ( x )
#S
T
"1
= v + B x " Cy " g ( y )
#y
=g
"1
($
y
y˙ + y ) " g
"1
( y)
Pseudo gradient ascent-
descent
%S
" x x˙ # $
%x
%S
" y y˙ #
%y
descent
ascent
• The components of these vectors have
the same sign.
!
True gradient ascent-descent
#S
x˙ = "
#x
#S
y˙ =
#y
descent
ascent
• When does this dynamics converge to
the solution of the minimax problem?
!
min max S ( x, y )
x
y
It depends
!
x2 y2
S=
"
2
2
S = xy
steady state
oscillations
!
Steady state
2
2
x
y
S=
"
2
2
#S
x˙ = " = "x
#x
#S
y˙ =
= "y
#y
Periodic behavior
S = xy
#S
x˙ = " = "y
#x
#S
y˙ =
=x
#y
Kinetic energy
2
2
#
S
#
S
T
T
˙
T = " x˙
x˙ + y˙
y˙
2
2
#x
#y
x˙ 2 y˙ 2
T=
+
2
2
!
• lower bounded
• nonincreasing if
!
" 2S
positive definite
2
"x
" 2S
negative definite
2
"y
Proof
# 2S
# 2S
x˙˙ = " 2 x˙ "
y˙
#x
#x#y
#S
˙x = "
#x
#S
y˙ =
#y
# 2S
# 2S
y˙˙ =
x˙ + 2 y˙
#x#y
#y
T˙ = x˙x˙˙ + y˙ y˙˙
2
! # S
2
#
S 2
2
= " 2 x˙ + 2 y˙
#x
#y
!
!
The saddle function could
either increase or decrease
dS
T "S
T "S
T
T
= x˙
+ y˙
= # x˙ x˙ + y˙ y˙
dt
"x
"y
!
Lyapunov function
L = T + rS
2
2
$
'
$
'
#
S
#
S
T
T
L˙ = " x˙ & 2 + rI ) x˙ + y˙ & 2 + rI ) y˙
% #x
(
% #y
(
!
!
" 2S
+ rI positive definite
2
"x
" 2S
+ rI negative definite
2
"y
choose r to satisfy these conditions
and keep L lower bounded
!
Legendre transform pairs
Legendre transformation
F !######" F
F !( x ) = f ( x ) F !( x ) = f
{
"1
(x )
}
F ( x ) = max px ! F (p)
p
!(p, x ) = 1T F (p)" p T x + 1T F ( x )
Generalized kinetic energy
! x x& + x = f (u + Ax " By )
1
2
! x x& 2 "
"# ! x$1%(u + Ax $ By , x )
!(p, x ) = 1T F (p)" p T x + 1T F ( x )
!(p, x ) # 0
!(p, x ) = 0 for f (p) = x
likewise, ! (q , x ) = 1T G(q )" q T x + 1T G ( x )
Lyapunov function
F "( x ) = f ( x ) F "( x ) = f #1 ( x ) G"( x ) = g( x ) G "( x ) = g#1 ( x )
!
kinetic
energy
saddle
function
!
"( p, x ) = 1T F ( p) # pT x + 1T F ( x )
$(q, x ) = 1T G(q) # qT x + 1T G ( x )
S = "uT x " 12 x T Ax + v T y " 12 y T Cy
+1T F ( x ) + y T BT x "1T G ( y )
1
1
Lyapunov
L = #( u + Ax $ By, x ) + %(v + BT x $ Cy, y ) + rS
function
"x
"y
!
Need to verify that L is lower bounded
!
Sufficient conditions for
stability
T
"1
"1
˙
˙
L˙ = x˙ T Ax˙ " y˙ T Cy˙ " (# "1
+
r
x
f
#
x
+
x
"
f
(
)
( x )]
) [ x
x
[
T
"1
"1
˙
˙
+( r " # "1
y
g
#
y
+
y
"
g
( y)
(
)
)
y
y
sufficient condition for L˙ " 0
!
T
max
a,b
(a # b) A(a # b)
(a # b)
T
( f (a) # f (b))
#1
#1
" 1+ r$ x
T
min
a,b
!
(a # b) C (a # b)
(a # b)
T
(g (a) # g (b))
#1
#1
% r$ y #1
]
Excitatory-inhibitory pair
• inhibitory feedback causes oscillations
• self-excitation required to sustain them
"
#
y
x
!"
Competitive network
local excitation
!
x
1
global inhibition
x
!
y
2
x˙ i + x i = f ( ui " y + #x i )
x
3
!
&
)
$y˙ + y = g( % x i +
' i *
Sufficient conditions
T = $ [ F ( ui + "x i # y ) # ( ui + "x i # y ) x i + F ( x i )]
i
%
1 2
V = $'#ui x i # "x i + F ( x i ) + G
&
2
i
(
(
$i x i *)
)
L =T +V +
L˙ = ${"x˙ i2 # (+ #1 + 1) x˙ i [ f #1 ( x˙ i + x i ) # f #1 ( x i )]}
i
!
Conclusion
• excitatory-inhibitory network
• dynamics on a saddle
– gradient ascent/descent
– shape of saddle determines behavior
Download