neural_networks

advertisement
NEURON
Dendrite
Axon Terminal
Soma
Node
10mm
Axon
Nucleus
Schwann Cell
Myelin
Soma: central part of neuron. Contains nucleus protein synthesis
Dendrites: input, some output
Axon: carries signal away from soma
Axon terminal: releases neuro-chemical to communicate with other
neurons
Excitatory neurons: excite their target neurons. Excitatory neurons
in the brain are often glutamatergic.
Inhibitory neurons: inhibit their target neurons. Inhibitory neurons
are often interneurons. The output of some brain structures
(neostriatum, globus pallidus, cerebellum) are inhibitory. The
primary inhibitory neurotransmitters are GABA and glycine.
Modulatory neurons evoke more complex effects termed
neuromodulation. These neurons use such neurotransmitters as
dopamine, acetylcholine, and serotonin.
Image of pyramidal neurons in mouse cerebral cortex expressing greeen
fluorescent protein. The red staining indicates GABAergic interneurons.
(Chemical) SYNAPSE: Electrical signal causes releases of chemicals
that flow across
synapse and
generate
electrical signal
in next neuron.
20 nm
Complexity of human brain and enormous number of possible states:
Synapses in brain: young children 1016 (10,000,000,000,000,000)
Stabilizes in adulthood by ½ to < 5 x1015
TERMINOLOGY
Synapse is asymmetric:
PRE-SYNAPTIC neuron secretes neurotransmitter chemicals
POST-SYNAPTIC neuron binds neurotransmitters at receptors
Pre-synaptic neuron
Release of neurotransmitter triggered by the arrival of a nerve impulse
Action potential
Cellular secretion, exocytosis, is very rapid
Pre-synaptic nerve terminal has docked vesicles docked at
Membrane containing neurotransmitter
Arriving action potential produces influx of calcium ions through
voltage-dependent, calcium-selective ion channels.
Calcium ions trigger biochemical cascade: vesicles fuse with presynaptic
membrane and release their contents to the synaptic cleft.
Post-Synaptic neuron
Receptors on opposite side of synaptic gap bind neurotransmitter
molecules
Respond by opening nearby ion channels in the post-synaptic cell
Ions rush in or out and change local transmembrane potential
Resulting change in voltage is called postsynaptic potential
Result is excitatory, in the case of depolarizing currents
inhibitory in the case of hyperpolarizing currents
Whether a synapse is excitatory or inhibitory depends on what type(s) of
ion channel conduct the post-synaptic current display(s), which in turn is
a function of the type of receptors and neurotransmitter employed at the
synapse.
Following fusion of the synaptic vesicles and release of transmitter
molecules into the synaptic cleft, the neurotransmitter is rapidly cleared
from the space for recycling by specialized membrane proteins in the
pre-synaptic or post-synaptic membrane.
This “re-uptake" prevents “desensitization" of post-synaptic receptors
Re-uptake ensures succeeding action potentials will elicit the same size
post-synaptic potential (PSP)
Necessity of re-uptake and the phenomenon of desensitization in
receptors and ion channels means that the strength of a synapse may
diminish as a train of action potentials arrive in rapid succession
Frequency dependence synapses
Sounds bad but nervous system exploits this property for computational
purposes, and tunes synapses through
Protein phosphorylation; size, number and replenishment rate of vesicles
Example: serotonin re-uptake inhibitors (SSRIs) inhibit serotonin reuptake
Neurotransmitters
Categorized into three major groups:
(1) amino acids (primarily glutamic acid, GABA, aspartic acid &
glycine)
(2) peptides (vasopressin, somatostatin, neurotensin, etc.)
(3) monoamines (norepinephrine NA, dopamine DA & serotonin 5-HT)
plus acetylcholine (ACh).
The major "workhorse" neurotransmitters of the brain are glutamic acid
(=glutamate) and GABA.
Neurotransmitters can be broadly classified into small-molecule
transmitters and neuroactive peptides.
Around 10 small-molecule neurotransmitters are known: acetylcholine, 5
amines, and 3 or 4 amino acids (depending on exact definition used),
Purines, (Adenosine, ATP, GTP and their derivatives) are
neurotransmitters.
Over 50 neuroactive peptides have been found, among them hormones
such as
LH or insulin that have specific local actions in addition to their longrange signalling properties.
Single ions, such as synaptically-released zinc, are also considered
neurotransmitters by some.
It is important to appreciate that it is the receptor that dictates the
neurotransmitter's effect.
Some examples of neurotransmitter action:
•Acetylcholine - voluntary movement of the muscles
•Norepinephrine - wakefulness or arousal
•Dopamine - voluntary movement and emotional arousal
•Serotonin - memory, emotions, wakefulness, sleep and
temperature regulation
•GABA (gamma aminobutyric acid) - motor behaviour
•Glycine - spinal reflexes and motor behaviour
•Neuromodulators - sensory transmission-especially pain
Serotonin
Acetylcholine
Dopamine
Neuronal Patterns
1.
How a neuron can act as a decision-making device;
2.
In what sense does a neuron recognizes a pattern;
3.
4.
How do input weights and threshold interact to determine the set of patterns recognized by
a neuron;
Limitations of the linear discrimination performed by neurons.
Model of a neuron: McCulloch-Pitts neuron
Not biologically neurons per se, but embody fundamental properties of biologically based
information processing (e.g., the weighted summation of excitatory inputs is converted to a
binary signal).
Able to recognize patterns in its environment based on the features of a given stimulus
Cognitive processes that are modeled using these neurons are pattern recognition and decisionmaking.
 Understanding of how the brain can produce cognition.
Use a geometric space (Euclidean distances) to quantify features and patterns
A feature (e.g., a person’s weight, height, hair length, bank account, or grade point average) is a
quantitative aspect of the total pattern we associate with that individual.
A feature can be an axis in a higher dimensional geometric space.
Qualitative features can also be valued, as zero or one, for the absence or presence of a feature
(e.g., ‘has a driver’s license’)
We will only use non-negative values to quantify a feature that corresponds to simple
interpretations of a neuron’s output, i.e., spike or no spike, spike frequency, or inverse interpulse
interval.
Ordered set of such features is called a n-dimensional vector as a point in n-dimensional space.
Re: vector is a point in a geometric space and an ordered set of scalar values
Because these values are ordered, a vector can be called a pattern
Any pattern can be expressed as a vector
Terminology
what goes into a network: input, input pattern, input vector
neural network : an organized system of neurons
Operate on patterns in very high-dimensional spaces—e.g., many neurons in the neocortex
receive 5,000-20,000 excitatory inputs and thus process patterns in 5,000-20,000 dimensions.
Each neuron is part of a network that processes information in an even higher dimensional space.
For organized, functionally delimited brain regions, such as human forebrain
the dimension ranges from 500,000 to 5,000,000 (high dimensional vector)
but
Similarity still based upon separation distance
Individual Neurons Make Decisions.
Synapses made between individual presynaptic inputs and postsynaptic neuron transform that
input.
Neuron’s threshold for action potential generation plays important role in determining a neuron’s
decision.
Assume neuron is dedicated to recognizing a family of closely related similar) patterns
Each active input to a postsynaptic neuron contributes to the excitation of that postsynaptic
neuron in proportion to the synaptic strength or weight, which connects that presynaptic input to
the postsynaptic cell.
Triggering of a postsynaptic neuron: neuronal decision-making is a voting system
Each presynaptic input neuron casts its vote for the firing of the postsynaptic neuron, either by
firing its own action potential or by not firing.
The postsynaptic neuron tallies the votes over the set of inputs—a ‘no’ vote is a zero and a ‘yes’
vote has some positive value (= 1 in simplest model).
If tally is large enough, postsynaptic neuron declares ‘yes’ and it fires
If the tally is not large enough, the postsynaptic neuron says ‘no’ and does not fire.
But, not all votes are equal  neuronal vote is not a true democracy.
Some ‘yes’ votes count more than others do.
Example, an affirmative decision by a postsynaptic neuron may require only a relatively small
number of ‘yes’ votes from its heavily weighted inputs.
Use McCulloch-Pitts neurons to study pattern recognition
Neurons function in discrete time intervals and have no memory of their past excitation.
i.e. assume that successive input sets, and the postsynaptic decisions they drive, occur in discrete,
sequential time intervals (an input set contains one or more input patterns).
Postsynaptic neuron makes a decision at each time interval based solely on its current input
weights, its active inputs, and its threshold.
Net synaptic excitation of a McCulloch-Pitts neuron is a linear function of its inputs.
Each synaptic connection scales its axon’s value and then the postsynaptic neuron adds up these
scaled values.
Each postsynaptic neuron has its own distinct summation process, which converts the current set
of active synapses into a scalar value termed the internal excitation or net input excitation (or just
excitation for short).
However, the output of such a McCulloch-Pitts neuron is either zero or one.
Crude binary encoding of the internal excitation:
neuron fires (output of one) when its net input excitation exceeds a certain value called
‘threshold.’
Threshold is the minimum value of the sum of the weighted active inputs needed for the
postsynaptic neuron to fire.
All neurons in model operate on same temporal cycle.
One computational cycle of a network for the weighted summation of active inputs and the
thresholding to occur.
On the next computational cycle of the simulation, this process is repeated.
Remember that a postsynaptic neuron does not remember what happened on the last cycle (note
that we use the phrase ‘computational cycle’ to describe computations by the network model)
NEURON NETWORK
Assess performance on pattern recognition problems
Look at simple network:
Single postsynaptic neuron receives connections from three different presynaptic inputs.
Computational elements of a McCulloch-Pitts neuron
Neuron, including its inputs, is a parallel computational device 
A neuron with n inputs performs n multiplications and n-1 additions at each discrete
time step.
The multiplications are performed in parallel—each synapse multiplies the incoming
excitation from each input by its weight.
McCulloch-Pitts neurons: these products are summed instantaneously with the total
ending up in the cell body.
Neuron then compares this summated excitation with its threshold, and this comparison
decides whether or not the neuron fires.
On the next time step, the whole process begins again.
Re: a McCulloch-Pitts neuron forgets its summated internal excitation from one time
step to the next.
The minimum, quantitative description of a McCulloch-Pitts neuron requires:
1. The set of all synaptic weights which defines the potency of each presynaptic input
to the neuron
2. The threshold for firing this neuron
Recapitulate: a McCulloch-Pitts postsynaptic (output) neuron
1. Has no memory from one time step to the next
2. Within a time step, it linearly sums its inputs as scaled by their respective synaptic
weights
3. It fires (output=1) if the value of its internal excitation exceeds a preset threshold
value of excitation, otherwise it does not fire (output=0).
presynaptic
postsynaptic
dendrite
A single McCulloch-Pitts neuron with:
three inputs (1, 2, and 3 that take on values x1, x2, and x3, respectively).
Each synapse is a connection formed between a presynaptic axon (input line) and the
single postsynaptic dendrite (indicated by the darkened circular symbol, .
The
 indicates the summation of internal excitation (analogous to the altered
polarization of neuronal cell body).
The output of the postsynaptic neuron is another axon that transmits any action
potentials of the postsynaptic neuron.
Summarize: McCulloch-Pitts neuron consists of:
 Its synaptic weights that scale its excitatory inputs,
 A summation process that adds up the scaled inputs,
 A spike generator that crudely encodes the value of this summation.
Two Types of Variables in Simulations
1. Variables whose values you choose
2. Variables whose values the simulation generates
In the first case: we parameterize the computational elements
In the second case: variables are continuously changed at successive time steps by the
computational steps of the simulation.
Parameterized variables that we choose for each postsynaptic neuron includes:
1. the value of each synaptic weight, and
2. the firing threshold.
The category of the calculated, changing values of a neuron includes time itself and
1. the internal excitation of each neuron, and
2. the neuron’s output (fire or not fire).
Simple Simulation: computational parameters that you specify are fixed for any one
simulation (pattern recognition), and there is no synaptic modification (no learning).
Internal excitation—a linear computation
What happens when an input vector arrives at a postsynaptic neuron?
The internal or postsynaptic excitation of a McCulloch-Pitts neuron is the linearly scaled
sum of its inputs
Individual synaptic weights specify the scaling on a connection-by-connection basis (the
terms synaptic weight and synaptic strength are interchangeable).
Example: input at time step one
x 1
1
0
0 .5
synaptic weights of a postsynaptic neuron named j
Figure on next slide 
wj
0 .1
0. 3
0.2
A single McCulloch-Pitts neuron
On one time step the inputs take on values xi where i{1,2,3}
Each synapse, wij, is formed by a presynaptic input line i connecting to postsynaptic
neuron j.
Here each input xi(t) makes a single synapse with the postsynaptic neuron called j.
Inputs x depend on state of presynaptic neurons and are time dependent.
To keep notation simple, the time notation, t, has been suppressed in diagram.
Excitation  (next slide)
EXCITATION
yi(1): the internal excitation of postsynaptic neuron i at time step one (reverse subscripts
meaning)
Excitation: the sum of the inputs scaled by their respective synaptic weights
3
yi ( 1 ) =  x j ( 1 )wij = 1  0.1+ 0  0.3 + 0.5  0.2 = 0.1+ 0 + 0.1 = 0.2.
j=1
n
yi =  x j wij
Time notation is suppressed
j=1
Notation for a synapse: wij  woutput,input (note subscript notation)
Synapse is the connection between two neurons  requires two integers to specify
Subscript j indicates a specific input neuron
Subscript i indicates a specific output neuron
This example has three synapses:
w11, the synapse connecting input neuron 1 with output neuron 1;
w12, the synapse between input neuron 2 and output neuron 1;
w13, the synapse between input neuron 3 and output neuron 1.
THRESHOLD
Threshold: parameter of a neuron that defines the minimum excitation needed for that
neuron to fire.
Logical operation that compares two values
if excitation of neuron is greater than its threshold value (logical true) then that neuron
fires an action potential;
if it is false, the neuron does not fire.
PATTERN RECOGNITION
Pattern: x-vector giving state (x values) of group of pre-synaptic neurons
Pattern Recognition: Post-synaptic neuron fires in response to input from x-vector
THOUGHT PROCESS (Consciousness?)
Parameterize the synaptic weights and the threshold of the post-synaptic neuron to
recognize a pattern(s) successfully
First Example (trivial)
All synaptic weights (w) are positive and threshold=0
x1
w1
x2
x3
w2
w3
x4
w4
y
 neuron fires in response to all input patterns (even x=0)
Trivial and uninteresting because neuron can only make one decision —
it recognizes everything (fires at every time step)
 neuron transmits no information (response is independent of input pattern x)
Neuron is not wrong, just boring
How to parameterize model to recognize different patterns
Example: Two different simple patterns
x1>x2
Pattern recognition: neuron y1 fires if x1>x2
neuron y2 fires if x1<x2
x1
w11
x2
w21 w12
y1
Re:
y  w .x
or x1<x2
w22
y2
n
i.e.
yi   wij x j
i 1
To recognize x-patterns, Parameters that can be varied are:
weights (w) and thresholds (h)
Define problem: which parameter is fixed versus which do you need to determine
Example: assume w are pre-determined based upon considerations such as
physiological information
Method for determining thresholds h1 and h2 to recognize patterns
Extremely simple example:
Normalize w and x each to 1 i.e.  xi2 =1 and i wij2 = j wij2 = 1
(This is not a real restriction, just normalization)
Simple weights
 w11 w12  = 1 0

  
 w21 w22  01 
Neuron x1 can only excite y1  w11=1, w21=0
Neuron x2 can only excite y2  w12=0, w22=1
x1
x2
w11
w22
y1
y2
Click mouse on screen for slide to evolve:
Draw unit circle and place w1j and w2j
Next: on circumference, locate point that corresponds to boundary between two patterns
Boundary conditions xb: x1=x2 and x12+x22=1  x12+x12=2x12=1
 x1=1/2=0.707  x2=0.707  tan=.707/.707=1  =45o
Drop perpendicular from xb to w1 vector: intersection is h1=(1) cos
Drop perpendicular from xb to w2 vector: intersection is h2=(1) sin
w2=(0,1)
xb
h2= 1/2 =.707
x2
45o
x1
h1= 1/2 =.707
w1=(1,0)
Graphical Method becomes complicated for wij0, and hard to visualize for > 2 input
neurons (> 2-D graph)
General Mathematical Way for Determining Thresholds
Relax restrictions that required i wij2 =j wij2 =1 but maintain  xi2 =1 (can relax later)
Still have basic relationship between input (x), weights w, and output excitation y
n
y=w•x
i.e.
yi   wij x j
j 1
Calculating thresholds h requires knowledge of boundary input vector xb
Iff xb is known then thresholds h are calculated from
n
h=w•xb
i.e.
hi   wij xbj
j 1
xbi are elements of boundary vector
Example: next page
Click mouse on screen for slide to evolve:
Determine h1 and h2 for a network
y1 to recognize the pattern (fire) if : x1>2x2
and

y2 to recognize the pattern (fire) if : x1<2x2
xb =
but
.894 
 
.447 
New wiring diagram (weights) with all (forward) connections
x1
w11
x2
w21 w12
y1
h1=w11xb1+ w12xb2
Example: w11=0.75 w12=0.23 
w21=0.45 w22=0.62
w22
y2
h2=w21xb1+ w22xb2
h1 = (0.75)(.894)+(0.23)(0.447) = 0.773
h2 = (0.45)(.894)+(0.62)(0.447) = 0.679
Check: x1=0.80 x2=0.60  x1<2x2
y1= (0.75)(.80)+(0.23)(0.60) = 0.738 <h1  does not fire
y2= (0.45)(.80)+(0.62)(0.60) = 0.732 >h2  fires
x1=0.95  x2=0.31: x1>2x2
y1=0.784 fires
y2=0.62 no fire
Download