Sei sulla pagina 1di 12

Biol. Cybern.

87, 404–415 (2002)


DOI 10.1007/s00422-002-0353-y
Ó Springer-Verlag 2002

Mathematical formulations of Hebbian learning


Wulfram Gerstner1 , Werner M. Kistler2
1
Swiss Federal Institute of Technology Lausanne, Laboratory of Computational Neuroscience, EPFL-LCN,
1015 Lausanne EPFL, Switzerland
2
Department of Neuroscience, Erasmus University, Rotterdam, The Netherlands

Received: 1 February 2002 / Accepted: 28 March 2002

Abstract. Several formulations of correlation-based correlation-based learning rules are now generally called
Hebbian learning are reviewed. On the presynaptic side, Hebbian learning.
activity is described either by a firing rate or by Hebb formulated his principle on purely theoretical
presynaptic spike arrival. The state of the postsynaptic grounds. He realized that such a principle would help to
neuron can be described by its membrane potential, its stabilize specific neuronal activity patterns in the brain.
firing rate, or the timing of backpropagating action If neuronal activity patterns correspond to behavior,
potentials (BPAPs). It is shown that all of the above then stabilization of specific patterns implies learning of
formulations can be derived from the point of view of an specific behaviors (Hebb 1949). Even though Hebb’s
expansion. In the absence of BPAPs, it is natural to statement was essentially a theoretical one, he did not
correlate presynaptic spikes with the postsynaptic mem- formulate it himself in mathematical terms. In this paper
brane potential. Time windows of spike-time-dependent we review several mathematical formulations of Heb-
plasticity arise naturally if the timing of postsynaptic bian learning. We start in Sect. 2 with rate-based de-
spikes is available at the site of the synapse, as is the case scriptions and turn then in Sect. 3 to spike-based models
in the presence of BPAPs. With an appropriate choice of of Hebbian plasticity. In both cases we focus on a single
parameters, Hebbian synaptic plasticity has intrinsic synapse wij that transmits signals from a presynaptic
normalization properties that stabilizes postsynaptic neuron j to a postsynaptic neuron i (Fig. 1). Large
firing rates and leads to subtractive weight normaliza- portions of the text of the present paper are based on a
tion. recent book on spiking neuron models (Gerstner and
Kistler 2002).

2 Rate-based Hebbian learning


1 Introduction
In rate-based neuron models, the activity of a given
Over the last 50 years, a large body of experimental and neuron i is described by its firing rate mi which is related
theoretical work on synaptic plasticity and learning has to the membrane potential ui by a nonlinear monoton-
been inspired by Hebb’s postulate: ‘‘When an axon of ically increasing function g:
cell A is near enough to excite cell B or repeatedly or
persistently takes part in firing it, some growth process mi ¼ gðui Þ : ð1Þ
or metabolic change takes place in one or both cells such
that A’s efficiency, as one of the cells firing B, is The membrane potential in turn can be calculated from
increased’’ (Hebb 1949). Today this famous postulate is the presynaptic P firing rates mj and the synaptic weights
often rephrased in the sense that modifications in the wij , i.e., ui ¼ j wij mj . In the following we assume that
synaptic transmission efficacy are driven by the correla- the firing rates mi and mj of the pre- and the postsynaptic
tions in the firing activity of the pre- and the postsy- neuron are constant during one trial of an experiment.
naptic neuron. Even though the idea of learning through For several classical experiments on long-term potenti-
correlations dates further back in the past (James 1890), ation (LTP) this is a reasonable assumption (for reviews
see Brown et al. 1989; Bliss and Collingridge 1993). LTP
can, for example, be introduced by high-frequency trains
of presynaptic pulses at several synapses that are
Correspondence to: W. Gerstner maintained over a time T. In such a situation the
(e-mail: wulfram.gerstner@epfl.ch, temporal resolution is rather coarse, and a description of
Tel.: +41-21-6936713, Fax: +41-21-6935263) pre- and postsynaptic activity by fixed rates is appro-
405

requirement of simultaneous activity of the pre- and the


postsynaptic neuron. How can we implement this
information in the function F in (2)? Since F is a
function of the rates mi and mj , we may expand F about
mi ¼ mj ¼ 0. An expansion to second order in the rates
yields
d post pre
wij  ccorr 2 2
2 ðwij Þmi mj þ c2 ðwij Þmi þ c2 ðwij Þmj
dt
þ cpre post
1 ðwij Þmj þ c1 ðwij Þmi þ c0 ðwij Þ
Fig. 1. The change at synapse wij depends on the states of the
presynaptic neuron j and the postsynaptic neuron i, and the present þ OðmÞ : ð3Þ
efficacy wij , but not on the state of other neurons k
The first term on the right-hand side of (3) picks up the
correlations between pre- and postsynaptic activity. In
priate. Time-dependent stimulation paradigms will be fact, it is this bilinear term mi mj – namely, a term that is
treated in the context of the spike-based formulation sensitive to the correlations between pre- and postsy-
outlined in Sect. 3. naptic activity – which makes Hebbian learning a useful
We consider the weight change Dwij during one concept. The simplest implementation of Hebbian
learning trial. Since the total weight change during a trial plasticity would be to require ccorr 2 > 0 and set all other
depends on the duration of the trial, we focus on the rate parameters in the expansion (3) to zero:
of change dwij =dt ¼ Dwij =T. d
wij ¼ ccorr
2 ðwij Þmi mj : ð4Þ
dt
2.1 Model requirements Equation (4) with fixed parameter ccorr 2 > 0 is the
prototype of Hebbian learning. A learning rule with
According to Hebb’s postulate, the weight should ccorr
2 < 0 is usually called anti-Hebbian. We note that, if
increase if, during an experimental trial, both neurons we continue the expansion on the right-hand side of (3),
are active together. Can we give a more precise more and more complex learning rules can be construct-
mathematical description of Hebb’s ideas? Our aim is ed. The next terms would be of order m3 : terms of the
to formulate the weight change during Hebbian learning form mi m2j , mj m2i , etc.
as a function of the activity of the pre- and the
postsynaptic neuron. From our point of view, at least 2.1.3 Synaptic depression. Hebb’s original proposal
six aspects are important for the formulation of a useful gives no rule for a decrease of synaptic weights, but
plasticity model, as described in Sects. 2.1.1–2.1.6. only refers to the conditions under which a strengthen-
ing of synapses should occur. It is clear, however, that a
2.1.1 Locality. The learning rule for the synapse wij system where synapses can only increase and never
connecting neuron j to neuron i should depend only on decrease is bound to be useless. An option for decreasing
the activity of j and i and not on the state of other the weights (synaptic depression) is therefore a necessary
neurons k 6¼ i; j. In a rate model the state of neuron i is requirement for any useful learning rule. This can be
fully characterized by its firing rate mi . Given mi , the achieved, for example, by a weight decay, i.e., we take
membrane potential ui follows from (1), i.e., ui ¼ g1 ðmi Þ, the parameter c0 in (3) as
where g1 denotes the inverse of the transfer function g.
The only other variable in a rate model that is locally c0 ðwij Þ ¼ c0 wij ð5Þ
available at the synapse is the present value wij of the with c0 > 0. Synaptic depression can also be implement-
synaptic efficacy itself. Mathematically, we may there- ed by several other combinations of the factors cpost pre
1 , c1 ,
fore write d
and c0 . For example, the rule dt wij ¼ ðmi  mh Þmj could be
d implemented by the choice ccorr
2 ¼ 1; cpre
1 ¼ mh < 0, and
wij ¼ F ðwij ; mi ; mj Þ ð2Þ all other parameters equal to zero. Such a rule is called
dt
presynaptically gated, since presynaptic activity is a
where F is a yet-unknown function (Sejnowski and necessary requirement for any change; the activity level
Tesauro 1989; Brown et al. 1991; Kohonen 1984). of the postsynaptic neuron determines the direction
Hebb’s formulation is clearly consistent with the locality of the change. An overview of various possibilities of
requirement. implementing synaptic depression in the framework of
(3) is given in Table 1.
2.1.2 Cooperativity. Hebb’s formulation ‘takes part in
firing it’ implies that both the pre- and the postsynaptic 2.1.4 Boundedness. In realistic rules, weights at excit-
neuron must be active to induce a weight increase. It atory synapses should remain bounded in a range
furthermore suggests a causal relationship between the 0  wij  wmax , where wmax is the maximal weight value
firings. We will return to an implementation of causality that is sustainable by the biochemical machinery
in Sect. 3. At the moment we restrict ourselves to the implementing the synaptic connection. To achieve
406

Table 1. The change dtd wij of a synapse from j to i for various Hebb Hebb with decay, presynaptic gating, postsynaptic gating, and
rules as a function of pre- and postsynaptic activity. ON indicates a covariance rule. The parameters are 0 < mh < mmax and
neuron firing at maximal rate (m ¼ mmax ), whereas OFF means an 0 < c0 < ðmmax Þ2 . Angular brackets denote an expectation value
inactive neuron (m ¼ 0). From left to right: standard Hebb rule,
d d d d d
Post Pre dt wij / dt wij / dt wij / dt wij / dt wij /
i j mi mj mi mj  c 0 ðmi  mh Þmj mi ðmj  mh Þ ðmi  hmi iÞðmj  hmj iÞ
ON ON + + + + +
ON OFF 0 – 0 – –
OFF ON 0 – – 0 –
OFF OFF 0 – 0 0 +

boundedness, we make use of the dependence of the to constantly changing input. If systems continue to
parameters in (3) upon wij . Since F on the right-hand remain adaptive, care must be taken that previously
side of (2) is a function of wij , the expansion coefficients learned information is not lost. Grossberg has coined the
post pre post pre
ccorr
2 ; c2 ; c2 ; c1 ; c1 , and c0 will also depend on the term ‘stability–plasticity dilemma’ for this problem
weight wij . A suitable choice of the wij dependence of the (Carpenter and Grossberg 1987; Grossberg 1987). A
positive parameters guarantees that the weight cannot simple overwriting of previously stored information,
become larger than an upper bound wmax ; similarly, a sometimes called the ‘palimpsest property,’ should be
suitable choice of the negative parameters assures that avoided.
the wij cannot decrease below a minimum value of zero To approach this problem, Fusi et al. (2000) have
(Oja 1982; Kohonen 1984; Miller and MacKay 1994). studied the problem of the consolidation of weights.
In a simple formulation of a ‘saturating’ weight They argue that consolidation of previously learned
dependence, we take the parameter ccorr 2 in (3) as items is possible with a weight dynamics that converges
to binary weight values wij ¼ 0; 1. In our framework,
ccorr
2 ðwij Þ ¼ g0 ðw
max
 wij Þ ð6Þ such a dynamics can be implemented by setting
with a constant g0 > 0. The factor ðwmax  wij Þ can be c0 ðwij Þ ¼ cwij ð1  wij Þðwh  wij Þ ð7Þ
seen as an implementation of ‘soft’ bounds. The closer a
weight is to its maximum, the smaller the effect of an where 0 < wh < 1 and c > 0. Small weights wij < wh
experimental trial with an LTP-induction protocol. As a decay to 0; large weights wij > wh increase towards 1.
modeling alternative to the ‘soft’ bounds, we can instead If (7) is inserted into (3), then the effects of learning
use ‘hard’ bounds: growth of the synaptic weights has a persist (or are even increased) after the end of a learning
constant factor ccorr
2 as long as wij < wmax and stops if trial. A combination of (7) with other linear and second-
max
wij ¼ w . Hence, hard bounds correspond to the order terms in the plasticity equation (3) can therefore be
replacement ðwmax  wij Þ ! Hðwmax  wij Þ, where considered as a model of consolidation of synaptic

Þ denotes the Heaviside step function. Similarly, in plasticity. In most current formulations of synaptic
a hard-bound formulation the weight decrease would plasticity, the problem of weight consolidation is disre-
simply stop at wij ¼ 0, whereas in the soft-bound garded.
formulation all negative factors would be taken as
proportional to wij .
2.2 Discussion: relation to other approaches
2.1.5 Competition. A further useful feature of learning
rules is competivity. If some weights grow, they do so at the While the requirements listed in Sects. 2.1.1–2.1.6 pose a
expense of others that must decrease. Ideally, competivity number of constraints for the formulation of learning
should be a consequence of the learning rule (3) and should rules, the framework sketched in (3) is general enough to
not require any additional assumptions. One specific classify various well-known learning rules. The Bienen-
implementation of competivity relies on the normaliza- stock–Cooper–Monroe (BCM) rule, for example,
tion of the set of weights wij of all synapses converging d
onto the same postsynaptic neuron. While, at a first wij ¼ g/ðmi  mh Þmj ð8Þ
glance, such a normalization step would seems to violate dt
the requirement of locality, it can in fact be replaced by is obtained if we expand the function F in (2) to linear
purely local rules. An example is Oja’s rule (Oja, 1982) order in the presynaptic rate mj while keeping all higher-
which is found from (3) if we take cpost 2 ¼ g0 wij , order terms in the postsynaptic variable mi . The function
ccorr
2 ¼ g0 > 0, and set all other parameters to zero: / can in fact be identified with dF =dmj evaluated at
c0 ¼ cpre pre
1 ¼ c2 ¼ c1
post
¼ 0. mj ¼ 0. In the BCM theory, / and mh are chosen so that
/ð0Þ ¼ 0 and /0 ð0Þ > 0; that is, /ðxÞ has a zero crossing
2.1.6 Long-term stability. Most of the learning theories at x ¼ 0 with positive slope (e.g., /ðxÞ ¼ x  x3 þ . . .). It
concentrate on the induction of weight changes. Once the is easy to demonstrate that an output rate mi ¼ mh is an
‘learning session’ is over, weights are taken as fixed unstable fixed point under the dynamics (8). To see this,
parameters. Most neural systems, however, are subject we simply note that for fixed input rates mj > 0 and
407

monotonic gain function g, all weights increase without capacitor and resistor is interpreted as the membrane
bounds if mi > mh . In order to avoid unlimited growth of potential. The voltage scale is chosen so that ui ¼ 0 is the
weights, mh is therefore taken in the BCM theory as an resting potential. The temporal evolution of ui is
adaptive parameter mh ¼ hmi i, where hmi i is a short-term
average of the output rate. The notion of a running dui
sm ¼ ui þ RIi ðtÞ ð13Þ
short-term average goes beyond the current framework, dt
since it requires some memory, i.e., it is nonlocal in time. where sm ¼ RC is the membrane time constant of the
We will see, however, that it can be incorporated in the neuron. Spikes are formal events. We say that neuron i
more general framework that is developed in Sect. 3 for has fired a spike if ui reaches at a time t ¼ tif a threshold
spike-based learning rules. #. The form of the action potential is not described
The short-term average plays also a role in the explicitly, but to mark the spike we can add a d
covariance rule of Sejnowski and Tesauro (1989): function dðt  tif Þ. Immediately after spike firing, the
d   potential ui is simply reset to a value ureset < #.
wij ¼ ðmi  hmi iÞ mj  hmj i : ð9Þ Integration of (13) is then resumed with ureset as the
dt
initial condition.
It is based on an expansion of F about a running average In a network of neurons, the input Ii to neuron i is
of mi and mj , rather than on an expansion about zero. For due to the spikes of presynaptic neurons j. In the sim-
fixed hmj i and hmi i, an identification with (3) is possible, plest model of a synapse, each presynaptic spike arrival
which yields c0 ¼ hmi ihmj i, cpre post
1 ¼ hmi i, c1 ¼ hmj i, evokes a postsynaptic current with a standard time
corr
and c2 ¼ 1. course a. The total input to neuron i is then
Instead of an expansion in the postsynaptic rate mi , an X
expansion in the membrane potential ui would also be Ii ¼ wij aðt  tjf Þ ð14Þ
possible. As mentioned in Sect. 2.1.1, ui and mi are re- j;f
lated by (1). Hence, we can define a new function F~ as
where the sum runs over all firing times tjf of all
F~ðwij ; ui ; mj Þ ¼ F ðwij ; gðui Þ; mj Þ ð10Þ presynaptic neurons. The factor wij is the synaptic
efficacy of a connection from a presynaptic neuron j to a
with Taylor expansion postsynaptic neuron i.
Since (13) is a linear equation, the integration can be
d
wij ¼ c0 þ cpre post corr
1 mj þ b1 ui þ b2 ui mj þ . . . : ð11Þ performed for each term in the sum of (14) separately.
dt The total membrane potential is then the sum of all the
Here we have assumed that the voltage scale is chosen so postsynaptic potentials caused by presynaptic firing plus
that the resting potential is at ui ¼ 0. The idea of an the refractory effect of a negative reset potential. Given
expansion in the membrane potential will be used again the last firing time ^ti of neuron i, the result of the inte-
in Sect. 3. gration is therefore of the form (for t > ^ti )
X
ui ðtÞ ¼ gðt  ^ti Þ þ wij ðt  ^ti ; t  tjf Þ ð15Þ
j;f
3 Spike-based learning
where
The approach taken in this section can be seen as a ðt  ^ti Þ
generalization of the Taylor expansion in the rate model gðt  ^tÞ ¼ dðt  ^ti Þ þ ureset exp½ 
of Sect. 2 to the case of spiking neurons. We recall that sm
we started our formulation of rate-based Hebbian and
learning from a general formula
Z t^ti
d ðt  ^ti ; sÞ ¼ aðs  s0 Þ expðs0 =sm Þds0 :
wij ¼ F ðwij ; mi ; mj Þ ð12Þ
dt 0

where weight changes are given as a function of the The firing time ^ti is updated whenever ui reaches the
weight wij as well as of the pre- and postsynaptic rates mj threshold # from below. Equation (15) is the equation of
and mi (see Eq. 2). The essential assumption was that the spike response model (SRM; Gerstner and van
neuronal activity is characterized by firing rates that Hemmen 1992a; Gerstner 1995a; Kistler et al. 1997), a
change slowly enough to be considered as stationary. generalization of the leaky integrate-and-fire neuron. In
Hebbian rules followed then from a Taylor expansion of the general SRM, the function gðt  ^ti Þ describes the
(12). In the following, we keep the idea of an expansion, form of the action potential and the hyperpolarizing
but drop the assumption of rate-coded neurons. spike afterpotential that induces neuronal refractoriness.
A simple model of a spiking neuron is the leaky Each term ðt  ^ti ; t  tjf Þ describes the time course of a
integrate-and-fire neuron (Lapicque 1907; Stein 1965). postsynaptic potential caused by presynaptic spike
In its simplest form an integrate-and-fire neuron i con- arrival at time t  tjf . Because of refractoriness, the
sists of a resistor R in parallel with a capacitor C driven postsynaptic potential  has, in the general SRM, a
by an external current Ii . The voltage ui across the nontrivial dependence upon the last postsynaptic firing
408

time. In the following we neglect the dependence of the voltage scale so that urest ¼ 0. A Volterra expansion of
postsynaptic potential upon t  ^ti and write (17) yields
Z 1
ui ðtÞ ¼ gðt  ^ti Þ þ hi ðtÞ ð16Þ dwij
¼ c0 ðwij Þ þ apre pre
1 ðwij ; sÞuj ðt  sÞds
P dt
where hi ðtÞ ¼ j wij ðt  tjf Þ is the total postsynaptic Z 1 0

potential. Thus, the internal state of spiking neurons is þ apost 0 post


ðt  s0 Þds0
1 ðwij ; s Þui
characterized by the membrane potential u which in turn 0
Z 1Z 1
depends on the last output spike and the total postsy- 0 pre post
naptic potential (see Fig. 2). Equation (16) defines the þ acorr
2 ðwij ; s; s Þuj ðt  sÞui ðt  s0 Þds0 ds
0 0
simple spiking model SRM0 .
We now return to the problem of Hebbian synaptic þ ... : ð18Þ
plasticity. As before, we start our formulation of Hebbian The next terms would be quadratic in upost or upre
i j , and
learning with the locality requirement. While the neuro- have been neglected. Equation (18) is the central result
nal state of rate neurons was characterized by their firing of this section. It provides a framework for the
rate m, the internal state of spiking neurons is character- formulation of spike-based learning rules and may be
ized by their membrane potential u. The generalization of seen as the generalization of the general rate-based
(12) to the case of spiking neurons is therefore model that we have derived in Sect. 2. The consider-
d ations summarized in the points discussed in
wij ðtÞ ¼ F ½wij ðtÞ; upost
i ðt0 Þ; upre 00
j ðt Þ ð17Þ Sects. 2.1.3–2.1.6 apply also to the case of spike-based
dt
learning rules, and will not be repeated here.
where F is now a functional of the time course of the pre- In order to establish a connection with various other
and postsynaptic membrane potentials at the location of formulations of spike-based learning rules, we consider
the synapse. Our notation with t0 and t00 is intended to the time course of the pre- and postsynaptic membrane
indicate that the weight changes do not only depend on potentials in more detail. At the presynaptic terminal,
the momentary value of the pre- and postsynaptic the membrane potential is most of the time at rest, ex-
potentials (at time t), but also on their history for cept when an action potential arrives. Since the duration
t0 < t and t00 < t. The weight value wij and the local of each action potential is short, the presynaptic mem-
values of pre- and postsynaptic membrane potentials are brane potential can be approximated by a train of d
the essential variables that are available at the site of the functions:
synapse to control the up- and downregulation of X
synaptic weights. In detailed neuron models, F would upre
j ðtÞ ¼ dðt  tjf Þ ð19Þ
depend not only on the weight wij and membrane f
potentials, but also on all other variables that are locally
available at the site of the synapse. In particular, there where tjf
denotes the spike arrival times at the presy-
could be a dependence upon the local calcium concen- naptic terminal.
tration. In the following we adopt the point of view that The situation at the postsynaptic site is somewhat
the calcium concentration is largely determined by the more complicated. For the simple spike response model
previous firing history, so that there is no need to SRM0 , the membrane potential can be written as
introduce an additional explicit variable for calcium. upost ðtÞ ¼ gðt  ^ti Þ þ hi ðtÞ ð20Þ
i
In analogy to the approach taken in Sect. 2, we now
expand the right-hand side of (17) about the resting state where ^ti is the last postsynaptic firing time. In contrast to
upost
i ¼ upre
j ¼ urest . For the sake of simplicity we shift the the usual interpretation of terms on the right-hand side
of (20), the function g is now taken as the time course of
the backpropagating action potential (BPAP) at the
location of the synapse. Similarly, hi ðtÞ is the local
postsynaptic potential at the synapse.
For a further simplification of (18), we need to make
some approximations. Specifically we will explore two
different approximation schemes. In the first scheme, we
suppose that the dominating term on the right-hand side
of (20) is the BPAP, while in the second scheme we ne-
glect g and consider h as the dominant term. Let us
discuss both approximations in turn.

Fig. 2. The membrane potential of a spiking neuron


uðtÞ ¼ hðtÞ þ gðt  ^tÞ can be seen as a superposition of the total 3.1 Sharply peaked backpropagating action potential
postsynaptic potential hðtÞ generated by presynaptic spike arrival and
the action potentials of the postsynaptic neuron. The time course of
the action potential is gðt  ^tÞ, where ^t is the firing time of the last We assume that the BPAP is sharply peaked, i.e., it has a
output spike. A backpropagating action potential could convey large amplitude and short duration. In this case, the
information about postsynaptic spike firing to the site of the synapse membrane potential of the postsynaptic neuron is
409

dominated by the BPAP, and the term hðtÞ in (20) can be respectively, then the expected weight change, (calculat-
neglected. Furthermore g can be approximated by a d ed from Eq. 18 without the correlation term or higher-
function. The membrane potential at the postsynaptic order terms) is
site reduces then to a train of pulses
X hdwij =dti ¼ c0 ðwij Þ þ cpre post
1 ðwij Þmj þ c1 ðwij Þmi þ . . .
upost
i ðtÞ ¼ dðt  tif Þ ð21Þ
f
which is identical to the corresponding terms in (3).
For the correlation term we exploit the invariance
where tif denotes the postsynaptic firing times. Equa- with respect to time translation, i.e., the final result
tion (21) is a sensible approximation for synapses that should only depend on the time difference tjf  tif . The
are located on or close to the soma, for which the full weight update occurs at the moment of the postsynaptic
somatic action potential is ‘felt’ by the postsynaptic spike if tjf < tif , and at the moment of the presynaptic
neuron. For neurons with active processes in the spike if tjf > tif . Hence, the assumption of instantaneous
dendrite that keep the BPAP well focused, (21) is also update yields two terms:
a reasonable approximation for synapses that are
further away from the soma. A transmission delay for f f
acorr
2 ðwij ; t  ti ; t  tj Þ !
backpropagation of the spike from the soma to the site (
of the synapse can be incorporated at no extra cost. W ðwij ; tjf  tif Þdðt  tjf Þ for tif < tjf
If we insert (19) and (21) into (18), we obtain ð25Þ
W ðwij ; tjf  tif Þdðt  tif Þ for tif  tjf .
dwij X pre
¼ c0 ðwij Þ þ a1 ðwij ; t  tjf Þ Thus, for sharply peaked BPAPs and rapid weight
dt f changes, we arrive at the notion of a time window W
tj
X post for Hebbian synaptic plasticity (see Fig. 3). Such a
þ a1 ðwij ; t  tif Þ ‘learning window’ has been used in numerous models of
tif spike-time-dependent plasticity (Gerstner et al. 1993,
XX f f 1996; Kempter et al. 1999; Roberts 1999; Kistler and van
þ acorr
2 ðwij ; t  ti ; t  tj Þ þ . . . : ð22Þ
Hemmen 2000; Roberts and Bell 2000; van Rossum et al.
tjf tif 2000; Song et al. 2000; Rubin et al. 2001) and also been
measured in experimental preparations (Bell et al. 1997;
We emphasize that weight changes are continuous. An Magee and Johnston 1997; Markram et al. 1997; Bi and
isolated presynaptic spike at time tjf , for example, will Poo 1998, 1999; Debanne et al. 1998; Zhang et al. 1998).
cause a weight change that builds up during some time Time windows have also been used in rate-based models
after tjf ; the time course of the weight change dwij =dt is with time-dependent firing rates (Herz et al. 1989; Minai
described by apre f
1 ðwij ; t  tj Þ. The total weight change and Levy 1993; Abbott and Blum 1996; Gerstner and
that Raccumulates for t!1 is therefore Abbott 1997). The relevance of the learning window
1
Dwij ¼ 0 apre 1 ðw ij ; sÞds. Similarly, a single postsynaptic W ðtjf  tif Þ is discussed in Sects. 4 and 5.
spike causes a weight change dwij =dt with time course To explore the relation between the learning window
apost f
1 ðwij ; t  ti Þ. Finally, a combination of pre- and and the correlation term in the rate equation (3), we
postsynaptic spikes within a short interval induces an assume that the firing times of the pre- and the postsy-
f f
additional weight change acorr 2 ðwij ; t  ti ; t  tj Þ. naptic neuron are generated by independent Poisson
In typical plasticity experiments, the synaptic weight process with constant rates mj and mi , respectively. The
is monitored every few hundred milliseconds, so that the expected weight evolution is then
exact time courses of the functions apre post
1 , a1 , and a2
corr

are not measured. We now assume that the weight d


h wij i ¼ c0 ðwij Þ þ cpre post
1 ðwij Þmj þ c1 ðwij Þmi
changes are rapid compared to the time scale of weight dt Z 1
monitoring. In other words, we make the replacements
þ W ðsÞds mi mj : ð26Þ
apre f pre f
1 ðwij ; t  tj Þ ! c1 ðwij Þdðt  tj Þ ð23Þ 0

Thus, the integral over the learning window plays the


apost f post f
1 ðwij ; t  ti Þ ! c1 ðwij Þdðt  ti Þ ð24Þ role of the correlation parameter ccorr in (3).
2
where
Z 1 3.2 No backpropagating action potential
cpre
1 ðwij Þ ¼ apre
1 ðwij ; sÞds
0 In the second approximation scheme, we assume that the
Z 1 membrane potential at the location of the synapse is
cpost
1 ðwij Þ ¼ apost
1 ðwij ; sÞds dominated by the slowly varying potential hi ðtÞ. This is,
0
for example, a valid assumption in voltage-clamp
are the accumulated weight changes. Note that we have experiments where the postsynaptic neuron is artificially
chosen the same symbols cpre 1 and cpost
1 as in (3). The kept at a constant membrane potential hpost . This is also
reason is that, if pre- and postsynaptic spike trains are a good approximation for synapses that a located far
generated by Poisson processes with rates mj and mi , away from the soma on a passive dendrite, so that the
410

Fig. 3A–C. Learning window. The change Dwij of the synaptic partially ‘causal’ in firing it. B An asymmetric biphasic learning
efficacy depends on the timing of pre- and postsynaptic spikes. A The window as introduced in model studies of delay selection (Gerstner
solid line indicates a rectangular time window as it is often used in et al. 1996). A synapse is strengthened (long-term potentiation, LTP)
standard Hebbian learning. The synapse is increased if the pre- and if the presynaptic spike arrives slightly before the postsynaptic one,
the postsynaptic neuron fire simultaneously with a temporal but is decreased (long-term depression, LTD) if the timing is reversed.
resolution Dt. The dashed–dotted line shows an asymmetric learning The biphasic learning window is sensitive to the temporal contrast in
window useful for sequence learning (Herz et al. 1989; Gerstner and the input. C Experimental results have confirmed the existence of
van Hemmen 1993). The synapse is strengthened if the presynaptic biphasic learning windows. Data points redrawn after the experiments
spike arrives slightly before the postsynaptic one, and is therefore of Bi and Poo (1998)

backpropagation of somatic action potentials is negligi- et al. 1991). If the postsynaptic neuron is modeled as a
ble. linear unit:
Let us consider a voltage-clamp experiment where X
hi ðtÞ is kept at a constant level hpost . As before, we mi ðtÞ ¼ wik mk ðtÞ ð28Þ
suppose that weight changes are rapid. If we insert k
P
upre
j ðtÞ ¼
f post
f dðt  tj Þ and ui ðtÞ ¼ hpost into (18), we then the standard Hebbian learning rule dwij =dt
obtain ¼ ccorr
2 mi mj yields an expected weight evolution
dwij X pre d X
¼ c0 ðwij Þ þ c1 ðwij Þdðt  tjf Þ h wij i ¼ ccorr
2 wik hmk ðtÞmj ðtÞi ð29Þ
dt f dt k

þ bpost
1 ðwij Þh
post
which is in the direction of the principal eigenvector of
X
þ bcorr post
dðt  tjf Þ þ . . . the spatial correlation matrix
2 ðwij Þh ð27Þ
f
Cjk ¼ hmk ðtÞmj ðtÞi : ð30Þ
where Angular brackets denote an expectation value. If
Z 1 learning is slow,1 then the actual weight vector in fact
cpre
1 ðwij Þ ¼ apre
1 ðwij ; sÞds stays close to the expected one so that the angular
0
Z 1 brackets on the left-hand side of (29) can be dropped.
bpost apost We may introduce a vector wi for the set of synapses
1 ðwij Þ ¼ 1 ðwij ; sÞds
that converge on the postsynaptic neuron i. The corre-
Z0 1 Z 1
0 0
lation between presynaptic neurons can be denoted by a
bcorr
2 ðwij Þ ¼ acorr
2 ðwij ; s; s Þds ds : matrix C with components Cjk . In matrix notation, (29)
0 0
is
Equation (27) is the starting point of the theory of spike-
based learning of (Fusi et al. 2000). Weight changes are d
wi ¼ ccorr
2 Cwi : ð31Þ
triggered by presynaptic spikes, and the direction and dt
value of the weight update depends on the postsynaptic Since the correlation matrix is positive definite, the
membrane potential. We note the close relation to (11). weight vector grows exponentially for in standard
In our framework, (27) is a special case of the slightly Hebbian learning (ccorr > 0).
2
more general (18).

1
4 Analysis of Hebbian learning The weight vector should change only by a small amount during
the time needed to obtain a representative sample of the input
statistics. This can always be achieved by taking ccorr
2  1. We say
It is well known that rate-based learning rules are in this case that the time scale of learning and that of the input are
sensitive to the spatial correlations in the input (Hertz separated.
411

To avoid unlimited growth of weights and to intro- We assume that input spike trains have stationary
duce competivity between the synapses converging onto statistics with known mean and correlations. Specifi-
the same postsynaptic neuron, (29) is usually combined cally, input spikes at a given synapse j are generated by a
with a suitable normalization procedure (Oja 1982; doubly stochastic point process. Spikes are generated by
Miller and MacKay 1994). As an example, we consider an inhomogeneous Poisson process with instantaneous
subtractive normalization (Miller and MacKay 1994). rate mj ðtÞ. The rate itself is drawn from a distribution
To see how this works, we assume that during a single with constant expectation value hmj ðtÞi ¼ mpre and cor-
trial of duration T, the standard Hebbian rule relations hmj ðtÞmk ðt0 Þi ¼ Cjk ðt  t0 Þ. We suppose that all
dwij =dt ¼ ccorr
2 mi mj would yield a ‘raw’ weight change presynaptic spike trains have identical properties. In
pre
D
P ~
w ij . In order to guarantee that the sum of the weights particular, we
1
PNrequire that hmj ðtÞi ¼ m independent of
k w ik does not change, all
P weights are reduced a pos- j, and N k¼1 Cik ðsÞ ¼ CðsÞ independent of i. Since
teriori by an amount N 1 k Dw ~ ik where N is the num- the expected input rates are P constant, the expected out-
ber of synapses converging onto the same postsynaptic put rate mpost  hmi ðtÞi ¼ mpre j wij ðtÞ is a slow function
neuron i. Overall the two steps (i.e., ‘raw’ change and of time which changes only as the synaptic efficacies
subsequent reduction) amount to a new learning rule: adapt.
In order to discuss the weight dynamics we start from
d XN
the general spike-based learning equation that has been
1 corr
wij ¼ ccorr
2 m m
i j  N c 2 m i mk : ð32Þ
dt developed in Sect. 3 for the case of sharply peaked
k¼1
BPAPs (see Eq. 22). Throughout the following we as-
The first term on the right-hand side of (32) is the sume that weight changes Dwij =jwij j are small during the
standard learning rule, and the second term the sub- time that is necessary to approximately sample the input
tractive normalization
P term. If we write the postsynaptic statistics. In this case we can separate the time scale of
rate as mi ¼ k wik mk and take as before the ensemble learning from that of the neuronal dynamics. The right-
average, we find hand side of (22) is then ‘self-averaging’ (Kempter et al.
1999), so that the evolution of the weight vector (22) is
d given by (Kempter et al. 1999; Kistler and van Hemmen
w ¼ ccorr
2 ðC  CÞw ð33Þ
dt 2000; Roberts 2000c)
P
where C is a matrix with components C kj ¼ n Cnj . d
Equation (33) is the analog of (31) for the case of wij ðtÞ ¼ c0 þ cpre post
1 hhSj ðtÞii þ c1 hhSi ðtÞii
dt
subtractive weight normalization. It is usually combined Z1
with hard bounds 0  wij < wmax .
Note that the rule (32) is nonlocal, since the weight þ W ðsÞhhSj ðtÞSi ðt  sÞiids : ð35Þ
change at the synapse from j to i depends on the firing 1
rate of other neurons k. We show in Sect. 2 that implicit
Here we have introduced
subtractive normalization can also be achieved by local
Z 1
rules via a stabilization of postsynaptic firing rates. pre
c1 ðwij Þ ¼ apre
1 ðwij ; sÞds
0
Z 1
4.1 Spike-based learning cpost
1 ðwij Þ ¼ apost
1 ðwij ; sÞds
0

An analogous argument can be developed for spike- and


based learning. The linear rate model in (28) is replaced Z 1
by a stochastically spiking neuron model with instanta- W ðwij ; tjf  tif Þ ¼ acorr 0 f 0 f 0
2 ðwij ; t  ti ; t  tj Þdt
neous firing rate: maxðtif ;tjf Þ
X X
mi ðtÞ ¼ wik ðt  tkf Þ ð36Þ
k f
Z The double angular brackets in (35) denote the average
X 1
over the input statistics. The notation with double
¼ wik ðsÞSk ðt  sÞds ð34Þ
0 angular brackets emphasizes that the underlying process
k
is a doubly stochastic one: we first have to calculated the
where tkf denotes the time of presynaptic spike arrival at expected number of spikes, given the rates; and then we
theP synapse from neuron k to neuron i. Sk ðtÞ have to average over the rates. The double angular
¼ f dðt  tkf Þ is the presynaptic spike train. The time brackets on the left-hand side have been dropped,
course of the postsynaptic potentialR 1is described by the because of the self-averaging property.
function , which is normalized to 0 ðsÞds ¼ 1. We im- For an interpretation of (35) we start with the terms
plicitly assume that the right-hand side of (34) is positive that are linear in the spike trains. We first calculate the
(which is guaranteed if wik  0 and   0). The left-hand expected number of spikes given the rates and take then
side can then be taken as the instantaneous rate (or the expectation of the rates:
stochastic intensity) of an inhomogeneous Poisson
process. We call (34) a linear Poisson neuron. hhSj ðtÞii ¼ hmj ðtÞi ¼ mpre ð37Þ
412

X Z 1
rates as well as the spatial and temporal correlations of
hhSi ðtÞii ¼ hmi ðtÞi ¼ wij ðsÞhhSj ðt  sÞiids
0
the input.
j
X
¼ wij ðtÞmpre ¼ mpost ðtÞ ð38Þ
j 4.2 Implicit subtractive weight normalization
Finally, the term hSj Si i on the right-hand side of (35) In this section we want to show that, for a suitable
describes the correlation between input and output at the choice of parameters, synaptic plasticity leads to an
level of spikes. We may summarize (35) by saying that intrinsic stabilization of the postsynaptic firing rate and,
the evolution of the weight vector is driven by the hence, to a normalization of weights that is akin to
expected firing rates and by correlations on the time scale subtractive normalization (Kempter et al. 1999, 2001;
of the learning window. Song et al. 2000). To do so we proceed in two steps (in
The correlation term depends not only on the input Sects. 4.2.1 and 4.2.2). As a ‘warm up’ we focus in a first
statistics, but also on the dynamics of the neuron model step on Poisson spike trains without correlations and
under consideration. Since we have assumed that show that for a suitable choice of parameters the rate of
learning is a slow process, the correlation term can be the postsynaptic neuron is stabilized. In a second step,
evaluated for constant weights wij , 1  i  N . For the we extend this result to the more interesting case
linear Poisson neuron model, the correlations between including correlations in the input and show that rate
pre- and postsynaptic spike trains can be written as stabilization implies subtractive weight normalization.
(Kempter et al. 1999)
hhSj ðtÞSi ðt  sÞii ¼ hmi ðt  sÞmj ðtÞi þ hmj iwij ðtÞðsÞ 4.2.1 No correlations. First, we show that for a suitable
choice of parameters, the output rate approaches a
ð39Þ stable fixed point. To do so, we consider a linear Poisson
with a postsynaptic firing rate mi ðtÞ given by (34). neuron that receives input from N presynaptic neurons
Substituting (37)–(39) into (35) we obtain with spike activity described by independent Poisson
processes with rate mpre . The postsynaptic
P neuron is thus
d firing at a rate mpost ðtÞ ¼ mpre Nj¼1 wij ðtÞ. Since the input
wij ðtÞ ¼ c0 þ cpre
1 m
pre
þ cpost
1 m
post
ðtÞ
dt is described as independent Poisson processes, the
2 1 3 correlations Qjk in (40) vanish, so we obtain
Z
þ4 W ðsÞds5mpost ðtÞmpre d
wij ðtÞ ¼ c0 þ cpre
1 m
pre
þ cpost
1 m
post
ðtÞ
1 dt
2 3 þ W mpre mpost ðtÞ
X
N Z0
þ wik ðtÞ4Qjk þ djk mpre W ðsÞðsÞds5 þ wij ðtÞmpre W ð43Þ
k¼1
1
R1 R1
where W ¼ 1 W ðsÞds and W ¼ 0 xðsÞW ðsÞds. In
ð40Þ this particularly simple case the weight dynamics is
characterized by a fixed point for the postsynaptic firing
where rate mpost ðtÞ ¼ mFP . To see this we multiply (43) by mpre
and sum over j. The left-hand side becomes dmpost =dt,
Z1 Z1
which we set equal to zero to find the fixed point:
Qjk ¼ dsW ðsÞ ds0 ðs0 ÞCjk
0
ðs þ s0 Þ : ð41Þ
1
c0 þ cpre
1 m
pre
0 mFP ¼  : ð44Þ
cpost
1 þ mpre W þ N 1 W
The important factor in (41) is the spatiotemporal input
covariance function This fixed point is attractive if the denominator is
negative. Since mpost is a firing rate, we have the additional
0
Cjk ðsÞ ¼ h½mj ðtÞ  mpre ½mk ðt  sÞ  mpre i ð42Þ requirement that mFP  0. Altogether we thus have two
conditions for the parameters of the learning rule: (i)
which is convolved with the learning window W and the cpost
1 þ mpre W þ N 1 W < 0, and (ii) c0 þ cpre 1 m
pre
 0.
postsynaptic potential . Thus the correlations between Note that we would obtain – apart from the term
pre- and postsynaptic neuron in (35) has been trans- ðN 1 W Þ – a completely analogous result from the rate
formed into spatiotemporal correlations in the input. formulation in (3) if we identify ccorr
2 ¼ W . Furthermore,
To summarize this section, we have solved the dy- note that the linearity is not essential for the stabilization
namics of spike-time-dependent plasticity under the as- of the postsynaptic rate. Any model where the output
sumption that learning is slow compared to the rate is a monotonous function of the sum of the synaptic
variations in the input. For the linear Poisson neuron, weights yields qualitatively the same result.
i.e., a stochastically spiking neuron model, the spike-to-
spike correlations between pre- and postsynaptic firings 4.2.2 Input correlations. We now extend the above
can be evaluated. The final result is a learning equation results and show that output rate stabilization implies
where weight changes are driven by the expected input weight normalization even in the presence of input
413

correlations. As before we introduce the notations additional spike-to-spike correlations that appear in
R1 R0 spiking neuron models give rise to the second term on
W ¼ 1 W ðsÞds and W ¼ W ðsÞðsÞds. We ex-
1 the right-hand side of (49). This tends to stabilize
pand and reorder the terms on the right-hand side of synapses that have large weights, and decrease synapses
(40) into the following form: with small weights. If we neglect the spike-to-spike
 correlations, the dynamics of the weight vector is
d
wij ðtÞ ¼ c0 þ cpre
1 m
pre
þ cpost
1 m
post
ðtÞ dominated by the eigenvector of the matrix Q  Q with
dt the largest eigenvalue. Equation (49) can be seen as the
þ W mpre mpost ðtÞ þ W N 1 mpost ðtÞ generalization of the simple Hebbian learning rule (31)
 to the case of spike-based learning.
mpost ðtÞ We recall that the coefficients c0 , cpre post
þQ pre 1 , and c1 , and
m the learning window W depend, in general, on the cur-
  rent weight value wij . In the above derivation we have
þ W wij ðtÞmpre  N 1 mpost ðtÞ assumed that these values are constant. It is, however,
X  
þ wik Qkj  Q : ð45Þ possible to set upper and lower bounds for the synaptic
k
efficacies wij (i.e., weight changes are zero) if wij > wmax
P or wij < 0. It is straightforward to extend the above
Here Q ¼ N 1 k Qik with Qik given by (41). For a arguments to this case (Kempter et al. 1999). More re-
discussion of (45), let us first
P consider the expected alistically, we could also assume an explicit weight de-
postsynaptic rate mpost ðtÞ ¼ j wij ðtÞmpre . Its temporal pendence where all positive terms have a saturating
derivative can be found by taking the sum over j on the factor ðwij  wmax Þ and all negative terms a decay factor
right-hand side of (45) and multiplying by mpre . The wij . With these dependencies it is again possible to to
terms outside the curly braces in (45) cancel each other analyze the weight dynamics (Kistler and van Hemmen
after summation, and the terms inside the braces yield a 2000; van Rossum et al. 2000; Rubin et al. 2001; Kistler
linear equation 2002).
dmpost ðtÞ
¼ cðmpost  mFP Þ ð46Þ
dt 5 Discussion: learning windows
with a fixed point
The theoretical framework developed above shows that
c0 þ cpre
1 m
pre time windows of Hebbian synaptic plasticity arise
mFP ¼  : ð47Þ naturally in spike-time-dependent plasticity. In this
cpost
1 þ mpre W þ N 1 W þ Qðmpre Þ1
section, we place the learning window into the context
The fixed point of the mean postsynaptic rate is found to of the experimental and modeling literature. For a
be stable if modern review of experiments on Hebbian plasticity, see
Bi and Poo (2001).
c ¼ cpost
1 þ mpre W þ N 1 W þ Qðmpre Þ1 < 0 : ð48Þ In early experimental paradigms neurons were stim-
ulated by bursts of presynaptic spikes, and so the tem-
To proceed with the analysis let us suppose that c  0. poral resolution was not sufficient to measure the time
This can always be achieved if either cpost1 or W is course in detail. Nevertheless, even then it was clear that
sufficiently negative. In this case, the effective time there were temporal contiguity requirements for Heb-
constant seff ¼ 1=c is short, so that the fixed point is bian learning (Levy and Stewart 1983; Gustafsson et al.
attained rapidly. After convergence
P towards the fixed 1987; Debanne et al. 1994). Recent advances in elec-
point, the summed weights j wij remain constant and trophysiological methods allow intracellular recordings
the terms within the curly brackets in (45) cancel each to be obtained from several neurons simultaneously, so
other. The remaining terms on the right-hand side of that the timing windows of Hebbian synaptic plasticity
(45) determine the evolution of the weight vector. If we have been measured with great detail (Bell et al. 1997;
switch to matrix notation we have Markram et al. 1997; Magee and Johnston 1997; Bi and
d     Poo 1998, 1999; Debanne et al. 1998; Zhang et al. 1998).
w ¼ Q  Q w þ W mpre w  N 1 mFP n ð49Þ One of the surprising findings has been that depending
dt
on the exact timing between pre- and postsynaptic
where n ¼ ð1; 1; . . . ; 1ÞT and Q denotes the matrix where spikes, synaptic weights are either increased or decreased
all elements have the same value Q. The result is similar (Fig. 3C). At excitatory synapses, a presynaptic spike
to subtractive weight normalization (see Eq. 33) with a that arrives slightly before the postsynaptic spike induces
few specific differences – the spatial correlation matrix C potentiation. This is in agreement with the ‘causality’
that appears in rate-formulations of Hebbian learning that is implicit in Hebb’s formulation: if a presynaptic
(Miller and MacKay 1994), has been replaced by the spike arrives slightly before a postsynaptic one, it is
matrix Q that describes the spatiotemporal covariance likely that this spike ‘takes part in firing’ the postsy-
on the time scale of the learning window and postsy- naptic one. However, a presynaptic spike that arrives
naptic potential. Thus, the learning rule behaves similar slightly after postsynaptic firing leads to synaptic de-
to Sejnowski’s covariance rule (see Eq. 9). The pression. Thus the learning window W ðtif  tjf Þ is not
414

only asymmetric [which means that the maximum of the (Kempter et al. 1999; Roberts 1999, 2000; Kistler and
function W ðxÞ is not centered around x ¼ 0] but also van Hemmen 2000; van Rossum et al. 2000; Song et al.
biphasic (which means that it contains both excitation 2000; Xie and Seung 2000; Kempter et al. 2001; Rubin
and depression). et al. 2001; Senn et al. 2002). Competition between
A similar asymmetric biphasic learning window has, synapses has been understood to arise from an intrinsic
in fact, been introduced in the modeling literature stabilization of postsynaptic firing rates (Gerstner et al.
(Gerstner et al. 1996), well before precise experimental 1997b; Roberts 2000; van Rossum et al. 2000; Song et al.
results have been available (Fig. 3B). Even though bi- 2000; Xie and Seung 2000; Kempter et al. 2001). It has
phasic learning windows with LTP and long-term de- been shown that a biphasic learning window can be
pression (LTD) are not a strict prediction of the theory, mapped to rate models where learning is driven by the
they have several useful properties that make them at- derivative of the postsynaptic firing rate (Roberts 1999;
tractive candidates for modeling. What are these prop- Xie and Seung 2000). Spike-time-dependent plasticity is
erties? thought to play a major role in sensory image cancel-
It has been recognized early on that asymmetric lation of electric fish (Roberts 2000) and in the tuning of
learning windows where the maximum is shifted to a delay lines in the barn owl auditory system (Gerstner
value x < 0 (i.e., synapses are optimally strengthened if et al. 1996). While most models of spike-time-dependent
the presynaptic spike arrives slightly before the postsy- plasticity start with a given form of the learning window,
naptic one) are ideally suited to sequence learning (Herz the time course of the learning window itself can also be
et al. 1989; Gerstner and van Hemmen 1993; Minai and understood in terms of microscopic models (Gerstner
Levy 1993; Abbott and Blum 1996). The reason is that et al. 1998; Senn et al. 2001; Shouval et al. 2002).
such a shifted learning window implements the ‘causality’
condition (Fig. 3A). If, during learning, neuron j was
systematically firing a few milliseconds before neuron, References
the synapse from j to i is strengthened since j ‘takes part
in firing’ neuron i. After learning, the activity of neuron j Abbott LF, Blum KI (1996) Functional significance of long-term
helps to induce activity of the postsynaptic neurons i, so potentiation for sequence learning and prediction. Cereb Cor-
that the sequence ‘first j then i’ is recalled. In these tex 6: 406–416
Bell C, Han V, Sugawara Y, Grant K (1997) Synaptic plasticity in a
models the width of the learning window (i.e., the range cerebellum-like structure depends on temporal order. Nature
of x during which the value W ðxÞ is significantly above 387: 278–281
zero) was on the same timescale as the sequence itself. In Bi G, Poo M (1998) Synaptic modifications in cultured hippo-
rate models where the activity evolves on a timescale of a campal neurons: dependence on spike timing, synaptic
few hundred milliseconds, the learning window was strength, and postsynaptic cell type. J Neurosci 18: 10464–
taken to have a width of a hundred milliseconds (Abbott 10472
Bi G, Poo M (1999) Distributed synaptic modification in neural
and Blum 1996); in spiking models where the spatio- networks induced by patterned stimulation. Nature 401: 792–
temporal sequence of spike patterns evolves on a time 796
scale of 1 or 2 ms, the width of the learning window was Bi G, Poo M (2001) Synaptic modification of correlated activity:
taken equal to 1 ms (Gerstner and van Hemmen 1993). If Hebb’s postulate revisited. Annu Rev Neurosci 24: 139–166
the width of the learning window were to be taken much Bliss TVP, Collingridge GL (1993) A synaptic model of memory:
broader, then the learning rule would strengthen not only long-term potentiation in the hippocampus. Nature 361: 31–39
Brown TH, Ganong AH, Kairiss EW, Keenan CL, Kelso SR
the ‘correct’ synapses that help to recall the sequence, but (1989) Long-term potentation in two synaptic systems of the
also many other synapses such that the success of hippocampal brain slice. In: Byrne J, Berry W (eds) Neural
learning would be weakened or even wiped out. models of plasticity. Academic, San Diego, pp 266–306
A biphasic learning rule that contains both potenti- Brown TH, Zador AM, Mainen ZF, Claiborne BJ (1991) Hebbian
ation and depression introduces competition between modifications in hippocampal neurons. In: Baudry M, Davis J
the synapses (Gerstner et al. 1996, 1997; Kempter et al. (eds) Long-term potentiation. MIT Press, Cambridge, Mass.,
pp 357–389
1999; Kistler and van Hemmen 2000; Song et al. 2000). Carpenter G, Grossberg S (1987) Art 2: Self-organization of stable
For an appropriate choice of parameters, only those category recognition codes for analog input patterns. Appl Opt
synapses that encode the optimal timing between pre- 26: 4919–4930
and postsynaptic neuron are strengthened, while a Debanne D, Gähwiler BH, Thompson SM (1994) Asynchronous
‘typical’ synapse is depressed. As a result, the learning pre- and postsynaptic activity induces associative long-term
rule is sensitive to temporal structure in the input that depression in area CA1 of the rat hippocampus in vitro. Proc
Natl Acad Sci USA 91: 1148–1152
can be fast compared to the width of the positive phase Debanne D, Gähwiler B, Thompson S (1998) Long-term synaptic
of the learning window (Gerstner et al. 1996). It is thus plasticity between pairs of individual CA3 pyramidal cells in rat
the causality condition (LTP for a timing ‘first pre- then hippocampal slice cultures. J Physiol (Lond) 507: 237–247
postsynaptic neuron’) in combination with competition Fusi S, Annunziato M, Badoni D, Salamon A, Amit DJ (2000)
(induced by LTD for a sequence ‘first post- then presy- Spike-driven synaptic plasticity: theory, simulation, VLSI
naptic neuron’) that makes biphasic learning windows implementation. Neural Comput 12: 2227–2258
Gerstner W (1995) Time structure of the activity in neural network
attractive candidates for learning. models. Phys Rev E 51: 738–758
During the last 5 years, asymmetric Hebbian learning Gerstner W, Abbott LF (1997) Learning navigational maps
rules with biphasic learning windows have therefore at- through potentiation and modulation of hippocampal place
tracted a large interest in the modeling community cells. J Comput Neurosci 4: 79–94
415

Gerstner W, van Hemmen JL (1992) Associative memory in a Magee JC, Johnston D (1997) A synaptically controlled associative
network of ‘spiking’ neurons. Network 3: 139–164 signal for Hebbian plasticity in hippocampal neurons. Science
Gerstner W, van Hemmen JL (1993) Coherence and incoherence in 275: 209–213
a globally coupled ensemble of pulse emitting units. Phys Rev Markram H, Lübke J, Frotscher M, Sakmann B (1997) Regulation
Lett 71: 312–315 of synaptic efficacy by coincidence of postysnaptic APs and
Gerstner W, Kistler WK (2002) Spiking neuron models. Cambridge EPSPs. Science 275: 213–215
University Press, Cambridge Miller KD, MacKay DJC (1994) The role of constraints in Heb-
Gerstner W, Ritz R, van Hemmen JL (1993) Why spikes? Hebbian bian learning. Neural Comput 6: 100–126
learning and retrieval of time-resolved excitation patterns. Biol Minai A, Levy WB (1993) Sequence learning in a single trial. In:
Cybern 69: 503–515 Proceedings of the INNS World Congress on Neural Networks
Gerstner W, Kempter R, van Hemmen JL, Wagner H (1996) A II, Portland, Oregon, 11–15 July, pp 505–508
neuronal learning rule for sub-millisecond temporal coding. Oja E (1982) A simplified neuron model as a principal component
Nature 383: 76–78 analyzer. J Math Biol 15: 267–273
Gerstner W, Kempter R, van Hemmen J, Wagner H (1997) A Roberts P (1999) Computational consequences of temporally
developmental learning rule for coincidence tuning in the barn asymmetric learning rules: I. Differential hebbian learning.
owl auditory system. In: Bower J (ed) Computational neuro- J Comput Neurosci 7: 235–246
science: trends in research 1997. Plenum, New York, pp 665– Roberts P (2000) Dynamics of temporal learning rules. Phys Rev E
669 62: 4077–4082
Gerstner W, Kempter R, van Hemmen JL (1998) Hebbian learning Roberts P, Bell C (2000) Computational consequences of tempo-
of pulse timing in the barn owl auditory system. In: Maass W, rally asymmetric learning rules: II. Sensory image cancellation.
Bishop CM (eds) Pulsed neural networks. MIT Press, J Comput Neurosci 9: 67–83
Cambridge, Mass., pp 353–377 Rossum MCW van, Bi GQ, Turrigiano GG (2000) Stable Hebbian
Grossberg S (1987) The adaptive brain. Elsevier, Amsterdam learning from spike timing-dependent plasticity. J Neurosci 20:
Gustafsson B, Wigstrom H, Abraham WC, Huang Y-Y (1987) 8812–8821
Long-term potentiation in the hippocampus using depolarizing Rubin J, Lee DD, Sompolinsky H (2001) Equilibrium properties of
current pulses as the conditioning stimulus. J Neurosci 7: 774– temporally asymmetric Hebbian plasticity. Phys Rev Lett 86:
780 364–367
Hebb DO (1949) The organization of behavior. Wiley, New Sejnowski TJ, Tesauro G (1989) The Hebb rule for synaptic plas-
York Hertz J, Krogh A, Palmer RG (1991) Introduction to the ticity: algorithms and implementations. In: Byrne JH, Berry
theory of neural computation. Addison-Wesley, Redwood WO (eds) Neural models of plasticity. Academic, San Diego,
City, Calif. pp 94–103
Herz AVM, Sulzer B, Kühn R, van Hemmen JL (1989) Hebbian Senn W, Tsodyks M, Markram H (2001) An algorithm for modi-
learning reconsidered: representation of static and dynamic fying neurotransmitter release probability based on pre- and
objects in associative neural nets. Biol Cybern 60: 457–467 postsynaptic spike timing. Neural Comput 13: 35–67
James W (1890) Psychology (briefer course). Holt, New York, Senn W, Schneider M, Ruf B (2002) Activity-dependent develop-
Chap 16 ment of axonal and dendritic delays or, why synaptic trans-
Kempter R, Gerstner W, and van Hemmen JL (1999) Hebbian mission should be unreliable. Neural Comput 14: 583–619
learning and spiking neurons. Phys Rev E 59: 4498–4514 Shouval HZ, Bear MF, Cooper LN (2002) A unified model of
Kempter R, Gerstner W, van Hemmen JL (2001) Intrinsic stabili- NMDA receptor dependent bidirectional synaptic plasticity.
zation of output rates by spike-based Hebbian learning. Neural Proc Natl Acad Sci USA 99: 10831–10836
Comput 13: 2709–2741 Song S, Miller K, Abbott L (2000) Competitive Hebbian learning
Kistler WM (2002) Spike-timing-dependent plasticity: a phenom- through spike-time-dependent synaptic plasticity. Nat Neurosci
enological framework. Biol Cybern 87: 416–427 3: 919–926
Kistler WM, van Hemmen JL (2000) Modeling synaptic plasticity Stein RB (1965) A theoretical analysis of neuronal variability.
in conjunction with the timing of pre- and postsynaptic Biophys J 5: 173–194
potentials. Neural Comput 12: 385–405 Xie X, Seung S (2000) Spike-based learning rules and stabilization
Kistler WM, Gerstner W, van Hemmen JL (1997) Reduction of of persistent neural activity. In: Solla SA, Leen T, Müller K-R
Hodgkin–Huxley equations to a single-variable threshold (eds) Advances in neural information processing systems 12.
model. Neural Comput 9: 1015–1045 MIT Press, Cambridge, Mass., pp 199–205
Kohonen T (1984) Self-organization and associative memory. Zhang L, Tao H, Holt C, Harris WA, Poo M-M (1998) A critical
Springer, Berlin Heidelberg New York window for cooperation and competition among developing
Lapicque L (1907) Recherches quantitatives sur l’excitation retinotectal synapses. Nature 395: 37–44
electrique des nerfs traitée comme une polarization. J Physiol
Pathol Gen 9: 620–635
Levy WB, Stewart D (1983) Temporal contiguity requirements for
long-term associative potentiation/depression in hippocampus.
Neurosci 8: 791–797

Potrebbero piacerti anche