Professional Documents
Culture Documents
Biological Modeling: George Reeke
Biological Modeling: George Reeke
Biological Modeling: George Reeke
77
George Reeke
Brief History
From the earliest days when it became clear that the brain is the organ that controls
behavior, and that the brain is an incredibly complex system of interconnected cells
of multiple types, scientists have felt the need to somehow relate the information so
laboriously gathered regarding the physiology and connectivities of individual cells
with the externally observable functions of the brain. Thus, one would like to have
answers to questions like these:
• How are objects and events in the world categorized by creatures with brains?
• How does the brain select an appropriate behavior at each moment in time from
the available repertoire of practiced and novel behaviors?
• How is the desire to perform a particular behavior converted into an effective
sequence of motor acts that will carry it out?
• Ultimately, can we understand the basis of consciousness in the operations of
this complex physical system without needing to invoke nonphysical processes
that would not be accessible to scientific study?
For a long time, these kinds of questions were considered to be in the realm of
psychology, or, in the case of the last one, philosophy, and not neuroscience, largely
because neuroscientists had their hands full just trying to understand sensory and
single-cell physiology. And until the advent of the computer, psychologists con-
tented themselves with developing quantitative relationships between stimulus and
response (e.g., the Weber-Fechner law) and with reducing learning to simple
paradigms (classical and operant conditioning) that could be studied in lab animals
as well as sometimes in humans. This mode of study reached a limit in the
behaviorist school, exemplified by the work of Edward L. Thorndike, John B.
Watson, and B. F. Skinner, which deliberately removed the brain from the
G. Reeke
Rockefeller University Laboratory of Biological Modeling, New York, NY, USA
e-mail: reeke@mail.rockefeller.edu
Fig. 77.1 Schematic diagrams illustrating how some common binary logic circuits could be
implemented with neurons. (a) One unit time delay, (b) OR, (c) AND, (d) NAND. Triangles
represent neurons; small circles, excitatory connections; short heavy bar, inhibitory connection.
A neuron is assumed to require a net input of two excitatory connections to fire. These diagrams
obviously resemble common symbols used in logic circuit diagrams (Redrawn from McCulloch
and Pitts (1943), p. 130, using more modern symbols)
Artificial Intelligence
Excitatory connection
Inhibitory
connection
R1
R2
Sensory units
Fig. 77.2 Schematic diagram of a simple perceptron with sensory units, association units, and
two response units. Small circles represent sensory neurons, triangles represent higher-level
neurons, solid lines represent excitatory connections, dashed lines represent inhibitory connec-
tions. Only a few representative examples of each type of connection are shown. A categorical
response occurs when the inputs to one of the response units exceed the inhibition received from
other units. The system is trained by modifying the connection strengths to optimize the generation
of correct responses (Redrawn from Rosenblatt (1958), p. 28, to clarify the distinct types of
neurons and connections)
multiplied each by a weight and summed the products to obtain outputs, one unit for
each category to be sensed. The output units were mutually inhibitory. The cell with
the highest output would indicate the predicted category. The weights were adjusted
during training to maximize the number of correct responses.
Thus, the perceptron was really little more than a machine to carry out the dot
product of an input vector with a weight matrix. In their book “Perceptrons,”
Marvin Minsky and Seymour Papert pointed out the shortcomings of this arrange-
ment: because it was essentially a linear device, the perceptron could only distin-
guish categories that were separated by hyperplanes in the space defined by the
input variables. Even though it might have seemed obvious at the time that there are
many possible approaches to overcoming this linearity problem (e.g., by placing
a nonlinear filter on each input line before the linear summation stage), the Minsky
and Papert critique discouraged further research in this area for some time. Further,
77 Biological Modeling 2337
Minsky later published an influential book, “The Society of Mind,” that laid out,
using a terminology invented by Minsky that never caught on, how a mind could be
constructed from interconnected logical units, where the idea that the units might in
some way correspond to neurons was entirely absent.
These early efforts were mostly trained by ad hoc methods, although a few
researchers saw the need for a general training procedure. As with the perceptron,
this problem was first solved without reference to how brains might learn, but with
a method more suited to the type of network of simple summating units that was
being explored at the time. The very influential series of books, “Parallel Distrib-
uted Processing,” by James McClelland, David Rumelhart, and a group of authors
who styled themselves “The PDP Research Group,” first popularized this learning
algorithm, known as “back-propagation of errors” or simply “back propagation” for
short. This algorithm works by computing the partial derivatives of the activities of
units in a network with respect to the various connection strengths between them.
Then, when an error occurs, each connection can be corrected by an amount that
will just bring the output as close as possible to the value specified by the teacher.
This adjustment procedure must be repeated, usually many times, because correc-
tions that are optimal for one stimulus generally do not also reduce the errors in
responses to other stimuli. Inasmuch as no analog for the calculation of the
derivatives needed for back propagation has been found in the brain, this algorithm,
while useful for certain practical applications of artificial neural networks, appar-
ently plays no role in biological learning processes.
While PDP systems and related neural networks all indeed incorporated “neurons,”
an emphasis on functionalist descriptions and mathematically tractable analyses of
learning, along with the high cost of computation with the computers available at
the time, essentially restricted these investigators to using the very simplest possible
caricature neurons. In parallel with these developments, more biologically oriented
investigators were developing more realistic models of single neurons. These
models could be more complex at the cell level because they did not take on the
computational burden of network interconnections and weight-adjustment rules for
supervised learning. These models include those of Hodgkin and Huxley, Fitzhugh
and Nagumo, Hindmarsh and Rose, Morris and Lecar, and others that have been
covered in previous chapters.
Only recently has it become possible to connect numbers of realistic single-
neuron models to form large-scale networks. A signal development in this field,
which first indicated the possibilities inherent in the approach, was the model of
hippocampal region CA3 by Roger Traub and Richard Miles, which incorporated
9,900 cells of 3 different types (excitatory pyramidal cells and two types of GABA
inhibitory cells). This model was used to study the conditions under which cells in
such a complex network could synchronize to generate wavelike oscillatory pat-
terns similar to those seen in the real hippocampus in vitro. However, Traub and
2338 G. Reeke
Miles had access to the largest available computers at IBM, and their models and
others of the kind required heroic amounts of computer time; hence, later workers
continued to use simpler single-neuron models. These may generally be classified as
“spiking” or “rate-coded” models. The former simulate the action potentials generated
by neurons at times computed from their inputs; the latter simply provide a graded
output level that is taken as a proxy report of the mean firing rate of the cell in some
time interval. In the main text, these modeling styles are described in a bit of detail, and
some examples of each are mentioned with their advantages and disadvantages.
In this chapter, some of the considerations that go into choosing a suitable modeling
approach for a particular neuroscience problem are discussed. Different questions
suggest different types and scales of neural system models. In order to understand
some of the implications of these choices, the reader should first be familiar with
a broad sampling of the types of models that are currently being proposed. It will be
apparent that the predominant consideration is always a trade-off between scale and
complexity: when more neurons and more connections are needed to explain
a particular behavior, the level of detail that can be devoted to each neuron and
each connection, even on the fastest available computers, is less. When multiple ion
channel types or neurons with complex geometries are needed, fewer can be
included in a model. When activities at very different time scales are important,
models must be kept simple to allow long simulations. Some of the details that go
into making these trade-offs are discussed next with examples.
Single-Area Models
At the simplest level are single-area models, for example, a model of early visual
cortex or of hippocampal place cells. These models can be used to investigate
questions such as:
• What are the relative roles of the different cell types in this area of cortex?
• Why do cells in this area need six types of potassium channels?
• How can the responses of this visual area remain invariant when saccades occur?
• Why do amputees often experience touch in what feels like the missing part
(“phantom limbs”)?
Many other examples could be cited. Models of this type are still the most
common, and for good reason: they are already extremely challenging, both in
terms of the expertise needed to implement them and the computational resources
needed to do an adequately realistic job. Many questions of the sort indicated still
do not have adequate answers, and many of them can be attacked at the single-area
level. This should be the first choice whenever the available data suggest that
a particular area is mainly or solely responsible for the phenomena of interest.
77 Biological Modeling 2339
Large-Scale Models
When the questions being investigated appear to involve the interactions of two or
more brain areas, single-area models are no longer sufficient. An example would be
a model to try to understand activations seen in functional MRI experiments when
subjects are performing a particular cognitive task. If one has the software tools to
construct a single-area model with more than one cell type and arbitrarily specified
connections between the cells, then similar tools can be used to model multiple cell
types in multiple areas. This kind of paradigm is exemplified by the “synthetic
neural modeling” approach of Gerald Edelman and colleagues. This term is
intended to imply that a model should be constructed to contain an entire synthetic
nervous system, not just selected components, for reasons discussed under “Pit-
falls” below. However, a main limitation of this approach is that one often lacks the
detailed knowledge of neuroanatomy and channel physiology that is necessary to
specify the interconnections and cell properties in such a model. One must often be
content to model a set of interacting brain regions, but not an entire nervous system.
Examples of such models include the thalamocortical model of Eugene Izhikevich
and Gerald Edelman, the SyNAPSE modeling effort sponsored by DARPA (the US
Defense Advanced Research Projects Agency), and the Blue Brain project, which
has already modeled a single rat cortical column and aims eventually to have
enough computer power to model the entire human cerebral cortex.
Robotics
When the aim is to understand aspects of neural system function that involve
interactions of a creature with its environment, a further difficulty arises. While
one can simulate simple environments in a computer, and software is available to
allow objects in such simulations to move and interact in accord with ordinary laws
of Newtonian physics, still, the real world is more complex than has yet been
captured in any simulation. Objects and creatures in the real world undergo
unexpected interactions that can not only modify their shapes, locations, orienta-
tions, and surface properties, but these interactions occur in real time and creatures
with brains must respond on appropriate time scales. The world contains multiple
actors, each operating according to its own programmed control system or animal
brain, which multiplies the complexity. Furthermore, modelers would rather put
their development efforts into the design of their neural system rather than into
a very detailed simulation of that system’s environment. And it has been noted that
access to the world in fact relieves some of the computational burden on the brain.
For example, it is only necessary to look at a familiar object to discern details of its
shape that would otherwise have to be memorized. For all these reasons, a number
of modelers have chosen to use model neuronal systems as robotic controllers.
Robots can be equipped with a variety of sensors and effectors to emulate those of
an animal, and interaction with the real world replaces the simulated world and
simulated interactions of other models. Early proponents of this approach, working
independently, were Grey Walter and Valentino Braitenberg, who conceived of
simple vehicles with nearly direct connections between sensory receptors and
motor effectors, and almost no memory. Behavioral rules were encoded in
2340 G. Reeke
the wiring diagram of the robot. More recent examples, in which responses are
learned rather than hardwired, include the “Darwin” series of models from the
group of Gerald Edelman at the Neurosciences Institute, the walking insects of
Holk Cruse and colleagues, and the self-organizing robots of JunTani, among many
others.
Choosing what kind of model to construct is the first and perhaps the most
important question that must be addressed in any modeling project. The choice
depends first of all on the hypothesis to be tested but also on the state of existing
knowledge about the physiology of the system in question (e.g., what types of ion
channels are prominent in the cells of the system?; is it safe to mix data from
different animal species in a single model?). Thus, one would like to include all the
components that are considered necessary for function according to ones hypoth-
esis, but there may not be enough information available to do so. One is then faced
with the necessity to use simplified models of essential components and hope that
the simplifications do not on the one hand remove the ability of the model to carry
out the target function or on the other hand eliminate complications that might
gainsay the applicability of the original hypothesis. An often quoted “rule” is that
a model should be just as complicated as necessary, but not more complicated.
However, extraneous considerations such as the amount of available computing
resource also enter into these decisions. Accordingly, only some general guidelines
can be offered. A few of the more common trade-offs that must be considered in
designing a systems model are:
Multicompartment models in practice have been most useful for studying single
cells, for example, the very complex cerebellar Purkinje cell. They have two serious
disadvantages for work with large-scale models. First, one must somehow come up
with the geometrical and physiological information needed to derive the compart-
mental structure of a cell. Anatomy can be determined by computerized analysis of
micrographs of stained cells, but this work is very time-consuming, and it is even
more difficult to localize multiple types of ion channels within even a single cell.
For a large-scale model, one would not want all the cells to be identical, as that
would likely introduce artifacts in the form of unrealistic network oscillations or
make it difficult to train the system due to a lack of a priori response variability.
Therefore, one would either have to analyze many real cells or else use a computer
to introduce variants of a few measured cells. In the later case, it would be difficult
to validate whatever assumptions were made in introducing the variation. Secondly,
the calculations for multicompartment cells are sufficiently time-consuming as to
make them just generally impractical for use in large-scale models.
intensive, either because they contain complex expressions, for example, expres-
sions with exponential functions, or because they may require integration at very
small time steps for sufficient accuracy or both. Thus, it may not be possible, unless
exceptional computing resources are available, to construct large-scale networks of
physiologically realistic cells.
To deal with these issues, many authors have attempted to derive simplified cell
models that eliminate the complexity of channel dynamics, but retain one key
characteristic of real neurons, namely, that they appear to retain some physical
trace of their synaptic inputs over time and fire an action potential when some
criteria on those inputs are met, for example, when the integrated postsynaptic
potential reaches a threshold. The threshold itself can either remain fixed or vary
according to some slower-changing function of the inputs. Such models are gener-
ically known as “integrate-and-fire” models. In their simplest form, the state of the
cell is represented by a single variable, considered to be the membrane potential.
Excitatory and inhibitory inputs simply add or subtract a suitable increment to the
membrane potential, which is usually made to decay by a slow exponential function
to a fixed rest value in the absence of input. This is referred to as having a “leaky”
membrane. The action potential is reduced to a single spike that is applied for one
time step of the simulation, after which the membrane potential returns instanta-
neously to a resting value, where it may be forced to remain until a fixed refractory
period has passed. Thus, the waveform of the action potential is reduced to
a rectangular spike that rises instantaneously in one time step and falls back
instantaneously to a rest value in the next time step. This spike may be applied to
connected postsynaptic cells after a suitable time delay corresponding to an axonal
spike conduction delay. The waveform of the action potential is unimportant
because the effect on postsynaptic cells is merely instantaneously to increment or
decrement their own membrane potentials. Whether this is a realistic assumption
remains a subject of some disagreement in the field.
Eugene Izhikevich has shown that a large number of different neuronal cell types
found in the brain can be accurately modeled with integrate-and-fire models with
two additions to the basic expression for the change in membrane potential at each
time step: a term quadratic in the membrane potential, and a second cell-state
variable, a slowly adapting modification to the membrane potential that effectively
modifies the firing threshold as a function of past activity. The equations for the
simplest form of the Izhikevich model (2007 version) are as follows:
where v is the membrane potential, u is the adaptation variable, I is the sum of any
input currents, t is time, C is the membrane capacitance, vrest is the rest potential,
vpeak is the potential at the peak of a spike, and k, a, b, c, d, and vthresh are parameters
77 Biological Modeling 2343
that may be adjusted to match the properties of the particular cell type of interest.
It can be seen that with this model, the increments in v and u at each time step, dv/dt
and du/dt, are extremely simple to compute, involving only additions and multipli-
cations (once the constant C on the left-hand side is replaced with 1/C on the right).
Similarly, Romain Brette and Wulfram Gerstner have improved the basic inte-
grate-and-fire model by adding to the equation for dv/dt a term exponential, rather
than quadratic, in the membrane potential, as follows (with changes in notation
from the original to emphasize the similarities to the Izhikevich model):
where v, u, I, C, vpeak, vrest, b, and d are as for the Izhikevich model, gL is the leak
conductance, tW is an adaptation time constant which can be identified with (1/a) in
the Izhikevich model, and DT is an additional parameter that can be varied to fit
a particular cell type.
The Izhikevich model is almost as simple to compute as the basic integrate-and-
fire model and yields much more accurate results; the Brette-Gerstner model is
perhaps even more accurate, at the cost of evaluating one exponential function in
each time step. However, it is possible, as suggested by Nicol Schraudolph, with
low-level assembly or C-language coding, to derive a moderately accurate approx-
imation of the exponential function from the components of the floating-point
machine representation of the potential, using the formula:
expðyÞ 2k ð1 þ f cÞ;
ds=dt ¼ ks þ sðIÞ;
2344 G. Reeke
where cij is the strength of the connection from cell j to cell i, Dcij is the change in cij
at each time step, si and sj are the states (firing rates) of cells i and j, respectively,
and d is a small number, usually called the “learning rate,” that effectively controls
how rapidly changes in firing rate are reflected in the synaptic strength.
However, the rule in this intuitive form is unusable for several reasons. It does
not provide for teaching signals, value signals, or other modulatory influences;
2346 G. Reeke
hence, systems based on this rule can only enhance or refine responses that already
occur spontaneously. Because si and sj are always either positive or zero, this rule
can only increase cij and never decrease it. Furthermore, there are no thresholds,
caps, or other nonlinear stabilizing effects. Therefore, any network constructed with
this rule will eventually become unstable as the cijs increase without bound.
Providing an upper bound for cij does not fix this problem, as all the cijs eventually
reach the upper bound and the system stops learning. Alternatively, the cijs can be
made to decay by subtracting a term gcij from the expression for Dcij, where g is
a small constant, but this introduces “forgetting,” and the system must be contin-
ually retrained with old stimuli in order to retain responses to them. A better idea is
to provide active mechanisms to reduce cij values (as with LTD) under particular
conditions, depending on the hypothesis being tested and local conditions such as
the magnitudes of the pre- and postsynaptic potentials and any modulatory signals.
A particularly general form of this idea was introduced by George Reeke and
colleagues in 1992:
replace sj in the expression for Dcij with a term designed to reflect a trace of past
activity. When combined with a present m value reflecting modulation or reward
based on behavior in the recent past, this helps to modify connections that were
active when the behavior was generated and at the same time serves to make
learning more reflective of input trends rather than short-term fluctuations. In
another variant, if the input to the connection, sj, is replaced by sj – <sj>, the cell
can be made to respond only when an unexpected input is received, possibly as part
of a reinforcement learning network.
s small
s large
φ(s,s)
s
Fig. 77.4 Synaptic strength change (f) as a function of current (s) and average ( s) cell activity
according to the Bienenstock-Cooper-Munro rule for synaptic modification. The value f here is
comparable to the product of the terms j(|cij|) (siyI) in the Reeke et al. rule given in the text. When
average cell activity is low (left curve), f rises more quickly with s, and it is easier to strengthen the
connection; conversely, when average activity is high (right curve) the axis crossing of the modifi-
cation curve (ym) is moved to the right, and strengthening requires a higher level of current activity (s)
(Redrawn from Bienenstock et al. (1982), p. 36, to combine separate figures for the three cases shown)
100
Change in EPSC amplitude (%)
50
−50
−100 −50 0 50 100
Postsynaptic - presynaptic spike time (ms)
Fig. 77.5 Typical curve for spike-time-dependent synaptic plasticity (STDP), which is assumed
here to control the level of excitatory postsynaptic current (EPSC) for a constant input after
synaptic change occurs. The change is maximal in the direction of potentiation (positive y axis)
when the presynaptic impulse closely precedes the postsynaptic impulse (high values at small
positive time differences) and maximal in the direction of depression (negative y axis) when the
presynaptic impulse follows the postsynaptic one (negative time differences), when it could not
logically have affected cell firing (Simplified version of Fig. 77.1 from Standage et al. (2007),
omitting experimental data shown by those authors)
adaptive to increase its weight as one of the connections contributing to that cell’s
activity; on the other hand, if the presynaptic spike comes just after a cell has fired,
it cannot have been predictive of that event, suggesting that its future firing should
be disregarding in determining the activity of the postsynaptic cell.
77 Biological Modeling 2349
The situation is actually more complicated than suggested here, because, except
at low firing rates, a presynaptic spike in fact generally occurs before one postsyn-
aptic spike but after another one; the synaptic modification rule must take into
account the relative timings of all three events. A large number of mathematical
formulations have been proposed for rules consistent with these data and suitable
for use in network simulations. It is beyond the scope of the discussion here to
review these proposals. Anyone interested in modeling synaptic change in networks
of spiking neurons should first become familiar with these proposals and their
relative advantages and disadvantages.
Modeling Tools
A number of free software packages are available for neuronal systems modeling,
and reviews have appeared comparing their features. Some are more suitable for
multicompartment models, others for larger networks of rate-coded cells. All are
based on the same general idea: the differential equations describing cellular
activity and synaptic change are recast as difference equations; the model proceeds
in fixed or variable-length (event-driven) time steps; in each time step, stimuli are
presented; the state variables of each cell and its synapses (or only selected cells in
an event-driven model) are updated according to the specified equations; these new
states are transmitted to all relevant postsynaptic cells, possibly with axonal trans-
mission delays; data are recorded for later analysis or presented in online graphics;
and the process is repeated for as long as desired. In choosing a suitable package,
one should consider not only the types of cell and synaptic models that are available
but also the flexibility in specifying network connectivity, graphics, ease of making
local modifications, and of course, compatibility with available hardware. In the
case of particularly original modeling efforts, it may be necessary to write new
computer code to implement exactly the features that will be needed.
An important aspect of modeling software that should not be overlooked is the
methods that are available for specifying stimuli, their times and sequences of
presentation, sensory modalities affected by those stimuli, and whether the model is
intended to activate effectors that may in turn affect the order of stimulus presen-
tation, as in synthetic neural models. Sensors and effectors may be modeled as part
of an overall system simulation package or may be connected as “plugins” or even
supplied on separate computers via network connections. In the case of robotics
work, sensors and effectors will be real-world devices, for which suitable interfac-
ing software must be provided. This will generally require ad hoc programming
according to the exact types and numbers of devices that need to be interfaced.
While most modeling is today performed on commodity desktop systems, larger
models will usually require some sort of parallel-processing hardware and associated
software to provide adequate performance. This may involve networks of
interconnected standard microprocessors or purpose-built devices based on field
programmable gate arrays (FPGAs) or even custom chips. The widely distributed
neuronal simulation packages generally provide support only for the first of these
2350 G. Reeke
Electrical Synapses
Gap junctions and other putative types of electrical synapses have been generally
ignored in neuronal systems modeling up to the present, as they were thought not to
have the plasticity possibilities of chemical synapses. However, experimental data
increasingly indicate the importance of these structures, particularly in areas such as
the inferior olive. Some plasticity may be implemented by changes in the protein
structures implementing the junction. In these situations, the traditional treatment of
gap junctions in models as simple resistive coupling links between compartments on
different cells is inadequate and more sophisticated treatments are beginning to
appear.
Pitfalls
This chapter contains a brief outline of some of the many modeling styles that are
available for studying neuronal systems. If one thing is made clear by this survey, it
is that there is always tension between making a model too abstract, in which case,
properties of the real system that may play important roles in determining behavior
can be overlooked, versus making a model too detailed, in which case, even if it
replicates some observed behavior quite exactly, it may not be possible to tease out
exactly which features are the crucial ones for understanding that behavior. One
should beware of selecting a representation for the neuronal responses or their
2352 G. Reeke
Outlook
Further Reading
Bienenstock EL, Cooper LN, Munro PW (1982) Theory for the development of neuron selectivity:
Orientation specificity and binocular interaction in visual cortex. J Neurosci 2:32
2354 G. Reeke