Professional Documents
Culture Documents
Near Decomposability and Complexity PDF
Near Decomposability and Complexity PDF
Herbert A. Simon
Camegie-Mellon University. Department of Psychology, Pittsburgh. PA 15213
Erratum:
The title of this book contains four very potent words: "mind," "brain,"
"complex" and "adaptive." The main task I have set for myself here is to
consider the relation among the first three of these; I will have somewhat
less to say about the fourth. Of course, it is generally agreed that the mind
resides in the brain, and that mind and brain are complex (and adaptive)
systems. Is there more than this to be said about the matter? Let me begin
with complexity.
Shortly after World War II a great interest arose hi something that was
called "General Systems Theory." After a time, that interest faded, and for
a simple and compelling reason: It proved to be very difficult, indeed, to
find properties that were common to systems "in general," thereby to fill
general systems theory with any theorems or other content. Control theory,
which is more closely related to adaptivity than to complexity, split off and
became a solid but separate domain within engineering. Communication
theory, whose interest surely does focus on complexity, did flourish also as
a separate domain and contributed as well to the foundations of computer
science; but its ties to a general systems theory were lost, and the dream
of creating such a theory has nearly died.
26 H. A. Simon
COMPLEXITY
I will not undertake a formal definition of complexity.
For our purposes, we can
regard a system as complex if it can be analyzed into
man
relatively many relations among them, so that the beh y components having
avior of each component
depends on the behavior of others. That definition emb
race
of the systems in nature in which we are interested, rang s a very large number
ing
microbes (and smaller systems) to mammals, metropolises from molecules and
, and megapowers (and
even larger systems).
HIERARCHY
It has been observed for a very long time that most,
thou
tems we regard as complex share a very important arch gh not all, of the sys-
itec
constructed in levels, with each level resting on the one tural feature: they are
metaphor, they can be represented as sets of boxes nest below it. To change the
ing within sets of boxes,
nesting within..., often to a considerable number of repe
titio
ple is the molecule, composed of atoms which are compose ns. A familiar exam-
d
which are composed of elementary particles, which are of electrons and nuclei,
composed of quarks. An-
other example is the biological organism, which is com
pose
composed of cells, which contain organelles, which are com d of organs, which are
posed of molecules, and
so on. A third example is a human society, which is com
and religious organizations, these, in turn of subgroups, posed of economic, social,
down to the level of fami-
lies. This social example is more complex than the others,
since each individual may
belong to a number of the larger subgroups a family,
and so on. a business firm, a church,
It has sometimes been suggested that this hierachical form
lies in the eye of the beholder that we describe complex of natural structures
systems to ourselves and
to others in terms of subsystems in order to simplify
the
it comprehensible to the human mind. This may well be description and make
computer program line by line, without any hint of larg true. Examining a large
er components of structure
that organize these lines, is seldom a meaningful exer
cise
doctrine of "structured" or "top-down" programming is . The widely accepted
aimed at using hierarchy
Near Decomposability and Complexity 27
one of a series of blocks along the main diagonal, while 'only small coefficients wilf
appear in the remainder of the matrix.
If the matrix of a dynamic system is block diagonal, with only zero coefficients
outside the blocks, then one can solve independently for the values of the set of
variables within each block. If the matrix is nearly block diagonal, ignoring the
coefficients outside the blocks and solving the resulting equations will still provide
approximate solutions for the original matrix.16 What this means from a dynamic
standpoint is that if the variables in a block are disturbed from equilibrium, the
equilibrium among them will be restored rapidly; over any longer span of time,
each block will appear to move hi a rigid fashion, the variables remaining close to
equilibrium in relation to each other. Over some still longer period of tune, the
weaker interactions among different blocks will establish equilibrium of the entire
system. The weaker the between-block interactions compared with the within-block
interactions, the smaller will be the ratio of the tune required to restore within-block
equilibrium to that required to restore between-block equilibrium.
The argument can be extended to a hierarchy containing more than two layers.
The blocks at the lowest level will equilibrate over the shortest time span; those
at the next level, more slowly with a restoration time that may be one or more
orders of magnitude greater. The top level of the system will have the longest time
for restoration of equilibrium. (The mechanism here is very close to the mechanism
that underlies renormalization in physics.)
Visualize a building composed of many rooms, each room of many cubicles.
Assume that there is a radical disequilibrium of temperature hi the building, cubic
foot by cubic foot. The temperature within each cubicle will equalize at some aver-
age value very rapidly say, within a matter of minutes. The temperatures of the
cubicles within each single room, separated by partitions but not by thick walls, will
equalize a little more slowly say, hi a half hour. The temperatures of the various
rooms will equalize still more slowly as heat passes through the walls that separate
them. After some time, say, a few hours, all of the rooms of the building, all of the
cubicles in each room, and each cubic foot of air within each cubicle will have the
same temperature.
Near-decomposability is of great assistance in gaining an understanding of hi-
erarchical complex systems. Suppose we have a system of N variables. To predict
its behavior, we must solve N simultaneous equations. Suppose, however, that the
system is composed of m stable subsystems each containing k variables, where
mk — N. Then, we can predict the behavior of each subsystem by solving only
k simultaneous equations. Having found the solution for each subsystem, we can
replace the original system of equations by a simpler aggregate system (technically,
the system described by the principal eigenvalues and principal eigenvectors of the
subsystems). We then obtain the low-frequency response of the total system by
solving only m simultaneous equations, one for each subsystem. Roughly speaking,
the details of structure of the subsystems affect only their high-frequency behavior
in returning to internal equilibrium; they do not affect the low-frequency behavior
Near Decomposability and Complexity • 29
of the total system. With lower time resolution, the details of behavior of the sub-
systems do not "show through" to the behavior of the whole system. Only aggregate
or average properties of the subsystems are important to that behavior.
and (6) follow one course of action or another, depending on the outcome of such
a comparison. Matching symbol structures for identify or difference is the basis for
the important process called "recognition."
The physical symbol system hypothesis8 asserts that the necessary and suffi-
cient condition for a system to be capable of thinking is that it be able to perform
the symbolic processes that I have just enumerated. By thinking, I mean doing those
things that, if they were done by a human being, we would call thinking solving
problems, reading, playing chess, making an omelette, or whatnot.
If the hypothesis is true, several consequences follow. First, computers, appro-
priately programmed, are capable of thinking. Second, the human brain, since it
is capable of thinking, is (at least) a physical symbol system. These are empiri-
cal hypotheses, to be tested by spelling out some of their observable consequences
and checking whether they are true. To test sufficiency, we write computer pro-
grams that enable computers to perform and to learn to perform the tasks that
humans perform when they are thinking. Testing necessity is more difficulty: we
must check whether the processes people use when they are thinking are just the
symbol processes listed above.
Both parts of the symbol system hypothesis have, in fact, been tested exten-
sively over the past 37 years, and the evidence has consistently supported it over a
wider and wider range of human thinking tasks. [For samples of the evidence, see
Newell and Simon8 and Simon15 (chapters 3 and 4).] Of course, it has not yet been
conclusively demonstrated that computers can do all kinds of thinking of which
humans are capable. Plenty of work remains in order to test the full scope of the
hypothesis.
Whether the processes a computer uses in thinking are similar to the pro-
cesses people use depends on how the computer is programmed. Consider, for ex-
ample, computer programs that play chess. The most powerful of these, currently
Deep Thought, plays at grandmaster level, but it demonstrably does not play in
a humanoid way. It may examine as many as 50 million or more continuations
to a position before it makes its move, while there is excellent experimental evi-
dence that grandmasters seldom examine more than 100, and certainly never more
than 1,000 numbers four or five orders of magnitude smaller than Deep Thought's
numbers.
Deep Thought uses its computational power to "try everything" (of course,
not really everything, but a great many possibilities). The grandmaster, with far
weaker computational powers, uses chess knowledge to select the few lines of play
that need to be examined. It is a contrast between nearly exhaustive search (at
least to a moderate depth) and highly selective search.
On the other hand, there exist some chess programs less powerful than Deep
Thought that have considerable skill, especially in sharp tactical positions, in se-
lecting moves after searches no more extensive than the human ones. The PAR-
ADISE program, for example, can discover many deep mating combinations after
Near Decomposability and Complexity 33
exploration of less than 100 continuations.19 PARADISE can serve as a first ap-
proximation to a theory of the symbolic process in human chess play; Deep Thought
cannot.
OF THINKING
NEUROLOGICAL THEORIES
l rec en t pro gre ss in ne uro sci en ce, our clues as to how sym-
In spite of the substantia ev en ts, are sti ll lar ge ly limited to connections,
ke d to ne ura l
bolic events are to be lin siz ed sym bo lic fun cti ons (e.g., the operation of
referred to earlier, betw een hy po the
reg ion s of the bra in ide nti fie d as performing these functions.
working memory) and ory of the "e ng rai n" the basic mechanism of
We do not have a ph ysi olo gic al the
hi the bra in. No r do we un de rstand what biological mech-
storage of information ati on . We kn ow a great deal about how the
anisms operate on thi s sto red inf orm
the pio ne eri ng wo rk of Hu be i an d Wiesel gave glimpses of mech-
retina works, and ret ina l inf orm ation transmitted to the brain.
cti ng fea tur es fro m
anisms for extra at the sym bo lic level of the subsequent pro-
ex ten siv e kn ow led ge
Although we have ing up on inf ormation that has proceeded through
cesses of recog nit ion , an d of rea son
gn itio n pro ces s, the de tai led ne uro logy of these processes lies largely
the feature-reco
in terra incognita.
36 H. A. Simon
SERIAL OR PARALLEL?
Until quite recently, digital computers have been predomin
ately
That is, although they have very large memories holding muc serial machines.
processes that act on this information are of a serial, one-at-a- h information, the
time, kind. Computer
programs exercise rather close control over the sequence in
which these actions
are performed. On the other hand, in the billions of neurons
of
myriads of synaptic connections among them, there appears to the brain and the
be
widespread simultaneous activity. Some investigators (but not, almost continuous
by any means, all)
have concluded from this that the brain is predominately a
para
sometimes they have further concluded that for this reason a seria llel device,10 and
l computer cannot
simulate brain processes.
This argument does not take levels into account. Even exte
at the neural level would not imply parallelism at the sym nsive parallelism
bolic level. Since we
do not know, from a neurological standpoint, how informat
ion
stable form hi memory, much continuous and diffuse brain activ is maintained in
ity may be required
simply for memory maintenance. Moreover, the presence of
disp
the brain may mean that each event at the symbolic level requ ersed activity hi
neural transmissions, which may not all be localized in any sing ires a sequence of
brain. The temporal resolution of our observations of neural le small part of the
great enough to choose among these and other interpretationevents is simply not
s of the widespread
continuous activity that can be observed hi the brain.
Near Decomposabilrty and Complexity 37
On the other hand, the information that we do have about brain localization
gives substantial evidence of elaborate local specialization, and of sequential trans-
mission of signals. It certainly does not support a picture of a highly undifferentiated
and dispersed parallel network mechanism.
before the partial products can be added. The multiplier and the adder cannot
operate in parallel. Whenever there are many precedence constraints, it becomes
exceedingly difficult to design a parallel architecture in which all or most of the
parallel components can be kept busy simultaneously most of the time. Of course,
if they cannot, the purpose of parallelism has been defeated, at a considerable hard-
ware cost. Nature guided by natural selection would not build parallel architectures
where priority constraints caused such inefficiency.
There has been a long history of failure (beginning, perhaps, with the ILLIAC
IV) hi designing highly parallel general-purpose computers, for the reason just set
forth. Typically, general-purpose designs with thirty or forty processors operate only
four or five times as rapidly as a single processor a very expensive way to gain
speed. On the other hand, there is a growing list of successes hi designing parallel
computers that are specialized to handle particular classes of tasks, for example,
array computers or special-purpose chess machines like Hightech. In these special-
purpose architectures, the parallel components are matched with components of the
task that are relatively independent and, hence, unlikely to be left idle by unsatisfied
precedence constraints. Computer scientists are just beginning to construct a body
of theory that would explain the successes and the failures in general but accurate
terms.
yet been demonstrated that connectionist models can account for any wide range of
complex cognitive performances in the domains of problem solving, use of language
or reasoning. Almost all of the tested computer simulations today of these kinds of
human thinking are serial symbolic models.
There is certainly every reason to continue research on parallel models of brain
activities along both of the paths just mentioned. But on the basis of what we know
about the hierarchical structure of complex systems, and from what neurological
evidence tells us about brain localization and the temporal sequencing of signals,
it is improbable that such models will supersede models of the serial kind as ex-
planations of behavior at the symbolic level, except possibly for mechanisms at the
sensory/perceptual or motor peripheries of the system. A more promising goal for
research with connectionist or neural net models is to use them to explain the ba-
sic mechanisms of memory and of the elementary information processes: inputting,
outputting, storing, modifying, and comparing symbols. We will need this land of
a bridging theory in order to account for symbolic processes in neurological terms.
The bridging will not remove the need for theories at a more aggregated level any
more than nuclear physics has removed the need for theories of biochemistry.
ADAPTIVENESS
Time does not permit me to take up the topic of adaptiveness at any length. There
has, however, been a good deal of attention to the implementation of learning
within physical symbol systems, and some comparison of these learning systems
with human learning. I will mention three examples of learning mechanisms that
have been tested.
The first is the EPAM system, which was originally designed to simulate human
performance in rote learning tasks (e.g., paired associate and serial anticipation
learning) of the kinds that are so widely prevalant in the psychological literature.
EPAM, which is currently being extended to an even wider range of tasks, has
shown itself capable of simulating, quantitatively as well as qualitatively, a broad
range of human learning phenomena.1'9
The "front end" of EPAM, a discrimination net capable of growth through
learning, resides at the very boundary between perceptual processes that we would
expect to be parallel in nature, and those we would expect to be serial. Therefore,
special interest lies in the ability of EPAM to perform tasks that have also been
studied with parallel architectures, for example, the letter perception task that
McClelland and Rumelhart4 simulated in their early Psychological Review papers.
In that case9 both architectures gave good agreement with the experimental data,
and we were able to identify the features of the architectures that made this possible.
There is a need for much more experimentation along this parallel/serial boundary,
Near Decomposability and Complexity 41
CONCLUSION
In this chapter, I have described what it means to explain human thought at the
symbolic level, characterizing the brain as a physical symbol system. I have shown
why we need several levels of cognitive theory: a neurological level to deal with
events of less than a few hundred milliseconds duration, and a symbolic level to
deal with the more aggregated slower events. The hierarchical structure of psycho-
logical theories resembles the hierarchical structure of theories in the other sciences
because, in the brain as elsewhere, hierarchy has provided nature with powerful
means for building reliable complex systems by assembling collections of simpler
stable components. Neuroscience and symbolic cognitive science must collaborate
in order to create a comprehensive, well-formed psychology.
I have sketched briefly some of the leading generalizations about thinking that
have emerged at the symbolic level, especially problem solving by heuristic search
and problem solving by recognition. These processes are strongly molded by the
bottleneck of short-term memory which guarantees that most events with duration
of a half second or more will occur serially rather than in parallel. Symbolic models
that do parallel processing will probably have their main applications at the per-
ceptual and motor peripheries of the neural system rather than in the more central
portions that are involved in problem solving and other "higher" activities.
42 H. A. Simon
ACKNOWLEDGMENTS
This research was supported by the National Science
Fou
9121027; and by the Defense Advanced Research Pro ndation, Grant No. DBS-
ject
Defense, ARPA Order 3597, monitored by the Air Forc s Agency, Department of
contract F33615-81-K-1539. Reproduction in whole or e Avionics Laboratory under
in part is permitted for any
purpose of the United States Government. Approved
for public release; distribution
unlimited.
REFERENCES
1. Feigenbaum, E. A., and H. A. Simon. "EPAM-Lik
e Models of Recognition
and Learning." Cog. Sci. 8 (1984): 305-336.
2. Goldman-Rakic, P. S. "Working Memory and the
Mind." Sci. Am. 267(3)
(1992): 111-117.
3. Larkin, J. H., and H. A. Simon. "Why a Diagram
is (Sometimes) Worth
.10,000 Words." Cog. Sci. 11 (1987): 65-99.
4. McClelland, J. L., and D. E. Rumelhart. "An Inte
ractive Activation Model of
Context Effects in Letter Perception: Part 1. An Acc
ount of Basic Findings."
Psych. Rev. 88 (1981): 375-407.
5. Neves, D. M. "A Computer Program that Learns
Algebra
Examining Examples and Working Problems in a Tex ic Procedures by
of the Second National Conference of the Canadian Soctbook." In Proceedings
iety for Computational
Studies of Intelligence, 1978.
6. Newell, A. Unified Theories of Cognition. Cambrid
ge, MA: Harvard Univer-
sity Press, 1990.
7. Newell, A., and H. A. Simon. "Computer Science
as Empirical Inquiry: Sym-
bols and Search." Comm. ACM 19 (1976): 113-126.
Near Decomposability and Complexity 43