Professional Documents
Culture Documents
Leonelli (189-210) Understanding in Biologý PDF
Leonelli (189-210) Understanding in Biologý PDF
J:@<EK@=@:
LE;<IJK8E;@E>
pa rt title ver so
chap ter title ver so
J:@<EK@=@:
LE;<IJK8E;@E>
Philosophical Perspectives
< ; @K < ; 9 P
?\ebN%[\I\^k#
JXY`eXC\fe\cc`#Xe[
BX`<`^e\i
LE@M<IJ@KPF=G@KKJ9LI>?GI<JJ
pa rt title ver so
:FEK<EKJ
Preface ix
. Focusing on Scientific Understanding
henk w. de regt, sabina leonelli, and kai eigner
EVgi>#JcYZghiVcY^c\!:meaVcVi^dc!VcY>ciZaa^\^W^a^in
. Understanding and Scientific Explanation
henk w. de regt
. Understanding without Explanation
peter lipton
. Ontological Principles and the Intelligibility of Epistemic Activities
hasok chang
. Reliability and the Sense of Understanding
stephen r. grimm
. The Illusion of Depth of Understanding in Science
petri ylikoski
EVgi>>#JcYZghiVcY^c\VcYBdYZah
. Understanding in Physics and Biology: From the Abstract
to the Concrete
margaret morrison
. Understanding by Modeling: An Objectual Approach
tarja knuuttila and martina merz
. The Great Deluge: Simulation Modeling and Scientific
Understanding
johannes lenhard
EVgi>>>#JcYZghiVcY^c\^cHX^Zci^ÒXEgVXi^XZh
. Understanding in Biology: The Impure Nature of
Biological Knowledge
sabina leonelli
viii contents
Contributors
Index
chap ter title ver so
GI<=8:<
The idea for this volume was born in the aftermath of the conference “Philo-
sophical Perspectives on Scientific Understanding,” held in Amsterdam in Au-
gust . Our thanks go to all the conference participants and particularly to
Hasok Chang and Peter Lipton, whose engagement and vision were decisive
input for the making of this volume. Peter Lipton’s untimely death in Novem-
ber is an irreparable loss for the philosophy of science community. He
submitted the final version of his contribution to our volume in the summer of
; it is sad that he cannot see it in print anymore. We dedicate this volume
to his memory.
Both the conference and the realization of this volume were part of the
research program “Understanding Scientific Understanding,” sponsored by the
Netherlands Organization for Scientific Research (NWO). We are grateful to
NWO and to the Faculty of Philosophy of the VU University Amsterdam for
the support given to the project and for enabling us to organize a conference to
share our interest with like-minded philosophers.
We would also like to thank our colleagues at VU University, especially Hans
Radder, for their enthusiastic support and for their intellectual contributions to
our thinking. Sabina Leonelli is grateful for the support of her colleagues at the
ESRC Centre for Genomics in Society, University of Exeter, and the project
“How Well Do ‘Facts’ Travel?” London School of Economics.
Finally, we thank Cynthia Miller and her staff at the University of Pittsburgh
Press for their cooperation.
ix
pa rt title ver so
chap ter title ver so
J:@<EK@=@:
LE;<IJK8E;@E>
pa rt title ver so
chap ter title ver so
(
Focusing on Scientific Understanding
= : C @ L # 9 : G : < I! H 6 7 > C 6 A : D C : A A > !
6C9@6>:><C:G
de re gt, leo nelli, and eigner
by reviewing what philosophers of science have said about it, and by outlining
the main themes of the book. Rather than defending specific claims in detail,
a task that we leave to individual chapters, we focus on what we perceive to be
the overall position emerging from this collection.
A key feature of understanding is the fact that it involves a cognizing sub-
ject. In the case of scientific understanding this subject is typically a scientist
who tries to understand a phenomenon, for example by developing a theory.
There is an important difference between understanding and explanation:
while one may legitimately say that theory T explains phenomenon P, one can
only speak about understanding P by means of T if one invokes a subject (in the
case above, the scientist). In other words, while explanation may be viewed as
a two-term relation between an explanans (for example, a theory) and an ex-
planandum (the phenomenon), understanding is always a three-term relation
involving explanans, explanandum, and a subject. This feature of understand-
ing has important implications.
First of all, it entails that understanding belongs, at least partly, to the do-
main of pragmatics: when we want to analyze how scientific understanding of
a phenomenon is achieved, we cannot restrict ourselves to analyzing the syn-
tactic structure and semantic content of the statements (for example, theories,
explanations) involved in this activity. In addition, we should consider the pur-
pose and effects of these statements for the particular persons who use them.
This applies to understanding as the end-product of explanatory activity (the
cognitive state achieved, the goal of explanation), because it is the state of a
cognizing subject. And it applies even more to understanding the theories and
models used in these explanatory activities: such an understanding concerns
the usefulness and tractability of the theories and models as evaluated by a
specific group of peers.
Philosophers in the logical empiricist tradition, most notably Carl Hempel,
claim that for this reason the notion of understanding falls outside the scope
of the philosophy of science. Their argument is that a philosophical account of
science should be objectivist and steer clear of pragmatic elements. Accord-
ing to this viewpoint, understanding is nothing but a subjective feeling evoked
by explanations, which is irrelevant for philosophy of science for it can have
no epistemic import. Nowadays, however, many philosophers of science have
more liberal viewpoints regarding the relevance of pragmatics for philosophy
of science and see no reason to reject the notion of understanding out of hand.
Obviously, the authors in this volume are among this group: they agree that
understanding is a philosophically interesting and relevant notion, despite or
by virtue of its pragmatic aspects.
Another implication of viewing understanding as a three-term relation is
accepting that it may be context-dependent. Whether a particular explanation
de re gt, leo nelli, and eigner
JcYZghiVcY^c\^ci]ZE]^adhde]nd[HX^ZcXZ
Until quite recently philosophers of science have paid relatively little at-
tention to the topic of scientific understanding because, as noted above, un-
derstanding necessarily involves a subject. Many philosophers of science,
especially those raised in the objectivist tradition of logical empiricism, took
this to imply that understanding is thereby also subjective. They concluded that
it can be nothing more than a psychological by-product of scientific activity. In
Hempel’s words, “such expressions as ‘realm of understanding’ and ‘compre-
hensible’ do not belong to the vocabulary of logic, for they refer to the psycho-
logical and pragmatic aspects of explanation” (Hempel , ).
According to Hempel, every good scientific explanation is a logically valid
argument in which the explanandum is deduced from an explanans contain-
ing at least one universal law and relevant initial and background conditions.
This is the well-known deductive-nomological (D-N) model, which was first
proposed by Hempel and Oppenheim in .1 In this pioneering paper (re-
printed in Hempel ), the authors distinguished between “understanding in
the psychological sense of a feeling of empathic familiarity” and “understand-
ing in the theoretical, or cognitive, sense of exhibiting the phenomenon to be
explained as a special case of some general regularity” (Hempel , ).
They criticized the former as being neither a necessary nor a sufficient require-
ment for scientific explanations, and argued that explanations can only be said
fo c using on scientific under standing
scientific explanations. There have been two major approaches in the objectiv-
ist tradition: the unificationist approach, advanced by Friedman, Philip Kitcher,
and others, and the causal-mechanistic approach, developed by Wesley Salmon
and others. In addition, more in line with Scriven’s position is a pragmatic ap-
proach to explanation advanced by Bas van Fraassen and Peter Achinstein.
Friedman’s idea was that objective understanding is provided by explana-
tions that unify the phenomena: “science increases our understanding of the
world by reducing the total number of independent phenomena that we have to
accept as ultimate or given” (Friedman , ). Because of technical problems
with Friedman’s account, Philip Kitcher (, ) developed an alternative
approach to unification in which argument patterns are central. An argument
pattern consists of a schematic argument with filling instructions and a clas-
sification of the inferential structure of the argument. According to Kitcher,
“Science advances our understanding of nature by showing us how to derive
descriptions of many phenomena, using the same patterns of derivation again
and again” (, ). Since Friedman’s and Kitcher’s essays, much more work
has been done on unification. A notable account that refers especially to the
understanding-providing power of unification is that of Gerhard Schurz and
Karel Lambert ().
Meanwhile, following the severe critiques launched against the D-N model,
causal approaches to explanation were developed that offered an alternative
conception of explanatory understanding. The most important proponent of
this view is Wesley Salmon, who presented an elaborate theory of causal-mech-
anistic explanation in his Scientific Explanation and the Causal Structure
of the World. This theory was later refined and elaborated in discussion with
others (see Salmon , esp. chaps. and ). According to Salmon, we need
a causal theory of explanation because “underlying causal mechanisms hold
the key to our understanding of the world” (Salmon , ; italics added).
This is because “causal processes, causal interactions, and causal laws provide
the mechanisms by which the world works; to understand why certain things
happen, we need to see how they are produced by these mechanisms” (). In
his later work, Salmon has put even stronger emphasis on the importance of
achieving understanding in science (see Salmon , , , , and chap. ).
Notable alternative causal approaches are defended by Paul Humphreys ()
and James Woodward (). Mechanistic explanation is nowadays very popu-
lar in the philosophy of biology (Machamer, Darden, and Craver ; Bechtel
and Abrahamsen ).
A different approach to explanation and understanding, which falls out-
side the objectivist camp, is provided by van Fraassen () and Achinstein
(). In van Fraassen’s account of science, explanation is a pragmatic virtue
fo c using on scientific under standing
and does not constitute an epistemic aim of science. Van Fraassen analyzes
explanations as answers to why-questions, and argues that explanation is con-
text dependent. Although the issue of the character of scientific understand-
ing is not mentioned explicitly by van Fraassen, one may safely conclude that
his analysis of the nature of explanation as extrascientific holds a fortiori for
understanding. Achinstein () focuses more directly on understanding by
providing a critique of noncontextual approaches to explanation. In his view,
the act of explaining is an illocutionary act characterized by the intention to
make something understandable. Whether an explanation is “good” depends
on the values and interests of the audience to which it is directed. Achinstein
emphasizes that, due to this context dependency, there is a variety of different
forms of good explanation in science.
6heZXihd[HX^Zci^ÒXJcYZghiVcY^c\
The above overview shows that understanding has typically been viewed as
a by-product of scientific explanation. This volume provides a radically different
approach: the study of understanding is not only worthy of philosophical atten-
tion, but actually contributes new insights to a number of traditional debates
within the philosophy of science. The most important themes emerging from
this collection of papers are the following: the relation between understanding
and explanation; the feeling of understanding; the role of models and theories
as tools to achieve understanding; the notion of intelligibility; understanding as
tacit knowledge; and pluralism of understanding. This list highlights some key
results obtained through our philosophical investigation of scientific under-
standing; it also serves to guide the reader toward the specific chapters where
reflections on these themes are most thoroughly developed.
JcYZghiVcY^c\VcY:meaVcVi^dc
The objectivist position reduces understanding to an automatic conse-
quence of having a good explanation, where “good” indicates its unifying power
or causal-mechanistic nature. As discussed above, this view does not resolve
the issue of what understanding actually involves. We take issue with the facile
equation between possessing an explanation and possessing understanding of
a phenomenon. Gaining understanding through explanations is not an auto-
matic process, but rather a cognitive achievement in its own right—a point that
is stressed, in various contexts, by all contributors to this collection.
Once it is granted that deriving understanding from an explanation is a
matter of ability, the question arises of how that actually works. What are the
mechanisms through which scientists extract understanding from explanations
de re gt, leo nelli, and eigner
that they already possess? Confronting this question in the case of historiog-
raphy, Edwin Koster emphasizes the role of personal judgment in determining
which ideas and events are selected as part of an explanation. Henk de Regt
and Sabina Leonelli both note how judgment in the natural sciences involves
the exercise of epistemic skills of various kinds, which help scientists to make
sense of the conceptual as well as the empirical significance of explanations for
the phenomena to which they refer. Peter Lipton strengthens the point by chal-
lenging the very idea that explanations are needed to acquire understanding: he
discusses cases where individuals acquire understanding without recourse to
explanations. Further, Lipton distinguishes four types of cognitive benefits that
derive specifically from understanding a phenomenon: causal information, as
acquired through activities such as observation, experimentation, manipula-
tion, and inference; a sense of necessity deriving from the knowledge that, in
given conditions, an event, object, or process could not have been otherwise;
a sense of what is possible, which can spring even from potential or false ex-
planations; and the unification obtained by comparing phenomena through
mechanisms such as analogies and classifications, as in the case of Kuhnian ex-
emplars. Lipton shows that each of these cognitive benefits can be obtained in
the absence of explanations, indicating that “understanding is more extensive
and more varied in its sources than those who would equate explanation and
understanding allow.”
I]Z;ZZa^c\d[JcYZghiVcY^c\
Petri Ylikoski explores cases where the acquisition of scientific understand-
ing is at least partly illusory, insofar as it springs from a feeling of understanding
(“Aha!”) that has no relation to whether that understanding is reliable or truth-
conducive. This cognitive phenomenon is recognized by several authors in the
volume, who all agree on the necessity to distinguish the feeling of understand-
ing and understanding itself, but who disagree quite radically on whether the
feeling of understanding has any epistemic value. De Regt does not think that
this is the case. In his view, “the phenomenology of understanding has no epis-
temic function”: what counts, rather, are the epistemic skills and values used
to construct tools for understanding, such as intelligible theories and models.
Ylikoski goes even further by characterizing this feeling, which he dubs “illu-
sion of depth of understanding,” as having a downright pernicious effect on
various aspects of scientific research. He notes how several features of contem-
porary science have evolved as measures against illusory understanding felt
(and promoted) by individuals. For instance, the peer review system exposes
personal insights to critique, while the development of external representations
makes it possible to express individual intuitions so that they can be discussed
fo c using on scientific under standing
and tested. According to Ylikoski, scientists can improve their ability to dis-
cern reliable from illusory understanding by adopting strict standards for what
constitutes an adequate explanation—and philosophers of science are ideally
placed to contribute to this effort.
By contrast, Lipton and Stephen Grimm share a belief in the epistemic
value of the “aha” experience, which they both characterize as helpful rather
than detrimental to the search for scientific understanding. After illustrating
the differences between the feeling of understanding and other pleasant feel-
ings involved in research, such as surprise and expectation, Lipton notes how
the “aha” experience might plausibly work not only as an incentive, but also as
a guide to the human search for understanding. Lipton proposes that scientists
use their feelings of understanding as ground for their choice of the “loveliest”
explanation. This does not imply that understanding itself is hopelessly subjec-
tive, as Lipton makes clear by distinguishing the feeling of understanding from
actual understanding (as he states, “understanding is a form of knowledge; the
feeling of understanding is not”). Rather, it points to the unarticulated, and yet
epistemically relevant, grounds on which scientists judge their results. Grimm
complements this argument by pointing to the “conditional reliability” of the
feeling of understanding, which depends chiefly on the accuracy of the back-
ground beliefs of whoever experiences it. Scientists are most likely to feel that
they understand something when that bit of information coheres with the rest
of their beliefs about the world. This by no means guarantees the reliability of
the feeling of understanding, but it does imply that it can be checked by review-
ing one’s background knowledge (or, as Johannes Lenhard proposes, by employ-
ing simulations to check the empirical consequences of one’s assumptions).
BdYZah!I]Zdg^Zh!VcYJcYZghiVcY^c\
A crucial question raised in this volume concerns the relation between
models and theories as aids toward acquiring understanding. Models are the
subjects of part and their role toward the acquisition of understanding is
highlighted by virtually all contributors. There are, however, various interpre-
tations of the relation between models and theories and of how scientists might
use either or both of them as tools to understand phenomena.
Some authors support the view that theories are only useful for under-
standing when they are represented through models. Lenhard argues along
these lines in his discussion of understanding through computer simulations.
In his view, simulations constitute surrogates for theories: they allow scientists
to explore the implications of adopting specific theoretical assumptions, yet
at the same time the development of simulations requires scientists to black-
box several aspects of the underlying modeling and theories so as to enable
de re gt, leo nelli, and eigner
>ciZaa^\^W^a^in
One well-known case of “using” instruments to understand phenomena
is Galileo’s use of the balance to understand principles of equilibrium. As ar-
gued by Machamer and Woody (), using the balance crucially improved
the intelligibility of Galileo’s views on motion to other scientists. Indeed, un-
derstanding unavoidably involves making something intelligible. Yet what de-
termines intelligibility and how can intelligibility be improved? Premodern
rationalist philosophers held that knowledge must be based on self-evident
“intelligible principles.” Modern science has rejected this methodology, and
empiricist philosophy leaves no room for intelligibility in this sense. Kant’s phi-
losophy revived the idea of intelligibility in a different form, namely via the
forms of intuition (Anschauungsformen) that structure sensory input and the
categories that transform it into knowledge. However, in the early twentieth
century the alleged refutation of the Kantian system by relativity and quantum
physics led again to a rejection of intelligibility as a philosophically reputable
concept, especially by logical-positivist and logical-empiricist philosophers
(Frank provides a nice example of this tendency—this semipopular in-
troduction to the philosophy of science turns out to be one long attempt to
debunk intelligibility).
In his chapter, Hasok Chang gives new vitality to the Kantian insights by
viewing intelligibility as an epistemic virtue whose function is to harmonize
our actions with our basic beliefs about the world. An unintelligible action is
an action that cannot be performed because it betrays one of the principles
guiding all human epistemic activity—for instance, the principle that “a real
physical property can have no more than one definite value in a given situa-
tion,” which Chang calls the ontological principle of single value. The betrayal
of an ontological principle does not represent a mistake or a falsehood; rather,
it involves taking up a belief that makes no sense to us (such as the belief that a
physical property can have two different values in the same situation), because
we would not know how to act on its basis. Ontological principles constitute
the platform of common sense on which the whole scientific enterprise is built
through its various activities, including testing, experimentation, observation,
and the like. In a pragmatist fashion, intelligibility is thus defined as the per-
formability of an epistemic activity—and understanding as the ability to actu-
ally perform such an activity.2
de re gt, leo nelli, and eigner
JcYZghiVcY^c\VhIVX^i@cdlaZY\Z
Objectivist objections to viewing understanding as a reliable source of
knowledge might perhaps seem plausible if one focuses only on theoretical
knowledge. However, a precious lesson taught by the philosophical literature on
models concerns the importance of tacit (or embodied) knowledge for obtain-
ing and interpreting theoretical knowledge. It is only through the use of mod-
els, or indeed any other kind of object or representation, that scientists acquire
understanding of the world. The unarticulated knowledge required for success-
ful interaction with models, phenomena, and experimental settings constitutes
not only an important source of knowledge in itself, but also a strong constraint
to the types of understanding that are actually developed in science.
As Knuuttila and Merz make clear in their chapter, scientists typically seek
to extract evidence from objects with very specific characteristics. Models
provide understanding insofar as they constitute “concrete constructed ob-
jects with which one can interact in various ways.” In their view, understand-
ing springs from the scientists’ interaction with these objects of knowledge.
Leonelli defines tacit knowledge more broadly as “the awareness of how to act
and reason as required to pursue scientific research.” Such awareness is ex-
pressed not only through the handling of objects as models, but also through
the implementation of experimental procedures and protocols, the calibration
of instruments, and the ability to work in specific research environments such
as the lab or the field. As Leonelli notes, in the life sciences the development of
explanations and interventions is inextricably intertwined—a result that is not
surprising in the light of the link proposed here between understanding and the
ability to apply theories and models.
fo c using on scientific under standing
EajgVa^hbd[JcYZghiVcY^c\
As we noted, the history of science shows great variation in what is and
what is not deemed understandable. Even at one particular moment in his-
tory opinions about what is understandable often diverge. For example, in
physicists in the Copenhagen-Göttingen circle believed that atomic phenom-
ena could be understood with the theory of matrix mechanics, while most
other physicists—notably Schrödinger—disagreed (Beller ; de Regt ).
Authors in this volume have several ways to account for scientific pluralism
among ways of understanding. One option is to distinguish between embodied
and theoretical knowledge as sources for the skills and commitments used by
scientists to understand phenomena. Theoretical skills and commitments are
the ones involved in reasoning through available concepts, theories, and ex-
planations, while performative skills and commitments are developed through
physical interaction with research objects and settings. Leonelli uses this dis-
tinction to argue for the coexistence of three types of understanding in biology:
one prioritizing recourse to theoretical commitments and skills, one where
performative skills are of primary importance, and one deriving from a bal-
anced coordination of theoretical and embodied knowledge. A second form of
pluralism concerns the diversity of understandings that can be acquired from
using the same tools. This point is supported by Knuuttila and Merz, who stress
the multiplexity of models as tools for acquiring understanding: the same mod-
el can be interpreted in a variety of ways depending on the background skills
and knowledge of their users.
Yet another form of pluralism concerns the epistemic interests and purposes
guiding the search for understanding. Jeroen Van Bouwel shows how “the plu-
rality of epistemic interests is better served by a plurality of theoretical per-
spectives than by a unified one.” In Van Bouwel’s view, understanding derives
from explanations that are both empirically adequate and adequate to the epis-
temic interests of whoever acquires it. Given the existing diversity of epistemic
interests among scientists, understanding is therefore acquired from several
explanations that cannot be reduced to one another. Unification is thus seen as
limiting, rather than aiding in, our understanding of the world.
This view is shared by Dieks in his analysis of physical explanations of rela-
tivistic effects. As he shows, the same physical theory may be used to con-
struct different arguments, depending on “exactly what one wishes to know,
and from which point of view”—that is, on which kind of understanding one
wishes to acquire. Dieks’s work dispels any suspicion that Van Bouwel’s claims
might be uniquely related to the area he studies, that is, political science. As
these two chapters jointly demonstrate, both in the natural and in the social
sciences epistemic interests vary considerably depending on the context and
group adopting them.
de re gt, leo nelli, and eigner
I]ZE]^adhde]^XVa;jijgZd[HX^Zci^ÒXJcYZghiVcY^c\
The arguments presented in this volume bear wide implications for the phi-
losophy of science. All chapters in this volume firmly agree on linking under-
standing to cognition. Understanding is defined as a cognitive achievement, or
as an ability acquired through appropriate training and as the bearer of cogni-
tive benefits (such as, in Lipton’s words, “knowledge of causes, of necessity,
of possibility and of unification”) that scientists could not acquire solely from
explanations. This emphasis on the cognizant individual involves a reevalua-
tion of the epistemic role of human agency in producing, disseminating, and
using scientific knowledge. To understand scientific understanding, philoso-
phers must find ways to study and analyze scientific agency. This means taking
scientific practices seriously, for arguably a study of agency in science needs
to be based on knowledge of how scientists across various fields actually act.
The study of experimentation, the iconic scientific activity, has already yielded
fo c using on scientific under standing
CdiZh
. Hempel later added inductive-statistical explanation to cope with explanations in
which the explanandum cannot be deduced with absolute certainty from the explanans,
but only with a high degree of probability.
. An important outcome of Chang’s analysis is a strong argument for disconnecting
intelligibility from truthfulness (or one of its measures, such as empirical adequacy). This
is a step that several other authors are reluctant to take.
. Paying attention to scientific practices such as experimentation also involves col-
laboration with historians and sociologists of science, whose work has already illuminated
several characteristics of experimental practice (for example, Shapin and Schaffer and
Bruno Latour ).
GZ[ZgZcXZh
Achinstein, P. . The nature of explanation. New York: Oxford University Press.
Bechtel, W., and A. Abrahamsen. . Explanation: A mechanistic alternative. Studies in
History and Philosophy of Biological and Biomedical Sciences :–.
Beller, M. . Quantum dialogue. Chicago: University of Chicago Press.
de Regt, H. W. . Spacetime visualisation and the intelligibility of physical theories.
Studies in History and Philosophy of Modern Physics B:–.
Frank, P. . Philosophy of science: The link between science and philosophy. Englewood
Cliffs, NJ: Prentice-Hall.
Friedman, M. . Explanation and scientific understanding. Journal of Philosophy
:–.
Gooding, D. . Experiment and the making of meaning. Dordrecht: Kluwer.
Hacking, I. . Representing and intervening: Introductory topics in the philosophy of
natural science. Cambridge: Cambridge University Press.
fo c using on scientific under standing
G8IK@
JcYZghiVcY^c\!:meaVcVi^dc!
VcY>ciZaa^\^W^a^in
pa rt title ver so
chap ter title ver so
)
Understanding and Scientific Explanation
= : C @ L# 9 : G : < I
henk w. de r egt
ories not only summarize facts but also ipso facto uncover relations between
facts (sometimes from quite different domains of knowledge), and our grasping
of these relations may legitimately be called “understanding.”
From a hindsight perspective, Schrödinger’s words appear to be very in-
sightful. The early s were the time when the subject of explanation was be-
coming respectable again, after having been condemned by the nineteenth- and
early-twentieth-century positivist philosophers of science (see, for example,
Mach /, Pearson /). It was only in that Carl Hempel
and Paul Oppenheim put the topic on the philosophical agenda again, with
their pioneering paper “Studies in the Logic of Scientific Explanation.”1
But Hempel, whose deductive-nomological model of explanation was to
dominate the debate for the next two decades, was still reluctant to talk about
understanding. The reason was that “such expressions as ‘realm of understand-
ing’ and ‘comprehensible’ do not belong to the vocabulary of logic, for they
refer to the psychological and pragmatic aspects of explanation” (Hempel ,
). In Hempel’s logical-empiricist view, the aim of philosophy of science is to
give an account of the objective nature of science by means of logical analysis
of its concepts. Its psychological and pragmatic aspects may be of interest to
historians, sociologists, and psychologists of science (in short, to those who
study the phenomenon of science empirically) but should be ignored by phi-
losophers. Since Hempel, philosophers of science have gradually become more
willing to discuss the topic of scientific understanding, but it has remained out-
side the focus of philosophical attention until quite recently.
In this chapter, I argue that understanding is an important philosophical
topic and I offer an account of the role of understanding in science. The outline
of the chapter is as follows. In the next section, I discuss objectivist views of the
relation between explanation and understanding, in particular those of Hempel
and J. D. Trout. I challenge these views by arguing that pragmatic aspects of
explanation are crucial for achieving the epistemic aim of science. I then offer
an analysis of these pragmatic aspects in terms of “intelligibility” and give a
contextual account of scientific understanding based on this notion. Finally, the
last section presents a brief case study from the history of physics as an illustra-
tion of the proposed account of scientific understanding.
DW_ZXi^k^hiK^Zlhd[:meaVcVi^dcVcYJcYZghiVcY^c\
The hallmark of scientific knowledge is, in Hempel’s view, its objective nature,
and philosophers of science should therefore try to give an objectivist account
of science, and of scientific explanation in particular—and therefore they should
ignore pragmatic aspects such as understanding and intelligibility.
It should be stressed that Hempel was right that the notion of understanding
is pragmatic in the sense that it is concerned with a three-term relation between
the explanation, the phenomenon, and the person who uses the explanation
to achieve understanding of the phenomenon. One can use the term “under-
standing” only with—implicit or explicit—reference to human agents: scientist
S understands phenomenon P by means of explanation E. That understanding
is pragmatic in this sense implies the possibility of disagreement and variation
based on contextual differences.2 But, as I will argue below, Hempel was wrong
when he stated that pragmatic notions are philosophically irrelevant.
Hempel sometimes used the term “scientific (or theoretical) understand-
ing,” which he associated with scientific explanation in the deductive-nomo-
logical sense.3 In the concluding section of his essay he remarks that the
understanding conveyed by scientific explanation “lies rather in the insight that
the explanandum fits into, or can be subsumed under, a system of uniformities
represented by empirical laws or theoretical principles,” and that “all scientific
explanation . . . seeks to provide a systematic understanding of empirical phe-
nomena by showing that they fit into a nomic nexus” (Hempel , ).
In , Michael Friedman argued that philosophical accounts of scien-
tific explanation should make it clear how explanations provide us with under-
standing, and he claimed that Hempel’s D-N model of explanation failed to do
so. To be sure, Hempel had suggested how science may lead to understanding:
“The [D-N] argument shows that, given the particular circumstances and the
laws in question, the occurrence of the phenomenon was to be expected; and
henk w. de r egt
it is in this sense that the explanation enables us to understand why the phe-
nomenon occurred” (Hempel , ; italics in original; compare ). In
this way, “understanding” is reduced to rational expectation. But, Friedman
(, ) observes, well-known examples, like the barometer prediction of a
storm, show that rational expectation is insufficient for understanding: we do
not understand the occurrence of a storm by merely referring to the barome-
ter’s indication.
While Friedman agrees with Hempel’s objectivism, he argues that Hempel
wrongly identified “pragmatic as psychological” with “pragmatic as subjective.”
Friedman (, –) holds that understanding can be pragmatic in the sense
of being psychological without also being subjective. Accordingly, an objec-
tivist conception of scientific understanding (connected with explanation as
its product) should be possible. Friedman defends an objectivist account of
scientific understanding based on the idea of unification (which was further
developed by Kitcher ). Since Friedman’s pioneering essay, philosophers
of science have been less reluctant to discuss the notion of understanding,
and particular models of explanation are typically defended by pointing to the
understanding-providing power of the favored type of explanation (for example,
Kitcher on unification, and Salmon and on causal-mechanical
explanation). However, these philosophers typically refrain from presenting a
detailed analysis of the notion of understanding itself.
Moreover, some present-day philosophers still endorse the Hempelian view
that understanding (in the psychological, allegedly subjective sense) should be
banned from philosophical discourse. One of them is J. D. Trout (, ,
), who follows Hempel in arguing that philosophical analyses of scientific
explanation should be objectivist in the following sense: “What makes an ex-
planation good concerns a property that it has independent of the psychol-
ogy of the explainers; it concerns features of external objects, independent of
particular minds” (Trout , ). Therefore, Trout claims, philosophers of
science should eschew subjective notions such as intelligibility and understand-
ing. The latter are epistemically irrelevant if not dangerous (because hindsight
and overconfidence biases show that feelings of understanding are typically
misguided). Trout criticizes philosophers of explanation such as Friedman and
Salmon for (unwittingly) relating explanatory power to a subjective sense of
understanding, and he notes that even Hempel could not resist the tempta-
tion to give an account of how deductive-nomological explanations produce
a feeling of understanding (namely, as mentioned above, by showing us that
the explained event was to be expected). According to Trout, we should not
try to justify our theories of explanation by referring to alleged understanding-
providing features. The fact that a particular explanation of a phenomenon
under standing and scientific e xpl anation
I]Z:e^hiZb^XGZaZkVcXZd[H`^aah
;ZZa^c\d[jcYZghiVcY^c\;J I]Ze]ZcdbZcdad\n!ÆV]VÇ
ZmeZg^ZcXZVXXdbeVcn^c\VcZmeaVcVi^dc
JcYZghiVcY^c\Vi]ZdgnJI 7Z^c\VWaZidjhZi]Zi]Zdgn
JcYZghiVcY^c\Ve]ZcdbZcdcJE =Vk^c\VcVYZfjViZZmeaVcVi^dc
d[i]Ze]ZcdbZcdc
henk w. de r egt
In a similar vein, Bas van Fraassen (, , ) explicitly contrasts an epis-
temic and a pragmatic dimension of theory acceptance. The epistemic dimen-
sion contains the relevant beliefs concerning the relation between a theory and
the world, while the pragmatic dimension contains reasons scientists may have
for accepting a theory independently of their beliefs about its relation to the
world; these reasons typically pertain to the use and the usefulness of the the-
ory (van Fraassen , ). Both Hempel and van Fraassen see the epistemic
and the pragmatic as sharply distinguished domains. The pragmatic dimension
pertains to the relation between the theory and its users, that is, to the dimen-
sion that seems to be excluded from the epistemic dimension by definition. The
thesis that epistemic and pragmatic can and should be kept separate presup-
poses that the epistemic status of a theory only and uniquely depends on a di-
rect evidential relation with the phenomena it purports to describe or explain.
On this presupposition, pragmatic elements—such as virtues of a theory that
facilitate its use by scientists—are indeed epistemically irrelevant: they do not
carry any additional justificatory weight.
However, as the previous discussion of the role of skills in deductive rea-
soning makes clear, this presupposition is false. Establishing relations between
theories and phenomena, as happens in deductive-nomological explanation,
crucially depends on skills and judgment. It follows that although it is surely
possible and useful to distinguish analytically between the epistemic and the
pragmatic, the two are inextricably intertwined in scientific practice: epistemic
activities and evaluations (production and assessment of knowledge claims) are
possible only if particular pragmatic conditions are fulfilled.
Thus, the pragmatic dimension of explanation is epistemically relevant.
This is the case for D-N type explanations, but even more so for scientific ex-
planations that do not conform to the ideal D-N scheme. As mentioned above,
Hempel’s deductive-nomological model of explanation fails as a general ac-
count of scientific practice: actual explanations often are not of the D-N type
because usually there is no strictly deductive relation between explanans (the-
under standing and scientific e xpl anation
real system to a suitable model. But once we have constructed a model to which
the theory can be applied, equations for the behavior of the model system can
be deduced: “There are just rules of thumb, good sense, and, ultimately, the re-
quirement that the equation we end up with must do the job” (Cartwright ,
). Thus, the scientist must rely on skills and practical judgment, as has been
observed already by Hilary Putnam:
I]Z>ciZaa^\^W^a^ind[I]Zdg^Zh
tive measure, however, because a theory can be more fruitful for one scientist
than for another.13 Which theories are deemed intelligible can vary through
time or across disciplines. Scientists prefer a more intelligible theory over a less
intelligible one, not because it gives them a psychological sense of understand-
ing, but rather because they have to be able to use the theory.
The arguments presented in the section “The Epistemic Relevance of Skills,”
above, can be summarized by the thesis that scientists need intelligible theories
in order to achieve scientific understanding of phenomena. Only an intelligible
theory allows scientists to construct models through which they can derive ex-
planations of phenomena on the basis of the given theory. In other words, un-
derstanding scientific theories is a prerequisite for understanding phenomena
scientifically. This can be stated as a criterion for understanding phenomena
(CUP):
>ciZaa^\^W^a^inVcYi]Z@^cZi^XI]Zdgnd[<VhZh
A case study from the history of physics can illustrate the above account of
scientific understanding. The case study concerns nineteenth-century attempts
to explain experimentally observed properties of gases on the basis of the ki-
netic theory of gases. The kinetic theory, developed in the s, was based on
henk w. de r egt
the idea that heat is mechanical motion, and that gases consist of molecules
in motion behaving in accordance with the laws of Newtonian mechanics. In
order to explain and predict experimental facts, one needed specific models
of molecules. Among the phenomena to be explained were the well-known
experimental gas laws (Boyle-Charles, Gay-Lussac), transport phenomena (vis-
cosity, heat conduction, and diffusion), temperature independence of specific
heats, and ratios of specific heats at constant volume and pressure. Later, in the
s, still more experimental knowledge was accumulated, most importantly
regarding spectral emission of gases.
The first mature versions of the kinetic theory were published by Rudolf
Clausius (/) and James Clerk Maxwell (/). In both cases the
core of the theory consisted in an application of the laws of classical mechan-
ics to aggregates of particles (molecules), where the description of the system
did not consist in tracing the motion of each particle individually but in a sta-
tistical treatment. The elementary model used by both Clausius and Maxwell
was that of the molecule as a hard elastic sphere.16 Having modeled molecules
in this way, kinetic theory supplies equations for the behavior of the system.
One of these equations concerns the specific heat ratio γ, that is, the ratio be-
tween the specific heat at constant pressure and the specific heat at constant
volume. The observed variation in specific heat ratios for different gases can
be explained by modeling these gases as polyatomic molecules with different
energies of translation and/or rotation (see de Regt ). In , Maxwell
based his theoretical predictions on his equipartition theorem, which asserts
that there is a constant equal distribution of kinetic energy over the various
possible (translational, rotational, and/or vibrational) “modes of motion” of the
molecule. However, Maxwell discovered that for many common gases (oxygen,
nitrogen) the equation led to a discrepancy between theoretical predictions
and experimental results for specific heat ratios: theoretically, the ratio of total
kinetic energy to energy of translation was β = , while the experimental value γ
= . implied that β = . (Maxwell /, ). This problem became
known as the “specific heat anomaly.”
Some years later, in , Maxwell introduced the idea of “degree of free-
dom,” which provided a new way of modeling molecules.17 Maxwell assumed
that “the position and configuration of the molecule can be completely ex-
pressed by a certain number of variables,” and added: “Let us call this number
n. Of these variables, three are required to determine the position of the centre
of mass of the molecule, and the remaining n– to determine its configuration
relative to its centre of mass. To each of the variables corresponds a different
kind of motion” (Maxwell /, ). Subsequently, Maxwell () pre-
sented a new formula for the specific heat ratio:
under standing and scientific e xpl anation
+n+e
γ=
n+e (1)
;^\jgZ'#daiobVccÉhYjbWWZaa
bdYZad[Y^Vidb^XbdaZXjaZh#
This simple model thus provides an explanation, and thereby scientific un-
derstanding, of the anomalous specific heat ratios. However, as Boltzmann
(/, ) admits, the dumbbell model cannot be a completely realistic
description of diatomic molecules because real gas molecules do have vibra-
tional degrees of freedom, as is already shown by the fact that they are capable
of emitting spectral lines.
Boltzmann’s dumbbell model nicely illustrates that model building is a
henk w. de r egt
nondeductive process. The model does not follow deductively from the kinetic
theory, and it cannot be derived from experimental data either: approxima-
tions and idealization are necessary in order to achieve a fit between model
and data:
8dcXajh^dc
CdiZh
Thanks to Kai Eigner, Frank Hindriks, Tarja Knuuttila, Sabina Leonelli, and Hans
Radder for helpful discussion.
. “Studies in the Logic of Scientific Explanation” was reprinted with a new
postscript in Hempel ().
. For example, in physicists in the Copenhagen-Göttingen circle believed that
atomic phenomena could be understood on the basis of matrix mechanics, while most
other physicists—notably Schrödinger—disagreed. Such differences may be traced back
to different scientific, philosophical, or social backgrounds (see de Regt , ).
. See especially his discussions of explanation in the social and historical sciences,
where he contrasts it with empathic understanding. Hempel deems the latter neither
necessary nor sufficient for scientific understanding (see Hempel , –, –).
. In response to my discussion of his essay (de Regt ), Trout (, )
claims that he does not want to eliminate talk of understanding completely: it is only
allusions to the (psychological) sense of understanding that he criticizes. He allows
for “understanding proper” in contrast to “the mere sense of understanding,” and for
“genuine understanding” in contrast to “counterfeit understanding.” However, he does
not explain what this purely objectivist conception of understanding consists in. He
refers to “objectivist accounts of implicit understanding” () and associates this with
Kitcher’s internalization of argument patterns, but does not give a positive account of
genuine understanding ().
. Compare Koster, this volume, for an in-depth analysis of the role of judgment in
the field of history.
. Note, however, that there are widely different accounts of representation, and that
the target systems of models may be real, fictional, or ideal (Knuuttila ).
under standing and scientific e xpl anation
. One might argue that the kinetic theory consists only of the laws of Newtonian
mechanics (and some additional statistical laws and assumptions), and accordingly does
not contain a model of gases. But then it is hard to understand the kinetic theory as a
distinct theory and as a theory of gases—it seems that the particle model of gases forms
an essential part of the kinetic theory.
. Strictly speaking, the term “autonomous” is too strong: the model is not fully
independent of the theory.
. See Kuhn (, ); McMullin (, –); Longino (, ); and Lacey
(). These theoretical virtues are called “values” by Kuhn (, ), “epistemic
values” by McMullin () and “constitutive values” by Longino (). I use the term
“virtues” in order to avoid confusion: the virtue is a property of the theory, which may
be valued by scientists. Furthermore, it is difficult to draw a sharp distinction between
epistemic and pragmatic values/virtues (McMullin), or between constitutive and
contextual values/virtues (Longino). For my purposes, this is not necessary because my
account of scientific understanding implies that there is interaction between pragmatic
and epistemic dimensions of science.
. The production of accurate theories is more often stimulated than hindered by
reliance on other (allegedly nonepistemic) theoretical virtues. For example, McAllister
() argues, on the basis of historical evidence, that scientists usually rely on
established aesthetic canons (containing accepted aesthetic criteria for theory choice),
and that this conservatism is conducive to empirical progress.
. See de Regt () for examples from modern physics.
. It often happens that a theory can be represented in different ways (think of the
various formulations of classical mechanics), and each of these representations may have
its own specific virtue, which may be relevant to the intelligibility of the theory.
. There may be a discrepancy between scientists’ perception of the fruitfulness
of a theory and its actual fruitfulness for these scientists. Such errors of judgment will
hamper their work. I assume that they are the exception rather than the rule.
. This formulation differs slightly from that given in de Regt and Dieks ().
. To be sure, these requirements are included in the list of (scientific, epistemic,
constitutive) virtues that are subject to valuation by scientists. Accordingly, there may be
some variation in the way they are valued and applied in specific cases. However, they
are generally accepted among scientists, and explanations that completely fail to meet
them will be rejected as unscientific.
. As mentioned above, the boundary between theory and model is vague and
arbitrary to some extent. In the case of the kinetic theory there is good reason for
drawing the distinction as follows: the representation of gases as aggregates of molecules
in motion is part of the theory, while the representation of molecules as hard elastic
spheres is part of the modeling stage. This is because all kinetic theories share the former
henk w. de r egt
model of a gas, while the latter model of a molecule is not an essential part of the theory
(kinetic theory can operate on different molecular models).
. Maxwell introduced the concept but did not yet use the phrase “degrees of
freedom,” which was introduced by H. W. Watson (/, vi).
. One might ask why it was Boltzmann and not Maxwell who proposed the
dumbbell model. Maxwell had introduced the concept of degrees of freedom, and
surely had the skills to construct the model. When Boltzmann advanced his dumbbell
model, Maxwell rejected it because he objected to the idealizations regarding rigidity
and vibrational degrees of freedom. The difference between Maxwell’s and Boltzmann’s
attitudes toward these idealizations can be traced back to differing philosophical
commitments (see de Regt ).
GZ[ZgZcXZh
———. . Paying the price for a theory of explanation: De Regt’s discussion of Trout.
Philosophy of Science :–.
———. . The psychology of scientific explanation. Philosophy Compass :–.
van Fraassen, B. C. . The scientific image. Oxford: Clarendon Press.
Watson, H. W. /. A treatise on the kinetic theory of gases. nd ed. Oxford, NY:
Clarendon Press.
chap ter title ver so
*
Understanding without Explanation
E:I:GA>EIDC
peter lip ton
they may show that the phenomenon had to occur, they may show how the
phenomenon could have occurred, and they may show how the phenomenon
fits into a broader pattern.
The switch from identifying understanding with explanation to identifying
it with some of the cognitive benefits of an explanation may not seem dramatic,
but it makes this essay possible. For by distinguishing explanations from the
understanding they provide, we make room for the possibility that understand-
ing may also arise in other ways. It is that space that I will begin to survey here:
the possibilities for understanding without explanation. I will look briefly in
turn at the four cognitive benefits of explanation—causation, necessity, pos-
sibility, unification—and suggest how each of them might also be acquired by
routes that do not pass through explanations. If these suggestions are along the
right lines, then understanding is more extensive and more varied in its sources
than those who would equate explanation and understanding allow. This is of
considerable intrinsic interest, and it may also have repercussions for other is-
sues in the philosophy of science, for example, through the connection between
understanding and inference. I will go on to consider a fifth and rather different
consequence that consumers of good explanations enjoy. This is the distinctive
and satisfying feeling of understanding, the “aha” feeling. I will suggest that
although the “aha” feeling is not itself a kind of understanding, it may play an
important role in the way understanding is acquired.
8VjhVi^dc
any phenomenon is long and wide, and most of it will not provide real under-
standing in the context of a scientist’s particular cognitive interests. To account
for this, one needs a model of causal selection that shows which causes pro-
vide understanding and which do not (Lipton , –). But that is another
project. For my present purposes, all I suggest is that some of the tacit causal
knowledge gained from manipulation provides understanding.
Images, models, and manipulation are thus plausible sources of understand-
ing without explanation. How might they be resisted? One might say that, even
if the causal knowledge gained in these cases is partially tacit, the knower will
nevertheless be able to say something, however oblique, about the causes. This
may well be so, but what the knower can say in these cases does not exhaust
what the knower understands. The cases could also be resisted by either ex-
panding our notion of explanation or contracting our notion of understanding.
But neither maneuver is plausible here. A tacit explanation would be an expla-
nation one could not give, even to oneself, and this does violence to the notion
of explanation. And to deny that this tacit knowledge constitutes understand-
ing would be Procrustean. We want a way of marking the difference between
someone who knows that the phenomenon occurs but has no inkling why, and
someone who has a deep and subtle appreciation of its causes, though he or
she is not in a position to articulate that knowledge. “Understanding” seems to
be the right word for the difference. It is not ad hoc to insist that explanations
must be verbal but that understanding need not be.
CZXZhh^in
by showing necessity but without providing an explanation, and there are likely
to be others as well.
Edhh^W^a^in
For there is no doubt that the world was created right from the start
with all the perfection which it now has. . . . This is the doctrine of the
Christian faith, and our natural reason convinces us that it was so. . . .
Nevertheless, if we want to understand the nature of plants or of men,
it is much better to consider how they can gradually grow from seeds
than to consider how they were created by God at the very beginning
of the world. Thus we may be able to think up certain very simple and
easily known principles which can serve, as it were, as the seeds from
which we can demonstrate that the stars, the earth and indeed every-
thing we observe in this visible world could have sprung. For although
we know for sure that they never did arise in this way, we shall be able
to provide a much better explanation of their nature by this method
than if we merely describe them as they now are or as we believe them
to have been created. (Descartes /, )
Was Descartes sincere? He seems to have been serious about the existence and
power of God, or else he would not have given him a lynchpin role in the cen-
tral positive argument of his Meditations. But for our purposes the example
Descartes gives here is only illustrative. One might perhaps prefer a purely nat-
uralistic case, say the ability of a Newtonian explanation of planetary motion to
peter lip ton
provide actual understanding, even though from the general relativistic point
of view that explanation is fundamentally mistaken about what is really going
on. The general suggestion is just that an explanation known to be false may
provide actual understanding.
Here again we need to compare our standards for explanation and for un-
derstanding. Explanations known to be false will only occupy the gap between
explanation and understanding if they are not themselves actual explanations,
and yet the benefit they confer is actual understanding. So why not rather say
that these explanations provide only potential understanding? A natural gloss on
the notion of potential understanding is that you would have understood why
the phenomenon occurred if the world had been as the explanation describes
it to be. Thus, supposing that God did actually create all living things, the seed
explanation provides only potential understanding; it would only have given you
actual understanding if living things had in fact come from seeds. On this view,
these false explanations only provide understanding of a possible world.
But this does not in my view do justice to the power of potential explana-
tions: they may provide understanding of the actual world—which is to say ac-
tual understanding—not just of possible worlds where they are true. Descartes
is right about this because understanding is a modal notion. This is obvious
from the signal roles of causation and necessity in understanding, themselves
modally laden notions. Information about other worlds illuminates the actual
world. The fact that my computer would not have overheated if the cooling fan
had not broken helps to explain why my computer overheated, and the fact that
my computer would not have overheated if the fan had not broken is, I take it, a
fact about nonactual worlds. Now in this case, we also have a fact that is in some
sense about the actual world, namely that the breaking caused the overheating,
but it seems that information about the merely possible may contribute to un-
derstanding too. As Descartes wrote, we may improve our understanding as to
why a phenomenon is as it is not just by being given a more or less accurate ac-
count of how it actually came about, but by being given even a wildly divergent
account of how it could have come about. Actual understanding may arise from
being shown how something is possible (Nozick , –).
So we cannot plausibly exclude potential explanations from the gap between
explanation and understanding by saying that they provide no actual under-
standing. But we might try to squeeze from the other end. That is, we might ad-
mit that the understanding gained from these explanations is actual, but insist
that this is so only insofar as the explanation itself is actual too, or contains a
part that is actual, or approximates an actual explanation. On this view, there is
not really any gap exposed between explanation and understanding: the actual
understanding that merely potential explanations produce is simply a product
under standing withou t e x pl anation
of those aspects of the explanation that are true. After all, few if any high-level
scientific explanations are strictly true throughout, but the understanding we
get arises from their approximate truth, or from the parts that are true.
But cases such as Descartes’ explanation do not work in this way. The seed
explanation (on Descartes’ assumption that the real cause is divine) does not
work because it approximates the actual etiology of life. Rather, it works be-
cause, as I have already suggested, it gives modal information, showing how
the phenomenon in question could have arisen. This may generate understand-
ing in many different ways. Information about a possible mechanism may give
oblique information about the actual mechanism (telling us something about a
class into which it falls) by describing a mechanism in the same class, namely
the class of mechanism that can generate this phenomenon.
Perhaps Darwin’s account of the domestic selection of animals provides an
example. “Intelligent selection” illuminates natural evolution even though it is
false of it, since it shows how natural selection can be understood as a mecha-
nism in a larger class.2 Again, possible explanations may provide improved un-
derstanding of actual capacities, powers, and causal relations. Merely possible
explanations may also be failures to describe the actual mechanisms. In such
common cases, we may again gain oblique information about why the phenom-
enon occurred by learning that a particular way it could have occurred was not
the way it did occur.
A merely possible explanation may also give information about the modal
status of the actual explanation. Thus, it may show a degree of contingency in
the actual explanation by showing an alternative route to the same result, or
it may indicate the opposite, showing a degree of necessity to the actual ex-
planation by showing how far from actuality one would have to travel to find
an alternative mechanism. It may also show necessity in another way, by re-
vealing fail-safe overdetermination. Thus, if the merely possible mechanism is
such that it would have been actual, had the actual mechanism not been in
operation, then showing how the phenomenon would have been produced by
the alternative mechanism may help to show that the phenomenon had to oc-
cur, one way or another. Even though we already know how the phenomenon
actually came about, the merely possible explanation thus improves our under-
standing. Suppose that a boxing match between Able and Baker is rigged so
that Baker—though in fact the far better boxer—would take a dive in the tenth
round. Knowing this helps us to understand why Able won, even if as a matter
of fact Able floors Baker with a lucky uppercut in the fifth.
Potential explanations may provide actual understanding without approxi-
mating an actual explanation, and so seem to earn their place in the present
survey of routes to actual understanding that do not pass through actual ex-
peter lip ton
planation. The understanding involves a kind of cognitive gain about the actual
phenomenon, even though the proffered explanation is not true of the actual
phenomenon.
There is still a way for skeptical readers to insist that the explanations are
actual after all: they can change the explanandum. They may take it that what
is explained in these cases is not the original phenomenon, but rather its pos-
sibility. The effect of this explanandum switch is to convert a merely possible
explanation of actuality into an actual explanation of possibility. My short an-
swer to this maneuver is that it is not to the point. For even if it provided an
acceptable gloss on how these explanations work, it still supports the claim of
a gap between explanation and understanding. Since now explanation of one
thing provides understanding of another, a counterfactual explanation of how
the phenomenon could have come about (but did not) provides understanding
of how it actually happened. To deny the gap in these cases, one would have to
deny that these explanations of how the phenomenon is possible provide any
understanding of the actual phenomenon.
Jc^ÒXVi^dc
I have so far considered how understanding in the form of causal and modal
knowledge might be acquired without relying on actual explanations. Another
equally familiar conception of explanatory benefit is unification (Friedman
). It is an ancient idea that we come to understand the world by seeing the
unity that underlies the apparent diversity of superficial phenomena, and while
philosophers have found this idea difficult to analyze, it is clear that one of the
ways science improves our understanding of the world is by showing how di-
verse phenomena can share underlying similarities.
Unification is a signal achievement of many great scientific explanations;
can it also be achieved in other ways? Thomas Kuhn’s account of the dynamics
of normal science provides a important instance (Kuhn ). His central ex-
emplar mechanism suggests that scientific understanding through unification
yet without explanation is ubiquitous. The formative power of exemplars is a
central theme in Kuhn’s account of science, and appreciating its force shows
not just that the acquisition of theoretical know-how is a route to knowledge of
what a theory says and so to the nature of the world that the theory describes,
but also that what is gained from this process goes beyond the explicit content
of the theory.
One of Kuhn’s central projects is to account for the coordinated behavior of
researchers in a given specialty during periods of normal science. How can one
explain rulelike behavior in the absence of rules? What gives the problem its
bite is the extent of that coordination, involving not just agreement on accepted
under standing withou t e x pl anation
theory, but also agreement on which new problems to address, how to address
them, and what would count as solving them. Kuhn found the rough functional
equivalent of the coordinating rules of research that were not there in shared
exemplars. These are concrete problem solutions in the specialty—very much
like the solutions to the problems in the student’s problem sets—that are shared
by members of the community and serve as a guide to future work. Unlike
rules, exemplars are clearly available to the members of the relevant scientific
specialty. And although they are, again unlike rules, specific in their explicit
content, they are general in their import. For shared examplars set up perceived
similarity relations, and normal scientists go on to choose new problems that
seem similar to the exemplars, to attempt solutions that seem similar to those
that worked in the exemplars, and to judge the adequacy of proposed solutions
by reference to the standards the exemplars represent. Shared exemplars thus
perform the same function as would shared rules.
The exemplar mechanism provides a plausible example of a route to un-
derstanding without explanation. Kuhn is here providing an account of sci-
entific knowledge, of the scientists’ ability to select problems and to generate
and evaluate solutions. And these abilities correspond also to a knowledge that
goes beyond the explicit content of the theory. The exemplars provide knowl-
edge of how different phenomena fit together. The unarticulated similarity
relations that the exemplars support provide a taxonomy that gives informa-
tion about the structure of the world. They thus have the effect of unifying the
phenomena and so provide understanding, and they do this by analogy, not by
explanation.
For Kuhn himself, the unifying structure that exemplars forge is not some-
thing antecedently present in nature, waiting for scientists to disclose. Kuhn is
an antirealist. Like Kant, he holds that the world that scientists investigate is
a joint product for the things in themselves and the organizing activity of the
investigators. But where Kant holds that the human contribution to the empiri-
cal world can only take one form, Kuhn maintains that the contribution varies
across scientific revolutions, as the exemplars change. Kuhn is Kant on wheels.
But this does not spoil the appropriation of the Kuhnian mechanism as an ex-
ample of scientific understanding of the world. For again like Kant, Kuhn insists
that he is no idealist. Scientists do not make the world up any way they please;
indeed, anomalies—nature’s recalcitrance—play a central role in the Kuhnian
account of the dynamic of scientific development. Resolving anomalies is the
central puzzle-solving task of normal science, and the continued resistance of
certain anomalies is what prompts the transition from normality to crisis, and
sometimes to scientific revolution.
Thus, even in the context of Kuhn’s antirealism, the unification understand-
ing that exemplars provide is a genuine understanding of the world. Moreover,
peter lip ton
in my view, we need not see this as a package deal. The exemplar mechanism
does not demand an antirealist reading: it is compatible with the view that nat-
ural kinds mark mind-independent divisions in nature and that the exemplar
mechanism provides a way for us to explore those structures. The idea that
we may unify the phenomena (and so improve our understanding of them) by
constructing schemes of classification that do not in themselves provide expla-
nations is compatible with a conventional realist conception of science as being
in the truth business. The same holds for the other routes to understanding I
have highlighted in this chapter. The understanding we acquire may not pass
through an explanation, but it may nevertheless be understanding of an inde-
pendent and objective reality.
One natural way to show that a purely “explanationist” conception of un-
derstanding is excessively narrow is to point out the extent to which scientific
understanding is understanding how rather than understanding why, and thus
the extent to which understanding flows from abilities or skills rather than from
explanations. Abilities appear early in this chapter in my example of causal
knowledge coming from manipulative skills. They appear again in this section
since Kuhn’s exemplar mechanism is in the first instance an account of scien-
tific abilities, such as the ability to select appropriate problems and appropriate
candidate solutions. But while abilities appear prominently in my story, they are
not what my story is about. I am sympathetic to the broader conception of un-
derstanding that encompasses understanding how in addition to understanding
why, but my present purpose is to show that, even on a narrow conception of
understanding as understanding why, we may nevertheless get understanding
without actual explanation. That narrow conception takes understanding why
to be a form of propositional knowledge, and lets explanations set the standard
for what kind of knowledge counts as understanding. For my present purposes,
what is important about abilities is not the sui generis forms of understanding
they provide, but rather the conventional forms of understanding they support,
as conventional as knowledge of causes and unification.
In sum, I have argued in this section that understanding may come from ex-
emplar rather than from explanation. If Kuhn is right about exemplars, the abil-
ity to use a theory in research involves the construction of similarity relations.
This is excess content over what the theory explicitly says, and it yields a kind of
unification, a linking of cases, that is plausibly identified with understanding.
;ZZa^c\
the feeling of understanding.3 What is controversial about this feeling is not its
existence but its status. How does the “aha” feeling, which I take to be a genuine
and distinctive consequence of some explanations, fit into the economy of un-
derstanding? At one extreme, we might say that it is itself a form of understand-
ing; at the other we might take it to be purely epiphenomenal, the irrelevant
phenomenal steam that the brain lets off in the course of its serious cognitive
work. Both of these positions are incorrect: the feeling is neither a form of
understanding nor irrelevant to understanding (compare de Regt ).
How does this issue of the role of the feeling of understanding bear on our
general theme of the scope of understanding without explanation? The familiar
philosophical literature that has approached understanding entirely through
models of explanation has ignored the feeling. This is not surprising. Certainly
the phenomenology is unlikely to be foregrounded if one’s focus is on the logic
or the metaphysics of explanatory arguments. But once understanding is disas-
sociated from explanation, the “aha” feeling becomes more salient. It is certain-
ly associated with the alternative routes to understanding we have considered.
The acquisition of information about causes, necessity, and possibility and the
exercise of the similarity relations that Kuhn describes are all sources of the
feeling of understanding.
I am not the person to attempt an articulated analysis of the phenomenol-
ogy of the feeling of understanding, but I would like to begin by making a few
brief comparisons between it and other feelings—of surprise, familiarity, and
expectation—comparisons that may begin to locate the “aha” feeling in phe-
nomenal space. A feeling of surprise often prompts a why-question. Indeed,
even when what we explain is at one level familiar, a why-question is often
motivated by a process of defamiliarization. Thus, initially we find it unsurpris-
ing that the same side of the moon always faces the Earth, until we realize that
this requires not that the moon not spin on its own axis, but that it spin with a
period exactly the same as the apparently unrelated period of its orbit around
the Earth. (Try this out with a salt and pepper shaker if you need convincing.)
The phenomenon comes to seem very surprising, and this prompts us to ask
why those periods should be the same.
Surprise does have a phenomenological aspect, and in some ways that feel-
ing is a rough counterpart to the feeling of understanding, but the feeling of
understanding is not merely the absence of a feeling of surprise, as is shown by
the many mundane, unsurprising things around us that we do not understand.
Moreover, the two feelings seem compatible. Recall Galileo’s thought experi-
ment. One of the reasons for its extraordinary intellectual beauty is that it gen-
erates the feelings of “aha” and surprise together. Typically, however, the feeling
of understanding supplants the feeling of surprise: once we understand why
something happens, we are no longer surprised that it does. It is sometimes
peter lip ton
the feeling that one would have expected it. In fact, however, this does not
capture the feeling of understanding. For one thing, unlike the indicative feel-
ing of expectation itself, it is not clear that seeing that something would have
been expected qualifies as a feeling. In any event, it is certainly not a feeling of
expectation. On seeing my younger son drop a balloon filled with water from
the upstairs window, I expect that my older son below will very soon be in for a
surprise; but when I am myself the unsuspecting victim of this prank, my sub-
sequent thought that I would have expected it if I had been looking up at the
right time is not itself a form of expectation.
The feeling of understanding is thus distinct from a feeling of expectation,
and when you have the former you do not typically have the latter. (The obvi-
ous exception is understanding why something will occur in the future.) And
vice versa too: you may have an expectation without any understanding or feel-
ing of understanding, as many of the counterexamples to the deductive-nomo-
logical model of explanation show. The falling barometer leads me to expect a
storm, but provides no understanding nor feeling of understanding as to why
the storm occurs.
In sum, while the feeling of understanding is interestingly related to the
feelings of surprise, familiarity, and expectation, it is different from all of them.
What then is the relationship of the feeling of understanding to understanding
itself? Clearly the two cannot be identified. The feeling may be genuine even
when the understanding is not, and one may have the understanding without
the feeling. This is particularly plausible in cases of one’s own intentional ac-
tions. When I decide on a course of action, for example, I have an explanation
for what I will do before the fact, and in mundane cases of this sort there seems
to be no distinctive “aha” feeling. When I decide that I will leave a great pile of
roses on the kitchen table for my wife on Valentine’s Day, I know in advance
why those flowers will be there, but at no stage in the operation do I seem to
enjoy that distinctive “aha” feeling (though perhaps my wife will). Perhaps the
feeling requires a transition between ignorance and understanding that cases
of this sort do not supply. Understanding why is a kind of knowledge—whether
of causes or modal facts or connections—and a feeling is not a kind of knowl-
edge. So it is not a kind of understanding.
Thus, I endorse the intuitive position that the “aha” feeling is not under-
standing itself but rather a consequence of understanding. But it does not
follow that the feeling is epiphenomenal: this effect may serve an important
function. Here it is irresistible to introduce Alison Gopnik’s suggestion from
her memorably titled essay “Explanation as Orgasm”:
Gopnik’s point is that although the pleasant “aha” feeling is an effect of ac-
tivities that leads to understanding, it is also a motivating cause of those activi-
ties. Just as orgasm is nature’s way of encouraging reproduction, the feeling of
understanding is nature’s way of encouraging cognitive activities that generate
understanding. (Gopnik, in fact, identifies explanation with the “aha” feeling,
whereas on my gloss here I distinguish them. The way I see it, explanations are
one of the causes of feeling rather than the feeling itself.)
Gopnik, in fact, appeals to two feelings—a “hmm” feeling corresponding
to the why-question and the “aha” feeling corresponding to finding the an-
swer—and she emphasizes the role of those feelings as forces driving the in-
quirer to engage in the practices of intervention and inference needed to infer
underlying causal structure.4 But I want to focus on three points concerning the
general and tantalizing idea that the “aha” feeling’s function is to motivate the
search for understanding.
First, whatever one thinks about evolutionary psychology, it is plausible to
say that the desire for this pleasant sensation serves to motivate the search for
understanding. Scientific understanding in the various forms I have touched
upon in this chapter has instrumental value. Knowing about causal history,
about necessity in nature, about how things might have come about, and about
how to work one’s way around a theory may all make us better able to cope
with our environment. But the payoffs are often highly indirect and only in the
long term. So it is just as well that there is also such an immediate and personal
phenomenological payoff to keep us going.
The second point is that this picture of the role of the feeling of understand-
ing helps to justify my decision not to count the feeling as a part of the un-
derstanding itself. By giving the feeling a different, plausible, and conspicuous
role—one of understanding motivator rather than of understanding compo-
nent—it seems even more natural to distinguish understanding from the feel-
ing it causes.
The third point is that there is a natural though substantial extension of
Gopnik’s hypothesis concerning the causal role of the “aha” feeling that I
would like to promote. Her point is that the feeling motivates the activity of
hypothesis construction. My suggested addition is that the feeling may also
play a role in guiding that activity. For it is natural to link Gopnik’s hypothesis
under standing withou t e x pl anation
8dcXajh^dc
CdiZh
I am grateful to my fellow participants at the conference on Philosophical Perspectives
on Scientific Understanding that took place at the Faculty of Philosophy, Vrije Univer-
sity Amsterdam, in August . I learned a great deal there from talks, comments, and
questions, and I am particularly grateful to the Henk W. de Regt, Kai Eigner, and Sabina
Leonelli for their hard organizational work and to Henk and Sabina for exceptionally help-
ful suggestions for revising my essay. I presented some of the material in this chapter to the
Department of Science and Technology Studies at University College London in March
and to the Division of Humanities and Social Sciences at the California Institute of
Technology in May , and I thank the audiences there. In particular, I would like to
thank Jed Buchwald, Robert Bud, Hasok Chang, Moti Feingold, Marcus Giaquinto, Alison
Gopnik, Christopher Hitchcock, Jeremy Howick, Michela Massimi, Alex Pine, and Jim
Woodward for very helpful comments and suggestions.
.In the examples below, the first class was suggested to me by Christopher Hitchcock,
the second by Jed Buchwald.
. I owe this suggestion to Sabina Leonelli.
. The “aha” feeling is also discussed in the essays by Grimm and Ylikoski in this
volume.
. Gopnik’s idea is also discussed in Grimm’s essay in this volume.
. This latter view is one I have discussed at length elsewhere (Lipton ). Here I will
just say enough to make the bridge to Gopnik’s proposal.
GZ[ZgZcXZh
Brown, J. R. . The laboratory of the mind. London: Routledge.
de Regt, H. W. . Discussion note: Making sense of understanding. Philosophy of Sci-
ence :–.
Descartes, R. /. Principles of philosophy. In The philosophical writings of Des-
cartes, vol. , trans. J. Cottingham, R. Stoothoff, and D. Murdoch, –. Cam-
bridge: Cambridge University Press.
Duhem, P. /. The aim and structure of physical theory. Trans. P. P. Wiener. Prince-
ton: Princeton University Press.
Friedman, M. . Explanation and scientific understanding. Journal of Philosophy
:–.
Galileo Galilei. /. Dialogues concerning two new sciences. Trans. H. Crew and A.
de Salvio. New York: Macmillan.
under standing withou t e x pl anation
+
Ontological Principles and the
Intelligibility of Epistemic Activities
=6HD@8=6C<
ontol o gic al principle s and episte mic ac ti vitie s
AZ^Wc^oVhVc>che^gVi^dcVcYVAVYYZgid7Z@^X`ZY6lVn
C B
C B
9ZhXVgiZhÉ;^ghiGjaZd[8daa^h^dc
Descartes’s First Rule of Collision
C B
C B
9ZhXVgiZhÉHZXdcYGjaZd[8daa^h^dc
Descartes’s Second Rule of Collision
;^\jgZ)#	ZhXVgiZhÉildgjaZhd[Xdaa^h^dch#
Rule : if these two bodies (for example, B and C) were exactly equal
[in size] and moved with an equal velocity towards one another, then
they would rebound equally, and each would return in the direction it
came, without any loss of speed.
Rule : if B were somewhat larger than C, and if they met each other
at the same speed, then C would be the only one to rebound in the
hasok chang
Now, it might seem that Descartes’ second rule is patently false, and anyone
can refute it empirically by doing a very simple experiment. Leibniz the meta-
physician, however, chose to criticize it without recourse to experience. He put
Descartes’ two rules side-by-side and asked what would happen in the limiting
case in which the object approaching from the right-hand side (B in the figure)
was bigger than the left-hand one (C), but the difference approached zero. Ac-
cording to Descartes, as long as the difference is nonzero, however small, the
object coming from the right (B) should push straight through without any
change in its speed, but as soon as the difference becomes zero it is bounced
back in a U-turn.
Leibniz (/, –) thought this was a metaphysical absurdity,
violating the principle of continuity: how would a continuous change in the
circumstance of the experiment make such a discontinuous difference in the
outcome?3 Leibniz articulated his principle of continuity at least in two places.
In the Specimen Dynamicum of , he stated, “If one case continually ap-
proaches another case among the givens, and finally vanishes into it, then it is
necessary that the outcomes of the case continually approach one another in
that which is sought and finally merge with one another” (Leibniz , ). A
slightly different formulation can be found in a letter to Pierre Varignon from
: “The Law of Continuity demands that when the essential determinations
of one being approximate those of another, as a consequence, all the properties
of the former should also gradually approximate those of the latter” (Leibniz
, ). According to Leibniz, Descartes’ laws of collision could be rejected
with certainty, without a single experiment being performed, or even imagined
as a thought experiment. Descartes’ physics was not intelligible to Leibniz (pro-
ducing a “monstrous and inconsistent” result), and that was sufficient ground
for rejection (Leibniz /, ).
This case shows an eminent philosopher-scientist using the lack of intel-
ligibility as a cogent reason for rejecting a major piece of theory (part of the
physical foundation of Cartesian mechanical philosophy). And that was done
without any regard to empirical adequacy. The contrast to standard empiricist
schemes could not be clearer: in Bas van Fraassen’s (, –) view, for ex-
ample, something like intelligibility would probably be treated as a “pragmatic
virtue,” a merely subsidiary criterion that might function as a tiebreaker if a
choice were desired between two theories of equal empirical adequacy. With
Leibniz, it is clear that intelligibility was just as important as empirical adequa-
ontol o gic al principle s and episte mic ac ti vitie s
cy, if not more important. What is also nice about Leibniz’s thinking is that the
verdict of unintelligibility is based on a clearly articulated ontological principle
that seems beyond dispute at first glance. This is rationalist physics at its best.
On the other hand, the Leibniz story serves as a cautionary tale as well.
Why was he so sure about the principle of continuity? How was the principle
justified? At least in light of certain developments in modern science, Leib-
niz’s principle of continuity is unjustified. Granted, there might be some room
for debate about whether or not quantum mechanics (or chaos theory) really
shows the principle of continuity to be false. But the real point is that we can
imagine the universe to be a place in which the principle of continuity is vio-
lated. I would like to craft a concept of intelligibility that supports the positive
aspects of what Leibniz was doing, while avoiding the pitfalls in his thinking.
L]Viid9dVWdjii]ZJcXZgiV^cind[BZiVe]nh^Xh
The troubling thing about Leibniz’s thinking is that he was so sure about his
principle of continuity seemingly without having an argument to back up that
certainty. Subjective certainty, even for such an eminent thinker as Leibniz,
may not be worth very much.4 This will be a problem with rationalistic thought
in general, and also with any attempt to take intelligibility as an indication of
truth. Einstein pleased himself in his objection to the Copenhagen Interpreta-
tion of quantum mechanics by declaring that “God does not play dice,” but how
did he have this access to God, which was not only direct but apparently ex-
clusive, since Niels Bohr would seem to have been denied the direct telephone
line? Even Kant fell into the trap of thinking that things like Euclidean geometry
and Newtonian mechanics were necessarily true. Looking at Kant from a little
distance makes one thing regrettably clear: his location in the Newton-enrap-
tured eighteenth century must have exerted a strong hold on his imagination.5
Philipp Frank gives a sober assessment of the poverty of philosophical op-
position to new scientific ideas: what parades as a deep metaphysical truth is
often simply the petrified remains of an outdated scientific theory. In Frank’s
view, science needs to educate metaphysics, not vice versa (, –). Paul
Feyerabend (, ) describes the same phenomenon as “contamination by
older cosmologies,” in the context of his discussion of Galileo’s defense of Co-
pernicanism. In order to render the “tower argument” harmless for Coperni-
canism, Galileo had to dismantle the implicit assumption inherited from earlier
physics that apparent motion was real motion.
There are many other cases from the history of science showing the ques-
tionable effect of metaphysical beliefs that are leftovers from earlier scientific
theories: Einstein’s insistence on determinism against the Copenhagen inter-
pretation of quantum mechanics, Erwin Schrödinger’s objection to quantum
hasok chang
I]Z8dccZXi^dcWZilZZcDcidad\^XVaEg^cX^eaZhVcY
:e^hiZb^X6Xi^k^i^Zh
=dlH^c\aZKVajZ>hCZXZhhVgn[dgIZhi^c\"Wn"DkZgYZiZgb^cVi^dc
Let me now take a closer look at the nature of ontological principles. By an
“ontological” principle I do not mean a principle that dictates whether some-
thing exists. Rather, the term “ontological” indicates a concern with the basic
ontol o gic al principle s and episte mic ac ti vitie s
nature of the entities that we are thinking about. The entities in question may
be of any type (for example, physical or abstract), but in any case they will
have some basic properties, and the ontological principles dictate what kind of
properties they can have.
Coming back to the principle of single value, we need to ask: what are the
grounds of the validity of this principle? The first thing to note is that it is not an
empirical generalization. To see that point, imagine trying to test the principle
by experiment. If we went around taking measurements of various quantities
in order to check that we get a single value each time, that would seem as silly
as testing “All bachelors are unmarried” by collecting a bunch of bachelors and
checking each one carefully to see if he is married. Worse, if someone reported a
counterinstance (say, that his weight is both kilograms and kilograms ), our
reaction would be total incomprehension. We would have to say that such an ob-
servation in “not even wrong,” and we would feel compelled to engage in a meta-
physical discourse to persuade this person that he is not making any sense.9
The principle of single value is not a logical truth, either. One can easily
imagine variables that have multiple values, especially in the realm of non-
physical quantities and designations: for example, names of persons or places,
or multivalued functions in mathematics. Still, where we do recognize it as
valid, the principle clearly seems to have a necessity about it. What could be the
source of this necessity? In terms of conceptions that are commonly known, I
think the closest thing to ontological principles as I conceive them is the Kan-
tian synthetic a priori. But there is a significant difference from the Kantian
conception, as I will explain later. Especially given the presence of exceptions,
what we must conclude is that the necessity of the principle of single value is
not universal but conditional, holding only in some situations.
I would like to propose that the necessity of the principle of single value
comes from the requirements of testing.10 Let me elaborate on that idea. Take
an ordinary situation in which a theory is tested, in the simplified view given
by the hypothetico-deductive model. We derive a prediction from the theory
and check it against observation. If the predicted and observed values differ, we
decide that there is something false about the theory. This is usually regarded
as a matter of simple deductive logic, but we are actually making a tacit reliance
on the principle of single value. We predict that the temperature here at a.m.
today will be ° C; our observation says it is °C. If the temperature could be
both °C and °C, there would be no refutation here. To put it formally: there
is no logical contradiction between “T = °C” and “T = °C,” until we also put
in the principle of single value, which gives “If T = °C, then T ≠ °C.” Now
we do have a contradiction, because we have “T = °C” and “T ≠ °C.” With-
out the principle of single value, we cannot engage in this kind of testing.
More generally, the principle of single value is required for what I call test-
hasok chang
=dl8djci^c\GZfj^gZh9^hXgZiZcZhh
The above idea about the necessity of ontological principles can be articu-
lated more carefully with the help of an even simpler example. The ontological
principle that enables the activity of counting is that of discreteness. If we want
to engage in the activity of counting, then we have to assume that the things we
are trying to count are discrete. Let me highlight several aspects of this situa-
tion, which we can later try to generalize to other pairs of ontological principle
and epistemic activity.12
. It is obvious that the necessity here is conditional: it is only if we
count things that discreteness has to be presumed. (Whether or not we
ought to be engaged in counting is a separate question.)
. The principle of discreteness says nothing about whether the world
in itself is discrete or not, only that we need to take it as such if we are go-
ing to count things.
. What underlies the sense of necessity here is impossibility: we
can try, but we just cannot count anything in an utterly undifferentiated
continuum. The nature of that impossibility is interesting. It is a pragmatic
impossibility, which is a notion I will take as a primitive rather than try to
analyze it further in terms of other notions, which I think are less basic.13
. The difference between empirical truth and pragmatic necessity
should be clear. A given bit of the universe may or may not contain dis-
crete pieces; if it does, then that domain is, in fact, discrete, which would
be an empirical truth; that would be a fact, with no necessity about it. The
necessity of discreteness only comes with our commitment to count.
. The denial of an ontological principle, while we are engaged in the
activity that requires it, generates a sense of unintelligibility. If we are try-
ontol o gic al principle s and episte mic ac ti vitie s
ing to count things that we cannot grasp as discrete, that makes no sense;
we are engaged in an unintelligible activity.
. This sense of intelligibility pertains to activities in the first instance,
and only indirectly to statements, propositions, representations, or mod-
els. A formal system in itself is neither intelligible nor unintelligible; it is
what we do with it that may or may not have intelligibility.
Di]ZgEg^cX^eaZ"6Xi^k^inEV^gh
So far we have examined two epistemic activities and found an ontological
principle that is required for each activity (single value for testing, discreteness
for counting). This picture can now be broadened: to each well-defined, basic
epistemic activity, there is an associated ontological principle that enables it
and makes it intelligible.
The proposed one-to-one pairing of epistemic activity and ontological
principle might seem too neat, artificial, and contrived. But it will seem more
natural if we recognize two things. First, the one-to-one correspondence only
applies to the most basic epistemic activities; complex ones will require mul-
tiple principles, as I will discuss further below. Second, in the principle-activ-
ity pairing, the ontological principle and the epistemic activity are mutually
constitutive. A general link between ontology and epistemology will be easily
granted, in the sense that the appropriate method of studying something is
surely linked with the nature of that something. Here I am pointing to the pur-
est and strongest version of that link: a distinct type of epistemic activity has
its particular type of object, the essential characteristic of which is defined by
the fundamental ontological principle associated with the activity. At the same
time, the nature of the activity is shaped by the basic character of its objects.
Table . gives a list of basic epistemic activities and their associated onto-
logical principles that I have started compiling. Even a brief consideration of
these pairings will reveal some quite interesting consequences of this way of
thinking. About the first two activity-principle pairs I have already said enough.
There is just one brief note I must add on the second pair: there are other modes
of “testing,” especially in the sense of a “trial run,” which can be made without a
specific prior prediction of the outcome. This is why I must always specify what
I have in mind here as “testing-by-overdetermination.” The remaining activity-
principle pairs deserve some further comments.
IVWaZ)#eVgi^Vaa^hid[eV^g^c\hd[
Ze^hiZb^XVXi^k^i^ZhVcYdcidad\^XVaeg^cX^eaZh
:e^hiZb^XVXi^k^in Dcidad\^XVaeg^cX^eaZ
8djci^c\ 9^hXgZiZcZhh
IZhi^c\"Wn"dkZgYZiZgb^cVi^dc H^c\aZkVajZ
GVi^dcVaEgZY^Xi^dc Jc^[dgbXdchZfjZcXZ
8dcigVhi^kZ:meaVcVi^dc Hj[ÒX^ZcigZVhdc
CVggVi^dc HjWh^hiZcXZ
A^cZVgDgYZg^c\ IgVch^i^k^in
EgV\bVi^XGddihd[CZXZhh^inVcYDcidad\n
An ontological principle derives its necessity from its rootedness in an
epistemic activity. An ontological opinion, in contrast, has no such grounding
in the requirements of an activity, and may be discarded with no pragmatic
consequences. These reflections actually throw a new light on the concept of
necessity in general. Necessity is a concept rooted in human action, which we
project metaphorically onto the rest of nature (sometimes sensibly and some-
times not). The ignoring of this human, pragmatic context is the main reason
why we have such trouble understanding the notion of necessity, as we try to
render “necessary truth” as a kind of truth stronger than mere actual truth,
with the help of possible-worlds semantics and such. Rather, if something is
“necessarily true,” that means it needs to be taken as true for the purpose of
some epistemic action.
Necessity also does not imply universality. Ontological principles are not
truly universal, as their necessity remains conditional on our commitment to
engage in their corresponding epistemic activities. We may not see this usually,
but there are situations in which we would decline to engage in certain basic
epistemic activities, such as rational prediction, or counting, or explanation.
This is the key difference I see between my way of thinking and the legacy of
Kant. In my scheme there are no eternal and universal categories since we are
not compelled to engage in all the same basic epistemic activities all the time.
It is a mistake to treat the activity-dependent conditions of intelligibility as un-
conditional metaphysical truths. Even compared to William Whewell’s scheme,
in which the categories evolve with the general development of science (see
Kockelmans , –), or Michael Friedman’s system of the relativized a
priori in which revolutions are allowed (, part , chap. ), my scheme is
distinct in allowing more freedom to epistemic communities and even indi-
viduals in their choice of epistemic activities, and in taking the pragmatic per-
formability of activities as the ultimate basis of necessity.
It is satisfying to see how our basic ontological conceptions arise from the
way we engage with the world. The most fundamental part of ontology is not
abstracted from what we passively observe; rather, it emerges and becomes es-
tablished as an essential ingredient of our successful epistemic activities.17 A
whiff of scientific realism can be derived from that insight: if some of our epis-
temic activities are successful, that must be some sort of indirect vindication of
the ontological principles presumed in those activities.18 But this vague sense
of vindication is all we can have, and it only points to an inarticulable harmony
ontol o gic al principle s and episte mic ac ti vitie s
between the state of the world and our ontological principles. There is a Kan-
tian insight here: nature in itself (not what Kant called “nature,” which consists
of phenomena) is not the object of our epistemic activities; our activity-based
ontology is not a direct representation of nature (or, the Kantian “thing in it-
self ”). As Henri Bergson put it (/, ), “The bodies we perceive are,
so to speak, cut out of the stuff of nature by our perception, and the scissors
follow, in some way, the marking of lines along which action might be taken.”
When nature “speaks” to us, it is only through the outcomes of our epistemic
activities. That is not to deny that nature enters and rules our life by deter-
mining which epistemic activities are pragmatically possible. But ontological
principles do not give us the kind of direct representation of nature that the
correspondence theory of truth would prompt us to seek.
>ciZaa^\^W^a^inVcYHX^Zci^ÒXJcYZghiVcY^c\
standing, or to deny that there are other key factors involved in understand-
ing.) All this may strike the reader as an anticlimax, but if so, I think I will have
made a healthy deflationary move. There are so many things people mean by
“understanding” that it is going to be difficult to lay down any firm definition
of it. I think the best nonvacuous general conception we can have of it is what
I have just given.
Understanding, as I see it, is not some distinct quality or state of mind that
exists separately from the sense of knowing how to do epistemic things. Under-
standing is simply knowledge taken in the active sense. The feeling of under-
standing is the sense of performing an epistemic activity well: “I know what I
am doing.” The degree of understanding is the degree of success with which we
engage in an epistemic activity, as assessed by ourselves or by others.
It is futile to try to tie understanding down to any particular activities. For
example, as Peter Lipton argues convincingly in this volume, it is a mistake
to think that understanding only comes from explanation (unless one defines
“explanation” as whatever increases understanding, in the manner of Michael
Scriven).20 Henk de Regt and Dennis Dieks (, –) offer a seemingly
more liberal notion of understanding: “A phenomenon P can be understood if
a theory T of P exists that is intelligible (and meets the usual logical, method-
ological and empirical requirements).” The intelligibility of theories, in turn, is
defined as follows: “A scientific theory T is intelligible for scientists (in context
C) if they can recognize qualitatively characteristic consequences of T without
performing exact calculations.”21 Even this notion of understanding, I think, is
too narrow, in two senses: it requires the existence of an intelligible theory, and
the criterion of intelligibility is tied to the activity of qualitative prediction.
With these insights, I can finally tackle something that has been an endur-
ing puzzle to me: what is the difference between simply applying an algorithm
to solve a problem, and doing the same with a sense of understanding? Simply
following an algorithm provides no relevant understanding to someone who
is interested in some other epistemic activity, for example, visualizing what is
going on, or giving a mechanical explanation. But for someone whose goal is
to derive a prediction, there is surely the relevant sense of understanding in
knowing how to apply the right tricks to derive the answer. (When I was an
undergraduate, learning physics in the standard way gave me no sense of un-
derstanding because I went into the enterprise expecting something else. It was
my mistake to think that my classmates who were not dissatisfied like me were
in some way shallow or unreflective; they were simply happy with the epistemic
activity they were engaged in.)
Any well-performed epistemic activity can generate a sense of understand-
ing. Since people have different preferences about what kind of activities they
ontol o gic al principle s and episte mic ac ti vitie s
appreciate and want to engage in, what they consider “real” understanding will
vary. This creates an ineliminable dimension of subjectivity in the concept of
understanding, which is due to the differences in people’s epistemic aims. I
believe that this essential subjectivity is closely related to the contextuality in
scientific understanding that de Regt stresses.
I]ZLVn6]ZVY
Di]ZgEg^cX^eaZ"6Xi^k^inEV^gh
First of all, I think there are many other pairs of basic epistemic activity and
ontological principle yet to be discovered. The following are worth considering,
though I cannot yet offer them with confidence.
8dbeaZm:e^hiZb^X6Xi^k^i^Zh
So far I have only discussed very simple and elementary epistemic activi-
ties. But we hardly ever engage in such epistemic activities. What we do in real
life, even in the practice of mathematics or physics, are mostly very complex
and specific activities, such as, say, predicting the frequencies of molecular
radiation by solving the Schrödinger equation using certain techniques of ap-
proximation. How can we make sense of intelligibility and necessity pertaining
to these real activities? I think it is going to be a very interesting and informa-
tive challenge to extend my analysis to complex epistemic activities in various
areas. A complex activity will involve multiple basic activities, each with its
attendant ontological principle. Therefore, all of those ontological principles
have to be satisfied for the complex activity to be intelligible. Each particu-
lar epistemic activity will be shown to have a particular set of conditions of
intelligibility.
There are some very interesting and common conjunctions of basic epis-
temic activities that play essential roles in science and everyday life. I will only
discuss a few of them briefly here to give a flavor of the insights we might gain
from a thorough analysis of complex epistemic activities. The activity of giving
mechanical explanations can be seen as a synthesis of contrastive explanation,
narration, imagined intervention, and perhaps visualization as well. Visualiza-
tion, I think, is supported by a set of principles that form the basis of Euclidean
geometry. (This explains why non-Euclidean geometry is deeply unintelligible
to those of us who try to visualize what is going on there.)24 The construction
of physical objects is another crucial activity, but it cannot be conceived fully in
terms of the basic activities and principles I have discussed so far in this paper.
It involves first of all the assumption that every object is “located” in space and
ontol o gic al principle s and episte mic ac ti vitie s
time (“physicalism” as meant by Otto Neurath). We also presume that all sen-
sations attributed to one and the same object are harmonious with each other
(for example, a big sharp edge felt in gently running a finger along a completely
smooth-looking surface is not intelligible). Quite possibly there is a presump-
tion of causality, too, embodied in the notion of a physical object. We can see
that the analysis of a complex epistemic activity is a challenging task that will
require imagination as well as care.
It would be wrong to think of a complex activity as made up of basic activi-
ties in an atomistic way. All distinct activities to which we give definite char-
acterization are abstractions from what we actually do. Although it is often
instructive to think of a complex abstraction as constructed from simpler ab-
stractions, it would be a mistake to build up a whole universe of activities from
basic activities.25 Rather, the important thing to note is the relationships hold-
ing between the domains of applicability of various activity-descriptions. If an
act that qualifies as an instance of activity A always qualifies as an instance
of activity B as well, but not vice versa, we can say that activity B is more ba-
sic than activity A. Applying that idea to the basic epistemic activities that I
have discussed so far, I find that some of them are more basic than others. For
example, testing-by-overdetermination always involves inference (as well as
attempted overdetermination itself ). Therefore, inference is a more basic activ-
ity than testing-by-overdetermination, and the latter requires the principle of
noncontradiction in addition to the principle of single value. Similarly, hardly
any other activity can take place without identification.
JcYZghiVcY^c\!Igji]!VcYi]Z6^bhd[HX^ZcXZ
I will close with a very brief remark on intelligibility and understanding in
relation to the aims of science. In my earlier essay I was at pains to emphasize
that understanding was a distinct goal from description and prediction, and
that it was an important aim of science in its own right (Chang , –).
Now, with the broadened sense of understanding that I am proposing in this
chapter, understanding becomes part of all the aims of science and all other
epistemic activities since understanding simply comes from performing any
epistemic activity well.
What about truth, then? I will limit myself to a few suggestive sentences for
now. It would be nice if we had a simple activity of truth-finding, but frankly
I do not know what that would look like. What scientific realists should hope
for is that there might be some complex activities the successful performance
of which would allow us to have a reliable sense of attaining truth. I used to
believe that truth and understanding were separate and equal aims. Now I can
see truth is only a possible by-product of certain types of understanding.
hasok chang
CdiZh
I would like to thank Henk de Regt, Sabina Leonelli, and Kai Eigner for inviting me
to speak at the Amsterdam conference on scientific understanding, in , and Dennis
Dieks for chairing the session and allowing a great deal of discussion. I cannot list here all
of the conference participants who gave me highly encouraging and stimulating feedback.
I thank Roberto Torretti for highlighting in my previous work the idea that formed the
central core of this chapter, and for some detailed comments on an earlier draft. I would
also like to thank the following people for their helpful criticism and suggestions, most of
which I regret not having been able to make full use of: Henk Procee, Mieke Boon, Rasmus
Winther, Peter Morgan, Eric Schliesser, Maarten Van Dyck, Marcel Boumans, Bill Brewer,
Hans Radder, Matt Lund, Michela Massimi, and Grant Fisher.
. I use the term epistemic broadly to mean “in relation to knowledge,” without tying
it to the concept of truth, in order to avoid prejudging various important issues in favor of
standard scientific realism. This marks one of the few disagreements I have with Henk de
Regt, on his concession that being “epistemically relevant” means being “truth-conducive”
(de Regt , n. ).
. This text, titled “Critical Remarks Concerning the General Part of Descartes’ Prin-
ciples,” was composed in . It was a paragraph-by-paragraph refutation of Descartes’
Principles of Philosophy, intended for publication in an edition that juxtaposed it against
Descartes’ text. It only appeared posthumously.
. Descartes may have sought to avoid the continuity problem by the word “somewhat”
in the statement of Rule . In that case, Descartes’ problem would be incompleteness.
. Descartes should at least get due credit for being painfully aware of the difficulty of
establishing certainty anywhere. I am not going to debate here the correctness of Des-
cartes’ epistemology and metaphysics.
. The Kantian project could be interpreted in a way that avoids a commitment to the
necessary truth of the basic theories of science, and that is something I hope to achieve in
future work. See Chang (), for a start.
. See de Regt and Dieks (, –), for an instructive discussion of the context-
dependence of intelligibility.
. What I present here is a superior solution to the one I proposed in Chang (),
–.
. See Chang (, , ). In experimental practice, “disagreeing” means differing so
much as to go beyond the admitted margin of error. I have shown how this principle has
done useful work, although it might seem entirely vacuous (Chang , –).
. The nonempirical nature of the principle of single value allows it to support thought
experiments; for an example, see the discussion of Galileo’s thought experiment showing the
independence of acceleration on mass in free fall in Peter Lipton’s chapter in this volume.
. See Chang (, –). I thank Roberto Torretti for getting me to think about
this point in a fresh way (Torretti, private communication, March and , , and April
, ).
. Belatedly I discovered the work of C. I. Lewis, a system of “conceptual pragmatism,”
in which he advances a “pragmatic conception of the a priori”; see Lewis (/) for
ontol o gic al principle s and episte mic ac ti vitie s
the most definitive summary. For my reflections on Lewis and ontological principles, see
Chang ().
. By an “epistemic activity” I mean a system of mental or physical actions contribut-
ing to the production or improvement of knowledge in a particular way, and following
some discernible rules (though the rules may be unarticulated). This is not intended as a
precise definition. For a more extensive characterization of epistemic activities, see Chang
(forthcoming).
. I want to argue that other types of impossibility are actually grounded in pragmatic
impossibility, being metaphorical extensions of the latter. This is why it would be futile to
try to analyze pragmatic impossibility further. In working out the notions of necessity and
possibility sketched here, I would like to build connections to Roberto Torretti’s (,
chap. ) ideas on the subject.
. See Salmon (), from which I also take the phrase “rational prediction.”
. Depending on the exact formulations, it could be that the principle of sufficient
reason is just the contrapositive of the principle of uniform consequence.
. My thinking about this particular activity-principle pair was inspired by the work
of my former Ph.D. student Georgette Taylor on affinity tables in eighteenth-century
chemistry.
. Compare Carlson () for an insightful argument that there were signifi-
cant Kantian elements in William James’s thinking, despite the latter’s renunciation of
Kantianism.
. This is reminiscent of Duhem’s (/) argument that when we witness the
extensive organization of phenomena through a successful theory, “it is impossible for us
to believe that this order and this organization are not the reflected image of a real order
and organization” (–); he also cites Poincaré in that discussion.
. The last statement is really a tautology (and therefore certainly true!), since I have
defined intelligibility as performability.
. Scriven (, ) states: “The question we have to answer is how and when cer-
tain descriptions count as explanations. . . . Explaining therefore sometimes consists simply
in giving the right description. . . . The right description is the one which fills in a particular
gap in the understanding of the person or people to whom the explanation is directed.”
. For an earlier articulation of the intelligibility criterion, see de Regt (, ).
. Whether the will itself is free or not is a separate question. Along vaguely Kantian/
Jamesian lines, I would imagine that the freedom of the will is an ontological principle
concomitant to the activity of making moral judgments (compare Carlson , –).
. I thank Roberto Torretti for this suggestion.
. This is not to say that every Euclidean situation is visualizable.
. The limitations are similar for any other kind of compositional reductionism. See,
for example, the antireductionist arguments in Dupré (, chap. ).
GZ[ZgZcXZh
Bergson, H. /. Creative evolution. Translated by A. Mitchell. London: Macmillan.
Carlson, T. . James and the Kantian tradition. In The Cambridge companion to Wil-
hasok chang
,
Reliability and the Sense of Understanding
HI:E=:CG#<G>BB
stephen r . gr imm
The fact is, our history is littered with inaccurate explanations we con-
fidently thought were obviously true: the explanation for mental ill-
ness in terms of demonic possession, the humoral theory of illness,
and so on. The sense of understanding would be epistemically idle
phenomenology were it not so poisonous a combination of seduction
and unreliability. It actually does harm, sometimes making us squea-
mish about accepting true claims that we don’t personally understand,
and more often operating in the opposite direction, causing us to over-
confidently accept false claims because they have a kind of anecdotal
or theoretical charm. (, –)
tion, one that recognizes the presumed positive epistemic status of many of the
exercises of our sense of understanding while also making sense of the fact that
astrologers, conspiracy theorists, and the like can go so systematically wrong.
Before we begin, it will help to bring a bit more order to the notion of “the
sense of understanding.” When Trout claims that the sense of understanding
is unreliable, what he seems to mean by this is that the phenomenological re-
sponse we feel while entertaining certain explanatory stories rather than others
is poorly correlated with the truth of those explanatory stories. In other words,
that more often than not (or at least, not often enough for some other reliability
threshold Trout has in mind) when a certain explanatory account “feels right”
to us, it turns out to be wrong.
Like its cousins the “sense of sight” and the “sense of hearing,” however, it is
important to see that at a more basic level the “sense of understanding” seems
to refer to the exercise of an ability or faculty that undergirds the phenomenol-
ogy. Thus, when we seem to “grasp” or “see” what it is that accounts for the
thing we want to explain, even when this grasp turns out to be mistaken, the ex-
perience of grasping is nonetheless distinctively active. Perhaps, for example, as
James Woodward () suggests, it is active in the sense that it brings with it
the ability to answer what-if-things-had-been-different questions. Or perhaps,
as Henk de Regt and Dennis Dieks () claim, it brings with it the ability to
recognize the consequences of certain theories without needing to perform
exact calculations (see also de Regt in this volume). However we ultimately
characterize this ability, the basic point that bears emphasizing is that there is
some sort of competence in play here, and that its operation is what undergirds
the phenomenology emphasized by Trout.
Given this more layered view of the sense of understanding, we can there-
fore think of Trout’s claim along the following lines: to say that our sense of
understanding is unreliable is to say the connections or correlations that it
“grasps”—where the graspings are in turn accompanied by a distinctive phe-
nomenology—turn out to be false more often than not.
One final preliminary: since the purpose of this chapter is to evaluate
whether and in what sense our sense of understanding is reliable, it is impor-
tant to bear in mind how we normally assess reliability. If we are evaluating
whether a certain thermometer is reliable, for instance, it is no strike against
the reliability of the thermometer to point out that it would fail to give an ac-
curate reading on the sun.3 Or again, if we are evaluating whether Ichiro Suzuki
is a reliable hitter of baseballs, then it is no strike against Ichiro’s reliability to
point out that he would fail to make contact with the ball while in the dark
or while blindfolded.4 Judgments about reliability, in other words, are always
stephen r . gr imm
6gbX]V^g:kVajVi^dc
8a^c^XVa:kVajVi^dc
In other words, what studies by Gopnik and her colleagues suggest is that not
only do we have a cognitive system dedicated to constructing causal maps of
the world that is apparently hard-wired (given our ability to construct these
maps from a very young age), but that we also have a natural aptitude for the
task. Constructing accurate causal maps is the kind of thing, apparently, that
we are naturally good at doing.
Alongside Gopnik’s research with children, Mark Steyvers and his col-
leagues () have reproduced similar findings with adults. In order to weed
out potential differences in background information (a point that will reap-
pear as significant later), Steyvers and his coauthors designed an experiment
in which subjects were asked to identify causal relationships among three hy-
pothetical “alien mind-readers.” At the beginning of the experiment subjects
were shown a picture displaying the thoughts of three aliens (of the Hollywood
variety), and they were asked to identify which alien was reading the minds of
the others.8
What Steyvers found was that not only were subjects quite good at identify-
stephen r . gr imm
ing how the alien mind-reading system worked based on passive observation of
the system, but that when they were allowed to manipulate the interactions their
reliability increased dramatically. As Steyvers and his colleagues summarize,
Combined with Gopnik’s work, both studies suggest that the very meth-
odology that a mature science uses to identify dependencies in the world—of
searching for correlations in the data, of using control groups, and so on—is in
some sense the built-in method that we use in our attempts to understand the
world. Given that we think the scientific method is a reliable way to identify
whatever it is that accounts for what we want to explain, it should therefore
come as no surprise that our sense of understanding turns out to be as reliable
as it is.
How, if at all, does the phenomenology of understanding that Trout empha-
sizes enter into this picture? Here again Gopnik (, ) provides a nice
insight. As she points out, once one appreciates the potential power that comes
from grasping how things depend on one another in the way specified above, it
is not particularly surprising that, over the course of time, a special kind of phe-
nomenology should have come to be associated with the exercise of our sense of
understanding. As Gopnik suggests, just as it “made sense” for Mother Nature
to associate sexual orgasm with its own pleasurable phenomenology in order to
encourage reproduction, in the same way it would presumably have made sense
for graspings or seeings to acquire their own distinctive, and distinctively desir-
able, phenomenology: a phenomenology that Gopnik characterizes in terms of
the “aha” experience. Since the ability to grasp dependencies in this way—and
thus to potentially manipulate one’s environment—is obviously a fitness-en-
hancing characteristic, it is natural to suppose that, over time, evolution would
have found a way to impress the goodness of such graspings upon us.9
Whether or not Gopnik is correct in characterizing the phenomenology
that accompanies the sense of understanding in terms of the “aha” experience is
reliabilit y and the sense of under standing
something we can set to one side here. My point is much more modest: namely,
that there is something that is distinctively like to “see” or “grasp” (or, at least,
to seem to see or to seem to grasp) why things are one way rather than another.
Whether one prefers to describe this phenomenology in terms of something
as dramatic as the “aha” experience, or whether instead one prefers to think of
the phenomenology in more subdued terms, the basic idea is that at least some
phenomenology accompanies the exercise of our understanding system, a sys-
tem that is apparently well connected with how things stand in the world.
8dggjei^dch
One possible conclusion to draw from these studies is therefore that, contra
Trout, our sense of understanding simply is reliable at identifying dependen-
cies—at least characteristically and under normal conditions. We might think
of this as the “optimistic conclusion.” But the optimistic conclusion is mislead-
ing; at the very least, it needs to be qualified significantly. The optimistic con-
clusion suggests that even though our sense of understanding is susceptible to
unreliability (the astrologers and the alchemists have to be accounted for, after
all), the unreliability is the result of some sort of external corruption rather than
the result of a problem inherent in the normal operation of the system—just as,
we might think, although our visual system is susceptible to unreliability via ex-
ternal corruption (cataracts, mind-bending drugs, and the like), in the absence
of such external corruption it is basically reliable. To put the point another way,
what the optimistic conclusion suggests is that when our sense of understand-
ing is functioning properly in its normal environment, it will be largely reliable,
just as when our visual system is functioning properly in its normal environ-
ment, it too will be largely reliable.
What I want to propose instead is that even a properly functioning sense
of understanding in its normal environment might be thoroughly unreliable at
identifying dependencies. In addition to its intrinsic interest, another reason
this would be an important result, if true, is that it would help to explain why
eminent thinkers such as Ptolemy and Galen did identify and affirm extremely
strange dependencies, in a way that would not have to attribute to them some
sort of malfunction in their sense of understanding.
In the remainder of this chapter I will consider two reasons—or, as I will
put it below, “types” of reasons—why even a properly functioning sense of un-
derstanding might systematically produce bad results. The first of these has to
do with problems with intellectual vices and the second with problems with
inaccurate background beliefs.
stephen r . gr imm
IneZ&HdjgXZhd[JcgZa^VW^a^in/>ciZaaZXijVaK^XZh
The basic idea behind Type sources of unreliability is that even in normal
environments, and even when the information we have at our disposal is quite
good, the influence of so-called “intellectual vices” can significantly diminish
the reliability of our sense of understanding.10
By way of illustration, suppose that Lisa is tracking the progress of one hun-
dred cold sufferers who are trying a new product, Cold Away!11 As it happens,
all one hundred of the sufferers recover from their cold within a week. Based
on the strong correlation between the taking of Cold Away! and the timely relief
from their colds, Lisa concludes that it was the Cold Away! that made the differ-
ence to the group’s improvement.
This will be a bad inference, naturally, and the problem (roughly put) can
be traced to the lack of anything resembling a control condition. At a mini-
mum, in order to determine whether it was the Cold Away! that made a dif-
ference to their recovery, Lisa should have checked to see how cold sufferers
who did not take Cold Away! fared. In this case, what she would have found is
that even those who failed to take Cold Away!—that is, even those who took
nothing—would have likewise recovered within a week. Colds are simply the
kinds of things that, given a little time, run their course. But then, presum-
ably, it was not the Cold Away! that made the difference to the cold sufferers
after all.
Crucially, however, the source of the failure in this case does not seem to be
Lisa’s sense of understanding—it is not as if, for instance, her sense of under-
standing was misled by the equivalent of cataracts or occlusions—but rather
Lisa’s handling of her sense of understanding.12 To draw again on an analogy
with the visual system, the more time you give a properly functioning visual
system to take in the world, the more likely it will be to represent the world ac-
curately. By contrast, the more you rely on quick glimpses and the like, the less
likely it will be to represent the world accurately—hence the less likely that be-
liefs formed on the basis of such information will be accurate. In the same way,
however, the more information Lisa feeds into her sense of understanding—in
the form of wide and varied samples, control groups, and so on—the more like-
ly it will be that she identifies genuine dependencies. By contrast, the more she
fails to provide her sense of understanding with this information—the more
she jumps to conclusions based on a small amount of data, for example—the
less likely it will be able to identify genuine dependencies.
I have characterized the source of Lisa’s unreliability in terms of a kind of in-
tellectual carelessness, perhaps even a kind of intellectual laziness: she jumped
to conclusions without bothering to acquire all of the relevant data. Needless to
say, however, there are dozens of other intellectual faults that likewise can con-
reliabilit y and the sense of under standing
IneZ'HdjgXZhd[JcgZa^VW^a^in/7VY7VX`\gdjcY7Za^Z[h
likely conclude that it was, in fact, the prayers that were making the difference
on these occasions.
Both Albert and Paul are therefore interestingly like Ptolemy in one respect
at least. Presumably Ptolemy’s belief that it was the earth that stood still and
that it was the sun that moved was so well entrenched (after all, just take a
look! which one seems to be doing the moving?) that this belief automatically
ruled out any explanatory story that portrayed the earth as a mover. What is
more, one of the two—either Albert or Paul—is also like Ptolemy in the fol-
lowing sense: not only is one’s belief about God false (as Ptolemy’s geocentric
belief was false), but it will be false in a way that negatively affects many of the
other conclusions he draws. If God does not exist, then Paul will be identifying
countless dependencies that simply are not there (petitionary prayer being only
one of them), not to mention (perhaps) overlooking the dependencies that are
there. If God does exist, however, then it will be Albert who will be missing out
on countless dependencies: he will be failing to grasp what actually makes the
difference in all of these cases.15
8dcY^i^dcVaGZa^VW^a^in
Bearing all this in mind, our answer to the question, “Is our sense of under-
standing reliable?” should therefore be: it depends. More precisely, our answer
should be: the faculty is conditionally reliable, and this in several respects.16
One respect in which the faculty is conditionally reliable is clear-cut: if you
get bad data in you will almost inevitably get a bad grasping out. Specifically, if
the information that you are mining is full of inaccuracies, then the dependen-
cies you identify on the basis of that information will likewise be inaccurate. As
we saw with respect to Type cases, this is also true in a more subtle sense: if
you have bad background beliefs, then scores of accurate explanatory stories
will not even receive consideration from you. In both of these cases, however, it
would be a mistake to say that the problem lies with the faculty or ability per se.
Instead, the problem has something to do with the quality of the information
that is being fed into the faculty.
With this in mind, moreover, when we look back to the Gopnik-Steyvers ex-
periments above, we can now appreciate one of the reasons why the subjects in
the experiment, first the children reasoning about blickets and then the adults
reasoning about alien mind-reading, performed so uniformly well. For notice
that in both experiments the problem of false background beliefs was essentially
factored out. For the children, the background beliefs were doubly factored out:
first, because children lack a large stock of background beliefs to begin with,
and second because the novelty of the blickets experiment required them to
approach the situation without any preconceptions. With the adults, the pe-
reliabilit y and the sense of under standing
when a certain proposition seems true to me, and when I am thereby inclined
to assent to it, I should resist this inclination. If what I want is knowledge, then
I should only put my faith in reliable indicators of truth, and the “seems right”
test is demonstrably not such an indicator.
But is that really good advice? More to the point, is that advice that it is even
possible to follow? Suppose you tell me that I should distrust the way things
seem to me and instead depend strictly on some other sort of evidence (per-
haps, statistical modeling). This is the only reliable way to get to the truth, you
say. As it happens, moreover, suppose that what you say seems right to me. I
feel like you are on to something. But now the problem seems obvious: Should
I distrust this feeling of seeming right? This feeling that I am on to something?
It seems, in sum, that relying on how things seem to us is simply our lot
in life—a maze from which we cannot escape—and that understanding seems
to have it no worse than everyday forms of belief on this score. I would also
continue to insist that the faults that we sometimes find in particular exercises
of the sense of understanding are, for most of us, less pervasive than Trout
suggests: perhaps, for many people, about as pervasive as my temptation to
be duped by something like the gambler’s fallacy. In any case, not pervasive
enough to underwrite the deep concerns about the reliability of the sense of
understanding that worry Trout.
8d]ZgZcXZGZ\V^cZY4
Studies concerning hindsight bias and the like aside, it therefore seems—to
pick up again on our earlier line of thought—that our sense of understanding
will be reliable to the extent that (among other things) our background beliefs
are accurate and varied and to the extent that we have a good cognitive charac-
ter that uses these beliefs appropriately.
We can now make sense, moreover, of one final point about the nature of
our graspings or seeings—and thus about the nature of understanding that sev-
eral commentators working at the intersection of epistemology and the philos-
ophy of science—including Jonathan Kvanvig (), Wayne Riggs (), and
Catherine Elgin ( and )— stress: namely, that coherence considerations
have a significant role to play in the exercise of our sense of understanding.19
According to Riggs, for example, understanding a particular bit of information
is essentially, perhaps even exclusively, a matter of seeing how it “fits together”
or “coheres with” the rest of one’s beliefs:
CdiZh
Thanks to Michael DePaul, Henk de Regt, Armond Duwell, John Greco, Sabina
Leonelli, Peter Lipton, J. D. Trout, and Fritz Warfield for helpful comments on earlier ver-
sions of this essay.
. De Regt , to which Trout is a response, likewise addresses the question of
the reliability of the sense of understanding. Although my approach differs from de Regt’s,
we share similar concerns about Trout’s view.
. See also Trout (, , –), where “the epistemically unreliable relation
between the sense of understanding and genuine (that is, factive) understanding” () is a
central theme.
. Compare Alston (, ).
. Compare Greco (, –).
. As we will see, however, the reliability of someone’s sense of understanding will
depend heavily on the accuracy of the person’s background beliefs, and I will make no at-
tempt to specify what counts as normal with respect to background beliefs.
. See especially Gopnik and Glymour and Gopnik et al. ().
. For more on the blickets experiment, see Gopnik and Sobel () and Gopnik et al.
().
. The alien “thoughts” were displayed in thought bubbles, and were quite basic. Thus,
in one experiment the first alien was thinking “TUS,” the second “POR,” and the third
“POR.”
. See also Lipton (this volume) for a similar appeal to Gopnik on this issue.
. Several epistemologists, especially those working in the virtue epistemology tradi-
tion, have stressed the connection between the intellectual vices and unreliability (see, for
example, Zagzebski ; Hookway , ; and Baehr ).
. The basic example is from (Salmon , sec. ); see also Bishop and Trout (,
–).
. In reality, of course, it would be very hard to pin the blame on the character flaw
rather than the faculty or ability; for all we know she might have some cataract-equivalent
problem with her faculty. In this case, though, we can simply stipulate that she does not.
. This case is loosely based on real life. In , the American Medical Association’s
Archives of Internal Medicine published a better-designed study of nearly a thousand
consecutive patients who were newly admitted to the coronary care unit of a hospital
in Kansas City. The researchers created a thirty-five-item score sheet that was used to
measure what happened to the patients during a twenty-eight-day period in which fifteen
groups of five persons (“intercessors”) prayed individually for about half the patients. The
intercessors were given the patients’ first names and were asked to pray daily for “a speedy
recovery with no complications.” The prayed-for group had a – percent reduction
in total scores even though their average length of hospital stay was similar to that of the
“usual-care” group. For further the details see Harris et al. ().
. At least, not from a distance; the fact that patients who pray for their own health
tend to improve as a result has been widely documented, and is now generally accepted.
reliabilit y and the sense of under standing
Whether the improvement is due to the patient’s sense of optimism, or whether instead to
divine intervention, is an open question.
. Naturally, the unreliability of someone’s sense of understanding can be attributable
to more than one source. Thus, someone might be unreliable at identifying dependencies
both because he or she is intellectually careless and because he or she has bad background
beliefs.
. In Alvin Goldman’s () words, “A process is conditionally reliable when a suf-
ficient proportion of its output beliefs are true given that its input beliefs are true” ().
Deductive reasoning therefore counts as an example of a conditionally reliable process,
for example, in the sense that if you start with garbage (falsehoods), you will end up with
garbage (more falsehoods), whereas if you start with truth you will end up with truth.
. The opposite sort of person—namely, someone with a good and reliable cognitive
character—will therefore be someone who takes all of the data seriously, does not restrict
their attention to a limited range of information that might favor a certain pattern, does
not rush to conclusions from small sample sizes, and so on. In short, someone with a good
cognitive character will be someone who has essentially internalized the scientific method.
What’s more, as Gopnik () notes in the following passage, even if one has not internal-
ized the method in this way, one might still use the method as a kind of external corrective
to compensate for the unreliable tendencies in one’s intellectual character, much as one
corrects the results of an unreliable visual system by using glasses:
It appears that one of the differences, perhaps the most important cognitive
difference, between organized science and spontaneous theory formation is
precisely that science contains additional normative devices that are designed
to supplement the basic cognitive devices of the theory formation system, and
to protect them from error. We might think of science as a kind of cognitive
optometry, a system that takes the devices we usually use to obtain a veridical
picture of the world and corrects the flaws and distortions of those devices. The
fact that most people over forty wear glasses is not, however, usually taken as an
indictment of the visual system. In fact, the analogy might be taken even further,
perhaps science compensates for our deteriorating adult theory formation abili-
ties the way optometry compensates for our deteriorating adult vision” (;
italics mine).
. Indeed, Trout has objected in this way, in his comments on an earlier version of
this chapter at the Central APA meeting. Thanks to Trout both for his comments,
and for the ensuing helpful discussion.
. So too, to a lesser extent, with Ernest Sosa:
Beyond “animal knowledge” there is a better knowledge. This reflective
knowledge does require broad coherence, including one’s ability to place
one’s first-level knowledge in epistemic perspective. But why aspire to any
such thing? What is so desirable, epistemically, about broad coherence?
Broad coherence is desirable because it yields integrated understanding, and
also because it is truth-conducive, even if in a demon world broad coherence
fails this test and is not truth conducive. (, ; italics mine)
stephen r . gr imm
GZ[ZgZcXZh
Alston, W. . How to think about reliability. In Epistemology: An anthology, edited by E.
Sosa and J. Kim. Oxford: Blackwell.
Baehr, J. . Character, reliability, and virtue epistemology. Philosophical Quarterly
:–.
Bishop, M., and J. D. Trout. . Epistemology and the psychology of human judgment.
New York: Oxford University Press.
de Regt, H. W. . Discussion note: Making sense of understanding. Philosophy of Sci-
ence :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Elgin, C. . Considered judgment. Princeton: Princeton University Press.
———. . True enough. Philosophical Issues :–.
Goldman, A. . What is justified belief? In Epistemology: An anthology, edited by E.
Sosa and J. Kim, –. Oxford: Blackwell.
Gopnik, A. . Explanation as orgasm. Minds and Machines :–.
———. . Explanation as orgasm and the drive for causal knowledge: The function,
evolution, and phenomenology of the theory formation system. In Explanation and
cognition, edited by F. C. Keil and R. A. Wilson, –. Cambridge, MA: MIT Press.
Gopnik, A., and C. Glymour. . Causal maps and Bayes nets: A cognitive and com-
putational account of theory formation. In The cognitive basis of science, edited by P.
Carruthers, S. Stich, and M. Siegal. New York: Cambridge University Press.
Gopnik, A., C. Glymour, D. Sobel, L. Schulz, T. Kushnir, and D. Danks. . A theory
of causal learning in children: Causal maps and Bayes nets. Psychological Review
:–.
Gopnik, A., and D. Sobel. . Detecting blickets: How young children use informa-
tion about novel causal powers in categorization and induction. Child Development
:–.
Greco, J. . Putting skeptics in their place. New York: Cambridge University Press.
Grimm, S. R. . Is understanding a species of knowledge? British Journal for the Phi-
losophy of Science :–.
Harris, W. S., et al. . A randomized, controlled trial of the effects of remote, interces-
sory prayer on outcomes in patients admitted to the coronary care unit. Archives of
Internal Medicine :–.
Hookway, C. . Epistemic akrasia and epistemic virtue. In Virtue epistemology: Essays
reliabilit y and the sense of under standing
-
The Illusion of Depth of Understanding
in Science
E:IG>NA>@DH@>
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
take seriously the possibility that scientific cognition is also affected by IDU
and spell out some possible causes of explanatory illusions in science. I will
conclude this chapter by discussing how scientific explanatory practices could
be improved and how the philosophy of science might be able to contribute to
this process.
JcYZghiVcY^c\VcYi]Z>aajh^dcd[9Zei]
the relevant inferences would be about the future consequences of some event
or series of events. In this case we would not make inferences about the conse-
quences of our imagined or real interventions, but would predict what will hap-
pen. Some of these predictions are about the real world, and others are about
counterfactual situations. Again, the amount of understanding can be roughly
measured by the scope of our predictive abilities. Other things being equal, the
broader the range of circumstances our anticipatory understanding covers and
the more precise our predictions, the better is our understanding of the phe-
nomenon. However, other things are not always equal, so we cannot equate the
amount of explanatory understanding with the ability to predict. Anticipatory
understanding is just one criterion for understanding.
We might be able to interact with an object or predict how it will behave
without making this knowledge explicit, which is required in science. For this
reason, the above criteria are not sufficient in the case of scientific understand-
ing. Something more is required—the ability to explain, that is, to communicate
our understanding. This constitutes the third way to demonstrate understand-
ing. Tacit understanding might give the ability to make predictions and to in-
teract with an object, but science requires us to make our knowledge explicit
by stating the underlying principles. I will call this theoretical understanding.1
Scientific understanding is demonstrated by giving explanations, that is, by
communicating one’s understanding. Explanations are answers to questions of
why; more specifically, they are answers to questions of what if things had been
different (Woodward ), so again we are dealing with an ability to make
counterfactual inferences. We also have a rough measure of the amount of un-
derstanding: the more, and more diverse, explanation-seeking questions we are
able to answer, the better (or deeper) is our understanding. In this case, ex-
planatory knowledge and understanding go hand in hand.
These three ways to demonstrate one’s understanding constitute criteria
according to which people attribute understanding to one another.2 They are
concerned with external displays of their knowledge. This knowledge might be
tacit or explicit; the crucial thing is that it gives an ability to do certain things.
Some authors suggest that understanding involves having an internal mental
model of the object of understanding (see, for example, Waskan ). This is
an interesting suggestion, but it cannot be the whole story about understand-
ing. When we evaluate someone’s understanding, we are not making guesses
about his or her internal representations, but about the person’s ability to per-
form according to set standards. The concept of understanding allows that the
ability can be grounded in various alternative ways, as long as the performance
is correct. Furthermore, the correctness of the internal model is judged by the
external displays of understanding, not the other way around. This makes un-
derstanding a behavioral concept.
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
prised by their own ignorance. The general conclusion is that most people are
prone to feel that they understand the world with far greater detail, coherence,
and depth than they really do. According to Keil, this effect is distinct from the
general overconfidence effect found in many psychological studies. Within the
experimental setup described above, the illusion of having detailed and coher-
ent knowledge occurs primarily for explanatory understanding. In contrast,
people’s ratings of how well they know facts, procedures, or narratives are quite
well calibrated, and they are not surprised at what they actually know (Rozen-
blit and Keil ; Keil ).
What is behind this interesting phenomenon? Keil suggests four possible
contributing factors. One factor is confusion between what is represented in
the mind with what can be recovered from a display in real time. People may
underestimate how much of their understanding exists in relations that are ap-
parent in the object as opposed to being mentally represented. A second factor
may be a confusion of higher and lower levels of analysis. For example, while
explaining how a car works, one might describe the function of a unit, such as
the brakes, in general terms, and then describe the functions of subcompo-
nents, such as brake lines and brake pads, which in turn can be broken down
even further. The iterative structure of explanations of this sort may lead to an
illusion of understanding when a person gains insight into a high-level func-
tion and, with that rush of insight, falsely assumes an understanding of further
levels down in the hierarchy of causal mechanisms. A third possible factor is
that many explanations have indeterminate end states. One usually has little
idea of what the final explanation will look like, and the end state is largely
indeterminate from the posing of the question. This makes self-testing one’s
knowledge difficult. The final factor in Keil’s list is the rarity of production: we
rarely give explanations and therefore have little information on past successes
and failures (Rozenblit and Keil , –.)
Keil’s thesis about IDU looks quite similar to the claims made by J. D. Trout
(), so it is important to see their differences. Trout argues that the sense
of understanding is often influenced by the overconfidence bias and that this
makes it a highly unreliable source of information. However, he does not cite
any studies about explanatory cognition to support his argument. His argu-
ment against the epistemic relevance of the sense of understanding is based
on the idea that humans are generally biased toward overconfidence. This idea
of general overconfidence bias has been criticized on theoretical and meth-
odological grounds (Juslin, Winman, and Olsson ). If these criticisms are
right, Trout’s argument is in trouble. However, these criticisms do not apply to
Keil’s experiments. He claims that there is a specific overconfidence effect in
the assessment of understanding, but the general overconfidence effect does
not have any role in his argumentation. Similarly, his experimental set-ups are
pe tri ylikoski
not based on the assumptions that the critics of overconfidence research have
found problematic. Finally, the conclusions he draws are different: he does not
suggest that we should give up the notion of understanding, as Trout does.
Keil’s studies show that understanding and the sense of understanding do
not always go hand in hand. Could the sense of understanding be calibrated to
be a more reliable indicator of understanding? We simply do not know. There
are no empirical studies of possible ways of improving explanatory practices.
It is possible that if we focus more on our explanatory practices, and make our
explanations more explicit, the calibration of our sense of understanding would
improve. This would help to address two causes of IDU suggested by Keil: the
indeterminacy of the end state and the rarity of production. However, there are
some grounds for being skeptical of our prospects in calibration. The extensive
experimental literature on reading comprehension shows that the calibration
of comprehension is not easily achieved (Lin and Zabrucky ).
I]Z8VhZ[dg:meaVcVidgn>aajh^dch^cHX^ZcXZ
The studies by Keil and his associates concentrate on ordinary people’s illu-
sions of understanding, and they avoid discussing this in the context of scientif-
ic enquiry. I maintain that IDU is possible and indeed common in the sciences
as well, and I will give a number of reasons for this suspicion. My claims will
be empirical hypotheses about factors affecting scientists. I will not assert that
these problems are unavoidable; rather, I will argue that they are prevalent
enough to be grounds for taking seriously the possibility of IDU in science.
The first reason to suspect that IDU might be relevant to science is the
continuity between scientific and everyday cognition. Scientists use the same
cognitive mechanisms as everybody else. Although scientific standards for
evidence, instruments, and social practices make scientific cognition different
from lay cognition, we should not assume that their use automatically makes
the problems related to evaluating explanatory understanding disappear. Most
of the things that contribute to the differences between scientists and ordinary
people are not related to the assessment of explanations.
The second reason IDU may be relevant is the level of attention in the sci-
entific community about the articulation and evaluation of explanations. Al-
though many scientists and philosophers are enthusiastic about explanation as
a proper cognitive aim of scientific enterprise, it is surprising how little explicit
concern explanatory practices receive in the sciences. As an example, scientific
journals provide many guidelines for presenting data and methods; however,
they do not provide any direction for presenting explanations. Typically, scien-
tific journals are structured to report empirical results, and as a consequence
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
lectual authority is a valuable resource. For all these reasons, a scientist might
be too impressed about the understanding of others and consequently might
be prone to overestimate the collective level of understanding of the scientific
community.
The fifth reason IDU is relevant to science is the future-oriented nature of
scientific evaluation. Most of the time scientists evaluate their theories, mod-
els, and methods in the context of making decisions about future research.
Although most of the philosophical discussion about the problem of theory
choice is set in the context of acceptance, scientists make their choices mostly
in the context of pursuit. They are not primarily assessing the current achieve-
ments of their theories, but making informed guesses about their fruitfulness
in the future. In other words, they are not choosing which theory to accept, but
choosing which theory they are going to work with. In the context of pursuit
they do not have to make up their minds about acceptance, but they must place
their bets on the most fruitful approach.
This observation is also relevant to our discussion. Scientists assessing ex-
planations are not simply evaluating their current status in terms of explana-
tory virtues and evidential support; they are also making guesses about the
future prospects of these explanatory hypotheses. The real issue is the promise
of future understanding, not the things that have been delivered thus far. This
future orientation makes the product to be evaluated extremely ambiguous;
the point in time when the evaluation is supposed to take place is not speci-
fied. Furthermore, the future versions of hypotheses to be evaluated cannot be
spelled out, so one does not really know what one is evaluating, nor does one
know what one is evaluating it against.4 If assessing existing explanations is
difficult, the assessment of future explanatory potential just might be impos-
sible.5 Of course, the evaluation is not completely arbitrary. The compatibility
with other relevant theories is one criterion; another is the past track record of
the candidate. But these are still indirect criteria, and people can reasonably
disagree on how to use them.
I submit that, most of the time, scientists make their choices on much more
pragmatic criteria than the promise of explanatory excellence. They choose op-
portunistically—a line of work, for example, that promises doable problems
that can be solved by the expertise, skills, and research resources they have at
their disposal. These primary reasons do not have much to do with explana-
tion, but scientists might still attempt to justify their choices (for example, to
possible funding sources) by appealing to future epistemic virtues. This would
make explanatory argumentation more or less mere window-dressing. The
real reasons for the choices might be different from the ones that are publicly
presented.
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
HZkZcHdjgXZhd[>9J
The previous section already indicated reasons to expect that IDU is a real
danger in science. In this section I will investigate some factors that might make
scientists prone to miscalibrated assessments of the depth of their understand-
ing. My suggestions will be a bit speculative, but this is unavoidable, as I have
not found much empirical research that would be relevant to this issue. How-
ever, the individual hypotheses that I will present are all empirically testable. I
do not claim that the following list is exhaustive; there are probably also other
relevant sources of IDU. It is also possible that their relevance may vary in dif-
ferent scientific fields. I will not make guesses about how influential these fac-
tors might be, nor am I making judgments about their relative importance. As
well, I leave the direction of the miscalibration open: given the point made in
the previous section, it is probable that scientists are disposed to overestimat-
ing their understanding, but it is also possible that some of the following cause
an underestimation of the understanding (in some circumstances).
The first possible source of IDU is the simple failure to find things puzzling. It
is typical in everyday cognition that only anomalous or otherwise surprising ob-
servations raise the need for an explanation. Only the unexpected or abnormal
events challenge one’s existing scheme of categorization. The same is probably
also true for the cognitive life of scientists, at least most of the time. Of course,
it is an important characteristic of science that it also asks why-questions about
the normal (Ylikoski ), but the investigative focus is always directed at a
limited number of issues, so this difference does not change the basic point I
wish to make. How does a failure to find things puzzling contribute to IDU?
A puzzling phenomenon forces one to come up with an explanation and
at the same time to face the limits of one’s understanding. This situation is in
sharp contrast to familiar things that behave as they are expected to. They do
not call for explanation, and as a consequence do not challenge a person’s confi-
dence in the depth of his or her understanding. The ability to find a categoriza-
tion for the initially surprising observation brings about a sense of confidence
in one’s conceptual scheme. This restoration of confidence in the conceptual
scheme might be confused with an increase in understanding and regarded
as a consequence of explanatory insight. In this way, familiarity can sustain an
instance of IDU. But it can also give rise to one. The puzzling things that are
not understood are contrasted with the familiar and the expected. This contrast
can easily be mistaken for a contrast between being understood and not being
understood. Furthermore, an explanation aims to remove the mystery about
the phenomenon, so understanding an anomaly makes it appear normal. From
this it is easy to infer that the observations regarded as normal already have the
pe tri ylikoski
property of being understood. Of course, this inference is not valid, but that
does not mean that we can easily escape making it.
The second possible source of IDU is the confusion between explanation
and description. Quite often one gets the impression that people think that
they automatically contribute to the explanatory understanding of an event
simply by finding out and describing facts about its causal history.6 Further-
more, they might think that the more facts about the causal history you have,
the better is your understanding of the explanandum. This is an interesting
confusion. Causal history is naturally relevant for explanation, but the crucial
point is that not all facts about the causal history are relevant to a given ex-
planandum. Their relevance depends on the aspects of the explanandum one is
trying to make sense of.7 It might be that every single fact about the causal his-
tory is relevant to some explanation-seeking question about the explanandum
event. But we are not interested in all possible explanation-seeking questions
related to the explanandum. We are not looking for questions that would make
our findings explanatory, but for facts that would answer our explanation-
seeking questions. Despite this, some scientists seem to be involved in this kind
of “explanatory fishing”: they think that their findings are important because
they might be crucial in explaining something. What constitutes that something
is conveniently left unarticulated. This ambiguity helps to maintain the illusion
of gained understanding. The illusion itself might be the product of the hard
work of research, even if the findings turn out to be irrelevant: one likes to feel
that one has achieved something, in this case, explanatory understanding.
The third possible source of miscalibration of the sense of understanding is
the ambiguity of the notion of explanation. The first piece of evidence for the
ambiguity comes from the fact that despite extensive debate, philosophers of
science have not reached a consensus about the nature of explanation. This is
weak evidence, as typically philosophers cannot reach a consensus about any-
thing, but in this case the philosophical disagreement reflects a more general
uncertainty. Discussions with scientists show that they have quite different
ideas about the criteria or characteristics of explanatory understanding. Some-
times they present an account that sounds quite strange to an interlocutor
who is familiar with various philosophical theories, but sometimes scientists
present ideas and concepts that they (or their teachers) have learned during a
philosophy of science class. It seems that the notion of explanation is a kind of
metascientific notion that is not explicitly present in everyday scientific prac-
tice: the scientists manage to do their work without having an articulated or
even shared notion of explanation.
When comparing explanations, scientists and philosophers often appeal to
metaphors of explanatory power and depth. Very few have tried to articulate
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
what these notions actually mean. However, there are grounds for thinking that
these notions do not refer to a single attribute of explanations. For example, the
notion of explanatory power can refer to a number of different properties of
explanations. An explanation could be considered powerful when it () is cog-
nitively salient (that is, is easy to understand and use); () is factually accurate;
() gives a precise characterization of the explanandum; or () is robust with
respect to changes in background conditions (Ylikoski and Kuorikoski ).
These virtues of explanation sometimes can be in conflict, which means that
the evaluation of explanatory power is not one-dimensional. The important
point in this context is that these different dimensions are not usually articu-
lated in scientific practice and that it is probable that scientists confuse them,
at least sometimes. Consequently, this confusion can lead to a miscalibration of
the assessment of understanding.
The fourth source of IDU is another kind of ambiguity. One cannot explain
anything completely; rather, one always explains some aspects of the explanan-
dum phenomenon. But which aspects are addressed with a given explanation?
Quite often it is very hard to see what the precise explanandum is that scientists
are trying to address with their theories. The ambiguity of the explanandum is a
real problem that surfaces in scientific controversies. A similar point applies to
the ambiguity of the explanans. All explanations take some things for granted
and treat them as background assumptions. But which are dispensable parts of
the background and which are the essential ingredients of the hypothesis? This
is a difficult problem, even for the person suggesting the hypothesis. Scientists
usually have intuitive ideas about the explanandum and the explanans, but they
cannot fully articulate them. In these circumstances it is extremely difficult to
say what has actually been achieved with the explanatory hypothesis and how
much it was the hypothesis (in contrast to the background assumptions) that
did the explanatory work. This gives more room for our wishful thinking and
egocentric bias to operate.
The fifth source of IDU is circular reasoning. The identification of cases of
circular reasoning is difficult outside formal logic (Rips ), and it is plau-
sible that sometimes people can mistake a restatement or a presupposition of
the explanandum for the explanans. In highly theoretical (and philosophical)
contexts this fallacy might be quite common, as the concepts are often difficult
to define precisely. Circular inferences are logically valid, and identifying them
is quite hard without making the whole argument explicit. Furthermore, as
definitions of the concepts and structure of the argument are often ambiguous,
judgments concerning circular reasoning are often controversial. This makes
their identification and avoidance even more difficult.
The sixth possible source of IDU is the confusion between the explanatory
pe tri ylikoski
power of the hypothesis and evidential support for it. When people are evalu-
ating an explanation they are not only making judgments about its explana-
tory virtues, they also assess the reasons to believe it. Psychologists are quite
confident that there exists an “explanation effect” according to which people
estimate the probability of an event much higher when they have an explana-
tory story about it. A task, like giving an explanation, which requires that a
person treat a hypothesis as if it were true, strengthens the confidence with
which that hypothesis is held. People also seem to have difficulty distinguish-
ing between explanation and evidence (Koehler ; Brem and Rips ). Of
course, this is a philosophical issue, as will be clear to anyone who has followed
the discussions around Peter Lipton’s () distinction between likeliness and
loveliness. Loveliness might be an indicator of likeliness, as many friends of
inference to the best explanation believe. I do not wish to take a stand on this
issue here. My point is that the mere existence of this debate is evidence for
the claim that explanatory power and evidential support are difficult to keep
separate. It is not always clear to which group a given criterion of explanatory
assessment belongs.
I am suggesting here that the evidential considerations might influence the
evaluation of explanatory goodness. This claim differs from the psychologists’
claim that explanatory considerations influence the evaluation of evidential
support. If people do not explicitly distinguish between their confidence in the
understanding provided by an explanation and their confidence in the truth of
the hypothesis, it is possible that these will influence each other. As a conse-
quence, one might think that a hypothesis provides understanding because one
is so confident that it is true.
The last source of IDU arises from scientists who often use complex repre-
sentations to make sense of the phenomena of interest. In order to understand
the phenomenon, one must also understand the epistemic tools that are used
as the medium of representation. In cases like this, understanding is mediated:
an object is understood via a tool that also needs to be understood. How might
this contribute to IDU? Learning to use the relevant mathematics, modeling
assumption, and other ingredients of the representative medium requires a lot
of work; acquiring the relevant abilities is not a trivial operation, but a real
achievement for an aspiring scientist. At the same time, many of the objects
that scientists are attempting to understand are difficult to access. They quite
often are abstracted, idealized, or stylized in a manner that makes them difficult
to grasp without the representative means provided by the theory. (Think, for
example, of the phenomena most economic theories are trying to deal with.) In
these circumstances a scientist can confuse his or her understanding of the theo-
ry with his or her understanding of the phenomena that the theory is intended to
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
deal with. In other words, the scientist confuses the intelligibility of the theory
(de Regt and Dieks ) with the intelligibility of the phenomenon.
The confusion is understandable, because in this situation the intelligibil-
ity of the theory is a necessary condition for understanding the phenomenon.
Both involve understanding; however, understanding a theory is different from
understanding a phenomenon when one considers the ways the understanding
can be demonstrated. The understanding of a theory is displayed by making the
right sorts of inferences from it, by knowing what kinds of problems it can deal
with and by building models, whereas the understanding of a phenomenon is
ultimately displayed by making right predictions, successful interventions, and
by answering explanation-seeking questions about it. The latter cannot be done
without a theory, but it requires more than understanding the theory. After
all, one can also understand theories that are considered false, outdated, or ir-
relevant. Making sense of a theory does not guarantee that one makes sense of
the phenomenon.
=dlid6kd^Y>aajhdgnJcYZghiVcY^c\
8dcXajh^dc
In this chapter I have argued for a number of different theses about explan-
atory understanding. I first argued that understanding should be characterized
as an ability and that it should be distinguished from the sense of understand-
ing. This is a distinction that is not commonly made in philosophy of science
literature. When understanding is analyzed as an ability to make counterfactual
inferences in the contexts of manipulation, prediction, and explanation, its re-
lation to knowledge can be clarified. Understanding is only knowledge about
the relations of dependence, not something mysterious added to knowledge.
This characterization allows for the possibility of tacit understanding, but this
chapter focuses mainly on scientific understanding, that is, explanatory un-
derstanding. One has explanatory understanding when one is able to answer
explanation-seeking questions. The more questions one can answer about a
phenomenon, the better one’s understanding, which makes the connection be-
tween scientific understanding and explanation quite close. However, this does
not mean that these notions are equivalent. Explanations are best analyzed
as answers to explanation-seeking questions. The notion of understanding is
more appropriate for characterizing the epistemic aims of science: organized
knowledge that allows one to answer to a whole series of what-if-things-had-
been-different questions about the world. In short, understanding is an ability
to give explanations.
The second main theme in this chapter is the possibility of illusory under-
standing in science. The key here is the sense of understanding that has an
important metacognitive role in our cognition. Despite its importance, there
is no reason to assume that the sense of understanding is perfectly calibrated
to our understanding. This makes it possible for people to overestimate the
detail, coherence, and depth of their understanding. Following Keil, I call this
phenomenon the illusion of depth of understanding. My main claim has been
that IDU is a real possibility in science. Due to the continuity of lay and scien-
tific cognition, the lack of explicitness in explanatory practice, the lack of clear
benchmarks, the division of cognitive labor, and the future-oriented nature of
scientific cognition, there is every reason to suspect that scientists might over-
estimate the depth of their understanding.
What does the possibility of illusion of understanding tell us about scien-
tific understanding? The main message, I believe, is that we should take more
care in assessing the level of our understanding. We might understand less
than we think. This is not a reason for skepticism concerning the possibilities
of scientific understanding, but a call for an improvement in our explanatory
practices. The simple reliance on the sense of understanding is not a sufficient
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
CdiZh
I would like to thank the editors of this volume and Tarja Knuuttila, Tomi Kokkonen,
Jaakko Kuorikoski, Aki Lehtinen, Uskali Mäki, Päivi Oinas, and Samuli Pöyhönen for their
useful comments.
. Knowledge cannot be made completely explicit; even scientific knowledge is based
on the foundation of some basic skills. For similar reasons, theoretical understanding is
often not easily transformable into a more practical form of understanding. The acquisi-
tion of theoretical understanding might not be accompanied by the relevant practical
knowledge and skills.
. These criteria of understanding raise many questions: How are they related to one
another? Are the different measures of amount of understanding comparable? Could
these criteria be different in different cultures or historical periods? How does one spell
out the crucial notion of “favorable circumstance” employed in the above discussion? I do
not pretend to be able to answer all of these questions. Here my modest aim is to describe
what kind of an ability understanding is by describing some criteria according to which it
is usually attributed to people.
. The psychological research (Cleeremans et al. ) on implicit learning testifies to
the plausibility of acquiring understanding without a feeling of understanding. Although
the boundary between learning with and without awareness is tricky to draw, the research
at least shows that one can learn without the metacognitive feeling that one is acquiring
new knowledge.
. Kyle Stanford () draws from the history of science and argues it is always pos-
sible that there are equally well-confirmed and scientifically serious alternatives to current
best theories that are just not conceived.
. A couple of psychological factors might make the evaluation of the future potential
of theories even more difficult. First, the assessment of the fruitfulness of one’s pet theory
is probably often tainted by wishful thinking. From a motivational point of view, this source
of optimism might be instrumental in bringing about bold hypotheses, but it is not helpful
in increasing the reliability of the scientist as a source of explanatory evaluation. The other
factor is the egocentric bias in evaluation. If one is tempted to be too optimistic about one’s
own theories, the contrary might hold for the competing accounts. They tend to be seen as
less promising. One way to bias the evaluation is to evaluate one’s own theory by its future
promise, and the competition by its achievements thus far. (There are various ways to
rationalize this asymmetry to oneself in such a manner that the imbalance seems only fair.)
. A similar point can be made about noncausal explanations, like constitutive ex-
planation. In this case, people think that findings about the parts of a system contribute
automatically to the understanding of the capacities of the system.
pe tri ylikoski
. The accumulation of irrelevant details makes the explanation worse, as its recipient
might not see which pieces of information are relevant for understanding and which are
not. But this has at least one positive consequence: at least the recipient is not prone to
IDU since he or she fails to get the point.
GZ[ZgZcXZh
Baker, G. P., and P. M. S. Hacker. . Wittgenstein: Understanding and meaning, part ,
Essays. Oxford: Blackwell.
Brem, S. K., and L. J. Rips. . Explanation and evidence in informal argument. Cognitive
Science :–.
Clark, A. . Being there: Putting brain, body, and the world together again. Cambridge,
MA: Massachusetts Institute of Technology Press.
Cleeremans, A., A. Destrebecqz, and M. Boyer. . Implicit learning: News from the
front. Trends in Cognitive Sciences :–.
Davis, D. A., P. E. Mazmanian, M. Fordis, R. Van Harrison, K. E. Thorpe, and L. Perrier.
. Accuracy of physician self-assessment compared with observed measures of
competence. Journal of the American Medical Association :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Donald, M. . Origins of the modern mind. Cambridge, MA: Harvard University Press.
Giere, R. N. . Scientific cognition as distributed cognition. In The cognitive basis of sci-
ence, edited by P. Carruthers, S. Stich, and M. Siegal, –. Cambridge: Cambridge
University Press.
Gopnik, A. . Explanation as orgasm and the drive for causal knowledge: The function,
evolution, and phenomenology of the theory formation system. In Explanation and
cognition, edited by F. C. Keil and R. A. Wilson, –. Cambridge, MA: Massachu-
setts Institute of Technology Press.
Hutchins, E. . Cognition in the wild. Cambridge, MA: Massachusetts Institute of
Technology Press.
Juslin, P., A. Winman, and H. Olsson. . Naïve empiricism and dogmatism in confi-
dence research: A critical examination of the hard-easy effect. Psychological Review
:–.
Keil, F. C. . Folkscience: Coarse interpretations of a complex reality. Trends in Cogni-
tive Sciences :–.
Koehler, D. J. . Explanation, imagination, and confidence in judgment. Psychological
Bulletin :–.
Koriat, A. . The feeling of knowing: Some metatheoretical implications for conscious-
ness and control. Consciousness and Cognition :–.
Lin, L.-M., and K. M. Zabrucky. . Calibration of comprehension: Research and
implications for education and instruction. Contemporary Educational Psychology
:–.
Lipton, P. . Inference to the best explanation. nd ed. London: Routledge.
Mills, C. M., and F. C. Keil. . Knowing the limits of one’s understanding: The devel-
t h e i l lusi o n o f d e p th o f un d e r sta n d i n g i n s c i e n c e
G8IK@@
JcYZghiVcY^c\VcYBdYZah
pa rt title ver so
chap ter title ver so
.
Understanding in Physics and Biology
From the Abstract to the Concrete
B6G<6G:IBDGG>HDC
m ar g ar et mo r r iso n
feature that furnishes understanding of the physical world. What both of these
accounts have in common is that they privilege a particular type of explanation
as the vehicle for producing understanding. De Regt and Dieks () criticize
these monolithic accounts of understanding and instead define it in terms of
a criterion of intelligibility that involves having a theory where one can recog-
nize qualitatively characteristic consequences without performing exact calcu-
lations. This is achieved using a variety of conceptual tools that are relevant to
the problem at hand. Their point is that in some cases causal knowledge will be
relevant to understanding, and in some cases it will not—the tool kit contains a
number of different resources that, depending on the context, will produce un-
derstanding. An important feature of their account is that its pragmatic quality
in no way compromises the epistemic relevance of understanding for scientific
investigation.
My own sympathies lie with the type of contextual analysis provided by
de Regt and Dieks, but I would put the point in slightly more negative terms.
While I agree that what it means to “understand” depends on contextual fac-
tors such as the nature of the problem and the resources available to solve it,
I want to claim that it is neither possible nor desirable to formulate a “theory”
of understanding. That does not mean that we cannot explain what it is to un-
derstand why a phenomenon behaves as it does or what we mean when we say
someone understands a mathematical concept.1 Rather, there is no canonical
account of what it means to understand. I would suggest the same holds for
explanation. Very often in science we desire explanations that are causal, but
that is not to say that this is the only form of explanation that is acceptable, or
that we can give a “theory” of explanation that centers on causal relations. On
the contrary, whether something has been successfully explained will depend
on the question and the stock of available answers, answers that come from our
background theories.
We simply cannot specify in advance what qualifies as an explanation or
what form it will take. However, I do think explanation and understanding are
linked in the intuitive sense that one often accompanies the other; that is, we
demonstrate our understanding by being able to offer explanations of the ob-
ject/concept in question and the success of our explanations is typically a func-
tion of how well we understand what we are trying to explain. My position is a
minimalist one insofar as neither understanding nor explanation is capable of
being codified into a philosophical theory.2
Bound up with the question on how mathematical abstraction enables us
to understand physical phenomena is the role played by models. In Morrison
and , I articulate how models can provide knowledge by focusing on
what I call their mediating role. That is, models function as autonomous media-
tors between theory and applications (the model provides simplifications of the
under standing in ph ysic s and biol o gy
theory’s equations so that they can then be applied) or between theory and the
world (the model is an idealized or abstract representation of some phenom-
enon or physical system). In this latter case, the representation can then be
compared with a concrete object/system.
The distinction between idealization and abstraction I am adopting for
the purposes of this chapter is roughly the following: abstraction is a process
whereby we describe phenomena in ways that cannot possibly be realized in
the physical world (for example, infinite populations). In this case, the math-
ematics associated with the description is necessary for modeling the system
in a specific way. Idealization, on the other hand, typically involves a process of
approximation whereby the system can become less idealized by adding cor-
rection factors (such as friction to a model pendulum). In the latter case, the
idealization is used primarily to ease calculation since the factors that have
been idealized will not, in general, be relevant for the type of problem at hand.
In their original state both abstraction and the idealization make reference
to phenomena that are not physically real; however, because the latter leaves
room for corrections via approximations, it can bear a closer relation to a con-
crete physical entity.3 This distinction will become important in showing how
mathematical abstraction enables us to understand certain features of empiri-
cal phenomena in ways that are different from the more straightforward fea-
tures associated with idealization.
The role of mathematical abstraction in model construction figures impor-
tantly in characterizing the autonomy of models.4 For example, in cases where
the system or phenomena under investigation may be inaccessible due to large
distances (galactic phenomena) or size (microstructures), the model occupies
center stage and takes on the task of representing how we assume the system
is constructed. And, because direct comparison between the model and the
system is not possible, the model supplants the system as the object under
investigation.
Models of stellar structure in astrophysics fit this category. In these cases,
the model fulfills its mediating role by serving as a source of mediated knowl-
edge, which is less direct than knowledge of real systems. In these cases, as well
as in more general modeling instances, there is often a good deal of mathemati-
cal abstraction (and idealization), structure that itself becomes the focus of in-
vestigation. Hence, our knowledge of real world situations is typically mediated
to some degree by this structure. In the cases I discuss below the mathematical
model supplants the physical phenomena as the object of investigation, and
the degree to which the model produces understanding comes via its abstract
mathematical structure. The interesting question is how we should understand
the relation between this abstract structure and the concrete physical systems
that this structure purportedly represents.5
m ar g ar et mo r r iso n
Hardy-Weinberg law and how the mathematical abstraction built into this law
allows us to understand fundamental features about biological populations. I
conclude with an account of how the abstract mathematics used by R. A. Fisher
produced tremendous advances in our understanding of the genetics of natural
populations.
Ldgg^ZhVWdjiA^b^ih
Even without bringing mathematics into the equation the very notion of
what it means to “understand” is by no means straightforward. As I noted
above, philosophical attempts to understand understanding are usually linked
to the equally controversial notion of explanation, more specifically, the task
of conveying information in a way that will answer a why- or how-question.
Typically, scientific understanding, especially in physics, involves the applica-
tion of both general principles and more specialized models that enable us to
describe, predict, and explain specific types of behavior. This is not to say that
an appropriate answer to every question must invoke both strategies but simply
to recognize that certain kinds of questions demand certain kinds of answers.
In other words, the idea that explanation can be achieved solely on the basis
of fundamental laws and initial conditions, as described by the D-N model, is
now thought to be insufficient for understanding much of the physical world.
We need a variety of things— fundamental theory, different kinds of models, as
well as laws that constrain the kinds of behaviors that systems display.
The important question here is how the abstraction (as opposed to idealiza-
tion) built into theories, models, and laws enables or indeed prevents us from
understanding physical systems. Initially one might think that understanding is
inhibited by abstraction, but we only need to look to scientific textbooks to see
that this is clearly not the case. The kind of models that serve as exemplars in
the Kuhnian sense (for example, the quantum harmonic oscillator, the infinite
potential well, and so on), as well as entities like virtual particles, embody a
great deal of abstraction and yet are the cornerstone of understanding essential
features of certain kinds of physical systems. This is because they enable us to
conceive of systems as being of a particular type, exhibiting certain kinds of
behavior that allow us to classify them in terms of their general features. All of
these factors enhance our understanding.
But, understanding via abstraction is not simply limited to general features
of physical systems. In many of these cases, abstract representations also pro-
vide the kind of detailed knowledge required to answer causal questions. For
example, if we ask why a particular metal exhibits superconducting properties,
we can explain it in terms of zero resistivity and the accompanying thermo-
dynamic phase transition, something that is only possible in certain kinds of
m ar g ar et mo r r iso n
metals. And if we want to know the details of what happens at the subatomic
level, we can invoke the BCS model complete with its description of electron-
phonon interactions, Cooper pairing, and the BCS wave function. Although
these models involve abstraction and approximation, they allow for a rather
peculiar situation: from them we can derive exact results associated with su-
perconductivity—infinite conductivity, exclusion of magnetic fields, flux quan-
tization, and zero resistivity.
The question is how one can get these kinds of exact consequences from
models that are approximations. In other words, we want to understand why
the models work so well. The reason is because we can also understand a su-
perconductor as a material in which electromagnetic gauge invariance is spon-
taneously broken and these models contain, as a fundamental assumption, the
breakdown of electromagnetic gauge invariance. The detailed dynamic theories
and models like BCS are required to explain why and at what temperatures this
symmetry breaking occurs, but not to derive the kinds of consequences men-
tioned above. In other words, these properties can be derived directly from the
assumption of the spontaneous breaking of electromagnetic gauge invariance
and so are consequences of general principles rather than specific approxima-
tions embodied in the model.6
But how, exactly, is mathematical abstraction involved here? The situation
is far from straightforward. The spontaneous breakdown of electromagnetic
gauge invariance involves a phase transition that is associated with the super-
conducting state. The occurrence of phase transitions requires a mathematical
technique known as taking the “thermodynamic limit,” N ∞; in other words,
we need to assume that a system contains an infinite number of particles in
order to understand the behavior of a real, finite system. Very briefly, the situa-
tion is as follows: in thermodynamics (TD), phase transitions are accounted for
in terms of discontinuities in the thermodynamic potentials. However, once we
move to statistical mechanics (SM), the equations of motion that govern these
systems are analytic and hence do not exhibit singularities. As a result, there
is no basis for explaining phase transitions in SM. In order to recover the TD
explanation, singularities are introduced into the equations, and thus far the
only way to do this is by assuming the number of particles in the system is in-
finite. Note that the problem here is not that the limit provides an easier route
to the calculational features associated with understanding phase transitions;
rather, the assumption that the system is infinite is necessary for the symmetry
breaking associated with phase transitions to occur. In other words, we have a
description of a physically unrealizable situation (an infinite system) that is re-
quired to explain a physically realizable phenomenon (the occurrence of phase
transitions).
The question is whether this procedure yields the kind of explanation that
under standing in ph ysic s and biol o gy
produces understanding. Although one might want to claim that within the
mathematical framework of SM we can causally account for (explain) the oc-
currence of phase transitions by assuming the system is infinite, it is neverthe-
less tempting to conclude that this explanation does not help us to physically
understand how the process takes place since the systems that SM deals with
are all finite. Similar doubts have been expressed by Earman (), who argues
against taking the thermodynamic limit as a legitimate form of idealization:
“a sound principle of interpretation would seem to be that no effect can be
counted as a genuine physical effect if it disappears when the idealisations are
removed” (). In other words, we should not assume that phase transitions
have been explained and thereby understood if their occurrence relies solely on
the presence of an idealization. Initially this seems an intuitive and plausible
objection, but if we reflect for a moment on the way mathematical abstraction
is employed for explanatory purposes in scientific models or theories, it be-
comes clear that this line of reasoning quickly rules out explanations of the sort
we deem acceptable in other contexts. Here the distinction I introduced at the
beginning between abstraction and idealization becomes especially important.
Specifically, we need to distinguish between the kind of mathematical abstrac-
tion required for a theoretical representation and the more straightforward
kinds of mathematical idealizations that are used simply to facilitate calcula-
tion. In the former case, the abstraction becomes a fundamental part of how
the system is modeled or represented and consequently proves crucial to our
understanding of how it behaves.
For example, consider the intertheoretical relations that exist in fluid me-
chanics between Navier Stokes equations and the Euler equations, or between
theories like wave optics and ray optics, and classical and quantum mechanics.
Because of the mathematical nature of physical theories the relations between
them will typically be expressed in terms of the relations between different
equations or solutions. In each case, we are interested in certain kinds of limit-
ing behavior expressed by a dimensionless parameter δ. In fluid dynamics δ
is equal to /Re (Reynolds number), and in quantum mechanics it is Planck’s
constant divided by a typical classical action (ћ/S). But, in fluid mechanics (as
in the other cases listed above) the limit δ is singular and it is this singular-
ity that is responsible for turbulent flows. Similarly, in the ray limit where geo-
metrical optics accurately describe the workings of telescopes and cameras the
wavelength λ. Because ψ is nonanalytic at λ=, the wave function oscillates
infinitely fast and takes all values between + and – infinitely often in any
finite range of x or t.
A good deal of asymptotic behavior that is crucial for describing physical
phenomena relies on exactly these kinds of mathematical abstractions. What
we classify as “emergent” phenomena in physics such as the crystalline state,
m ar g ar et mo r r iso n
:migVXi^c\i]ZIgji][gdbÆ;VahZÇAVlh
erations to reach Hardy-Weinberg equilibrium. One can see, then, the relation
between the stability of the frequencies and Mendel’s law of segregation. With
random cross-fertilization there is no disappearance of any class whatever in
the offspring of the hybrids; each class continues to be produced in the same
proportion.11
But, and here is the important point, what is significant about the Hardy-
Weinberg law is not so much the binomial form of the genotype frequency and
the prediction of genotypes based on the stability of the population, but rather
what the stability actually shows or presupposes. Despite the unrealistic as-
sumptions, the stability allows us to comprehend something about Mendelian
populations that is significant for comprehending heredity and variation. In
other words, certain conditions must be present for the stability to be possible.
Thus, the predictive success of the law is intimately connected with certain ba-
sic claims about genetic structure. What the Hardy-Weinberg law says is that
if no external forces act, then there is no intrinsic tendency for the variation
caused by the three different genotypes that exist in a population to disappear.
It also shows that because the distribution of genotype frequencies is inde-
pendent of dominance, dominance alone cannot change genotype frequencies.
In other words, there is no evidence that a dominant character will show a
tendency to spread or allow a recessive one to die out. Instead, the genotype
frequencies are maintained in constant proportions. The probabilistic genetic
structure is conserved indefinitely; but should it be influenced by an outside
force, for example, mutation, the effect would be preserved in a new, stable
distribution in the succeeding generation.
This was crucial for understanding the problems with blending inheritance
as advocated by the Darwinians, and to that extent the claim that the law is false
in some sense misses the point if our concern is understanding and conveying
information. Under blending inheritance, variation was thought to decrease
rapidly with each successive generation, but Hardy-Weinberg shows that under
a Mendelian scheme it is maintained. This pointed to yet another fundamental
aspect of Mendelism, namely, the discontinuous nature of the gene and why
it was important for the preservation of variation required for selection. How
was it possible for the genetic structure to be maintained over successive gen-
erations? The reason for the stability could be traced directly to the absence of
fusion, which was indicative of a type of genetic structure that could conserve
modification. This condition was explicitly presupposed in the way the law was
formulated and how it functioned. In that sense, then, one can see the Hardy-
Weinberg law as the beginnings of a completely new understanding of the role
of mutation and selection and how they affect our understanding of evolution.
Appealing to the abstraction/idealization distinction can also clarify our
under standing in ph ysic s and biol o gy
not have been arrived at nonmathematically, such as the results in the multilo-
cus and the stochastic theories. Moreover, the fact that selection could operate
in Mendelian populations could only be shown using the kind of mathematical
abstraction that we frequently see in physics.
>cÒc^iZEdejaVi^dch/;gdbHiVi^hi^XVaE]nh^Xhid
BVi]ZbVi^XVa7^dad\n
viduals were affected would the effect of selection override random survival,
and even then only a small minority of the population would be affected. To
do this he introduced stochastic considerations and examined the survival of
individual genes by means of a branching process analyzed by functional itera-
tion and then set up the “chain-binomial” model and analyzed it by a diffusion
approximation.15 He was able to calculate the amount of mutation needed to
maintain the variability given a specific amount of selection and found that to
maintain variability in the case of equilibrium in the absence of selection, the
rate of mutation had to be increased by a very large quantity. So the presence
of even the slightest amount of selection in large populations had considerably
more influence in keeping variability in check than did random survival. Con-
sequently, the assumption of genotypic selection balanced by occasional muta-
tions fit the facts deduced from the correlations of relatives in humans.
So, by making assumptions about the large size of the population and its
high degree of genetic variability, Fisher was able to demonstrate how his
stochastic distributions led to the conclusion that natural selection acting on
genes (rather than mutation, random extinction, epistasis, and so on) was the
primary determinant in the evolutionary process. He found that mutation
rates significantly higher than any observed in nature could be balanced by
very small selection rates. The distribution of the frequency ratios for different
factors was calculated from the assumption that the distribution was stable.
The kind of statistical independence that figured prominently in the velocity-
distribution law was applicable to the effects of selection in Mendelian popula-
tions. In keeping with the idealization/abstraction distinction I have been mak-
ing we can see why Fisher’s assumptions and techniques fall into the category of
mathematical abstraction. Without the requirement of an infinitely large popu-
lation of genes and the mathematical techniques necessary for dealing with it in
the context of other aspects of the population, Fisher would have encountered
the kind of problems that Pearson envisaged using straightforward statistical
analysis. In that sense, mathematical abstraction became a necessary feature
for understanding basic features of Mendelian populations.
What can we say therefore about the understanding of selection that
emerges from Fisher’s analysis? The distribution of the gene ratio provides the
ultimate expression of selective effects because the gene remains the only trace
of the existence of an individual in a population.16 Given that selection acts
on the heritable, what is important is the mean effect of each allele. Although
we cannot know the state of each of the genes in the population, we can know
the statistical result of their interaction in the same way that gas laws can be
deduced from a collection of particles. Selection is mass selection, taking into
account only the additive effects of genes; stochastic factors can be ruled out
m ar g ar et mo r r iso n
like Mayr who argue that the methods of population genetics provide little in
the way of understanding evolutionary development. However, what I take my
analysis to have shown is that this is at least prima facie incorrect. While it is
true that our understanding of selection has departed from the traditional Dar-
winian one, it is also the case that a properly quantitative account of selection as
the mechanism of evolutionary change necessitated this departure.17
8dcXajh^dc
CdiZh
I would like to thank the Social Sciences and Humanities Research Council of Canada
for research support. I would also like to thank the editors for helpful comments.
. For example, we understand a mathematical concept if we can apply it to solve prob-
lems in ways other than simply reproducing the example used for learning the concept;
and we understand a phenomenon like the weather if we are able to make predictions
and explain the reasons for those predictions. The point here is simply that in some cases
prediction is a feature of understanding and not in others.
. I realize, of course, that de Regt and Dieks may not agree with my interpretation of
their position. My intention here is not to argue for my view of explanation or understand-
ing but simply to state how I see the relation between the two for the purposes of this
chapter.
. My use of the terms abstraction and idealization differs from many of the accounts
in the literature, for example, Cartwright (). Instead of defining abstraction as leaving
out properties that belong to the system in question and idealization as the distortion of
certain properties (for example, point particle), I have chosen, instead, to focus on how
certain types of abstract mathematical frameworks are necessary for representing physical
systems/phenomena. My reason for changing the terms of reference here relate to what
I see as a deficiency in the standard account where the motivation for idealization and
abstraction is typically related to modeling assumptions or ease of calculation. In cases
of both idealization and abstraction, we are able to add back properties or de-idealize in
order to make the system more realistic or concrete. However, it says nothing about how
we should understand cases where this is not possible—where the abstract mathematical
representation is necessary for representing the target system. Because of the importance
of this latter type of abstraction for mathematical modeling, we need to have some account
of the way it functions in our understanding of the model systems. For more on this way of
characterizing the distinction, see Morrison (b).
. In Morrison () I claim that what makes a model autonomous is not simply that
it is not derived from theory or phenomena but involves elements from a variety of dif-
ferent sources that, taken together, make the model a unique kind of entity. It is by virtue
of this uniqueness that they can and do play a mediating role. With models, and in other
situations as well, mediators need to occupy a relatively independent position if they are to
properly fulfill the function they are assigned. But, their role as a mediator can vary widely.
Indeed, in some cases the structures comprising the model fail to provide any clues about
the physical nature of the system in question, and the model also fails to point to specific
relations/reduction to some background theory. Instead, models may be used simply for
their predictive power or to account for specific kinds of behaviors without interpreting
them as having any realistic representational power whatsoever. In these latter cases, the
models do not really provide us with any understanding of the system/phenomena in ques-
tion; they save the phenomena without answering questions about why the system behaves
as it does. An example of this kind of model is the aether models developed by James Clerk
Maxwell in his early formulations of the electromagnetic theory. While I still believe these
considerations are important in characterizing a model’s autonomy, other features also
play a crucial role—one of which is the way the model represents the system of interest.
under standing in ph ysic s and biol o gy
. That is, the Mendelian mechanism ensures that although a population may be said
to have continued existence, the individuals that comprise it do not. The variability that
passes from one generation to the next through reproduction is related to but not identical
to phenotypic variability.
. For a longer discussion of these issues, see Morrison (b).
GZ[ZgZcXZh
Batterman, R. . Critical phenomena and breaking drop: Infinite idealizations in phys-
ics. Studies in History and Philosophy of Modern Physics :–.
Cartwright, N. . How the laws of physics lie. Oxford: Clarendon Press. .
———. . The dappled world. Cambridge: Cambridge University Press.
. How the laws of physics lie. Oxford: Clarendon Press.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Earman, John. . Curie’s principle and spontaneous symmetry breaking. International
Studies in the Philosophy of Science :–.
Edwards, A. W. F. . The fundamental theorem of natural selection. Biological Review
:–.
Ewens, W. . Mathematical population genetics, vol. , Theoretical introduction. New
York: Springer.
Fisher, R. A. . The correlation between relatives on the supposition of Mendelian
inheritance. Transactions of the Royal Society of Edinburgh :–.
———. . On the dominance ratio. Proceedings of the Royal Society of Edinburgh
:–.
Friedman, M. . Explanation and scientific understanding. Journal of Philosophy
:–.
Hempel, C. G. . Aspects of scientific explanation. In Aspects of scientific explanation
and other essays in the philosophy of science, –. New York: Free Press.
Kitcher, P. . Explanatory unification and the causal structure of the world. In Scientific
explanation, edited by P. Kitcher and W. C. Salmon, –. Minneapolis: University
of Minnesota Press.
Mayr, Ernst. . Where are we? Reprinted in Ernst Mayr, Evolution and the diversity of
life. . Cambridge, MA: Harvard University Press, –.
———. . The growth of biological thought: Diversity, evolution, and inheritance. Cam-
bridge, MA: Belknap Press.
Morrison, M. . Modelling nature: Between physics and the physical world. Philosophia
Naturalis :–.
———. . Models as autonomous agents. In Models as mediators: Perspectives on natu-
ral and social science, edited by M. S. Morgan and M. Morrison, –. Cambridge:
Cambridge University Press.
———. . Unifying scientific theories: Physical concepts and mathematical structures.
Cambridge: Cambridge University Press.
———. . Modelling populations: Pearson and Fisher on Mendelism and biometry. Brit-
ish Journal for the Philosophy of Science :–.
under standing in ph ysic s and biol o gy
/
Understanding by Modeling
An Objectual Approach
I6 G ? 6 @ C J J II > A 6 6 C 9 B 6 G I > C 6 B : G O
under standing by modeling
I]ZGZegZhZciVi^dcVakZghjhi]ZEgdYjXi^kZ6XXdjcid[BdYZah
where the design of new kinds of organisms on the one hand and new kinds of
computers on the other have converged—that the resulting “genetic comput-
ers” are not mere models in the traditional sense of the word: they are tools for
both material intervention and thinking. In much the same vein, Darden (,
) points out, “The model of DNA served as a model for investigating the
functions of the genetic material.” What is interesting about this remark is its
insight that “modeling of ” and “modeling for” proceed in tandem in scientific
research. The idea of models as tools for thinking is also present in concep-
tions that consider them to be surrogate systems (Suárez ). Investigating
economic models, which have often been considered to give notoriously false
depictions of their target systems, Mäki () characterizes them as “mini-
worlds” that are easier to study than the complex real world in which we live.
Consequently, as models enable further investigation of different phenomena,
a large part of their scientific and cognitive value is due to their workable and
tool-like character rather than to their success in representing some target phe-
nomenon more or less accurately.3
Considering models from a productive perspective requires one to address
them as autonomous, concrete objects. This is (partly) recognized by Morrison
and Morgan (), who focus in their account of models as investigative in-
struments on how we learn from them by constructing and manipulating them.
This, they claim, is made possible by the model’s partial autonomy, which they
interpret in terms of its relative independence from the traditional theory-data
framework. However, it seems that their view on modeling permits, and to be-
come efficient in fact requires, a more radical reading. If our aim is to stress how
models enable us to learn from the processes of constructing and manipulating
them, it is not sufficient that they are considered autonomous: they also need to
be concrete in the sense that they must have a tangible dimension that can be
worked on. This concreteness is provided by a material embodiment that gives
them the spatial and temporal cohesion that enables their workability, and that
preserves their identity across different sites and users. Consequently, once we
approach models as concrete things, the possible mediation that they provide
is no longer restricted to mediation between theory and data. Instead, scien-
tific models can then be viewed as concrete, human-made objects that draw
together numerous activities and widely different actors.4
The idea that models are concrete objects that have a material dimension
appears to run counter to the established philosophical view, which, apart from
treating models as representations, also tends to conceive of them as abstract
or conceptual things. However, it is difficult to see how one could work and
experiment with models and communicate one’s findings about them to oth-
ers without their concretization—in the form of diagrams or various symbols
under standing by modeling
6cDW_ZXijVa6eegdVX]idHX^Zci^ÒXJcYZghiVcY^c\
The idea that models are productive, concrete objects resonates nicely with
views that link scientific understanding to the abilities of scientists. Whereas
some writers claim that understanding is a kind of ability (Woodward ;
Ylikoski, this volume; Kuorikoski ), for others certain kinds of abilities
serve as a criterion for scientific understanding (de Regt and Dieks ). For
instance, Kuorikoski combines the Wittgensteinian nonmentalist conception
of understanding with the manipulative theory of causation and explanation
presented by James Woodward in his aim to show that understanding need not
be seen as any “special or privileged mental state” (Kuorikoski ). Instead,
he claims that it is constituted of our perceived abilities to infer and intervene,
which are due to our knowledge about the relevant causal dependencies. Thus,
the rationale behind likening understanding to abilities is that we typically tend
to attribute it to others to the extent that they are able to use their knowledge
successfully (de Regt ).5 These abilities are manifested in several ways: by
inferring, manipulating, and building successfully or by being able to account
for a phenomenon theoretically (see Ylikoski, this volume). Thus, understand-
ing appears as knowledge put into use, which might explain why understanding
and knowledge are often so difficult, if not impossible, to untangle.
One way of examining more closely the aforementioned entanglement of
knowledge and understanding is to ask how epistemically relevant abilities arise.
From the objectual perspective knowledge, understanding, and manipulative
abilities could arise in the very same process of interacting with external things.
This applies to modeling as well as to experimentation. While interventional
and manipulative abilities have been more readily ascribed to experimentation
than to theory (or theoretical models), our perspective on models depicts them
as objects through which we can make our theoretical ideas workable and “ex-
perimentable.” Thus, they also question the distinction between representing
and intervening (Hacking ), as they are entities constructed with the help
of different representational means for the purposes of manipulation.
tarja knuu t til a and m artina mer z
rial dimension of models and the diverse representational media they make
use of are crucial for promoting understanding through them. Consequently,
Knuuttila argues that, contrary to the philosophical tradition, one should take
into account the medium through which scientific models are materialized as
concrete, intersubjectively available objects (Knuuttila and Voutilainen ;
Knuuttila ). This media-specific approach focuses on their constraints and
affordances, which are partly due to the characteristics of the specific repre-
sentational means (diagrammatic, symbolic, -D, and so on) with which they
are imbued.9 There is evidence from studies in cognitive science that the rep-
resentational means used influences how well human subjects can grasp the
intended content. Zhang (), for instance, shows that different representa-
tions of the same abstract structures have different enablings as to how hu-
mans understand them. It seems that analogical and visual representations are
easier for humans to grasp than those in digital and numerical forms. This fits
well with the various pronouncements of scientists concerning the virtues of
visualization.
Apart from making use of representational media in an understanding-pro-
viding way, models typically also constrain the problem at hand, which renders
the initial problem situation more intelligible and workable. They provide us
with surrogate systems, the study of which furthers understanding by simpli-
fying or modifying the problems scientists deal with. Often the real world is
just too complex to study as such. Thus, scientific work typically proceeds by
turning the limitations built into the models into affordances; one devises the
model in such a way that one can learn and gain understanding from using or
“manipulating” it. Yet, as concrete objects, models have their own construc-
tion, which also explains how they can be doers: their construction resists some
uses and allows others. Moreover, they can lead to unexpected findings and
they also breed new problems.
We suggest that understanding through models comes typically by way of
building them, experimenting with them, and trying out their different alterna-
tive uses—which in turn explains why they are typically valued for their per-
formance and their results or output. Conceived of in this way, they may be
epistemically valuable even if they do not represent reality accurately. Purpose-
fully “wrong” models can give us knowledge that is commonplace in scientific
practice, which shows that models can be good inference generators even if
they do not represent their putative target objects accurately in any relevant re-
spect (this point is further discussed below in the parsing case; see also Suárez
). Moreover, the stress on artifactuality and materiality can accommodate
what Baird () calls “thing knowledge,” which is often tacit and bound more
to local environments and expertise than to explicit theoretical knowledge.
Furthermore, the artifactual view accommodates different kinds of epistemic
under standing by modeling
Ild8VhZhd[JcYZghiVcY^c\WnBdYZa^c\
:kZci<ZcZgVidgh^cEVgi^XaZE]nh^Xh
Models in particle physics come in a variety of disguises, of which the mod-
el-based practice of computer simulation is particularly prevalent.11 Extensive-
ly applied, simulation contributes to the production of scientific knowledge in
various ways. We suggest that simulation work is predominantly “understand-
ing work” targeted at distinct kinds of understanding tasks. It can be produc-
tively viewed through its objects, the simulation programs (or models), which
draw together an entire web of scientific activities, different physics communi-
ties, and temporal phases.
Simulation is rendered productive for the purpose of a collider experiment.
Such experiments target new types of elementary particles (for example, the
Higgs boson) and their properties, which are traced and measured by particle
detectors with the aim of learning more about the basic constituents of mat-
ter. The novel particles and properties are mediated by collisions between two
particle beams, each collision generating a profusion of debris particles whose
properties are at stake. Simulation maps and mimics the aforementioned ex-
perimental process, that is, the production and subsequent detection of ele-
mentary particles as they are thought to occur in a “real”12 experiment. For this
purpose, the two simulation models that assume the roles of the collider (the
“event generator”) and the detector are coupled, and the output of the event
generator is fed into the detector simulation.
Event generators come in varying forms and sizes. Of particular impor-
tance are the comprehensive, multipurpose event generators that are made
publicly available throughout the particle-physics community. Event genera-
tors are not based on a unique theoretical framework: the models embodied in
them are of different theoretical origins. They account for the successive stages
of the collision process and its consequences (models of hard processes, parton
distributions, initial- and final-state radiation, beam remnants, fragmentation,
and decays, for example) and contain a combination of exact calculations, ap-
proximations, corrections, and parameterizations. Another set of ingredients
consists of the values of quantities determined in earlier experiments (particle
masses, decay widths, couplings, and cross sections, for example). The random
nature of physical processes is accounted for by Monte Carlo techniques. In a
simulation program, all of these ingredients are coupled by way of algorithms
under standing by modeling
that turn the input (for example, type of colliding particles, beam energy) into
the output that consists of listings of the properties of all particles produced in
the event.
Due to their complex makeup and continuing adaptation to novel experi-
mental conditions and theoretical insights, event generators retain the status
of preliminary, unfinished, and unfolding objects. While their “authors” (that
is, constructors) continuously modify, improve, and update them to meet new
requirements, other physicists use them, at the same time, as tools for a range
of tasks. The object’s multiplex nature, which we describe below, goes hand in
hand with simulation as performed by members of different particle-physics
communities and thus produces different kinds of outcomes. The communities
comprise theoretical physicists (including the event-generator authors) and
experimental particle physicists (including most event-generator users), who
may be further differentiated in terms of their specific expertise (for example,
in detector systems or physics analysis). Event generators assist in coordinating
work across these communities. Below we outline three ways in which develop-
ing and working with simulations creates scientific understanding in particle
physics, in each case involving different actors, types of expertise, and simula-
tion roles.
physics potential of new collider features. The event generator is thus used as a
kind of processor that turns the new model features into numbers, the simulat-
ed events, which can then be scrutinized. This exemplifies what we introduced
above as the productive and interactive nature of models.
by way of interacting with its surrogate. In order to optimize the detector, they
vary the design features in the simulation and interactively try out which con-
figurations are the most sensitive. Simulation enables scientists to flexibly re-
consider a variety of design features with the possibility of isolating specific
aspects. Through its materialization as a concrete object the detector simula-
tion constitutes an enhanced environment that allows physicists to zoom in
on contested theoretical areas and test out a variety of potential solutions to
technical problems. In this application the simulation program thus becomes
an inference generator that assists physicists in negotiating competing physics
programs and technical design options.
Due to the brevity of our account we could only provide a glimpse into the
multifunctional nature of simulation work in particle physics, suggesting that
scientific models (in this case simulation programs) inhabit a field of tension
spanned by their epistemic productivity and their instrumental character. Thus,
both event generators and detector simulation programs are multiplex objects.
When they are coupled for the purpose of specific applications both can play
different roles as tools, scientific objects, and inference generators. This com-
plex interweaving makes it possible for simulation programs, as multiplex ob-
jects, to enhance understanding of entities as varied as conceptual objects (for
example, theoretical ideas), epistemic strategies (such as data analysis) and the
workings of technological artifacts (as in detectors).
EVghZgh^cAVc\jV\ZIZX]cdad\n
Caught between different disciplines, programming, and practical purpos-
es, language technology provides an interesting case for studying how techno-
logically oriented model-building that strives primarily for new applications
and efficient solutions can give us scientific understanding. In the following we
consider the case of the Constraint Grammar parser, which challenges many of
our established views on scientific knowledge and understanding. First, parsers
are both unrepresentational and atheoretical things in the sense that they are
not expected to be realistic depictions of human linguistic competence, nor are
they based on any autonomous formal linguistic theory. Thus, second, a large
part of the scientific understanding they provide is not theoretical but rather
inherent in the expertise that develops in the construction process. This un-
derstanding concerns language and cognition as well as the technologies used.
Third, like event generators, parsers produce many kinds of understanding, de-
pending on the purposes for which they are used.
A parser could be characterized as a computational model of a certain (as-
pect of a) language implemented as a computer program. To parse is to describe
a word or a sentence grammatically; in the context of language technology, it
refers to “the automatic assignment of morphological and syntactic structure
tarja knuu t til a and m artina mer z
(but not semantic interpretation) to written input texts of any length and com-
plexity” (Karlsson et al. , ). There are two primary approaches to parsing,
one of which is grammar based, linguistic, and descriptive and the other proba-
bilistic, data-driven, and statistical. The methods of the data-driven approach
include corpus-based learning rules, Hidden Markov models, and machine
learning. Both approaches are grounded basically in the notion of parsing be-
ing based on the information given by the neighboring words.
The words we use in natural language are typically ambiguous: the same
word form, say “bear,” can be read differently depending on the context. The
parser works by making use of the morphological and syntactic information of
the words in the vicinity of the word in question. For instance, the word form
“bear” has both nominal and verbal readings; thus, if it is preceded by an article,
the parser assigns it a nominal reading. This is not how we humans proceed,
since unlike parsers we also have the semantic dimension of language available
and understand the intended meaning of the words and phrases.
The Constraint Grammar parser (CG) (Karlsson et al. ) is a product
of long-term research in language technology carried out at the University of
Helsinki.13 It is a grammar-based parser built on hand-crafted rules made with
the help of the Constraint Grammar, which is a language-independent parsing
formalism devised for describing the rules of any language. In contrast to tra-
ditional and formal grammars, which are licensing grammars, a CG grammar
is reductive. Instead of defining the rules for the formation of correct expres-
sions in a language, the constraint grammar of a particular language specifies
the constraints that discard as many improper alternatives as possible. The CG
parser is firmly grounded in linguistic corpora. A parser for a certain language
is created in a piecemeal fashion involving continuous rounds of rule writ-
ing, implementing the rules as a computer program, and testing the grammar
against manually disambiguated benchmark corpora. For example, the EngCG
(Voutilainen et al. ; Karlsson et al. ), which is still one of the best pars-
ers for English, took a group of linguists and a computer scientist several years
to construct.
the developers put it. Thus, they are primarily productive things, being rather
models for than models of.
In addition to treating models as representations, the philosophical tradi-
tion also sees them as interpretations of formal theoretical templates. In fact,
from this perspective representation and interpretation are interwoven, since
interpretation gives a theoretical template empirical content and makes it thus
represent some target object; representation, in turn, is based on the structure
of the model. What is more, if this structure is formally simple and preferably
shared by many other theories, then from the unificationist perspective at least
it imparts scientific understanding: we have succeeded in identifying an under-
lying structure that is common to many other phenomena. All this, however,
does little to account for the part that theory plays in language technology. Due
to its corpus-based nature, the CG grammar carries constraints that are quite
different from the grammar rules described by formal syntactic theories.14 In-
stead of giving one general rule, it sets out a few dozen down-to-earth con-
straints that state bits and pieces of the phenomenon by ruling out what cannot
be the case in some specific context.
Yet, it would be misleading to characterize the CG parser as a nontheoreti-
cal entity. There is a theory behind constraint grammar parsing, but rather than
being an abstract description of the general mechanisms and laws prevailing
in external reality, it is a set of instructions for building a specific artifact (see
Karlsson et al. ). These principles remain on a very general level, how-
ever, and do not in themselves really amount to any guide as to how to build
a CG parser. What this shows is that a large part of knowledge concerning
parser construction can be attributed to a special kind of expertise rather than
to overall explicit theoretical understanding. Expertise has typically been at-
tributed to local, experience-bound, and tacit dimensions of science in recent
discussions, yet it has remained rather difficult to tell what expertise consists of
and how it is gained.15 The case of parsing suggests that it is firmly tied to work-
ing with certain specific methods and artifacts, taking into account the tasks to
be accomplished. The researchers in the field highly appreciate methodological
know-how—the ability of the researcher to (fore)see what can be done with
different methods.
a fact that is often neglected in the philosophy of science with its strong focus
on the representational aspects of models. Interestingly, the CG parser is a dif-
ferent kind of object to linguists and to computer scientists: while linguists
see parsers primarily as linguistic descriptions and tools for linguistic analysis,
computer scientists abstract from the linguistic description and consider them
in terms of their computational characteristics.
In their capacity as inferential devices, parsers function largely like any oth-
er models, and as such they are able to impart understanding about language
and cognition despite their primarily instrumental nature. Their understand-
ing-providing features derive from the fact that they are artifacts with known
principles, mechanisms, and structures (although we need not be in command
of the structure, as the machine-learning approaches to parsing show). Yet the
parsing case indicates that the structures themselves do not necessarily provide
any understanding. If we are to understand how language actually functions,
the description has to be made operative and, as a consequence, productive.
In the case of the CG parser, this is done by implementing the language de-
scription as a computer program—a possibility that has made grammars more
interesting as scientific objects. Traditional evaluation of grammars has relied
on conventional academic discussion, but it is practically impossible for hu-
mans to consistently follow such complex rule systems as set out in extensive
grammars. This is possible for computers, which can check whether such rule
systems are consistent, for example. In terms of implementation, the linguist
also has the possibility to learn from the performance of the model by finding
out which rules cause trouble and by trying out different ones: this emphasizes
the importance of the workability of models in creating scientific understand-
ing in the sense of licensing different inferences.
Interestingly, the instrumental fitness of the parser as a tool is also impor-
tant from the epistemic perspective. Once it functions well, it provides a start-
ing point for diverse interpretations and questions. For instance, it is possible
to study what properties distinguish successful models from unsuccessful ones,
and what assumptions can be made about human language faculties. Last but
not least, the making of language-technological artifacts has given us a new
understanding of language that we could not have gained by any other means.
Here, “understanding” is a more appropriate word than “knowledge,” since over
and above any explicit knowledge it has enabled researchers to grasp how poly-
semous, context-bound, and ambiguous our language use indeed is.
8dcXajh^dch
We began this chapter with the observation that models provide an impor-
tant locus of understanding in science. This observation is backed by the fact
under standing by modeling
of the understanding created is coupled with the specific uses and reactions of
the artifacts the scientists work with, then the special skills and experiences
of the scientists can be attributed to the features of the objects themselves.
In this respect, the material dimension of scientific objects is crucial: objects
are intersubjectively present to us only to the extent that they materialize as
recordings, experimental effects, or as purposefully constructed epistemic ob-
jects—in other words, models. Moreover, the objectual relations inherent in
scientific practice also explain the link between familiarity and understanding.
Working with certain kinds of objects makes them familiar, but this does not
mean that they have to be familiar in an everyday sense.
Finally, the affective and emotional component of scientific work—an
important issue we have not touched upon in this chapter—can also be ap-
proached from the objectual point of view. There exists a large body of litera-
ture in psychology on object relations, which are taken to be constitutive in the
creation of our individual selves from early childhood on. In conclusion, one
may wonder whether our scientific identities are also intricately tied to the ob-
jects we deal with in our scientific work, often over extended time spans.
CdiZh
We wish to thank the three editors of the present volume, Henk de Regt, Sabina
Leonelli, and Kai Eigner; and the members of the Philosophy of Science Group at the
University of Helsinki, Uskali Mäki, Petri Ylikoski, Jaakko Kuorikoski, Aki Lehtinen, Tomi
Kokkonen, and Jani Raerinne, for their insightful and fruitful comments on earlier drafts of
this chapter.
. This overall commitment of philosophers to the representational character of
models has gone as far as asserting that models are representations (see, for example, Frigg
; Hughes ; Teller ).
. The epistemic value of models refers here to the different ways in which they give
us knowledge. It has traditionally been ascribed to the context of justification as opposed
to the context of discovery. In our opinion, this distinction does not fit well with modeling
because part of the justification of models is typically built into them in that we tend to
trust separately in many of the ingredients they consist of (Boumans ). Moreover, the
question of representation has provided to several philosophers a way to evade the concept
of truth, which has been taken as the epistemic concept par excellence.
. We do not deny that, in many cases, scientific models do represent some target
objects, but would rather stress, in line with the pragmatists, that representation is always
also an accomplishment of representation users. On the level of actual scientific work, this
involves an extended process of building and working with models—the work often pro-
ceeding quite apart from the production of links to observable phenomena and data. The
respective practices of theoretical modelers and experimentalists often do remain rather
separate. Thus, the unilateral interest in representation tends to leave unnoticed a large
part of modeling work and its epistemic significance.
under standing by modeling
. In this, scientific models function like boundary objects, which according to Star
and Griesemer () manage the cooperation and understanding between different
“social worlds.”
. We have intentionally left out of our discussion the so-called feeling of understand-
ing. We agree with Ylikoski (this volume) that the feeling of understanding and under-
standing as a kind of ability should be distinguished, but from the objectual perspective it
is possible to explain from which kinds of experiences the feeling of understanding arises.
For example, Myers (, ) describes how scientists in the field of protein crystal-
lography cultivate a “feeling for the possible forms and movements of the protein in vivo”
through extended periods of building and navigating through protein structures on screen.
. By “objects,” we refer to entities that are both able to resist (“object to”) our interpre-
tations and manipulations, and that are variously materialized, thus also having a corporeal
and concrete dimension—if only through representations and measurements—and experi-
mental or simulated effects and outcomes.
. This is essentially a Kantian point, though historicized and materialized.
. This characteristic is also shared by computer models and simulations (see below),
which continually acquire new features and changes in appearance, and makes them
resemble theoretical entities.
. See de Chadarevian and Hopwood () for studies on the epistemic and didactic
importance of three-dimensional scientific models made out of diverse materials such as
wood, plastic, and wax. Klein is a beautiful study on the importance of the material
dimension of the different “paper tools” (such as diagrams, symbols, and formulas) that
scientists use. Leonelli () argues for the productivity of a multimodel approach that
considers the complementary use of material and theoretical models.
. This perspective is inspired by investigations of computer development (Woolgar
; Mackay et al. ).
. Our first case is based on the results of a long-term ethnographic study of particle
physics that one of us (M. Merz) conducted at CERN, the European Laboratory for Particle
Physics, in Geneva, Switzerland. The empirical material presented here draws on the more
detailed discussions in Merz (, , and ).
. Particle physicists label experiments, data, and so on “real” in contrast to their
“simulated” counterparts.
. This case is based on an empirical study by one of us (T. Knuuttila) conducted at
the Department of General Linguistics at the University of Helsinki and reported in several
articles in both Finnish and English (for example, Knuuttila ).
. Certain parsers are based on autonomous grammar theory, but their success in
parsing running text has been rather modest.
. See Collins and Evans () for an overview of this discussion.
GZ[ZgZcXZh
Bailer-Jones, D. M. . When scientific models represent. International Studies in the
Philosophy of Science :–.
tarja knuu t til a and m artina mer z
Klein, U. . Experiments, models, paper tools: Cultures of organic chemistry in the nine-
teenth century. Stanford: Stanford University Press.
Knorr-Cetina, K. . Sociality with objects: Social relations in postsocial knowledge
societies. Theory, Culture and Society :–.
———. . Objectual practice. In The practice turn in contemporary theory, edited by
T. R. Schatzki, K. Knorr-Cetina, and E. von Savigny, –. London: Routledge.
———. . Epistemic cultures. In Science, technology, and society: An encyclopedia, ed-
ited by S. Restivo, –. Oxford: Oxford University Press.
Knuuttila, T. . Models, representation, and mediation. Philosophy of Science
:–.
———. . From representation to production: Parsers and parsing in language technol-
ogy. In Simulation: Pragmatic constructions of reality, edited by J. Lenhard, G. Küp-
pers, and T. Shinn, –. Dordrecht: Springer.
Knuuttila, T., and A. Voutilainen. . A parser as an epistemic artefact: A material view
on models. Philosophy of Science :–.
Kuorikoski, J. . Simulation and the sense of understanding. Paper presented at Models
and Simulations , Charlottesville, VA, March –, . Available at http://philsci_
archive.pitt.edu/archive/.
Lenhard, J. . Surprised by a nanowire: Simulation, control, and understanding. Phi-
losophy of Science :–.
Leonelli, S. . What is in a model? Using theoretical and material models to develop
intelligible theories. In Modeling biology: Structures, behavior, evolution, edited by
M. D. Laubichler and G. B. Muller, –. Cambridge, MA: Massachusetts Institute of
Technology Press.
Mackay, H., C. Carne, P. Beynon-Davies, and D. Tudhope. . Reconfiguring the user:
Using rapid application development. Social Studies of Science :–.
Mäki, U. . Models are experiments, experiments are models. Journal of Economic
Methodology :–.
Merz, M. . Multiplex and unfolding: Computer simulation in particle physics. Science
in Context :–.
———. . Kontrolle—Widerstand—Ermächtigung: Wie Simulationssoftware Physiker
konfiguriert. In Können Maschinen handeln? Soziologische Beiträge zum Verhältnis
von Mensch und Technik, edited by W. Rammert and I. Schulz-Schaeffer, –.
Frankfurt a.M.: Campus.
———. . Locating the dry lab on the lab map. In Simulation: Pragmatic constructions
of reality, edited by J. Lenhard, G. Küppers, and T. Shinn, –. Dordrecht: Springer.
Morgan, M. S. . Models, stories and the economic world. Journal of Economic Meth-
odology :–.
Morrison, M. . Models as representational structures. In Nancy Cartwright’s phi-
losophy of science, edited by S. Hartmann, L. Bovens, and C. Hoefer, –. London:
Routledge.
Morrison, M., and M. S. Morgan. . Models as mediating instruments. In Models as
mediators: Perspectives on natural and social science, edited by M. S. Morgan and M.
Morrison, –. Cambridge: Cambridge University Press.
tarja knuu t til a and m artina mer z
Myers, N. . Animating mechanism: Animations and propagation of affect in the lively
arts of protein modelling. Science Studies :–.
Nagel, E. . The structure of science. New York: Harcourt, Brace, and World.
Rheinberger, H.-J. . Experiment, difference, and writing, part , Tracing protein syn-
thesis. Studies in History and Philosophy of Science :–.
———. . Toward a history of epistemic things: Synthesizing proteins in the test tube.
Stanford: Stanford University Press.
Star, S. L., and J. R. Griesemer. . Institutional ecology, “translations” and boundary
objects: Amateurs and professionals in Berkeley’s Museum of Vertebrate Zoology,
–. Social Studies of Science :–.
Suárez, M. . Scientific representation: Against similarity and isomorphism. Interna-
tional Studies in the Philosophy of Science :–.
———. . An inferential conception of scientific representation. Philosophy of Science
:–.
Teller, P. . Twilight of the perfect model model. Erkenntnis :–.
van Fraassen, B. C. . The scientific image. Oxford: Clarendon Press.
Voutilainen, A., J. Heikkilä, and A. Anttila. . Constraint grammar of English: A perfor-
mance-oriented introduction. Department of General Linguistics, publication no. .
Helsinki: University of Helsinki.
Woodward, J. . Making things happen: A theory of causal explanation. New York:
Oxford University Press.
Woolgar, S. . Configuring the user: The case of usability trials. In A sociology of mon-
sters: Essays on power, technology and domination, edited by J. Law, –. London:
Routledge.
Zhang, J. . The nature of external representations in problem solving. Cognitive Sci-
ence :–.
the gre at deluge
0
The Great Deluge
Simulation Modeling and Scientific Understanding
?D=6CC:HA:C=6G9
jo hanne s lenhar d
It is not the mere printout, the snapshot that my friend could have brought
with her, that delivers the information. The benefits of an ultrasound test are
gained through a complicated interaction: namely, the physician’s hand with its
changing position when guiding the probe and the accompanying changes on
the screen. It is only through a temporal sequence of controlled movement and
observation that the specialist can learn which imaged effects remain invariant,
which are perhaps only blurred shadings, and so forth. The full diagnostic po-
tential of the ultrasound scan can only be exploited through multidimensional
movement and through observing the relation between movement and image
over a certain period of time. The procedure draws on an essentially dynamic
aspect; hence, it is basically not a single representation, but a relational tableau
with which the physician gradually becomes familiarized through the tempo-
rally extended interaction of body movement, instrumental evaluation, and its
visualization—an orientation is gained through a sequence of images that the
physician can influence through his or her own movement. Dynamics, itera-
tion, and visualization are essential aspects of the instrumental access through
the ultrasound scan that could never be conveyed, even in principle, through
a single image.
This chapter considers which kind of understanding (a certain class of typi-
cal) computer simulations can provide, and how they do so. Discussing these is-
sues will reveal major parallels to the ultrasound scan scenario sketched above
because dynamics, feedback with visualization, and the orientation that can be
attained in a model characterize the understanding that researchers can obtain
in computer simulation as well.1
Models come in different types. Basically, we can distinguish between theo-
retical models (in mathematical language), simulation or computational mod-
els (for example, discretized versions of a theoretical model), and algorithmical
implementation (sometimes also called the computational model). Simulation
modeling involves going back and forth between all of them. The modeling
process, however, can be described with more levels of modeling (Winsberg
), with emphasis on experimental aspects (Dowling ; Winsberg ;
Lenhard ), or by analyzing whether theoretical models may be dropped
entirely in certain circumstances (Keller ). The details do not matter here
(see Lenhard et al. for a sample of the modeling discussion specializing in
simulation).
Simulations weaken the link between theory and understanding. In this
book, Tarja Knuuttila and Martina Merz’s “Understanding by Models” points
out the affinities between studies of modeling and of understanding. However,
if the investigation of scientific understanding should be located in the context
of modeling, what distinguishes simulation models from models in general, if
such characteristics exist at all? Knuuttila and Merz claim that the “workable
the gre at deluge
nature” of models determines their epistemic value, and part of this chapter
will spell out what this nature can be in simulation.
Simulation is leading to a pragmatic mode of scientific understanding. The
“subjective” concept of understanding is adapting to a new instrument—com-
puter simulation. Prediction and control on the one side and understanding
and insight on the other have formed the dual goal of science more or less since
its beginnings in modernity. Understanding is traditionally conceived as a mat-
ter of course in a theoretical sense. It is a deeply rooted point of view that those
who understand a phenomenon theoretically also have the appropriate means
to control it. Although one might well assign Platonic roots to this conception,
it had certainly become effective during the course of the seventeenth century
after Galileo and Newton had built the new science on the basis of mathema-
tization. Admittedly, work in the philosophy and history of technology reveals
that building and controlling devices is not entirely dependent on a foregoing
theoretical understanding, but this may well develop from engineering prac-
tice. I do not intend, however, to delve into that discussion. Instead, I want to
investigate an area that is a strong, arguably the strongest, case in favor of the
theory-based view of understanding and show that even in this area simula-
tions weaken the link between theory and understanding. Hence, this chapter
will focus on the mathematical instruments of science.
Although the computer in general and simulations in particular draw on
mathematics directly, simulation models usually reveal one particular feature
that threatens to shatter the traditional approach, namely, simulation models,
which are epistemically opaque (particularly in complex cases), so it is no lon-
ger possible to grasp in a real sense the ways in which the mathematical com-
ponents and algorithms interact. Instead, researchers have to test and study
the models in simulation experiments. Even when some parts of such models
have to be treated as a black box, simulation experiments can still render the
results of, for example, a whole field of parameters both visible and accessible
to experience. Researchers can establish a degree of orientation within a model,
a familiarity with how the model behaves.
This is the point at which the approach of the physician who is performing
an ultrasound scan meets up with that of a numerical optimizer who is using
simulation processes to explore the quality of a certain procedure, with the
methods used by a climate researcher who is varying parameters to try to find
out something about the stability of the Gulf Stream, or with the approach of
a researcher in nanotechnology who is looking for ways to control nanoscale
friction. In all these cases, an action-oriented understanding is constructed
that makes it possible to exercise control and make predictions, even when the
dynamics have not been grasped in theoretical terms. Hence, one can justifi-
ably talk about a pragmatic concept of understanding.
jo hanne s lenhar d
Yet, at the same time, it is questionable whether it is right to use the term
“understanding” in this context. The traditional concept of understanding is
linked to theoretical insight. Even if simulations do not make any understand-
ing possible at all in this sense, one should not interpret this as a weakness of
an instrument that, so to speak, waters down the possibilities of mathematics.
One could just as well see it as a strength: the instrument allows one to do with-
out a complete understanding of the theory—control and technology become
possible in other ways as well.
My argument starts with a closer consideration of the call by de Regt and
Dieks () to place scientific understanding on the philosophical agenda. It
will analyze the work of Gunter Dueck at IBM on the so-called Great Deluge
Algorithm. This case study on how simulation is used to deal with complex
problems in a thoroughly technological context of application will reveal that
some characteristics of simulation-based understanding do not fit easily into
the framework of theory-based understanding. Some scientists argue that the
latter understanding should be replaced by a model-based one. Two options for
balancing the perceived relation between simulation and scientific understand-
ing exist: either stick to the traditional “insight” concept of understanding and
take simulation as a new instrument that can provide control and options for
intervention without understanding, or enlarge (weaken) the concept of un-
derstanding and admit a pragmatic simulation-based mode of understanding
that works partly without the intelligibility of theory, but fits to an engineering
context of application.
JcYZghiVcY^c\VcY6eea^XVi^dc
this way. De Regt and Dieks argue convincingly that scientific understanding
is neither a simple derivative of explanation nor restricted to the realm of in-
dividual psychology. I agree with their account, especially when they stress the
pragmatic and context-dependent nature of understanding. Moreover, they are
well aware of the difficulties confronting their project of a general criterion of
understanding: “The history of science shows great variation in what is and
what is not deemed understandable. Even at one particular moment in history
opinions about what is understandable often diverge” (, ).
Hence, every philosophical standpoint on scientific understanding has to
find the appropriate level of analysis that does not dissipate into individualistic
criteria of single researchers nor get trapped by positions that are too general
and unable to take account of the observed variability in understanding. They
formulate a Criterion for Understanding Phenomena (CUP) that makes under-
standing dependent on the existence of an intelligible theory:
If I interpret de Regt and Dieks correctly, they are making a twofold assertion
here: First, understanding is the ability to recognize qualitatively characteristic
consequences. Second, this will be possible via intelligible theories. In my view,
their account is preoccupied with theory and underrates the role of models and
modeling.
Computer simulation is a new instrument of mathematical modeling that
can provide researchers with the ability to recognize and to interact with quali-
tatively characteristic consequences of modeling assumptions even though the
model dynamic remains partly opaque. I acknowledge the ability account of de
Regt and Dieks but do not see theory as a necessary ingredient. Furthermore,
simply replacing “theory” with “model” misses one crucial point: the ability to
draw consequences does not depend entirely on the intelligibility of the dynam-
ics (neither of model nor theory). The contribution of P. Ylikoski (this volume)
also analyzes understanding fundamentally as an ability. In my view, an attempt
to define understanding will have to take a closer look at how the mentioned
ability is brought about. This amounts to a kind of implicit definition—defining
jo hanne s lenhar d
H^bjaVi^dcVhi]Z<gZVi9Zaj\Z
Since the invention of the electronic computer in the middle of the twen-
tieth century, calculating speed has increased enormously. Work that, at one
time, required the most powerful computers in major research institutes is
nothing out of the ordinary for machines that can be purchased throughout
the world. However, mere calculating speed does not necessarily correspond to
a capacity to solve problems. Alan Turing, who worked out a machine on pa-
per that corresponds essentially to the modern computer, already offered fun-
damental insights into computability and its limitations. The vast majority of
real numbers, for example, cannot be computed. The question regarding which
types of problem cannot be solved with the calculating speed of computers (of
the Turing machine type) led to the explication of Kolmogorov-Chaitin com-
plexity (see Chaitin ). One also uses the term “computational complexity”
to describe the sort of complexity that measures how well a computer can cope
with processing a problem.
What makes this complexity so interesting is that a great number of prob-
lems emerge in scientific and technical practice that are complex in exactly this
sense. I should like to take the well-known “traveling salesman problem” (TSP)
as a prototype. A number of cities are given with the appropriate distances be-
tween them, and the problem is to find a route that links together all the cities
while simultaneously also being as short as possible. For each such problem
with n cities, there exists a solution—so much is obvious. Several routes may
well provide the shortest traveling distance. However, how can one find such an
optimal route? This question formulates a typical optimization problem that is
in no way simple to solve even with the enormous computing power available
today. Why is this so? Eventually, it would not be all that difficult to work out
an algorithm that would try out all the different routes, and (at least) one would
have to be the shortest. This approach is correct in principle, but not feasible
in practice.
The problem is that even a moderately large n will result in so many po-
tential routes that working through them all would overwhelm even the most
the gre at deluge
but does not go uphill continuously; every new step sets off in a randomly cho-
sen direction. The hiker moves according to a random process that takes no
account of the form of the mountains. Parallel to this, there is a second process
that gave the algorithm its name, namely, a great deluge that is causing the wa-
ter level to rise continuously. Only one restriction is imposed on the random
hiker: not to get wet feet. Whenever the random generator proposes a step
that would lead to a point below the water level, another, new step has to be
simulated. Evidently, this simulated hike ends at a local maximum as soon as all
neighboring points lie underwater. Because the hike can also lead downhill, the
starting point does not ascertain the final outcome in advance.
An obvious objection is that the simulated hiker is bound to become ma-
rooned on one of the islands that will inevitably become isolated at some point
by the deluge. In other words, is there any argument to explain why the local
maximum attained should be equal to or almost equal to the global maximum?
Nonetheless, practice—for example, finding good placements for components
of computer chips—shows that the algorithm delivers surprisingly good re-
sults, something that its inventor also notes:
How can one be confident that the outcome actually is “excellent”? Being
able to improve a solution that already works in industry would indicate that
the algorithm has done a good job. The most comfortable judgment situation is
when different approaches can be compared. Consider the example of locating
holes with a driller in a circuit board. The driller head has to be placed succes-
sively at all locations of holes; hence, it proceeds like a traveling salesman who
starts and finishes in a fixed town (the null position of the driller head). Finding
a good solution for this TSP is of obvious economic benefit, because it would
raise the number of items produced per time unit. Figure . shows the circuit
board and locations of all the holes to drill (towns to visit). Figure . shows the
starting point of the simulated route, that is, a random path connecting all the
holes (towns).
the gre at deluge
;^\jgZ.#^gXj^iWdVgYl^i]]daZh# ;^\jgZ.#'#6gVcYdbeVi]XdccZXi^c\
Cjaaedh^i^dcd[Yg^aaZg^ci]ZadlZgaZ[i" Vaa]daZh#9jZX`'%%)#GZegdYjXZYWn
]VcYXdgcZg#9jZX`'%%)#GZegdYjXZYWn eZgb^hh^dc[gdbHeg^c\Zg
eZgb^hh^dc[gdbHeg^c\Zg
Figure . shows the algorithm at an intermediate stage. Figure . displays the
end result of the Great Deluge Algorithm. It took about , steps, or a few
seconds, and produced a path with a length of about . inches. The optimal
result is known in this special case due to the theoretical efforts of Holland
(). It is . inches.
;^\jgZ.#(#>ciZgbZY^VgnhiViZ!XjggZci ;^\jgZ.#)#:cYgZhjaid[dcZgjcd[
aZc\i]d[eVi]^cY^XViZYVWdkZ#9jZX` i]ZVa\dg^i]b#9jZX`'%%)#GZegdYjXZY
'%%)#GZegdYjXZYWneZgb^hh^dc[gdb WneZgb^hh^dc[gdbHeg^c\Zg#
Heg^c\Zg#
jo hanne s lenhar d
JcYZghiVcY^c\BdYZahVcYH^bjaVi^dch
This point is crucial: the simulation does not have to start with a clear idea about
the gre at deluge
the structure of the solution space; on the contrary, this space is explored via
a simulated process. Hence, the simulation acts like a dynamic tentacle in our
example: a restricted random hike through unknown space. In this way, simula-
tions are “extending ourselves”—in line with the title of Paul Humphreys’s book
on simulation ()—in a very pragmatic way.
The particular case under consideration should not lead us astray regard-
ing the relevance of theory. Quite often, simulations have an elaborated theo-
retical basis. Consider nanoscience, which is devoted to the investigation of
phenomena brought about by many, but not too many, molecules. In principle,
the Schrödinger equation (quantum theory) governs their interactions; in prac-
tice, however, it is impossible to derive behavior theoretically for reasons of
complexity. Problems with complexity are endemic in the simulation context:
Knuuttila and Merz (this volume) discuss examples of parsing (language tech-
nology) and high-energy physics that would fit well here.
Lenhard () analyzes some phenomena of lubrication on the nanoscale
as observed in molecular dynamics simulations by Landman (). Landman’s
simulations make scientifically accessible—and render visible—phenomena of
increased friction in nano-sized dimensions that nanotechnology will have to
deal with. Feynman () already pointed to the plausibility of encountering
new phenomena of friction at the nanoscale. His theoretical ideas on volume
and surface, however, could not determine what these phenomena would look
like in concrete terms. Landman’s simulations do not just visualize the behavior
of molecules in thin layers; he has also experimented with small oscillations to
restore the desired and usual properties of lubricants. The simulation at least
indicates that this might be a promising approach in technology.
Comparable to the optimization case, the governing equations tell us noth-
ing about a complex situation. It is that kind of orientation gained in an ex-
perimental mode and assisted by visualization that indicated how increased
friction could perhaps be overcome. To avoid any misunderstanding: no strong
claim is made here about the representational accuracy of simulations. I agree
that validity is one of the main problems in the epistemology of simulation.
However, I hold that simulations can deliver an adequate dynamics, although
the model structure does not represent the mechanisms at work in the actual
phenomenon. Whether validity of performance is dependent on representa-
tional accuracy is a controversial issue: Winsberg (), for instance, argues
for “reliability without truth,” whereas Morgan () favors an ontological ar-
gument for the primacy of experiments over modeling approaches in general,
just because of their more direct representational force. The case of simulation
is particularly interesting, because artificial components, that is, deliberate de-
viations of the simulation model from the theoretical model, can contribute
positively to the model performance (compare Lenhard ; Winsberg ).
jo hanne s lenhar d
The view expressed by Feynman and many others was: “Understand the
equations, because when calculations begin, you will be lost in complexity and
complication.” The stance extracted from the simulation cases can be summa-
rized as: “Simulate and observe the outcome; repeat that procedure to explore
the models in order to eventually gain some understanding.” Hence, simulation
presents an instrument to deal with overwhelming calculations, even making
them invisible while, at the same time, the model dynamics tend to become
black-boxed—a diagnosis that has been termed “epistemic opacity” (Hum-
phreys ; compare also Lenhard ).
The long-standing epistemological promise of theory and theoretical
models is that they can be known and understood perfectly because they are
completely constructed. It has been argued that this promise is unfounded in
complex modeling situations that, however, are unavoidable in contexts of ap-
plication. The argumentation has brought up two aspects: first, the question
about the appropriate framework regarding how theory and models are related
to understanding in general; and, second, the problem of the relationship be-
tween understanding and simulation in particular. Admittedly, theory often is
the starting point, or even a main source, in the development of simulations.
Nevertheless, the crucial concepts are models and modeling, predominantly
because predictions are not derived from theory, as the previous discussion of
simulation has shown.
Hence, the general lesson is to frame the debate on scientific understand-
ing with a discussion of models as initiated by Nancy Cartwright (), with a
clear tendency against usability of theory, and exemplified by the more moder-
ate stance of Morgan and Morrison () and others. There should be con-
ceptual room for a distinctly model-based mode of scientific understanding.
This mode fits well with the aim of de Regt and Dieks () of developing a
context-dependent and pragmatic concept of scientific understanding and also
with related accounts in this volume (Knuuttila and Merz, Ylikoski).
What does the more specific lesson about simulation and understanding
consist of? Recall why theory was claimed to be essential. De Regt and Dieks
argued by opening up an alternative: a mere list of facts or predictions on the
one side and, on the other, a general theory from which one can derive pre-
dictions. Set aside the argument that the latter may be questionable. The list
would appear like an oracle that produces a correct prediction, but for reasons
one cannot understand: “In contrast to an oracle, a scientific theory should be
intelligible: we want to be able to grasp how the predictions are generated, and
to develop a feeling for the consequences the theory has in concrete situations”
(de Regt and Dieks , ).
Consequently, scientific understanding is introduced as the element that
grants usability. The investigation of simulations in this chapter has revealed,
jo hanne s lenhar d
6EgV\bVi^X6XXdjcid[JcYZghiVcY^c\
I have argued above that simulation fulfills this criterion. Hence, it achieves
what is commonly assumed to be achievable by understanding. In short, ac-
cording to (*), simulations provide understanding. However, a certain ambigu-
ity remains: the Great Deluge example shows how simulation can provide a
feeling for the consequences, even though Dueck and his colleagues recognized
a clear deficit in their understanding. Two options can resolve this ambiguity
by discerning a strong and a weak conception of understanding:
(i) Stick to the traditional “insight” concept of understanding; that is,
take (*) as a necessary but not sufficient criterion, and take simulation as
a new instrument that makes it possible to satisfy the criterion without
involving scientific understanding.
(ii) Stick to (*) as being sufficient, and enlarge (or weaken) the concept
of understanding. That means, adopt (*) as an implicit definition of un-
derstanding, implying that if simulation models are suitable instruments
to satisfy the criterion, they provide a mode of understanding that works
partly without the intelligibility of theory, thereby making epistemic opac-
ity and understanding partly compatible.
To vote for option (i) would mean to drop the concept of scientific understand-
ing from the agenda of investigating simulation and, instead, to analyze how
simulation permits intervention and manipulation. This would perhaps find
favor with many philosophers of science who are suspicious of understanding
because of its subjective connotations.
A vote for option (ii) would imply a more intricate standpoint. Think of a
science-studies-inspired endeavor to analyze whether and how the formation
of a simulation community takes place in science. This formation would po-
tentially include the development of a shared instrument-based way of under-
standing the world and point in the direction of option (ii). Simulation would
then come close to a paradigm in the sense of Rouse’s () interpretation of
Kuhn: paradigms form a shared way of seeing the world. The simulation-based
way would then be linked to understanding the world in the sense of (*). The
point is that for principled reasons of complexity, one cannot expect more than
this type of insight for many fields of science and technology in the context of
application. Thus, the lucidity of theory-based explanations is partly replaced
by the quasi-empirical access of simulations, constituting a genuinely novel
mode of scientific understanding.
This mode is a new one in the context of mathematical modeling in the sci-
ences. However, it is a well-known one in other contexts. Thus, expressed more
precisely, simulations expand a certain mode of understanding to mathemat-
ics-based modeling. What does this new yet old mode consist of? It is a thor-
oughly pragmatic one—it employs experimental exploration, visualization, and
jo hanne s lenhar d
CdiZh
I would like to thank the audience of the conference Philosophical Perspectives on Sci-
entific Understanding held in Amsterdam in , and the editors of this volume for highly
valuable comments and suggestions.
. My argument is located in the growing debate about models in science. I view
simulations as a kind of model, and I will discuss simulations and simulation models
interchangeably in this chapter. My view on models has much in common with Mary
Morgan and Margaret Morrison’s account of “models as autonomous mediators” (). In
particular, Morgan and Morrison strengthen the (partially) independent role of models in
relation to theory.
GZ[ZgZcXZh
Boumans, M. . The difference between answering a “why” question and answering a
“how much” question. In Simulation: Pragmatic constructions of reality, edited by J.
Lenhard, G. Küppers, and T. Shinn, –. Dordrecht: Springer.
Brooks, R. . Intelligence without reason. MIT Artificial Intelligence Lab memorandum
.
Cartwright, N. . How the laws of physics lie. Oxford: Clarendon Press.
Chaitin, G. J. . The limits of mathematics: A course on information theory and the
limits of formal reasoning. Berlin: Springer.
Chang, H. . Inventing temperature: Measurement and scientific progress. New York:
Oxford University Press.
the gre at deluge
G8IK@@@
JcYZghiVcY^c\^cHX^Zci^ÒXEgVXi^XZh
pa rt title ver so
('
Understanding in Biology: The Impure
Nature of Biological Knowledge
H67>C6A:DC:AA>
sabina leonelli
I]Z>bejgZCVijgZd[7^dad\^XVa@cdlaZY\Z
than not, different types of theories can each have relative significance toward
understanding a given phenomenon (Beatty , s). As highlighted by
Longino () and Mitchell (), this pluralism in theoretical approaches
is more conducive to the development of biological knowledge than a uniform
landscape centered on a few unifying laws would be.
Notably, theoretical results are often expressed and used by researchers
in forms other than the propositional statements contained in an explana-
tion. Apart from symbolic representations (such as mathematical equations),
we find diagrams, photographs, maps, and various three-dimensional objects
playing important roles not only in the context of discovery, but also in the
expression and justification of biological knowledge. Biologists refer to these
representational tools as models of biological phenomena.4 The philosophical
position that most closely accounts for biologists’ eclectic views on models was
put forward in by Morgan and Morrison, according to which a model is
defined not by its physical features, but by its epistemic function as a “media-
tor” between theories and phenomena. Models represent phenomena (through
processes including idealization, abstraction, and analogy) in ways that allow
scientists to investigate them. Their representational value depends on their
material features as well as on the interests and beliefs characterizing the re-
search context in which they are used (Morgan and Morrison , ). Model-
based reasoning, broadly defined as the use of representational tools toward
gaining epistemic access to natural processes, is widely recognized as playing a
crucial role in the production of scientific knowledge.
:meaV^c^c\Wn>ciZgkZc^c\
How precisely do biologists reason through models? The answer to this
question comes from studies of research practice in various biological disci-
plines that emphasize that model-based reasoning involves the manipulation
of models; that is, the reasoning has the possibility to modify them in order to
fit different research contexts. Models are tailored to pursuing specific research
purposes, which can vary depending on changes in the features of the phenom-
ena to be investigated as well as changes in the material, social, and institutional
settings of the researchers involved.5 Good models are ones that can be used by
the researchers to fulfill their goals in ways deemed appropriate by the relevant
research communities. If a researcher is unable to use a model because he or
she lacks the relevant skills, it will not matter that the model could in principle
be useful to the investigation: it cannot be used, and therefore cannot prove
useful toward the research. Consider the case of research on model organisms.
The worm C. elegans, the fruit fly Drosophila melanogaster, the plant Arabi-
dopsis thaliana, and the mouse Mus musculus are protagonists of experimental
sabina leonelli
nents of matter look and behave is worth huge investments, no matter what
the instrumental value of that knowledge turns out to be. Research of this kind
has been contrasted with research explicitly aimed at intervening on natural
processes and shaping them at will. Cryptography was born out of the military
need to send messages without fear of being intercepted. As such, cryptogra-
phy can be portrayed as a paradigmatically applied science: it does not produce
explanations of natural phenomena, but rather provides the means for efficient
and secure communication.
When considering the practices and goals characterizing medicine, molec-
ular biology, ecology, or evolutionary biology, one finds no obvious distinctions
between research aimed at producing ways of intervening in the world and
research aimed at producing explanations of biological phenomena. These two
types of outcome are part of the same investigation processes and are deeply
intertwined with each other. Take biotechnology, which has traditionally been
referred to as an “applied” field because of its explicit focus on enhancing tech-
nical expertise and devising methods to apply theoretical knowledge to real
processes. A well-known instance of biotechnological research is the develop-
ment of genetically modified organisms (GMOs), which requires scientists to
“apply” knowledge coming from molecular biology to producing plants and an-
imals with characteristics that are socially and/or economically advantageous.
The classification of GMO research as applied does little justice, however, to
the contributions to theoretical knowledge by biologists and bioengineers en-
gaging in it. Similarly, the vast majority of supposedly “basic” molecular biolo-
gists explicitly aims to produce techniques through which organisms can be
genetically modified. This is because the possibility to intervene on an organ-
ism generates opportunities to explain some of its features (as in knock-out ex-
periments, where researchers take away specific chunks of the DNA sequence
of an organism so as to gain insight on which genes play a role in the develop-
ment of which phenotypic traits). At the same time, the ability to explain a
given phenomenon is tightly connected with the ability to intervene on it: the
production of GMOs is informed by available explanations of how the genome
is structured.
Let me emphasize again that I am not contending that biologists aim at
intervening and manipulating nature rather than at explaining it. I am stressing
that scientific efforts at explaining the biological world are necessarily inter-
twined with efforts to intervene in the world.8 At least in the case of biology,
explaining and intervening are two sides of the same coin, both in terms of the
types of knowledge employed in research (that is, theoretical and applied) and
in terms of the goals that research is expected to fulfill (such as truthful expla-
nations of natural processes and control over those same processes).
sabina leonelli
I]Z>bedgiVcXZd[:bWdY^ZY@cdlaZY\Z
Already in the s, the intuition that theories and explanations consti-
tute but a part of scientific knowledge was put forward by philosophers Gilbert
Ryle and Michael Polanyi.9 Both men presented systematic reflections on what
I shall refer to as embodied knowledge—“know-how” in Ryle’s words () and
“personal knowledge” in Polanyi’s (/). They suggested that scientific
knowledge never involves merely a theoretical content, that is, that which is
known (as expressed in propositional or symbolic terms through theories and
explanations). It also involves the ability to apply such theoretical content to
reality and to interpret it in order to intervene effectively in the world as self-
aware human agents.
Polanyi sees such ability as an “unspecifiable art,” which can only be learned
through practice or by imitation, but which is rarely self-conscious, as it needs
to be integrated with the performer’s thoughts and actions without swaying
attention from his or her main goals (/, –). This is why Polanyi
refers to this knowledge as “tacit,” that is, as informing scientists’ performance
(actions as well as reasoning) in ways that cannot be articulated in words. Po-
lanyi distinguishes between two ways in which scientists can be self-aware: fo-
cal awareness, which is tied to the intentional pursuit of “knowing that”; and
subsidiary awareness, which concerns the scientists’ perception of the move-
ments, interactions with objects, and other actions that are involved in their
pursuit. In this view, focal and subsidiary awareness are mutually exclusive; one
cannot have one without temporarily renouncing the other. Switching from
“knowing that” to “knowing how” and back is a matter of switching between
these two ways of being self-aware. In fact, his characterization of know-how as
subsidiary awareness relegates this type of knowledge to a secondary position
with respect to “knowing that” (/, ). The acquisition of theoretical
content constitutes, for Polanyi, the central focus of scientists’ attention. Tacit
knowledge makes it possible for scientists to acquire such theoretical content,
but it is not valuable in itself: it is a means to a (theoretical) end.
While drawing inspiration from Polanyi’s emphasis on tacit expertise, I do
not think that his hierarchical ordering of types of knowledge (“knowing how”
being subsidiary to “knowing that”) does justice to the intertwining of “knowing
that” and “knowing how” that characterizes biological research. Polanyi bases
his reflections on the study of physical chemistry—and indeed, it might be that
some chemists value theoretical knowledge more highly than the practices and
skills used to construct it. Within most of biology, however, the development of
skills, procedures, and tools appropriate to researching specific issues is valued
as highly as—or, sometimes, even more highly than—the achievement of theo-
ries or data. The evolution of tools and procedures (as in the case of modeling)
under standing in biol o gy
Ryle does not specifically apply his account to the case of scientific knowledge.
Nevertheless, his characterization of agents “trying to get things right” by it-
eratively improving their performance fits my characterization of biologists’
practices and interests. As I illustrated, biologists do learn to use a variety of
tools in order to pursue their intellectual interests. They need to keep improv-
ing their ability to handle models and instruments, so as to reach their goals.
In Morgan’s words,
JcYZghiVcY^c\^c7^dad\n
6BViiZgd[H`^aa[ja6\ZcXn
Understanding natural phenomena by means of theoretical knowledge
does require some embodied knowledge; and, vice versa, embodied knowl-
edge needs to be coupled with theory-informed interpretations to provide un-
derstanding. To capture this insight, I define understanding as the cognitive
achievement realizable by scientists through their ability to coordinate theoreti-
cal and embodied knowledge that apply to a specific phenomenon.
By the verb coordinate, I mean the strategies that a scientist can learn to
use in order to () select beliefs, thought processes, and experiences that are
relevant to the phenomenon in question, and () integrate these components
sabina leonelli
with the goal of applying them to the phenomenon.13 Modeling is not the only
strategy used for gathering scientific understanding (reflection, observation, or
taxonomic exercises are good examples of other such strategies). Still, I empha-
size modeling here insofar as it is a prominent activity in contemporary biol-
ogy as well as a good exemplification of the means by which a coordination of
knowledge that is relevant to a specific phenomenon can be achieved.
Take again the case of model organism research. Bioinformaticians have
been collecting data concerning the most prominent model organisms (includ-
ing Arabidopsis, C. elegans, and Drosophila) into large databases, which can
be accessed by biologists across the globe. The purpose of these databases is to
facilitate the retrieval of available knowledge about these organisms, so as to
allow biologists to integrate the insights that they gain from their own research
with the understanding acquired by others. Notably, these databases make
abundant use of digital models (pictorial representations of varying degree of
abstraction, depicting entities as disparate as cells, metabolism, and gene ex-
pression) and even simulations (of processes such as flowering).
To enhance biologists’ understanding of the information that is provided,
the curators of these databases do not consider it sufficient to provide a text
describing what is known about a particular biological process together with
the relevant experimental data. Curators found that biologists can better assess
the significance of those data and descriptions when they are directly related
to the components of the organism that they are taken to provide evidence
for—for instance, by means of images.14 These images, which some researchers
call “virtual organisms,” constitute models with which biologists can work. By
tinkering with the images, modifying parameters, and comparing how differ-
ent data sets fit the models, researchers coordinate their theoretical knowledge
(for example, of the chemical composition of mRNA, the mechanisms through
which mRNA travels across the cell, and the available data sets documenting
this phenomenon) with their embodied knowledge (such as what mRNA actu-
ally looks like under the microscope and how those data sets are produced),
thus obtaining some understanding of processes such as gene expression.
Even more remarkably, the majority of experimental biologists that use da-
tabases insist on being able to interact not only with virtual representations of
organisms, but also with actual specimens. This is because physical interac-
tion with samples of the phenomena in question provides different information
from interaction with digital models: particularly, the embodied knowledge ac-
quired through the manipulation of a plant or a fruit fly is not the same as the
embodied knowledge derived through interacting with images of their meta-
bolic cycle. Laboratory specimens offer countless possibilities for exploration
and surprising results.15 Most importantly, they offer the possibility to match
the researcher’s sensory experience of a phenomenon (and of the system in
under standing in biol o gy
which the phenomenon manifests itself ) with his or her hypotheses and back-
ground knowledge. This is another example of the process that I wish to cap-
ture through the notion of “coordination.”16
It will be clear by now that, in my view, understanding a claim or explanation
is a largely subjective matter. Understanding is not an attribute of knowledge
itself, which can be measured quantitatively (as in “how much do you under-
stand?”). Rather, it is a cognitive achievement that is acquired by individuals in
a variety of ways and that can therefore take different forms depending on the
instruments that are used to obtain it. In the case of biological understanding,
the quality of the understanding displayed by a researcher will depend on his
or her acquisition of appropriate background knowledge as well as expertise in
handling the instruments, models, and theories that make it possible to pro-
duce and apply any scientific explanation. Importantly, both types of expertise
can only be obtained through participation and training in one or more scien-
tific communities.
It is indeed important to keep in mind that scientific understanding is not
the result of mere individual introspection. Its features are shaped by the ne-
cessity of intersubjective communication. The understanding acquired by one
researcher will not be accepted as a contribution to science unless he or she
is able to communicate insight to his or her peers, so as to make it vulnerable
to public scrutiny and evaluation. Individual understanding becomes scientific
only when it is shared with others, thus contributing to the growth of scientific
knowledge and partaking in the rules, values, and goals characterizing scien-
tific research.
I do not wish to argue that individuals can disseminate their understand-
ing of a phenomenon in an unmediated, direct way—for instance, by talking to
other individuals. This position would be paradoxical vis-à-vis my definition of
understanding as the cognitive achievement of individuals, based at least in part
on knowledge that cannot be articulated. I propose that individuals possessing
a specific understanding of a phenomenon enhance other individuals’ chances
of acquiring it in an indirect manner: that is, by constructing tools (including
models, explanations, experimental set-ups, and materials) that might enable
other individuals to experience the same kind of understanding. This indirect
sharing of understanding characterizes both the acquisition and the dissemina-
tion of understanding in scientific communities. On the one hand, biologists
seeking an understanding of a phenomenon are required to learn as much as
possible from similar efforts by other scientists, so as to use the understanding
accumulated by others as a vantage point for starting their own research. On
the other hand, biologists who acquire new understanding of a phenomenon
are required to contribute to the body of theoretical and embodied knowledge
available on that phenomenon in ways that will help other scientists acquiring
sabina leonelli
the same insight. The social processes through which biological understand-
ing is acquired and disseminated have a strong impact on the features of such
understanding.17 A biologist’s training and professional life constitute experi-
ences that are highly efficient in enabling him or her to understand a series of
phenomena in a number of specific ways.
:e^hiZb^XH`^aahVcYGZhZVgX]8dbb^ibZcih
Prima facie, epistemic skills may be broadly defined as the abilities to carry
out a number of activities in order to increase one’s understanding of reality.
They may be partly innate, such as the skill of drawing an object (which depends
to some degree on the talent of the individual attempting such action), yet they
are most often acquired through the imitation of others and/or through expe-
rience, for instance by trial and error. Skillful actions are instrumental in the
sense of being necessarily targeted toward the achievement of a goal, which in
the case of epistemic skills is an improved understanding of (some aspects of )
reality. At the same time, the notion of skill concerns also the means and man-
ner by which action is undertaken. An action can be judged as skillful even if
it does not result in the accomplishment of its intended goal. This is because
the successful achievement of an aim involves more factors than just the inten-
tions and ability of the individual acting to that aim: adverse conditions in the
environment, bad timing, social context, interference with other individuals’
goals and actions—all these factors can influence the outcome of an action, no
matter how skillfully that action is carried out.
The assessment of an action as skillful depends as much on the manner in
which the action is undertaken as it depends on its effectiveness in achieving
the intended goal. Possessing an epistemic skill implies more than the ability to
perform an action: it requires performing that action well. This means that we
need criteria to determine what it means for an action to be well performed.
However, these regulative criteria are context-dependent and arguably impos-
sible to classify in an analytic fashion. They are dictated by factors as disparate
as the nature of the goal to be achieved; the interests of the person perform-
ing the action; the social as well as the material context in which the action is
carried out; and the tools available to carry out the action. For instance, what
counts as adequate actions in designing an experiment on rats depends on the
phenomenon to be explored (for example, crowding behavior, which requires
a space where rats are kept together, or intelligence, which involves setting up
tasks for individual rats to perform); the eventual hypothesis to be tested (if
researchers want to prove that “high population density induces pathologies,”
they will test rat behavior in increasingly smaller communal cages over a long
period of time); current regulations for ethical treatment of laboratory animals
(and consequently, whether rats are given sufficient food and space to survive);
under standing in biol o gy
and, last but not least, whether animal ecologists are likely to consider the re-
sults of the experiment as representative for the behavior of rats in the wild.
This extreme context-dependence makes it uninteresting to list regulatory
criteria without looking at specific practices. Skillful actions in biological re-
search have only two features in common: () they involve the exploitation of
tools in a way deemed appropriate to effectively pursue a proposed goal; and
() judgment on what constitutes an “appropriate” course of action depends on
standards upheld within the relevant social contexts. The tools to be exploited
range across models, theories, experimental instruments, features of the en-
vironment, as well as samples of phenomena themselves, while the relevant
social context is constituted by the community of scientific peers in charge of
examining the methods and results of any research program. These insights
are captured by the following definition of epistemic skill: the ability to act in
a way that is recognized by the relevant epistemic community as well suited to
understanding a given phenomenon.
I wish to distinguish three types of epistemic skills. Theoretical skills involve
mastering the use of concepts, theories, and abstract models (that is, being able
to manipulate various expressions of theoretical knowledge) toward the un-
derstanding of a phenomenon. These skills enable biologists to reason through
given categories and classification systems according to specific inferential
rules, while at the same time judging the validity of those categories with refer-
ence to alternative theoretical frameworks. A second type of skill encompasses
the performative skills enabling biologists to exploit material resources toward
the acquisition of biological understanding of a phenomenon. These skills can
only be acquired through direct interaction with the environment (including
laboratory equipment and specimens). For instance, a biologist can be told how
to cultivate a plant so that it develops as required by a specific experimental set-
up. The corresponding performative skills, however, are acquired only through
practice, that is, by trying over and over again to act in the desired manner, thus
gradually adapting movements and sense perception to the tools and materi-
als used in an experiment, as well as to the standards enforced by the research
community of interest. After having experienced the results of sowing seeds in
the wrong type of soil, with too little or too much humidity and under varying
lighting conditions, the biologist will hopefully have acquired the ability to as-
sess the health of a growing plant and thus to produce specimens that fit his or
her experimental needs.
The ability to conform to existing standards can also be seen as an example
of the third type of epistemic skill, that is, of the social skills denoting the ability
of researchers to behave and express their insights in ways that are recognized
by their peers and/or other participants in their social context. Social skills
such as patience, charisma, and self-assurance enable scientists to secure an
sabina leonelli
BVcnIneZhd[JcYZghiVcY^c\
which can guide their actions even when they have no idea of why or how the
processes referred to as “symptoms” take place.
The third kind of understanding is the integrated understanding deriving
from a balanced exercise and coordination of theoretical and embodied knowl-
edge. Neither theoretical nor embodied knowledge are given a privileged role
in the understanding of a phenomenon. Rather, researchers are interested both
in acquiring a theoretical interpretation of a phenomenon and in obtaining
tools and methods that will enable them to match such theoretical interpre-
tation to the actual features of the phenomenon. To this aim, the embodied
knowledge used to acquire this type of understanding includes both theoretical
and performative skills: researchers exercise their ability to reason as much as
their ability to observe and/or intervene on the phenomena under scrutiny.
Each type of understanding is suited to different research contexts, since
the quality of understanding acquired by biologists depends largely on their
research settings and goals. For instance, theoretical biologists need to con-
centrate all their skills and resources on acquiring theoretical interpretations of
phenomena. This means committing to using mathematical modeling, that is,
models with high explanatory power but little empirical content. Arguably, this
leads to gaining insights that might not have been obtained if biologists insisted
on making empirical sense of their findings throughout their research. It thus
seems entirely justified to regard theoretical understanding as the most useful
kind of understanding in theoretical biology. Another example is the field of
natural history, where researchers commit to accumulating detailed knowledge
about the morphology of as many species as possible. They thus acquire em-
bodied understanding of what organisms look like, where they can be found,
and how they should be spotted, kept, and eventually embalmed. It does not
matter to them that the skills and commitments of natural historians greatly
limit their theoretical understanding of the differences among organisms (and
the causes of such difference): for their research purposes, the best understand-
ing of organisms is embodied rather than theoretical.
Do theoretical, embodied, and integrated understanding then have the
same epistemic value in biology? I think not. From a normative viewpoint,
integrated understanding constitutes the most desirable form of understand-
ing biological phenomena, since it involves considering theoretical and em-
bodied knowledge as equally relevant to understanding a phenomenon. As I
have claimed above, biology is about the exploration as well as the analysis of
phenomena: researchers who understand a phenomenon in one of these two
ways should ideally strive to understand it also in the other, thus balancing the
amount of theoretical and embodied knowledge used to this aim and acquiring
as rich as possible an understanding. The theoretical understanding of evolving
systems acquired by theoretical biologists could actually be viewed as very par-
sabina leonelli
tial, given that these researchers often have no idea of how their results could
apply to actual phenomena. Reliance on mathematical models often implies
losing the (embodied) sense of how the results acquired through these models
and the parameters used therein could be matched to observations, measure-
ments, statistics, and other empirical data, since what is taken to define an or-
ganism or a population in the models differs substantially from the complex
features of actual organisms and populations. Similar problems plague the field
of natural history, which has lost much of its eighteenth-century attractiveness
precisely because it does not provide ways to make theoretical sense of the ob-
served differences among species. In both examples, researchers seeking either
theoretical or embodied understanding seem to be missing out on something
important. In Hacking’s (, ) words, what biologists should strive for are
“happy families, in which theory and experiment coming from different direc-
tions meet.”
Whether biologists can actually pursue integrated understanding in prac-
tice, given the disunity and extreme specialization characterizing contempo-
rary research, is a matter of dispute. Cutting-edge fields such as system biology,
bio-ontology, and evolutionary-developmental biology are attempting to build
tools that might indeed enable biologists to better coordinate their theoretical
and embodied knowledge of the natural world (such as databases and in silico
experiments). Philosophers interested in the practice of understanding will
hopefully be in a position to contribute to the success of these efforts.
CdiZh
. Dupré (), Rosenberg (), Mitchell (), and Winther (), among oth-
ers, propose careful analyses of these elements.
. For the mechanisms through which specific tools come to be regarded as indispens-
able to an epistemic culture, see Knorr-Cetina () and Galison ().
. Mitchell critically surveys different notions of biological theory formulated by
philosophers of biology.
. The collections of essays edited by de Chadarevian and Hopwood (), Laubichler
and Müller (), and Suárez () provide examples of the many types of models used
to gain knowledge about natural phenomena.
. Mattila () discusses the significance of tailoring models.
. On the advantages of model organism research, see also Ankeny () and Leonelli
(a, b).
. A similar point is made in Hacking’s () defense of the significance of manipulat-
ing entities in science, even if his notion of intervention, focusing on experimental interfer-
ence with phenomena, is narrower than the one I propose.
. Woodward’s () manipulationist account of causal explanation takes a view that
is more general than mine—in the sense of applying to causal explanation in all scienc-
under standing in biol o gy
GZ[ZgZcXZh
Anderson, M. L. . Embodied cognition: A field guide. Artificial Intelligence
:–.
Ankeny, R. A. . Wormy logic: Model organisms as case-based reasoning. In Science
sabina leonelli
Radder, H., ed. . The philosophy of scientific experimentation. Pittsburgh: University of
Pittsburgh Press.
———. . The world observed / the world conceived. Pittsburgh: University of Pittsburgh
Press.
Rheinberger, H.-J. . Toward a history of epistemic things: Synthesizing proteins in the
test tube. Stanford: Stanford University Press.
Rosenberg, A. . Instrumental biology, or the disunity of science. Chicago: University of
Chicago Press.
Ryle, G. . The concept of mind. Chicago: University of Chicago Press.
Suárez, E. . Satellite DNA: A case-study for the evolution of experimental techniques.
Studies in History and Philosophy of Biological and Biomedical Sciences :–.
———, ed. . Variedad sin limites: Las rapresentationes en la ciencia. Mexico City:
Universida Nacional Autónoma de México y Editorial Limusa.
Winther, R. G. . Parts and theories in compositional biology. Biology and Philosophy
:–.
Woodward, J. . Making things happen: A theory of causal explanation. New York:
Oxford University Press.
m a r c e l b oum a n s
((
Understanding in Economics
Gray-Box Models
B6G8:A7DJB6CH
under standing in economic s
IVWaZ&&#&#I]gZZIneZhd[FjZhi^dch^c:Xdcdb^Xh
The subsequent sections will show that gray-box models are considered to
be the most appropriate answers to how’s that–questions in economics. I will
show that a validated explanation of an economic phenomenon outside the lab-
oratory will lead unavoidably to comprehensive but incomprehensible models.
under standing in economic s
:meaVcVi^dc
The way in which the factors xi might influence y can be represented by the
following equation:
The delta’s, Δ, indicate a change in magnitude. The terms Fi indicate how much
y will change proportionally as a result of a change in magnitude of factor xi.
Suppose we are trying to discover an invariant generalization that could
be used to explain the behavior of y. In principle, there are an infinite number
of factors x, x, . . . , that could influence the behavior of y, but we hope that it
may be possible to establish a constant and relatively simple relation between y
and a relatively small number of explaining factors, x. In a laboratory, we would
artificially isolate a selected set of factors from the other influences. In other
words, we would take care that ceteris paribus (CP) conditions are imposed:
Δxn+ = Δxn+ = . . . = 0, so that a simpler relationship can be investigated:
Thus, the relationship y = F(x, …, xn) explains the actual observed values of y,
provided that the factual influence of all the unspecified factors together was
very small as compared with the factual influence of the specified factors x,
. . . , xn.
The problem, however, is that it is not possible to identify the reason for the
factual influence of a factor, say xn+, being negligible, that is, Fn+·Δxn+ ≈ 0. By
passive observations alone, we cannot distinguish whether its potential influ-
m a r c e l b oum a n s
ence is very small, Fn+ ≈ 0, or whether the factual variation of this factor over
the period under consideration was too small, that is, Δxn+ ≈ 0. We would like
only to get rid of factors whose influence was not observed because their po-
tential influence was negligible to start with. At the same time, we want to re-
tain potential factors whose influence was not observed because they varied so
little that their potential influence was veiled.
The variation of xn+ is determined by other relationships within the sys-
tem. In some cases, a virtually dormant factor may become active because of
changes in the economic structure elsewhere. However, deciding whether a
factor should be accounted for in the relationship under investigation should
not depend on such changes, but only on its potential influence. The relation-
ship should be autonomous with respect to structural changes elsewhere. Au-
tonomous relations are those relations that should be expected to have a great
degree of invariance with respect to various changes in the economic structure.
However, this kind of invariance should not be equated with the observable
degree of constancy or persistence of a relation.
Haavelmo’s design rules for white-box modeling in econometrics were
considered as an alternative to the experimental methods of science (Mor-
gan , ). However, although researchers at the Cowles Commission2
adopted Haavelmo’s “blueprint” for econometrics (Morgan , ), they
scrapped the term “autonomy” because they believed that the theoretical rela-
tionships they were trying to measure were obviously autonomous. The reason
the Cowles Commission assumed this was because Haavelmo had pointed out
the possibility that the empirically found relationships may be simpler than
theory would suggest. This could lead researchers to discard potential influ-
ences, which could be avoided by building models as comprehensive as possi-
ble, based on a priori theoretical specifications. To explain phenomena outside
the laboratory, white-box models cannot discard the phenomena’s background
conditions; they should not only provide causal-descriptive statements about
the phenomena to be explained but also statements about the environments in
which they appear.
So, the Cowles Commission’s solution to the problem of autonomy was to
build increasingly comprehensive white-box models by building in as many po-
tential influences as possible. In the s, Lawrence Klein was commissioned
to build Cowles Commission–type models of the United States. The program’s
aim was to build increasingly comprehensive models to improve their predict-
ability so that they could be used as reliable instruments for economic policy.
Klein maintained this view throughout his life:
One of the major efforts of the s in this respect was the Brookings
econometric model of the United States (Dusenberry et al. ). This white-
box model was a joint effort of many individuals, and at its peak it contained
nearly four hundred equations. Although much was learned from this exer-
cise, the model never achieved the success that was initially expected, and it
was laid to rest around . Such an enormous model, of which the builders
only survey the part they have contributed, and of which the overall dynamic
behavior can only be revealed by computer simulations, does not provide any
understanding of the economic system it represents. Such models can only an-
swer policy-related questions, which can be labeled as what-if–questions (see
also Lenhard in this volume).
AjXVhÉhEgd\gVbd[<ZcZgVa":fj^a^Wg^jb:Xdcdb^Xh
At the end of the s a new practice of simulations arose which has had a
lasting impact on how to assess models. Klein together with Goldberger ()
had arrived at a model of twenty-five difference equations with a correspond-
ing number of endogenous variables; it was nonlinear in character and some of
the difference equations were of the fifth order—at that time clearly the most
advanced macroeconometric model. The model had been applied to annual
projections of economic activity in the United States with some success, but its
dynamic properties were only analyzed using highly simplified assumptions.
The innovative element of the research carried out by Adelman and Adelman
() was that, rather than making simplifying assumptions, the complexity of
this white-box model was left intact and the equations were programmed for a
computer and simulated for one hundred annual periods.
Adelman and Adelman were interested in an explanation of the persistent
business fluctuations of the United States. They found that random shocks su-
perimposed on the extrapolated values of the exogenous quantities and ran-
dom shocks introduced into each nondefinitional model equation induced
cycles with three- to four-year periods and amplitudes that were “reasonably
realistic.” That the amplitudes and the periods of oscillations observed in this
model were “roughly the same as those which are found in practice” (Adelman
m a r c e l b oum a n s
and Adelman , ) was seen by the Adelmans as “merely a necessary con-
dition for an adequate simulation of the cyclical fluctuations of a real industrial
economy” (). The question now was whether the shocked model could pro-
duce business cycles in the “technical” sense: “If a business cycle analyst were
asked whether or not the results of a shocked Klein-Goldberger computation
could reasonably represent a United States-type economy, how would he re-
spond? To answer these questions we shall apply to the data techniques devel-
oped by the National Bureau of Economic Research (NBER) for the analysis of
business cycles” ().
A comparison between the characteristics of the cycles generated by the
shocked model and the business-cycle characteristics summarized in NBER3
publications was considered to be “quite a stringent test of the validity of the
model” by the Adelmans (, ). The striking result was that when random
shocks of a “realistic order of magnitude” were superimposed on the Klein-
Goldberger model equations, the characteristics of the resulting cyclical fluc-
tuations appeared to be similar to those observed in the U.S. economy. The
Adelmans concluded, “it is not unreasonable to suggest that the gross char-
acteristics of the interactions among the real variables described in the Klein-
Goldberger equations may represent good approximations to the behavioral
relationships in a practical economy” ().
To Lucas, the Adelmans’ achievement signaled a new standard for what it
means to “understand” business cycles: “One exhibits understanding of busi-
ness cycles by constructing a model in the most literal sense: a fully articulated
artificial economy which behaves through time so as to imitate closely the time
series behavior of actual economics” (Lucas , ). To see that the Adel-
mans’ test works as a stringent test, Lucas understood that the facts that are
to be reproduced should be a list of characteristics providing as much detail as
possible, to be found in the various NBER empirical studies.
However, Lucas paraphrased the Adelmans’ question above as follows: “The
Adelmans posed, in a precise way, the question of whether an observer armed
with the methods of [the NBER] could distinguish between a collection of eco-
nomic series generated artificially by a computer programmed to follow the
Klein-Goldberger equations and the analogous series generated by an actual
economy” (). This paraphrasing is important because the characteristics test
of the Adelmans is thereby reinterpreted as a Turing test. This test was original-
ly described by Turing () as an “imitation game” to investigate the question
“Can machines think?” Today, a Turing test is generally described as follows:
Reports based on the output of the quantitative model and on measurements
of the real system are presented to a team of experts. When they are not able to
distinguish between the model output and the system output, the model is said
to be valid (see, for example, van Daalen, Thissen, and Verbraeck ). As one
under standing in economic s
can see, the test is in principle the same as the Adelmans’ test: an observer (in-
terrogator or expert) has to decide whether a distinction can be made between
a computer output and output from the “real” world.
The enormous advantage of Turing’s approach to artificial intelligence is
that it freed scientists from building replicas of the human mind to achieve ma-
chine thinking that meets the standard of human intelligence. In the same way,
the Adelmans’ characteristics test freed macroeconometricians from having to
build white-box models: “detailed, quantitatively accurate replicas of the actual
economy” (Lucas , ). As mentioned above, economic white-box models
of macroeconomic systems are inevitably too complex and too comprehensive.
Turing testing legitimized Lucas to work with very simple (and therefore un-
realistic) models. Lucas’s approach was not to aim at models as “accurate de-
scriptive representations of reality”: “Insistence on the ‘realism’ of an economic
model subverts its potential usefulness in thinking about reality. Any model
that is well enough articulated to give clear answers to the questions we put to
it will necessarily be artificial, abstract, patently ‘unreal.’ ” (Lucas , ).
According to Lucas, the model assumptions need not be assertions about
the world:
the parameters may change with shifts in the policy regime (Lucas ). Lu-
cas’s essay is perhaps the most influential and most cited paper in mac-
roeconomics (Hoover ), and it contributed to the decline in popularity of
the Cowles Commission approach. The Lucas Critique was an implicit call for
a new research program that amounted to the aim and justification of develop-
ing black-box and gray-box models. This alternative to the Cowles Commission
program involved formulating and estimating macroeconometric models with
parameters that are invariant under policy variations and can thus be used to
evaluate alternative policies. The only parameters Lucas assumed to be invari-
ant under policy changes are the parameters describing “tastes and technology”
(Lucas , ). By “tastes” Lucas referred to microeconomic characteristics
like individual preferences, and by “technology” he referred to stable institu-
tional characteristics.
Lucas’s “general equilibrium program” was most successfully carried out
by Kydland and Prescott (). In a special symposium, Computation Experi-
ments in Macroeconomics, they explained their general equilibrium program
methodology. Their “experiment” was an implementation of Lucas’s program
that ran a simulation experiment on an artificial economy. According to Kydland
and Prescott (, –), any economic computational experiment involves
the following methodology: The purpose of a computational experiment is to
derive a quantitative answer to some well-posed question. Therefore, what is
needed is a general-equilibrium theory. A theory is not a set of assertions about
the actual economy; rather, following Lucas (), it is defined as an explicit set
of instructions for building a mechanical imitation system to answer a question.
A model can be judged only relative to some given question. The features of a
given model may be appropriate for some specific question (or specific class of
questions, like how’s that–, how much– or what if–questions), but not for oth-
ers. Generally, some economic questions have a known answer, and the model
should give an approximately correct answer to them if we are to have any con-
fidence in the answer given to the question with an unknown answer. Kydland
and Prescott called this way of model assessment “calibration.” Thus, data are
used to calibrate the model economy so that it mimics the world as closely as
possible along a limited but clearly specified number of dimensions.
Kydland and Prescott’s specific kind of assessment is similar to Lucas’s idea
of testing, although he did not call it “calibration.” I argued above that Lucas’s
idea of testing is similar to a Turing test. It should be noted that a Turing test
does not make sense if you raise questions to which you yourself do not know
the answers. To have confidence that a computer is intelligent, it should give
approximately correct answers to questions to which you know the (approxi-
mately correct) answer in advance. To test models as “useful imitations of real-
ity,” we should subject them to shocks to “which we are fairly certain how actual
under standing in economic s
<gVn"7dmBdYZah
In the general equilibrium literature (see, for example, the two standard survey
volumes, Cooley and Stokey and Lucas with Prescott ), two modules
are always part of the models: a Cobb-Douglas production function, which has
the form: F(kt, ht) = ktθht1-θ, where kt denotes capital and ht labor; and a house-
hold utility function defined by u(ct, –ht) = (–α)log ct + α log (–ht), where
ct denotes consumption. They are parameterized, that is, incorporate a list of
required stylized economic growth facts, and need only be slightly adapted to
meet the wishes of the customer.
In systems engineering, an important step in the design process is param-
eterization: “Parameterization consists of identifying the numerical quantities
that specify each element in the system and their permissible ranges. A sys-
tem specification consists of a set for the system parameters together with the
system configuration. Alternatively, we could view this as the specification of
a set of parts and the instructions to assemble and put these to use” (White
under standing in economic s
The fact, then, that many complex systems have a nearly decompos-
able, hierarchic structure is a major facilitating factor enabling us to
understand, to describe, and even to “see” such systems and their
parts. Or perhaps the proposition should be put the other way round.
If there are important systems in the world that are complex with-
out being hierarchic, they may to a considerable extent escape our
observation and our understanding. Analysis of their behavior would
involve such detailed knowledge and calculation of the interactions
of their elementary parts that it would be beyond our capacities of
memory or computation. (Simon , )
8dcXajh^dch
The economic system is a whole of which all the parts are connected
and react on each other. . . . It seems, therefore, as if, for a complete
and rigorous solution of the problems relative to some parts of the
economic system, it were indispensable to take the entire system into
consideration. But this would surpass the powers of mathematical
analysis and of our practical methods of calculation, even if the values
of all the constants could be assigned to them numerically. (Antoine
Augustin Cournot, quoted in Hoover , –)
The first revision is that economists use models, instead of theories, to un-
derstand economic phenomena.5 Theories are considered to act only as guides
to find the main causal factors or, more generally, as a set of instructions to
built models. The second revision is that for phenomena outside the labora-
tory, the methodological requirements imply that these models are modular
designed gray-box models (the logical requirements), which are validated by
structure-oriented behavior tests and behavior pattern tests (the empirical re-
quirements). Simon’s work on nearly decomposable systems shows that econo-
mists can recognize qualitatively characteristic consequences of these kinds
of models. These two revisions lead to the following two adjusted criteria for
economics:
. Criterion for Understanding Economic Phenomena (CUEP): An
economic phenomenon P can be understood if a model M of P exists
that is intelligible.
. Criterion for the Intelligibility of Models (CIM): A scientific
model M is intelligible for economists (in context C) if it is a gray-box
under standing in economic s
CdiZh
. I will assume here that for phenomena investigated in a laboratory, how’s that–ques-
tions can be answered by white-box models. In other words, for laboratory phenomena,
explanations can also provide understanding.
. The Cowles Commission for Research in Economics was set up in and funded
by Alfred Cowles specifically to undertake econometric research. The journal Economet-
rica, in which Haavelmo’s essay appeared, was run from the commission. The Cowles
Commission’s econometric approach, developed in the s and s, became the
standard approach as presented in econometric textbooks.
. Founded in , the National Bureau of Economic Research (NBER) is a research
organization focused on empirical investigations of how the U.S. economy works.
. Boumans () discusses several economic examples for which Simon applied his
framework of hierarchical systems.
. See also the quotation from Mankiw () at the beginning of this chapter, which
is from a standard textbook in macroeconomics used worldwide. A similar suggestion for
a model-based understanding is also proposed by Lenhard, and Knuuttila and Merz in this
volume.
GZ[ZgZcXZh
Adelman, I., and F. L. Adelman. . The dynamic properties of the Klein-Goldberger
model. Econometrica :–.
Ando, A. . On the contribution of Herbert A. Simon to economics. Scandinavian
Journal of Economics :–.
Barlas, Y. . Formal aspects of model validity and validation in system dynamics. Sys-
tem Dynamics Review :–.
Boumans, M. . A macroeconomic approach to complexity. In Simplicity, inference and
modelling, edited by A. Zellner, H. A. Keuzenkamp, and M. McAleer, –. Cam-
bridge: Cambridge University Press.
Cooley, T. F., ed. . Frontiers of business cycle research. Princeton: Princeton University
Press.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
m a r c e l b oum a n s
Dusenberry, J. S., G. Fromm, L. R. Klein, and E. Kuh, eds. . The Brookings Quarterly
econometric model of the United States. Chicago: Rand McNally.
Franklin, A. . Calibration. Perspectives on Science :–.
Haavelmo, T. . The probability approach in econometrics. Supplement to Economet-
rica .
Hoover, K. D. . The new classical macroeconomics. Oxford: Basil Blackwell.
———. . The problem of macroeconometrics. In Macroeconometrics, developments,
tensions, and prospects, edited by K. D. Hoover, –. Boston: Kluwer.
Klein, L. R. . My professional life philosophy. In Eminent economists: Their life philoso-
phies, edited by M. Szenberg, –. Cambridge: Cambridge University Press.
Klein, L. R., and A. S. Goldberger. . An econometric model of the United States,
–. Amsterdam: North-Holland.
Krohs, U., and W. Callebaut. . Data without models merging with models without
data. In Systems biology: Philosophical foundations, edited by F. C. Boogerd, F. J. Brug-
geman, J.-H. S. Hofmeyr, and H. V. Westerhoff, –. Amsterdam: Elsevier.
Kydland, F. E., and E. C. Prescott. . The computational experiment: An econometric
tool. Journal of Economic Perspectives :–.
———. . A response from Finn E. Kydland and Edward C. Prescott. Journal of Eco-
nomic Perspectives :–.
Lucas, R. E. . Econometric policy evaluation: A critique. In The Phillips Curve
and labor markets, edited by K. Brunner and A. H. Meltzer, –. Amsterdam:
North-Holland.
———. . Understanding business cycles. In Stabilization of the domestic and in-
ternational economy, edited by K. Brunner and A. H. Meltzer, –. Amsterdam:
North-Holland.
———. . Methods and problems in business cycle theory. Journal of Money, Credit,
and Banking :–.
Mankiw, N. G. . Macroeconomics. rd edition. New York: Worth.
Morgan, M. S. . The history of econometric ideas. Cambridge: Cambridge University
Press.
Morrison, M., and M. S. Morgan. . Models as mediating instruments. In Models as
mediators: Perspectives on natural and social science, edited by M. S. Morgan and M.
Morrison, –. Cambridge: Cambridge University Press.
Nagel, E. . The structure of science: Problems in the logic of scientific explanation. Lon-
don: Routledge and Kegan Paul.
Orcutt, G. H. . Simulation of economic systems. American Economic Review
:–.
Simon, H. A. . The architecture of complexity. Proceedings of the American Philosophi-
cal Society :–.
Stokey, N. L., and R. E. Lucas, with E. C. Prescott. . Recursive methods in economic
dynamics. Cambridge, MA: Harvard University Press.
Turing, A. M. . Computing machinery and intelligence. Mind :–.
van Daalen, C. E., W. A. H. Thissen, and A. Verbraeck. . Methods for the modeling
under standing in economic s
()
Understanding in Physics
Bottom-Up versus Top-Down
9:CC>H9>:@H
under standing in ph ysic s
tion was first explicitly introduced—this will, quite fittingly, help to make clear
what its role and status are.
Shortly after the First World War, in , Albert Einstein unexpectedly
rose to worldwide public fame. In the final version of his General Theory
of Relativity had appeared, securing his reputation in the academic world. Even
during the war years preparations had started in English university circles to
test one of the most significant predictions of this new theory, namely the bend-
ing of light by massive bodies. It is true that Newton’s theory of gravitation also
predicts such an effect, if light is conceived as a stream of particles attracted
by gravity, but the numerical value of the deflection predicted by Einstein was
significantly different. In the case of star light bent by the sun general relativity
yields a deflection about twice as big as the Newtonian value, which suggests
the possibility of a crucial experiment.
So it happened that in two sun eclipse expeditions were sent off from
England in order to measure the actual magnitude of the light deflection. The
outcomes were presented at a special joint meeting of the Royal Astronomi-
cal Society and the Royal Society of London on November of the same year,
where the majority of those present considered the results to favor Einstein’s
theory most, although there was no unanimity. The next day, November , ,
however, the London Times carried an extensive article about the meeting, with
the headline “Revolution in Science: New Theory of the Universe.” This marked
the beginning of Einstein’s role as a public hero and genius. The editors of the
London Times invited Einstein to write a popular piece about his new theory,
a request he gladly agreed to, and on November his now-famous article My
Theory appeared (Einstein /, ).
In addition to its role as one of the milestones in the public perception of
Einstein, this article has become known for the methodological consider-
ations with which Einstein prefaced his explanation of relativity theory. Here,
Einstein made a distinction between two ways of constructing theories and
giving explanations in physics, one “bottom-up” and one “top-down.” As he
put it,
Thus, constructive theories and explanations start from the basic con-
stituents and elementary processes that build up a phenomenon, like when
we explain the pressure exerted by a gas from the collisions of the gas mol-
ecules against the walls of the container. By contrast, a principle theory begins
with postulating some general principle, suggested by experience. Properties
of individual processes are subsequently deduced from the requirement that
these processes should behave in accordance with what the general principle
stipulates.
Einstein’s introduction of this methodological dichotomy evidently served
an immediate purpose: he was facing the important task of explaining his own
theory of relativity, which he indeed had presented and developed starting from
general principles. The latter is true both for the special theory of relativity,
published in , and the general theory, published in . In particular the
special theory, on which I shall focus here, closely follows the axiomatic-deduc-
tive model. It starts with two clearly stated principles: () the relativity principle
(the physical laws have the same form in all inertial frames of reference), and
() the light principle (the speed of light is independent of the velocity of the
emitting source) (Einstein /, –). As it turns out, these two simple
and general starting points suffice for the prediction of many concrete physical
phenomena, like time dilation and length contraction. As we shall shortly see, a
striking and important point is that these predictions are possible without go-
ing into any detail about how the clocks and rods exhibiting the dilations and
under standing in ph ysic s
contractions are built up from atoms and molecules, and without any specific
information about how these elementary constituents interact. Everything fol-
lows from very general considerations: the type of explanation is top-down.
In his London Times article, Einstein is not very outspoken about the
relation between the two types of physical explanations that he had distin-
guished: he cites advantages on both sides. If anything, he seems to favor the
constructive approach, given his statement that “when we say that we have suc-
ceeded in understanding a group of natural processes, we invariably mean that
a constructive theory has been found which covers the processes in question.”
But given the context, namely the purpose of making the theory of relativity
understood and acceptable to a general audience, and given that Einstein him-
self had developed his theory from general principles, it is only natural that the
constructive approach is mentioned in the beginning of the London Times
article but does not receive much further attention.
The theory of relativity soon acquired enormous fame and came to be con-
sidered as the paradigm of modern physical thinking, on a par with or even
surpassing Newton’s achievements. Given this course of events, it is under-
standable that both in circles of physicists and of philosophers of science the
idea has not infrequently taken root that at least as far as relativity theory is
concerned, top-down, principle explanations possess a privileged status. In-
deed, it is not difficult to find claims in the literature that bottom-up explana-
tions of typically relativistic effects like time dilation and length contraction are
inappropriate or even impossible (see below for a sample of such claims).
However, quite generally both types of account—bottom-up and top-
down—are viable. The typically relativistic dilations and contractions can
surely also be understood in a constructive, bottom-up fashion. This does not
deny that the top-down derivations that have become standard in the literature
possess explanatory value and lead one to understand why these effects must
exist according to relativity theory. Indeed, the choice between the different
explanatory strategies has a pragmatic character and depends on contextual
factors. There is no clear-cut and general difference between the two types of
explanation with regard to their power to generate understanding because the
notion of understanding is contextual in the same way explanation is.
As already mentioned, for concreteness I shall focus on the choice between
bottom-up and top-down explanations in special relativity. However, I think it
will become clear from the general line of argument that with respect to these
possible explanatory options there is no important relevant difference between
relativity theory and other physical theories. Moreover, bottom-up and top-
down explanations are just two examples from a wider gamut of possible expla-
nation forms (compare de Regt ). Quite generally, physics is pluralistic as
far as explanations and ways of obtaining understanding are concerned.
dennis diek s
HeZX^VaGZaVi^k^inVhVI]Zdgnd[Eg^cX^eaZ
In the very first sentence of his “On the Electrodynamics of Moving Bod-
ies,” the essay in which he first formulates the special theory of relativity,
Einstein (/, ; Stachel et al. , ) sets the tone for both his own
article and for most of the work on relativity that was to follow later. He draws
attention to the fact that the hitherto usual theoretical treatments of electrody-
namic phenomena contain distinctions and asymmetries that are absent from
these phenomena themselves. This rather abstract observation, about possible
“excess baggage” in the form of theories, then motivates Einstein’s first pos-
tulate, the principle of relativity. As still formulated in the introduction of his
essay, this principle tells us that the laws of electrodynamics and optics, like the
laws of mechanics, hold good in their same standard form in all inertial frames
of reference—a little bit later on in the article it becomes clear, however, that
this is taken to be valid for all laws, even for yet to be discovered nonelectrody-
namical and nonmechanical ones. To this postulate Einstein immediately adds
a second that stipulates that light is always propagated in empty space with a
definite velocity c, independent of the state of motion of the emitting body. As
Einstein promises us, still in the same introductory remarks, these two pos-
tulates will suffice for the attainment of a simple and consistent theory of the
electrodynamics of moving bodies.
Here we see the principle-theory approach in full swing: the starting point
is the formulation of very general requirements, suggested by empirical data.1
Einstein’s two starting points are very general in character, even to the point
that one wonders how they could lead to such concrete and specific predictions
as changes in the lengths of moving rods and retardations in the pace of mov-
ing clocks. But Einstein is able to honor his promise and indeed deduces these
effects from his axioms.
This remarkable feat is achieved by Einstein through his insistence that
lengths and times are not merely abstract concepts, but correspond to what is
indicated by concrete measuring instruments, rigid rods and clocks. At the end
of the introductory section of his essay, Einstein famously declared, “The
theory to be developed is based . . . on the kinematics of the rigid body, since the
assertions of any such theory have to do with the relationships between rigid
bodies (systems of co-ordinates), clocks, and electromagnetic processes.” Thus,
coordinates are identified with notches in rigid material axes, distance is what
is measured by rigid measuring rods, and time corresponds to what is indicated
by the hands of synchronized clocks. It certainly appears at first sight that a
strong operationalist flavor emanates from these statements, a theme that was
soon picked up by logical positivists and operationalists (see Reichenbach
and Bridgman )—although it is not so clear whether Einstein himself was
under standing in ph ysic s
I]ZE]nh^XVaGZVa^ind[GZaVi^k^hi^X:[[ZXihVcYi]ZHiVijhd[
GdYhVcY8adX`h
given rise to questions about the nature of the relativistic effects. Even if it is
clear that the existence of contractions and dilations can be logically derived
from the postulates, does this also mean that they represent physically real
phenomena, real material changes in the rods and clocks? It has repeatedly
been suggested that this is not the case: that the changes have to do more with
the theoretical description given to the rods and clocks, or with the conditions
of observation than with things happening in the bodies themselves; that in
this sense the effects are more apparent than real. Indeed, no connection with
bottom-up accounts in terms of molecules, and forces between them, is pro-
vided so that a “mechanical” understanding of the effects is lacking in the deri-
vation from the postulates. Some authors have been misled by this, even to
the extent of claiming that the relativistic effects are more psychological than
physical in character. Among early examples are the American physicists Gil-
bert Lewis and Richard Tolman, who wrote in that in Einstein’s theory
“the distortion of a moving body is not a physical change in the body itself,
but is a scientific fiction”; the changes in the units of space and time are “in a
certain sense psychological.” This they contrasted with older proposals by H.
A. Lorentz according to which there is “a real distortion of the body” caused
by deforming forces (Lewis and Tolman ). In the same vein, Laub ()
commented on Lorentz’s theory that in it, “observed changes in a moving body
are of an objective nature,” implying that in Einstein’s theory the situation was
different. Similarly, von Ignatowski () wrote, “measurements on a moving
body yield only an apparent value.” And the mathematician Vladimir Varičak
() commented that the relativistic “contraction is, so to speak, only a psy-
chological and not a physical fact.”
These allegations did not remain unchallenged. Already in Paul Eh-
renfest intervened with a thought experiment that showed how the relativistic
contractions can give rise to undoubtedly physical effects, like the explosion
of a solid body. Ehrenfest’s () article invites us to consider a solid material
cylinder that is set into rotating motion around its axis. As a result of the rota-
tion, strains will arise in the material: the elements of the cylinder will tend to
undergo relativistic contraction into the direction of their rotational motion
but will remain undeformed in the directions perpendicular to their rotational
velocity. For the cylinder as a whole this leads to conflicting tendencies that
may result in the destruction of the cylinder if the velocity of rotation becomes
great enough. There can therefore be no doubt that causal processes, having to
do with the forces between the atoms and molecules, are involved in the con-
tractions and dilations.
Also Einstein himself participated in the discussion with a reply to Varičak’s
paper (Einstein ). Opposing the idea that the Lorentz contraction is subjec-
tive or an artefact of conventions of measurement, he described a method for
under standing in ph ysic s
determining quite objectively that moving ideal rods are shorter, even without
the use of synchronized clocks. Moreover, in a letter of to the philosopher
Joseph Petzoldt, Einstein commented on Ehrenfests’s thought experiment: “It is
well to remark that a rigid circular disk at rest must break up if it is set into rota-
tion, on account of the Lorentz contraction of the tangential fibers and the non-
contraction of the radial ones. Similarly, a rigid disk in rotation (produced by
casting) must explode as a consequence of the inverse changes in length, if one
attempts to put it at rest” (Stachel , ). In a preserved draft of a letter from
, Einstein likewise pointed out that to obtain a rotating rigid disk without
relativistic strains and tensions, one would have to first melt a disk at rest, then
put the molten disk into rotation and finally solidify it while rotating ().
Clearly then, the principle-theory approach should not lure us into the mis-
taken belief that no bottom-up, causal stories about the relativistic effects are
possible. In fact, at several points in his career Einstein in retrospect gave vent
to related reservations about his original introduction of special relativity. An
early example is in his lecture “Geometry and Experience” (Einstein /,
), delivered not long after the just-mentioned discussions about the physical
reality of the contraction effects. In “Geometry and Experience,” Einstein states,
“The idea of the measuring rod and the idea of the clock in the theory of relativ-
ity do not find their exact correspondence in the real world. It is also clear that
the solid body and the clock do not in the conceptual edifice of physics play the
part of irreducible elements, but that of composite structures, which must not
play any independent part in theoretical physics” (/, ).
This is very relevant to our topic: as we have seen, Einstein’s original deri-
vation of the contraction and dilation effects proceeded through the identifi-
cation of distances and periods with the indications given by ideal rods and
clocks, without mentioning anything about the causal processes going on in
the interior of these devices. Nothing at all was said about their atomic or mo-
lecular constitution or about the forces that keep them together, and as we
have seen this could easily create the false impression that no ordinary causal
processes are involved at all in the contractions and dilations. But in the quoted
passage, Einstein emphasizes that rods and clocks are ordinary bodies with a
microscopic structure and therefore determined in their macroscopic features
by what occurs at the microscopic level.
In the same lecture Einstein continues: “It is my conviction that in
the present stage of development of theoretical physics these concepts [that is,
rods and clocks] must still be employed as independent concepts; for we are
still far from possessing such certain knowledge of the theoretical principles
of atomic structure as to be able to construct solid bodies and clocks theoreti-
cally from elementary concepts” (/, ). Note how the last part of
the quote resonates with Einstein’s description of the constructive method in
dennis diek s
his newspaper article. It is pretty clear that Einstein never thought that
general principles about rods and clocks should replace considerations about
atomic constitution and causal processes. The axiomatic approach was intro-
duced as an alternative that was particularly appropriate, given the context of
the situation Einstein was facing. This situation is that we are accustomed to
using rods and clocks for making space and time coordinates physically con-
crete, but are not able to directly describe these devices in terms of their atomic
and molecular constitution; this makes it expedient to introduce them as inde-
pendent concepts. Moreover, we are interested in questions of a very general
character, not pertaining to particular bodies or particular causal processes.
The axiomatic approach is fitting here: it is able to explain why the contractions
and dilations must be there in a quite general way, whatever the details of the
underlying causal processes.
Einstein later characterized the top-down approach as a practical decision,
made for the time being. As soon as a direct causal characterization via funda-
mental physical theory becomes available, Einstein tells us, this treatment will
have to replace the axioms-about-rods-and-clocks account. We find this state-
ment at various places in Einstein’s writings. Thus, at the end of his career he
writes in his autobiographical notes,
One is struck by the fact that the theory introduces two kinds of physi-
cal things, i.e., () measuring rods and clocks, () all other things, e.g.,
the electromagnetic field, the material point, etc. This, in a certain
sense, is inconsistent; strictly speaking measuring rods and clocks
would have to be represented as solutions of the basic equations (as
objects consisting of moving atomic configurations), not, as it were, as
theoretically self-sufficient entities. (Schilpp , )
He then immediately goes on to explain that the appeal to measuring rods and
clocks should be seen as a makeshift procedure for the time being, as long as
no complete fundamental bottom-up treatment of rods and clocks is available,
“with the obligation, however, of eliminating it at a later stage of the theory.”
In fact, Einstein seems overcautious here. He could have resorted to Her-
mann Minkowski’s work: in his famous address Minkowski already
showed how special relativity can be built up in terms of fundamental physical
laws alone, without any appeal to rods and clocks as independent entities.
B^c`dlh`^VcYi]ZEg^cX^eaZ6eegdVX]EZg[ZXiZY
from the introductory statement, “Henceforth space by itself and time by it-
self are doomed to fade away into mere shadows, and only a kind of union of
the two will preserve an independent reality,” has acquired proverbial status.
Still, there are aspects to Minkowski’s ideas that so far have not received the
attention they deserve (Dieks ). I here shall focus on how Minkowski
is able to avoid an appeal to rods and clocks in his presentation of special
relativity.
The impression created by Minkowski’s just-quoted winged words, namely
that space-time is an entity existing independently of matter, should be treat-
ed with caution. Reading Minkowski’s essay, rather, suggests that his position
was sympathetic to Leibnizean relationism, and that he was close to Einstein’s
sympathies—although both authors were not explicit about their philosophy
of space and time, writing as they did on problems in physics rather than phi-
losophy. How Minkowski viewed the ontological status of space-time becomes
clear if we look at his discussion of how we can lay down space-time coordi-
nates. Unlike Einstein, Minkowski did not engage in a discussion of rods and
clocks in order to define such coordinates; he suggested a very different proce-
dure. This alternative procedure has not been the subject of serious study in the
philosophy of physics literature—rather surprisingly, for Minkowski’s proposal
implements a way of constructing the edifice of special relativity in which only
general features of fundamental physical laws play a role, without the need of
invoking macroscopic measuring instruments. In a nutshell, Minkowski’s sug-
gestion is to do without any pre-given interpretation of coordinates in terms of
distances and clock settings, and to start with a theoretical account of elemen-
tary physical phenomena in terms of completely arbitrarily chosen variables.
The form of the regularities formulated in this way will most likely be extremely
complicated, but by going to new independent variables, that is, by performing
mathematical coordinate transformations, one may simplify. As the final result
of this process of successive simplifications, a preferred set of space-time co-
ordinates will be found as that “system of reference x, y, z, t, space and time, by
means of which these phenomena then present themselves in agreement with
definite laws” (Minkowski /, ).
As can be expected, the “definite laws” Minkowski explicitly refers to are
Maxwell’s equations of electrodynamics in their standard form. But later in his
article he assumes that all fundamental laws of nature, including those yet to be
discovered and responsible for the stability of matter, should exhibit the same
symmetry properties as the equations of electrodynamics. This boils down to
executing Einstein’s program of doing away with rods and clocks; in spite of
the absence of complete knowledge of all laws and in spite of a lack of de-
tailed insight into how macroscopic measuring devices should be described by
fundamental theory—as Minkowski’s elegant mathematical treatment shows,
dennis diek s
HeZX^VaGZaVi^k^inVhV8dchigjXi^kZI]Zdgn
Typically relativistic effects like length contraction and time dilation can
give rise to phenomena that one at least is inclined to describe in terms of mi-
croscopic causal processes. In the case of the rotating and eventually exploding
cylinder, for example, it is natural to ask about the forces that are responsible,
the amounts of energy that are involved, and so on. Likewise, one may won-
der about the detailed mechanisms that make rods shrink when a velocity is
imparted to them. Also in this latter case effects may arise like the breaking of
objects, which seems to cry out for causal explanation. A notorious example is
furnished by John Bell’s (, ) two rockets problem in which two rockets
undergo equal accelerations in such a way that their mutual distance remains
the same. A rope connecting the rockets will tend to Lorentz contract during
the process, because it acquires a velocity. As a consequence, the rope will be-
come tighter and may eventually even break when the velocity becomes high
enough.
As Bell recounts, he experienced that many physicists are amazed when
confronted with this outcome of the thought experiment. Bell attributes their
confusion to the top-down manner in which the theory of relativity is usu-
ally taught, and pleads for a different approach in which the focus is shifted
from top-down to bottom-up, constructive and causal accounts of relativistic
phenomena. I will comment on the relation between these two approaches in
the next section; here, I merely want to emphasize that such a constructive ap-
proach is indeed possible.
I mentioned above that Einstein already had warned against the idea that
macroscopic bodies like rods and clocks are not susceptible to bottom-up anal-
ysis: these bodies evidently consist of atoms and molecules, so their behavior
should be determined by the laws that govern these submicroscopic constitu-
dennis diek s
ents and their interactions. I mentioned in passing that before the advent of
relativity theory, H. A. Lorentz had already developed a theoretical scheme
where he tried to explain that moving bodies contract.3
Lorentz’s (/, ) approach focused on the forces that hold mac-
roscopic bodies, like measuring rods, together. These forces change when the
bodies are set in motion, and this has the effect of changing the mutual dis-
tances between the atoms and molecules. Lorentz proved that this leads to the
right macroscopic contraction in the direction of motion (the same contraction
we encounter in relativity theory) if it is assumed that under motion all possible
forces change in the same way as the electromagnetic ones. This latter assump-
tion anticipates Minkowski’s requirement that all laws should be invariant un-
der Lorentz transformations, but still, the type of explanation here is clearly
causal and bottom-up. The account works by determining the equilibrium
positions of the atoms and molecules in the interatomic and intermolecular
force fields, and by finding out how these equilibria are altered when the forces
change due to motion.
Exactly the same explanations can be given in relativity theory. If we know
what the fundamental building blocks of matter are, and know how the forces
between them vary when a system is set into motion, it becomes possible in
principle to calculate how the macroscopic features of these bodies will change
when they start moving. Actually, it seems pretty obvious that bottom-up ex-
planations of this type are possible for everything that happens according to
relativity theory.
I]ZGZaVi^dcWZilZZc7diidb"JeVcYIde"9dlc
Minkowski’s approach makes it easy to confront and compare our two ex-
planatory strategies. In Minkowski’s presentation of special relativity the basic
ingredients are matter and fundamental laws governing matter, and although
Minkowski’s own approach is top-down, these ingredients also set the stage for
a Lorentz-like bottom-up account in terms of atoms, molecules, and the forces
between them. Minkowski’s scheme further contains the general principle that
all laws are invariant under Lorentz transformations. The way Minkowski in-
troduces his scheme—starting from concrete phenomena, investigating their
regularities and thus arriving at laws, and finally inquiring about the invari-
ance properties of these laws—might create the impression that this principle
cannot be an independent component of the theory: once the laws have been
found, whether or not they are Lorentz invariant is a matter of inspection and
does not need independent stipulation. However, the idea obviously is that it is
part of relativity theory to suppose that new laws that may be found as a result
of future research will possess these same invariance properties. Interpreted
under standing in ph ysic s
(1–v²
c²)
between the unprimed coordinates (assuming that the velocity v is in the x di-
rection). So as judged from the first frame the length of the moving body is less
than : it has the Lorentz contracted value.
As will be clear from this sketch, the relativity principle does all the work
in securing that the right contraction and dilation factors will result. The mi-
dennis diek s
8dcXajh^dc
Since its introduction, the special theory of relativity has usually been re-
garded as a typical principle theory in which the direction of explanation is
top-down. In accordance with this, it is not uncommon to find statements in
the literature to the effect that attempts at bottom-up, causal explanations are
out of place in relativity. However, there have certainly been many physicists,
among them the old masters Einstein, Minkowski, Ehrenfest, and their like,
who never fell prey to this misunderstanding. The possibility of bottom-up
strategies in relativity, which I have stressed here, is therefore not something
dennis diek s
new, but it is still worth stressing in view of the continuing dominance of the
top-down approach both in the teaching of relativity and in philosophical ac-
counts of the theory. Bell () was right in resisting this dominant attitude by
drawing attention to the possibility of bottom-up accounts. From my point of
view, however, “he struck a spark but threw no light”: he overshot the mark by
suggesting that the bottom-up approach is the understanding-providing one.
The issue has been receiving growing attention in the philosophy of physics
literature, especially after the publication of Harvey Brown’s () book Physi-
cal Relativity, which defends a position very similar to that of Bell. In the wake
of Bell’s and Brown’s publications, a serious dispute has arisen among philoso-
phers of physics about the correct way of explaining relativistic effects. On the
one hand there are those who defend the orthodox viewpoint: these argue that
the top-down approach is appropriate since it alone fully captures and reflects
the general space-time structure of the world according to special relativity
(see, for example, Janssen , ; Balashov and Janssen ). On the other
hand, there are those who maintain that this top-down approach ignores the
physical mechanisms that are at work (for example, Brown and Pooley ,
; Brown ), and therefore is unable to provide us with genuine explana-
tions and understanding.
What I have argued here is that this dispute derives from a misconception.
There is no uniquely best way of explaining the relativistic effects. The differ-
ences between the explanations are differences in the use we make of one and
the same theoretical scheme. They relate to decisions about where to put the
emphasis and where to begin and end the analysis; these decisions in turn are
conditioned by what we are interested in. In other words, the difference is one
of pragmatics. Explanation and understanding are relative to questions we ask
and interests we have. Explanation and understanding, not only in relativity
theory but also in theoretical physics and even in theoretical science in general,
are inherently pluralistic.
CdiZh
. That no difference between inertial frames can be detected as far as the form of
applicable theoretical principles is concerned, and that the velocity of light is always the
same, had been suggested by several nineteenth-century experiments performed precisely
with the purpose of finding empirical counterparts to the theoretical asymmetries men-
tioned by Einstein in the first sentence of his article. It was the failure of these experiments
that provided positive empirical input for Einstein’s postulates.
. Many of the ideas can already be found in a lecture Minkowski gave a year earlier
(Minkowski ).
. Even before special relativity, the contraction idea had gained currency as a way of
accommodating the famous null experiments, such as the Michelson-Morley experiment.
under standing in ph ysic s
. See de Regt and Dieks () for more on the relevance for understanding of the
intuitive accessibility of such theoretical pictures.
GZ[ZgZcXZh
Balashov, Y., and M. Janssen. . Presentism and relativity. British Journal for the Phi-
losophy of Science :–.
Bell, J. S. . How to teach special relativity. Speakable and unspeakable in quantum
mechanics, –. Cambridge: Cambridge University Press.
Bridgman, P. W. . Einstein’s theories and the operational point of view. In Albert Ein-
stein: Philosopher-scientist, edited by P. A. Schilpp, –. LaSalle: Open Court.
Brown, H. R. . Physical relativity: Space-time structure from a dynamical perspective.
Oxford: Oxford University Press.
Brown, H. R., and O. Pooley. . The origins of the spacetime metric: Bell’s Lorentzian
pedagogy and its significance in general relativity. In Physics meets philosophy at the
Planck scale, edited by C. Callender and N. Huggett, –. Cambridge: Cambridge
University Press.
———. . Minkowski space-time: A glorious nonentity. In The ontology of spacetime,
edited by D. Dieks, –. Amsterdam: Elsevier.
de Regt, H. W. . Wesley Salmon’s complementarity thesis: Causalism and unification-
ism reconciled? International Studies in the Philosophy of Science :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Dieks, D. . The adolescence of relativity: Einstein, Minkowski, and the philosophy of
space and time. Forthcoming in Minkowski spacetime: A hundred years later, edited by
V. Petkov. New York: Springer.
Ehrenfest, P. . Gleichförmige Rotation starrer Körper und Relativitätstheorie. Physika-
lische Zeitschrift :.
Einstein, A. /. On the electrodynamics of moving bodies. In The principle of
relativity, by H. A. Lorentz, A. Einstein, H. Minkowski, and H. Weyl, –. London:
Methuen.
———. . Zum Ehrenfestschen Paradoxon. Physikalische Zeitschrift :–.
———. /. My theory. In Ideas and opinions, –. New York: Crown.
———. /. Geometry and experience. In Ideas and opinions, –. New York:
Crown.
Janssen, M. . Reconsidering a scientific revolution: The case of Lorentz versus Ein-
stein. Physics in perspective :–.
———. . Drawing the line between kinematics and dynamics in special relativity. Stud-
ies in History and Philosophy of Modern Physics :–.
Laub, J. . Über die experimentellen Grundlagen des Relativitätsprinzips. Jahrbuch der
Radioaktivität und Elektronik :–.
Lewis, G. N., and R. C. Tolman. . The principle of relativity and non-Newtonian me-
chanics. Philosophical Magazine :–.
Lorentz, H. A. /. Electromagnetic phenomena in a system moving with any
dennis diek s
(*
Understanding in the Engineering Sciences
Interpretative Structures
B>:@:7DDC
I]Z:c\^cZZg^c\HX^ZcXZh
mieke b o o n
JcYZghiVcY^c\E]ZcdbZcV
preting must be distinguished that cannot be reduced to one another, and play
complementary roles in the development of science.
8g^iZg^V[dgJcYZghiVcY^c\
De Regt and Dieks () have proposed a method for explicating the role
of understanding by articulating criteria for understanding—thereby avoiding
a merely psychological interpretation. Their basic assumption is that under-
standing is acquired by scientific explanation, and their two criteria, CUP and
CIT, aim to provide a general account of what that means. CUP is their crite-
rion for understanding phenomena: “A phenomenon P can be understood if a
theory T of P exists that is intelligible” (, ) which is related to their cri-
terion for the intelligibility of a theory (CIT): “A scientific theory is intelligible
for scientists (in context C) if they can recognize qualitatively characteristic
consequences of T without performing exact calculations” (, ). In these
formulations, de Regt and Dieks are not explicit about what they mean by “the-
ory.” I will assume that their use of “a theory of P” means “a scientific explana-
tion of P,” which by my account is synonymous with “a scientific model of P.”
De Regt and Dieks have several objectives. First, by articulating CUP and
CIT, they aim to provide a more precise interpretation of the nature of skills in-
volved in achieving scientific understanding. In further explicating these skills,
they put emphasis on the qualitative and intuitive dimensions of understand-
ing. For instance, “theories may become intelligible when we get accustomed to
them and develop the ability to use them in an intuitive way” (, ). From
the perspective of the engineering sciences, their emphasis on intuitive aspects
is problematic for, in many cases, scientists do understand theories in terms
of how and where to use them, whereas they often do not have an idea of their
qualitative consequences, that is, the outcomes of explanations or predictions
that make use of the theory. This is particularly the case when scientists, in de-
veloping an explanation of a complex phenomenon, PC, integrate their scientific
understanding of physical phenomena P, . . . , Pi for which respective scientific
explanations T, . . . , Ti are available.
One example stems from the way scientists have developed an explana-
tion of the phenomenon of sono-luminescence, which is the emission of a light
pulse from imploding bubbles in a liquid when excited by sound. Brenner and
colleagues (, ) state that “an enormous variety of physical processes
is taking place inside this simple experiment [that produces the phenomenon
of sono-luminescence], ranging from fluid dynamics, to acoustics, to heat and
mass transfer, to chemical reactions, and finally to the light emission itself.”
What scientists usually do in generating predictions and explanations is to in-
terpret a complex phenomenon PC (for example, the emission of a light pulse
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
involves understanding at least two different types of theories. This also results
in two types of explanatory models of PC, which are a causal-mechanistic and
a nomo-mathematical model of PC (compare Boon ). The first model-type
allows for causal-mechanistic reasoning, whereas the second allows for nomo-
mathematical reasoning about the phenomenon PC. Although the two types
of models are related and explain different aspects of the phenomenon, they
cannot be reduced to each other. Therefore, in current scientific practices, dif-
ferent types of theories are combined in order to develop explanations and pre-
dictions of phenomena. Usually this is not a matter of preference, but related
to the intended uses of explanations and predictions of the phenomenon—for
example, models that predict how to create the phenomenon, models for cal-
culating optimal conditions for the phenomenon to occur, models used in the
design of processes that manufacture materials, and models that represent how
new properties of materials are generated.
Hence, de Regt and Dieks’s () two criteria of understanding cannot suf-
ficiently account for the engineering sciences. In particular, their proposals that
understanding is acquired by scientific explanation, and that understanding a
theory consists in recognizing qualitatively characteristic consequences of T,
are unsatisfactory as an answer to the question, “What is it about understand-
ing a theory that allows for using it?” One problem with their account is that
the proposed criteria do not explicate why and how an explanation provides
understanding, for example, how it is possible that a scientist recognizes the
qualitative consequences of a scientific explanation in concrete conditions, and
why does understanding of a scientific explanation allow for generating new
explanations and predictions. In brief, it does not sufficiently explicate how
scientific understanding does the work it is supposed to do in the intellectual
activities and abilities of scientists.
My account of scientific understanding expands on de Regt and Dieks
() and aims to overcome the mentioned difficulties. My leading questions
are: “What is it about an explanation that allows for understanding it?” and
“What is it about understanding an explanation that allows for using it?”
>ciZgegZiVi^kZHigjXijgZh
the ability to know where and how to apply the theories of a scientific field, and
how to infer consequences from them. I have illustrated this with the explana-
tion of sono-luminescence above, which involves the ability to recognize that
certain general theories (such as thermodynamics, fluid dynamics, acoustics,
theories of heat and mass transfer, and theories of chemical reactions) may
apply to this phenomenon and how to apply these theories to the case at hand.
Hence, criterion CITa must account for the role of scientists’s understanding
of a scientific field within which the phenomenon is relevant. This results in
a more general criterion for understanding scientific fields (CUSF): scientists
understand a scientific field if they have acquired the ability to use scientific
knowledge of the field (for example, fundamental and general theories, and
interpretative structures of relevant phenomena) in developing explanations
and predictions of other phenomena that are relevant to the field (including the
ability to meet the methodological criteria of the field).
These new criteria CUPa and CUSF, as alternatives to CUP and CIT, still
do not explicate why scientists understand interpretative structures, and what
it is about an interpretative structure that allows for using it in generating ex-
planations and predictions in new situations. I propose that scientists are able
to understand an explanation of P if this explanation presents an interpretative
structure of P. This implies that scientists understand theories and explanations
because theories and explanations are interpretative structures. But why is that?
My brief answer is, because interpretative structures represent how scientists
conceive of objects and relations among these objects. In general, conceiving of
relations among objects allows for making inferences. Hence, the specific char-
acter of explanations is that explanations are interpretative structures that rep-
resent how scientists conceive of relations between objects, which allows for
using explanations, for example, in making inferences under new conditions.
The idea that understanding T consists of scientists’ ability to use T in their
reasoning—where such reasoning is possible because interpretative structures
represent relations from which scientists can draw inferences—also avoids the
rather intuitive aspect of CIT, which is that “scientists recognize qualitative
consequences of T.”
I]Z6eegZci^XZ8VW^cZibV`Zg/68VhZd[<ZdbZig^XVaHigjXijg^c\
HigjXijg^c\VcY>ciZgegZi^c\Å>ciZgegZiVi^kZHigjXijgZhÅ
>ciZgegZiVi^kZ;gVbZldg`h
Structuring and interpreting can be done in terms of different interpretative
frameworks. These frameworks determine the type of phenomena that scien-
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
tists discern and the type of relations they “see,” for instance, in terms of logical,
geometrical, nomo-mathematical, causal-mechanistic, or statistical relations.
I will present a case that aims to show how the development of an interpre-
tative structure of a phenomenon works, and why an interpretative structure
allows for understanding a phenomenon. I also will clarify how understanding
the field (for example, Euclidean geometry)—that is, the ability to use the fun-
damental theories of the field articulated in CUSF—is involved when develop-
ing an interpretative structure that explains the phenomenon. Additionally, the
case of the apprentice cabinetmaker aims to illustrate that mathematics can
do the explanatory work, that is, that a mathematical explanation can provide
understanding. This goes against those who believe that only causal or causal-
mechanistic explanations are satisfactory in that sense.
An apprentice cabinetmaker has done his best to make a door that perfectly
fits into the opening of a cupboard, but when he hangs it on its hinges, he finds
that the door does not fit. He does not understand what has gone wrong since
he is absolutely certain that he has made the correct measurements. The master
cabinetmaker explains the observed phenomenon of this nonfitting door, PO,
as follows. First, the master cabinetmaker structures and interprets PO with-
in the framework of geometry. She abstracts from properties such as weight,
color, or properties of the wood, and focuses on proportions and dimensions
of PO, which she projects onto a two-dimensional geometrical space. She also
abstracts from the concrete values of the dimensions and proposes to assume
that the width of the opening of the door is W, the width of the door is D, and
the thickness of the door is d. She represents the structure thus obtained in fig-
ure .. When seeing figure ., most of us will immediately recognize how
the explanation works. But that is not the point of this case. What I aim to show
are several, more general points relevant to my argument.
W
Door
Door
D
;^\jgZ&(#&#AZ[i^aajhigVi^dcgZegZhZcihVc^ciZgegZiV"
i^kZhigjXijgZ>Hl^i]^ci]Z[gVbZldg`d[:jXa^Y^Vc
\ZdbZignd[i]Ze]ZcdbZcdcEi]ViVYddgZmVXian
Òihi]ZdeZc^c\d[VXjeWdVgY#G^\]i^aajhigVi^dc
gZegZhZcih>HD"<d[ED#
mieke b o o n
derstanding how PG has been constructed, that is, understanding how to struc-
ture and interpret the observed phenomenon PO within a geometrical space.
Accordingly, structuring and interpreting PO in the framework of geometry
has produced PG. In the second step of developing an explanation of PO, PG can
be structured and interpreted in terms of the axioms and laws of Euclidian ge-
ometry, producing ISG-G. Figure . represents how this works.
r=W
Dmax W–a
Door
W W
Dmax a W–a a
d d
W W
Dmax = √(W2–d2) (W–a) = √(W2–d2)
;^\jgZ&(#'#AZ[i^aajhigVi^dcgZegZhZcih>H<"<&d[
E<#G^\]i^aajhigVi^dcgZegZhZcih>H<"<'d[E<#
In a traditional language we would say that PG “obeys” the axioms and laws
of Euclidean geometry; in the modern language of the semantic conception of
theories, we would say that PG “satisfies” these laws (see, for example, Suppe
). The master has different possibilities in developing interpretative struc-
tures ISG-G to explain PG. I will call them ISG-G() and ISG-G(). For instance, by
applying Pythagoras’ theorem, she shows that the width of the door, D, should
not exceed √(W 2-d 2). If the apprentice understands this explanation ISG-G() of
PG, he is able to infer two consequences from it (see fig. ., left). First, as the
thickness, d, of the rectangle “door” exceeds zero, the width of this rectangle, D,
must be less than the width between the fixed rectangles, W. Second, the maxi-
mum width of “door” can be calculated from the formula: Dmax = √(W 2-d 2). Ac-
cordingly, understanding ISG-G() as an explanation of PG within the framework
of Euclidean geometry allows for inferring consequences from ISG-G(), which
in this case consists of making predictions about the effects of interventions,
such as making the rectangle “door” smaller by removing a slice with thickness
a.
However, the cabinetmaker has a specialty: she prefers to make her doors
skewed, since removing slice a would cause a gap between rectangle “door”
and the rectangle to the right. Therefore, she expounds on her explanation
mieke b o o n
ISG-G() by telling the apprentice that although he has done a good job, he is not
yet finished. He should cut off a triangle from the edge of the door by drawing
a line between the outside corner to a point, a, on the inside width-line of the
door. Since the apprentice understands this explanation, and because he also
understands geometry, he can infer how to calculate point a. He infers that
point a must satisfy the formula (W-a) = √(W 2-d 2). Also this second formula
has been derived from structuring PG in terms of Pythagoras’ theorem, which
involves a further refinement of ISG-G().
On the basis of his understanding of geometry, the apprentice also under-
stands that he can easily construct point a by applying the law to which all
circles obey, which is that the distance is equal from the center of a circle to
any point on the circle. This is how a second interpretative structure ISG-G() of
PG is developed within the framework of geometry. He constructs point a by
drawing a circle with the dot in its center and taking W as the length of radius
r. A rounded edge is constructed by means of a circle with a radius of r=W. In
order to make the “door” fit, he must then draw a straight line from point a to
the outside corner of the door in order to mark the triangle that he should cut
away. In this way, PG has been structured and interpreted in terms of the law
that defines a circle. Inspired by these arcs, the apprentice decides that he will
develop his own specialty once he becomes a fully qualified cabinetmaker. This
specialty will be rounded edges, since, on the basis of the axioms of geometry,
he understands that the door can only be closed if the distance from the hinge
to the edge of the door at all points along d is exactly equal to W. He finds this
solution (right-hand side in fig. .) the most elegant.
8g^iZg^dc[dgJcYZghiVcY^c\HX^Zci^ÒX;^ZaYh8JH;
By further analyzing this case, I will illustrate the abilities involved in under-
standing a scientific field. How does the criterion for understanding scientific
fields (CUSF) apply to the case at hand? Understanding Euclidean geometry
involves the ability of the cabinetmaker to structure and interpret the phenom-
enon of the nonclosing door, PO, within the framework of Euclidean geometry,
producing PG. Understanding the interpretative structure of PG, ISG-G, guides
the apprentice and the cabinetmaker in exploring the effects of possible inter-
ventions with the door’s geometrical structure. The cognitive abilities of the
apprentice and his master thus comprise their being able () to recognize that
the observed phenomenon PO can be structured within the framework of Eu-
clidean geometry; () to abstract from properties of the material door that are
irrelevant within that framework; () to construct a geometrical structure of
PO, ISO-G, that represents how the cabinetmaker conceives of the phenomenon
in terms of geometrical relations between rectangles; () to conceive of ISO-G as
a phenomenon in a geometrical space, PG; () to know which theorems are fit-
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
8VjhVa"BZX]Vc^hi^XVcYCdbd"BVi]ZbVi^XVaHigjXijg^c\
HigjXijg^c\VcY>ciZgegZi^c\
As illustrated in the case above, the first step in an explanation involves
structuring and interpreting the observed phenomenon, PO, in terms of a cer-
tain type of interpretative framework. A simplified example may illustrate how
simultaneous structuring and interpreting works within a causal-mechanistic
(CM) and a nomo-mathematical (NM) framework works.
Assume a Boyle type of experiment, in which a correlation is measured
between the pressure and the volume of a gas in an air-tight cylinder. The ob-
served phenomenon PO is that “decreasing the volume of the cylinder requires
someone to increase the force on the piston.” This phenomenon is framed in
terms of a CM framework, that is, “decrease of the volume of the gas causes
an increase of the pressure of the gas,” which describes a causal-mechanistic
phenomenon PCM. The measured values of the pressure, P, and the volume, V,
are framed in terms of an NM framework, where nomo-mathematical means
“mathematical relations between physical variables.” For instance, the set of
data points plotted in a P-V diagram represents the nomo-mathematical phe-
nomenon PNM. Accordingly, within the CM framework, the relation between
pressure and volume observed in this experiment is conceived as a causal-
mechanistic relation, whereas within the NM framework the relation between
the measured values of pressure and volume is conceived as a mathematical
relation between data points.6
Analogous to the example of the cabinetmaker, the next step consists of
developing interpretative structures of PCM in terms of CM relations, and of
PNM in terms of NM relations. A CM interpretation employs (or introduces
new) theoretical entities, such as gas molecules, and properties of these enti-
mieke b o o n
adequacy of ISNM-NM must be tested for observed phenomena of real gases. The
ideal gas law, for instance, is only empirically adequate for ideal gases.12 A next
step in scientific research is to explain deviations between conclusions inferred
from ISCM-CM, on the one hand, and measurements, on the other. Such devia-
tions are structured and interpreted in terms of ISCM-CM, for instance, in terms
of van der Waals’s forces between gas molecules. This is how science develops:
structuring and interpreting proceeds ever further, which involves developing
interpretative structures ISNM-NM of NM phenomena within NM frameworks,
and interpretative structures ISCM-CM of CM phenomena in CM frameworks, in
an ongoing mutual interaction.
HX^Zci^ÒXJcYZghiVcY^c\
Scientists understand interpretative structures because these represent
how they conceive of objects and relations between objects. This understand-
ing allows for inferring conclusions from these structures. For instance, un-
derstanding Hooke’s law as an interpretative structure for the behavior of a
spring allows for predicting the stretch of a spring that results from exerting
a force. Developing interpretative structures involves an understanding of rel-
evant scientific fields, which in turn implies the involvement of specific cog-
nitive abilities. First, an observation or a measurement must be represented
adequately within an interpretative framework, such as a causal-mechanistic
or a nomo-mathematical framework, in order to produce a PCM or PNM, respec-
tively. Second, scientists must decide on the right fundamental principles for
structuring and interpreting a PNM. Third, when structuring and interpreting
within a causal-mechanistic framework, scientists must envisage physical enti-
ties and properties relevant to PCM. Fourth, deductions from fundamental laws
must be mathematically correct. Often, approximations and simplifications are
needed in order to make the mathematical equations manageable. An example
is neglecting viscosity terms in the Navier-Stokes equation. Doing this correct-
ly usually involves scientific understanding in terms of a causal-mechanistic
framework. These are the kinds of abilities that determine someone’s scientific
understanding of a scientific field.
:e^hiZbdad\^XVa>hhjZh/DW_ZXihVcYGZaVi^dchWZilZZcDW_ZXih
The core aspect of interpretative structures is that they represent objects
and relations between objects. These structures represent relations in a similar
way as to how judgments represent relations between disparate objects and/or
events. For instance, “the cat is on the mat” represents a relation in space, or an
individuation of objects, that is, the cat and the mat; “the cat was on the mat”
represents as a relation in time between two events; “if the cat sits on the mat,
then the dog is out” represents a logical relation; “the cat is skinny” represent a
mieke b o o n
relation between an object and a property; “the cat has dirtied the mat” repre-
sents a causal relation between two events; “the cat sits one meter beside the
mat” represents a mathematical relation—in this case, the length of a straight
line between two points; and “this is the path that the cat always takes” repre-
sents a relation between locations on a geographical map.
The basic idea of these examples is that in observing the world, humans
must structure and interpret their observations in order to arrive at even the
most elementary judgments, for instance, PG, PCM, or PNM. In doing so, they
relate what they observe in terms of interpretative structures. As may be clear
from the examples, different kinds of interpretative structures are involved,
which result in different types of relations, such as relations in space and/or
time, relations that individuate (discern) objects or events, relations between
objects and properties, causal relations or interactions, mathematical, and logi-
cal relations. Instead of conceiving our “observations” of relations as something
that exists independent of us, we must conceive of “observed” relations as ways
in which we have structured and interpreted our observations.13
Knowledge represented in theories and scientific explanations must be
understood as similar in form to knowledge represented in judgments. Ac-
cordingly, theories and scientific explanations also represent objects and rela-
tions between objects, such as logical, geometrical, causal-mechanistic, and/or
nomo-mathematical objects and relations. These are not objects and relations
that scientists could somehow observe if they had the proper means, but rela-
tions in terms of which scientists structure and interpret what they observe.
Why do we have a basic interest in structuring and interpreting the world?
Several reasons can be given. First, structuring and interpreting the world is a
basic cognitive need, because without this activity the world would appear cha-
otic and arbitrary to us. Second, interpretative structures often have an aesthet-
ic value. Third, structuring and interpreting the world according to ever-higher
epistemic standards is an intellectual challenge. Fourth, we need interpretative
structures in order to think rationally about our acting in, and intervening with,
the world. Again, knowledge of these relations is not gained by observing them,
but by means of “structuring and interpreting” our observations in terms of
certain frameworks. These frameworks guide how to construct or invent rela-
tions between objects or events.
8dcXajh^dch
Why are the engineering sciences such an interesting case with regard to
perspectives on scientific understanding? I propose that considering the engi-
neering sciences as science presents us with consequences on how to conceive
of the laboratory sciences in general.
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
CdiZh
I would like to thank Henk Procee, Bas van Fraassen, Tarja Knuuttila, Isabelle Pe-
schard, Henk de Regt, Sabina Leonelli, and Kai Eigner for their constructive contributions.
This research is supported by a grant from the Netherlands Organisation for Scientific
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
Research (NWO). I also wish to thank the ZIF Centre for Interdisciplinary Research, Biele-
feld University, for their financial support, and the members of the ZIF research group
(–), Science in the Context of Application, for their intellectual support.
. My view on science stands in the tradition of the New Experimentalists. Authors
who have worked at the line between philosophy of science and history of science, such as
Ian Hacking (), Alan Franklin (), Peter Galison (), Robert Ackerman (),
Margaret Morrison (), and Hasok Chang (), give reasons to believe that the tradi-
tional aim of science is also too narrow for most of the present-day “basic” laboratory sci-
ences. A severe critique of this traditional picture shared by these authors is that it neglects
the crucial role of experiments and instruments in science. In the laboratory sciences,
experiments and instruments are not mere spectacles on the world. They are more than
just tools for observing the world; they also have a life of their own. For instance, instru-
ments produce phenomena that do not exist outside the laboratory, and those phenomena
become topics of scientific and technological interest. An example is the phenomenon of
superconductivity, which can only exist in artificially produced superconductor materials
at extremely low temperatures.
. Petri Ylikoski (this volume) defends a similar thesis. He refers to Wittgenstein (Phil-
osophical Investigations, ), who claims that understanding is an ability, since, when a
person understands something, he or she is able to do certain things. Michael Dummett
(/, –), to whom I am indebted, stands in the same tradition. Against the
generally accepted idea that understanding the meaning of a word or a sentence consists in
its method of verification, or in its truth conditions, Dummett suggests that understanding
a word or sentence involves knowing how to use it. Similarly, I propose that understand-
ing a theory does not depend on its truth, which would mean that a scientist understands
the theory because it provides a literally true story of what the world is like (compare van
Fraassen ). Rather, understanding a theory involves the ability to use it.
. My notion of interpretative structures is much broader than the common use of
structures in philosophy of science, which indicates logico-mathematical structures that
represent logical and mathematical relations.
. The distinction between the “observable” and “unobservable” world refers to an
intuitive distinction that we usually make between the world that is directly accessible for
sense experiences, and the world behind the phenomena. Van Fraassen () used this
distinction for explicating his antirealism, and his work has initiated a lot of philosophi-
cal discussion on this subject. Nevertheless, it is still a clarifying distinction in our daily
language, which is how I aim to use it here. My point in this chapter is that observability is
problematic for another reason, which is that we cannot make sense of “directly” observing
the relations presented in our judgments of the “observable” world. This is why I have put
these terms in quotes.
. Radder () analyzes in great depth the claim that observation presupposes
conceptual interpretation. Radder’s emphasis is on the idea that concepts structure our
observations of the world. My emphasis is on the idea that cognitive activities structure our
observations of the world; these activities consist of structuring and interpreting in terms
of specific types of objects and relations between objects.
mieke b o o n
. The semantic view of theories gives a refined account of how data points are
converted into a data model, (see, for example, Mayo , chapter ). I agree with this
account, but will not use “data-model” in this chapter.
. The formula ISNM-M means an interpretative structure IS of measured data points
in a nomo-mathematical framework NM. Data points are related in terms of an arbitrary
mathematical equation M (for example, a polynomial function).
. My proposal that the axioms of fundamental theories present a theoretical frame-
work within which scientists structure and interpret phenomena agrees with the semantic
view of theories of van Fraassen (, –), and Giere (, ). From this per-
spective, Giere (, ) argues against Cartwright (). According to Giere, funda-
mental “laws cannot tell lies about the world, because they are not really statements about
the world. . . . They are, as Cartwright herself sometimes suggests, part of the characteriza-
tion of theoretical models, which in turn may represent various real systems.” Opposite
to this view stands structural realism. An overview of the constructive empiricism versus
structural realism debate can be found in Psillos ().
. The engineering sciences involve at least five topics that are considered as general
basic theory, necessary for taking a scientific approach to the development of technology,
which are typical examples of how nomo-mathematical structuring works. These topics
are mechanics, flow phenomena, electricity and magnetism, thermodynamics and statisti-
cal mechanics, and quantum mechanics. My basic assumption is that, like the axioms
of Euclidean geometry, the axioms of fundamental theories are not true or empirically
adequate, but fitting for structuring and interpreting certain types of phenomena. For in-
stance, Newton’s three laws of motion and the law of gravity play a central role in mechan-
ics. Conservation laws, such as conservation of mass, energy, momentum and chemical
substance play a central role in flow phenomena. Electricity and magnetism draw on
Maxwell’s laws; thermodynamics on the basic laws of thermodynamics; statistical mechan-
ics on the Boltzmann equation and the laws of probability theory, and quantum mechanics
on the Schrödinger equation.
. With the complementarity of interpretative frameworks, I mean something differ-
ent than Salmon’s () complementarity thesis, which claims that explanations in terms
of causal-mechanisms, and explanations in terms of derivations from fundamental laws are
complementary. De Regt’s () analysis of this claim shows that Salmon interprets his
causal conception of scientific explanation as an “ontic” claim: “in contrast to ‘epistemic’
conceptions which regard explanations as arguments, ontic conceptions assert that ‘events
are explained by showing how they fit into the physical patterns found in the world’ ” (de
Regt , ). My conception of science as developing interpretive structures within
interpretative frameworks aims at avoiding ontic or metaphysical connotations.
. More correctly, the phenomenon is interpreted and structured in analytic geom-
etry. Scientific explanations of phenomena are often structured within a framework that
integrates Euclidian geometry and analytic geometry that uses Cartesian coordinates for
defining the space. As a result, the IS can be represented in mathematical spaces that are
no longer strictly geometric, but constructed from combinations of x, y, z, t, and physical
variables such as the velocity, v. Introduction of principles from analytic geometry allows
for instance for the construction of differential equations. Another aspect that I will not
u nd e r stan d i n g i n th e e ngi ne e ri ng s c ience s
explain in more detail is that the deduction of mathematical structures does not usually
start from the fundamental principles, but from equations already deduced from them,
such as the Maxwell-Boltzman equation, or Navier-Stokes equation.
. Clearly, ideal gases are defined in this way: a gas is ideal if it obeys the ideal gas
law. Therefore, the belief that the ideal gas law is empirically adequate for ideal gases is
problematic.
. On this matter I am deeply indebted to Kant’s criticism of Hume’s empiricism;
compare Kant (/) and Henry Allison’s () interpretation of Kant’s work.
Rather than using Kant’s notion of categories, I propose the notion of interpretative
frameworks.
GZ[ZgZcXZh
Ackerman, R. . The new experimentalism. British Journal for the Philosophy of Science
:–.
Allison, H. E. . Kant’s transcendental idealism: An interpretation and defense. New
Haven: Yale University Press.
Boon, M. . Technological instruments in scientific experimentation. International
Studies in the Philosophy of Science :–.
———. . How science is applied in technology. International Studies in the Philosophy
of Science :–.
Boon, M., and T. Knuuttila. . Models as epistemic tools in engineering sciences. In
Handbook of the philosophy of science, vol. , Philosophy of technology and engineering
sciences, edited by Anthonie Meijers, –. London: Elsevier Science.
Cartwright, N. . How the laws of physics lie. Oxford: Clarendon Press.
———. . Nature’s capacities and their measurement. Oxford: Clarendon Press.
Chang, H. . Inventing temperature: Measurement and scientific progress. New York:
Oxford University Press.
de Regt, H. W. . Wesley Salmon’s complementarity thesis: Causalism and unification-
ism reconciled? International Studies in the Philosophy of Science :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Dummett, M. /. The philosophical basis of intuitionist logic. In Truth and other
enigmas, –. London: Duckworth.
Franklin, A. . The neglect of experiment. New York: Cambridge University Press.
Galison, P. . How experiments end. Chicago: University of Chicago Press.
Giere, R. N. . Explaining science: A cognitive approach. Chicago: University of Chicago
Press.
———. . Science without laws. Chicago: University of Chicago Press.
Hacking, I. . Representing and intervening: Introductory topics in the philosophy of
natural science. Cambridge: Cambridge University Press.
———. . The self-vindication of the laboratory sciences. In Science as practice and
culture, edited by A. Pickering, –. Chicago: University of Chicago Press.
Kant, I. /. Critique of pure reason. Cambridge: Cambridge University Press.
mieke b o o n
Mayo, D. G. . Error and the growth of experimental knowledge. Chicago: University of
Chicago Press.
Morrison, M. . Unifying scientific theories: Physical concepts and mathematical struc-
tures. Cambridge: Cambridge University Press.
Psillos, S. . Is structural realism possible? Philosophy of Science :s–s.
Radder, H. . The world observed / The world conceived. Pittsburgh: University of Pitts-
burgh Press.
Salmon, W. C. . Scientific explanation: Causation and unification. In Philosophy of
science: Contemporary readings, edited by Y. Balashow and A. Rosenberg, –.
London: Routledge.
Suppe, F. . The semantic conception of theories and scientific realism. Urbana: Univer-
sity of Illinois Press.
Trout, J. D. . Scientific explanation and the sense of understanding. Philosophy of Sci-
ence :–.
van Fraassen, B. C. . The scientific image. Oxford: Clarendon Press.
———. . Laws and symmetry. New York: Oxford University Press.
chap ter title ver so
(+
Understanding in Psychology
Is Understanding a Surplus?
@6>:><C:G
k ai eigner
logical positivism, which included, among other things, the requirement that
all subjective elements should be banished from science. Because of their posi-
tivistic attitude, neobehaviorists were called “black-box psychologists”: they
considered the organism as a black box with observable stimuli as input and
observable responses as output, and they rejected any allegedly metaphysical
speculation about the mechanisms inside this black box. Metaphorically speak-
ing, opening the box would not yield “insight.” Instead, they declared as the aim
of psychology to find functional relationships that describe the correlations be-
tween stimuli and responses. This positivistic tendency makes neobehaviorism
highly relevant to a study on intelligibility and understanding. Insofar as un-
derstanding and intelligibility, with their subjective connotations, were allowed
to play a role at all by the neobehaviorist doctrine, this was only in the context
of discovery and not in the context of justification. By focusing on two leading
figures of neobehaviorism, Edward C. Tolman (–) and Clark L. Hull
(–), I will show that, despite their positivistic ideas, in practice even
the neobehaviorists aimed at insight and implicitly embraced intelligibility as
an epistemic value. What might seem to be one of the strongest counterex-
amples of my claim appears in fact a corroboration of it, and as such provides a
strong case in favor of my claim concerning intelligibility and understanding.
In this chapter, I will offer a philosophical account of intelligibility and un-
derstanding based on the work of Henk de Regt, but with a shift in focus from
theories to models. I will then present the case study of neobehaviorism and,
by means of an analysis of the work of Tolman and Hull, in which I will use my
philosophical account of intelligibility and understanding, I will demonstrate
my claim about the constitutive value of intelligibility.
JcYZghiVcY^c\E]ZcdbZcVWnBZVchd[>ciZaa^\^WaZBdYZah
6eean^c\I]ZdgZi^XVaEg^cX^eaZhidEVgi^XjaVgE]ZcdbZcV
The first aspect of the ability to use theoretical knowledge is the ability to
make a connection between theory and the world. The relation between scien-
tific theories and the world—which is epistemologically relevant for the empir-
ical justification of theoretical knowledge—is a topic in philosophy of science
that has received much attention. Since the s we have seen a gradual shift
in the philosophical ideas concerning this relation. The traditional theory-
dominated view was superseded by a view in which models play an important
role in representing aspects of the world. For instance, Cartwright, Shomar, and
Suárez () hold the view that the connection between scientific theory and
the world is accomplished by constructing models that form representations of
phenomena. Constructing these models is a skillful activity that involves evalu-
ations and decisions. Along the same line, de Regt and Dieks (, ) write
that the ability of scientists to apply a theory to a particular phenomenon de-
pends both on their skills and on the pragmatic virtues of the theory, such as
its visualizability or simplicity. According to de Regt and Dieks, this activity of
connecting theory and the world, which has a pragmatic nature, has epistemic
relevance because it is a necessary ingredient for achieving the epistemic aims
of science. Although I agree with this, the argumentation for it needs more
sophistication in order to avoid the objection that the activity of constructing
models—and thus forming the connection between theory and the world—
u nd e r stan d i n g i n p s yc h ol o gy
belongs to the context of discovery, and that the pragmatic dimension of it does
not play a role anymore once the connection between theory and phenomena
has been established. An important step in circumventing this objection is to
formulate more precisely what we mean by concepts like theory, model, and
representation. For this I will make use of the representational view of models
of Giere (, ), which, similar to the view of Cartwright and colleagues,
shares the contemporary idea mentioned above that models play an essential
role in representing aspects of the world.
Giere (, ) prefers to speak of “theoretical principles” instead of
“theories” or “laws” because the latter terms are used quite broadly and even
ambiguously, both in scientific practice as well as in philosophical discussions
about the sciences. Examples of theoretical principles are Newton’s laws of me-
chanics, the evolutionary principle of natural selection, and, as we will see in
this chapter, Hull’s principles of behavior. Although such principles have often
been interpreted as empirical statements, Giere sees them merely as defini-
tions of the abstract terms that compose theoretical models. For instance, in
Newton’s three laws of motion, quantities called force and mass are related
to well-understood quantities such as position, velocity, and acceleration. The
laws, however, do not spell out in more specific terms what might count as a
force or mass (Giere , ). According to Giere, the function of these theo-
retical principles is to act as general templates for the construction of models.
Giere (, ) calls his view on models representational because it takes
models as tools for representing the world. He is aware that the models used in
scientific practice seem to form a heterogeneous class, including physical mod-
els, scale models, analogue models, and mathematical models, but he argues
that with his account it is possible to understand models in a way that usefully
encompasses this heterogeneity to a large extent. “What is special about mod-
els is that they are designed so that elements of the model can be identified
with features of the real world. This is what makes it possible to use models
to represent aspects of the world” (Giere , ). In short, models can be
used to represent phenomena, and theoretical principles constitute definitions
of “abstract objects” that can be used to construct these models. As we will
see, this view of the relation between theory and phenomena fits nicely with
the practice of the neobehaviorists, who dealt with principles of behavior (for
example, Hull a). These principles functioned exactly as Giere described
in his representational view, as definitions of the abstract objects that compose
theoretical models.
In order to use a model to represent some aspect of the world, one should
be able to pick out relevant features of the model that are similar in certain
respects and degrees to what is modeled. Giere notices that there are no objec-
tive rules for picking out some specific features of the models. Moreover, he
k ai eigner
argues that the concept of similarity is context dependent (Giere , ) and
that there is no objective measure of similarity between the model and the real
system.
9ZkZade^c\>ch^\]i^c8dchZfjZcXZhd[I]ZdgZi^XVaEg^cX^eaZh
The second aspect of the ability to use theoretical knowledge, which I will
discuss in relation to the intuitions about scientific understanding as found in
de Regt and Dieks (), is the ability to develop qualitative insight into the
consequences of theoretical principles in concrete situations. According to de
Regt and Dieks (, ), this ability requires conceptual tools that make it
possible to circumvent the calculating stage and make an intuitive jump to the
conclusion. In their account, the virtues of the theory can be used as such tools
u nd e r stan d i n g i n p s yc h ol o gy
if they realize “the right combination” with the skills of the scientist (de Regt
, ). As mentioned above, I regard this as a valuable insight for an ac-
count of understanding, although in my view these virtues are better to be seen
as virtues of models instead of theories.
An illustration provided by de Regt and Dieks is the way Ludwig Boltzmann
introduced the kinetic theory of gases. Boltzmann () devoted the introduc-
tory section to a purely qualitative analysis of gases in which he concluded that
a gas can be pictured as a collection of freely moving molecules in a container.
This model of the gas provides qualitative insight in the behavior of gases. For
instance, it is straightforward to explain in a qualitative way that a gas exerts a
pressure on the walls of the container. If a gas molecule collides with a wall of
the container, it gives a little push, and the total effect of the pushing produces
the pressure. This picture also makes clear that a decrease of volume results
in an increase of pressure. A decrease of volume will cause an increase of the
number of molecules per unit of volume, and this causes an increase in the
number of impacts per unit of time on the wall surface, and thus an increase
in pressure. As de Regt and Dieks argue, in this way qualitative understanding
about gases is obtained which is not based on calculations:
It is important to note that the above reasoning does not involve any
calculations. It is based on general characteristics of the theoretical
description of the gas. Its purpose is to give us understanding of the
phenomena, before we embark on detailed calculations. Such calcu-
lations are subsequently motivated, and given direction, through the
understanding we already possess. (de Regt and Dieks , –)
means picking out the relevant features of the model and judging the similari-
ties between the behavior of the model and the real gases—due to the causal
nature of Boltzmann’s model, which in light of our skill of making causal infer-
ences is a virtue that renders the model intelligible.
We now have the ingredients of the account of understanding used in
this chapter. In order to understand a phenomenon, an intelligible model is
needed, which is a model that has virtues skillful scientists can use for two
purposes, namely for “doing the representing” of the phenomenon and for
seeing intuitively the characteristic consequences of the model. I will use this
account to give an analysis of the role of intelligibility and understanding in
the case study about neobehaviorism, where I will mainly look at the work of
Tolman and Hull, who are considered to be the main representatives of the
methodology of neobehaviorism. I will use the work of Tolman to get an idea
of the character of the skills and virtues that play a role in neobehaviorism and
I will use the work of Hull in order to demonstrate the epistemic relevance of
these virtues and skills, and thus of the epistemic relevance of intelligibility
and understanding.
7Z]Vk^dg^hiBZi]dYdad\n
The two basic ideas of classical behaviorism, which can already be found in
this manuscript of Watson, were that the (almost exclusive) subject matter of
u nd e r stan d i n g i n p s yc h ol o gy
psychology should be overt behavior, and that the major goals of psychology
should be the prediction and control of behavior instead of the description and
explanation of states of consciousness (Hergenhahn , ).
When logical positivism arose in the s it heavily influenced behavior-
ism, and since then we speak of neobehaviorism. Although logical positivism
offered the behaviorists a philosophical foundation for their objectivist views,
it was not evident how this body of thought could be translated into recom-
mendations for scientific practice. Therefore, leading figures of neobehavior-
ism, such as Tolman and Hull, initiated a behaviorist methodology that was
meant to give shape to an objective psychology in which there was no room for
subjective elements. A major ingredient of this methodology was operational-
ization, a procedure by which the theoretical terms used in psychology were
given objective definitions and that was meant to eliminate possible subjective
connotations, originating, for instance, from prescientific insight.
We will see, however, that, in spite of their efforts, the meaning of the the-
oretical terms used by Tolman and Hull surpassed their objective definition:
apparently the terms had indispensable “surplus meaning.” An analysis of this
surplus meaning, in which I make use of my account of understanding, will
show that its function is to render intelligible the scientific models in which
these terms appear. From the indispensable character of this surplus meaning I
will argue that actually, despite their aversion to subjective elements in science,
the neobehaviorists aimed at constructing models that are intelligible. My con-
clusion from this will be that even in the case of neobehaviorism intelligibility
functioned as a significant epistemic value.
:YlVgY8#IdabVc
In the s, Tolman developed a methodology for neobehaviorism that
he called “operational behaviorism” (Tolman /). A motive for Tolman
to develop this methodology was his interest in the logical positivism of the
Vienna Circle. After a sabbatical in Vienna, Tolman expressed the intention
to apply the ideas of logical positivism to psychology. “The ‘logical positivists,’
that is such men as Wittgenstein, Schlick, Carnap in Europe and Bridgman,
C. I. Lewis, Feigl and Blumberg in this country, have already done the task for
physics. But no one as yet, so it seems to me, has satisfactory done it for psy-
chology” (Tolman /, ). In a logical positivistic spirit he defined the
aim of psychology to find functional relationships between the behavior of an
organism and its environmental “determiners,” labeled by him respectively as
“dependent” and “independent variables.” He schematized this in the following
manner (Tolman /, ):
For complex behavior, finding this function f is a difficult task, and therefore
Tolman proposed to break up this complicated function into more manageable
component functions, f and f, which led to the introduction of theoretical
terms, or “intervening variables.” Tolman schematized the new situation as fol-
lows (Tolman /, ):
Independent variables ——— f ——— Intervening variables ——— f ——— Dependent variables
“WRONG” “CORRECT”
OGL OL OR OGR
BL OC BR
;^\jgZ&)#I"bVoZl^i]h^c\aZX]d^XZed^ci#D82i]Zed^cid[
X]d^XZ0DA$G2i]ZXdbeaZmd[hi^bjajhdW_ZXihbZi\d^c\Ydlci]ZaZ[i$
g^\]iVaaZn0D<A$G2i]Z\dVaVii]ZaZ[i$g^\]i07A$G2i]ZaZ[i$g^\]i"ijgc^c\
WZ]Vk^dgIdabVc&.(-$&.++!&*%#
worthy of note, and not other aspects, such as whether the rats moved quickly
or slowly. One may wonder why he decided to measure the tendency of the rats
to enter the left alley, that is, the alley where there is no food. Maybe his inten-
tion for choosing this as a dependent variable was to give the impression of
being unprejudiced and therefore objective. For independent variables, Tolman
proposed to look at the number of previous trials in the maze made by the rats
and the period of food deprivation. He gave no reasons for this choice either;
apparently, he considered this choice of variables to be obvious. That it was
based on a prescientific understanding of the behavior of rats becomes visible
when we look at Tolman’s choice of intervening variables.
In Tolman’s methodology, intervening variables are introduced in the pro-
cess of breaking down the function between the independent and dependent
variables into component functions. In his example, Tolman introduced the
variable “demand” that intervenes between the period of food deprivation and
the percentage of rats that enters the left alley, and also the variable “hypoth-
eses” that intervenes between the number of trials and the percentage of rats
that enters the left alley. Although intervening variables were not supposed
to be seen as representing entities or processes, it is obvious that this way of
breaking down the function is based on the ideas that the rats are getting hun-
gry if they are deprived of food for a long time, that during the trials in the maze
the rats form hypotheses about where food can be found, and that hungry rats
will use these hypotheses in order to find food. Apparently Tolman used his
prescientific understanding of the behavior of rats as a heuristic guide in his
choice of dependent, independent, and intervening variables.
As second step, let us look at the formulation of the f functions between
k ai eigner
the independent variables and the intervening variables. These functions form
the operational definitions of the intervening variables. Formulating these
functions was a central element in Tolman’s methodology of operational be-
haviorism. In this procedure of operationalization, Tolman made use of what
he calls “standard experiments” in which he measured the functional relation
between a specific independent and a specific dependent variable while keep-
ing the other independent variables at a constant (standard) value. The results
of these experiments formed the basis for the definition of the intervening vari-
ables. The rationale behind this second step was that fixing the definitions of
the intervening variables using an objective methodology removes the surplus
meaning. However, we will see that the definitions Tolman provided were not
entirely based on his methodological rules. This follows, for instance, already
from their dependency on “standard values,” which are not specified in Tolman’s
methodology. In most experimental situations this is not a problem because it
is obvious what the standard values are. For instance, the rats used for train-
ing experiments in a maze should not be completely underfed or overstuffed.
Determining the right (standard) level of nourishment is a judgment that is not
based on general methodological rules.
In his example, the operational definition of the intervening variable “de-
mand” was based on measurements of the relation between the period of food
deprivation and the percentage of rats that entered the left alley, where the
amount of training of all the rats was set to a standard value. The result of this
standard experiment is visible in figure ..
BL
BL + BR
%
50
M- (hours
0 12 24 36 48 60 72 since
feeding)
;^\jgZ&)#'#EZgXZciV\Zd[gVihZciZg^c\i]ZaZ[iVaaZn
Vh[jcXi^dcd[i]ZeZg^dYd[[ddYYZeg^kVi^dc#IdabVc
&.(-$&.++!&*-#
u nd e r stan d i n g i n p s yc h ol o gy
With the increase of the number of hours passed since feeding, the number
of rats that entered the left alley decreased (and later increased). On the basis
of figure ., Tolman formulated a definition of the intervening variable “de-
mand.” To turn this figure into a definition of “demand,” Tolman argued, one
modification was necessary, namely that “demand” should really be defined as
the inverse of this ratio (Tolman /, ). This led to the following op-
erational definition of “demand” (see fig. .):
DEMAND
%
100
M- (hours
0 12 24 36 48 60 72 since
feeding)
;^\jgZ&)#(#9ZbVcYVh[jcXi^dcd[i]ZeZg^dYd[[ddY
YZeg^kVi^dc#IdabVc&.(-$&.++!&*.#
The inversion (and translation) of the graph, which is not based on meth-
odological rules, indicates that these rules are not sufficient for formulating
the operational definitions. Apparently Tolman regarded it to be essential that
demand increases with the time since feeding—at least for the first fifty hours.
That it does not increase any more and even decreases after that amount of
time is presumably an unwanted effect of this operationalization. Although
Tolman made no remark about it, he did make the judgment to stop the experi-
ment after about eighty hours, which suggests that he regarded this last period
to be irrelevant or uninteresting. This is again an example of a decision that is
not based on methodological rules.
The notable act of transforming graphs without methodological motiva-
tions is not exceptional. Similar transformations can be found in the case of
other intervening variables. For the development of the operational definition
of the intervening variable “hypotheses,” Tolman measured the function be-
tween the number of previous trials made by hungry rats and the percentage
of the rats that entered the left alley, and it appeared that with the increase of
k ai eigner
the number of previous trials, the number of rats that entered the left alley
decreased. He transformed this function into the definition of “hypotheses” by
taking the inverse of this function. The result was, as Tolman argued, “no more
than our old friend, the learning curve” (Tolman /, ), which is a
frequently used function in neobehaviorism (see, for example, fig. . below).
In this way he provided “hypotheses” with an objective, operational definition
in terms of the number of previous trials. Tolman did not explain why both “de-
mand” and “hypotheses” should be defined as an inverse of the graphs on which
they were based. This raises the question about why Tolman had to apply these
transformations. As will be clear when we look at the next step regarding the
formulation of the f function, Tolman applied these transformations, which
were based on his prescientific understanding of the behavior of rats, because it
enabled him to attach surplus meaning to the terms in order to make his model
intelligible. Instead of ensuring that the meaning of the theoretical terms in his
behavioral model did not exceed their operational definition (as he ought to
have done, according to the philosophical doctrine he preached), Tolman in
fact ensured that these terms had surplus meaning.
As the third step, I will examine the formulation of the f function, about
which Tolman wrote, “It is by means of this f function (if we but knew what it
was) that we would be able to predict the final outcome for all possible values
of the intervening variables” (Tolman /, ). Finding the f function
is easier said than done. Tolman confessed that this is the part of his doc-
trine about which he was “haziest” (). He proposed to follow a strategy of
anthropomorphism:
that together constitute the model of the behavior of rats in mazes. Tolman
used his understanding of rats to construct this model. This understanding was
based on certain skills, of which imagining was the most important. Tolman
imagined how, if he were a rat, he would behave. This is imagining in the sense
of empathizing. By means of this skill, Tolman was able to exploit a fruitful an-
thropomorphic conceptualization. He was able to cast concepts like “demand”
and “hypotheses” into a mold such that he could derive useful hunches from
everyday experience.
Obviously the transformation of the graphs mentioned above was already
part of this framing process. The objective definitions of the theoretical terms
used in Tolman’s behavioral model of the rat in the maze were shaped in such a
way that Tolman was able to give an anthropomorphic interpretation of them.
The meaning of the theoretical terms therefore exceeded their objective defini-
tion; these terms had acquired surplus meaning. By using these terms in his
model, Tolman ensured that the model had the virtue of being interpretable in
the above-mentioned way, which made it intelligible (provided that one pos-
sessed the skill of imagining).
For Tolman this model was intelligible. First, he was able to apply the model
to the behavior of the rats, or, in other words, he was able to “do the repre-
senting.” By putting himself in the position of a rat, he was able to judge the
similarities between the model and the phenomenon of rats in mazes and pick
out the relevant features of the model (for example, from everyday life he knew
that demand is a relevant feature). Second, Tolman was able to develop quali-
tative insight into the consequences of the model in concrete situations. His
skill of imagining enabled him to derive useful preliminary hunches that he
could use to see intuitively how, in concrete situations (such and such demands
combined with such and such hypotheses), the rats were expected to behave.
The model is intelligible for Tolman, and thus provides understanding of the
phenomenon of rats in mazes.
However, one might question whether this understanding can be labeled
as scientific understanding. Tolman’s example seems to be rather unscientific,
without reference to theoretical principles. The model does not seem to be
composed of abstract objects defined by theoretical principles, as in Giere’s
() representational view of scientific models. Therefore, one could argue
that this model does not enable Tolman to use theoretical principles in the case
at hand. However, Tolman’s remark—that the graph that formed the basis for
the definition of “hypotheses” was actually the learning curve—reveals, in fact,
that Tolman’s model does incorporate theoretical principles. According to the
learning rule, which is an important principle in neobehaviorism, the learning
curve describes the function between the strength of a stimulus-response con-
nection and the number of “reinforcements” (rewards and punishments). By
k ai eigner
identifying the hypotheses graph with the learning curve, Tolman apparently
considered the hypotheses to be a stimulus-response connection of which the
strength is defined by the principle of the learning rule.
The above analysis of Tolman’s example corresponds with the claim that in-
telligibility is an epistemic value of science. However, Tolman’s assertion—that
he would eventually try to translate the outcome of his methodology into ob-
jective and respectable-sounding terms—indicates that his striving for intel-
ligible models is merely an intermediate step in the development of objective
scientific claims. The final step would consist of a translation of his claims into
objective and respectable-sounding terms without surplus meaning. If Tolman
would succeed in removing the surplus meaning of the terms, the virtues of
the models as described above would also be gone, and accordingly the “right
combination” between the virtues of the model and the skills of the scientists
would be lost.
In actual scientific practice, do neobehaviorists like Tolman rigorously
perform this last step? And, if not, is there a special reason why the surplus
meanings should not be annihilated? In the article “On a Distinction between
Hypothetical Constructs and Intervening Variables,” which caused ardent dis-
cussion among neobehaviorists and logical positivists, MacCorquodale and
Meehl () dealt with these questions. They ascribed to Tolman the position
that theoretical terms should not have surplus meaning, whereas they ascribed
to Hull the opposite position. Tolman () contributed to this discussion
with an article in which he advocated—quite remarkably according to some of
his colleagues—that theoretical terms should have surplus meaning: “I am . . .
convinced that ‘intervening variables’ to which we attempt to give merely op-
erational meaning . . . really can give us no help unless we can also imbed them
in a model from whose attributed properties we can deduce new relationships
to be looked for” (). As he admitted, even Tolman himself was surprised by
his own statement, because for many years he had objected to it. Nevertheless,
he came to realize that it is “inevitable and to be desired” to provide the theo-
retical terms with surplus meaning by framing them in a model that provides a
conceptual substrate—“a substrate which is endowed by its author with certain
intrinsic properties of its own.” (–)
The first part of my case study, about Tolman, highlights the important role
of the surplus meaning of theoretical terms used in models for enabling the
right combination between the virtues of the model and the skills of the scien-
tist, which brings about understanding of the phenomena in the case at hand.
The second part of the case study, about Clark L. Hull, will show that this sur-
plus meaning cannot be removed; it is essential for scientific knowledge, thus
revealing the epistemic relevance of intelligibility.
u nd e r stan d i n g i n p s yc h ol o gy
8aVg`A#=jaa
Hull’s methodology can be seen as an elaboration of Tolman’s objective be-
haviorism. Hull used Tolman’s idea to operationally define theoretical terms
and, like Tolman, he constructed ingenious theoretical models of the adaptive
behavior (or learning behavior) of organisms. Hull’s methodology was very in-
fluential. At the peak of his career Hull was the most widely known behaviorist,
and his methodology dominated psychology (Smith , ). In , Hull’s
magnum opus Principles of Behavior appeared, which was not only “the begin-
ning of an attempt to sketch a systematic objective theory of the behavior of
higher organisms” (Hull a, ) but also a clarification of Hull’s view of sci-
ence and methodology.
Hull’s ideas about science were in many respects closely similar to those of
the logical positivists. He shared the interest in the use of formal logic and was
attracted to British empiricism. Hume’s Treatise of Human Nature especially
fascinated him (Smith , ). Hull regarded behaviorism as a direct de-
scendant from British associationism, of which Hume’s work was exemplary,
and he shared Hume’s admiration for Newton’s scientific method. In a lecture
for the international congress of the Unity of Science Movement of in
Paris, Hull described his scientific procedure as “logical empiricism.” In this
lecture, and also in the lecture for the international congress of the Unity of
Science Movement of in Chicago, where he was an invited speaker, Hull
emphasized the kinship between logical positivism and behaviorism.
Hull’s work can be seen as an attempt to apply the methodology of logical em-
piricism in psychology. In his view, by adopting the logical positivistic ideal
of objectivity, psychology could become a natural science. To put it simply, as
Hull did in his lecture, this “methodology of logical empiricism” has three
k ai eigner
100
UNITS OF HABIT STRENGTH ( S H R )
80
60
40
20
0 5 10 15 20 25 30
SUCCESSIVE REINFORCEMENTS (N)
;^\jgZ&)#)#I]ZaZVgc^c\XjgkZ#=jaa&.)(V!&&,#
16
AMPLITUDE OF GALVANIC SKIN REACTION (A)
14
12
10
0
0 8 16 24 32 40 48
NUMBER OF REINFORCEMENT REPETITIONS (N)
;^\jgZ&)#*#BVc^[ZhiVi^dcd[]VW^ihigZc\i]^cVcZmeZg^bZcil^i]
]jbVchjW_ZXih#=jaa&.)(V!&%(#
k ai eigner
14
12
10
0
0 10 20 30 40 50 60 70 80 90
NUMBER OF REINFORCEMENT REPETITIONS (N)
;^\jgZ&)#+#BVc^[ZhiVi^dcd[]VW^ihigZc\i]^cVcZmeZg^bZcil^i]
VaW^cdgVih#=jaa&.)(V!&%+#
view, the theoretical term “habit strength” can successfully be applied to dif-
ferent domains: he regards his experiment with the rats as a replication of the
experiment with the electric shock—they are both experiments that measure
the functional relationship between habit strength and the number of rein-
forcement repetitions, even though the material realization of the experiments
differs completely. Indeed, the experiments produce similar graphs, but this is
not enough to call them replications of each other. This also requires certain
judgments, for instance, that it is reasonable to suppose that obtaining food
after depressing a bar can be seen as reinforcement, and that it is reasonable to
see the number of unreinforced reactions made by the rats as corresponding to
habit strength. In my view, Hull was able to make these judgments, by means
of which he could apply his behavioral model in different domains, by virtue of
the surplus meaning of the terms in this model.
This goes beyond the neobehaviorist analysis, for in that analysis the con-
nection between model and learning phenomenon is not allowed to depend on
judgments of the user of the model, but instead should be stated in objective
terms by means of operational definitions. For every situation, the neobehav-
iorist ought to provide operational definitions for “stimuli,” “responses,” and
“reinforcement.” In a specific experimental situation concerning humans hear-
ing a noise, “reinforcement” is to be defined as receiving a mild electric shock
after hearing the noise. In another experimental situation concerning rats de-
u nd e r stan d i n g i n p s yc h ol o gy
obtained food after pressing a bar, Hull was able to identify the shocks and the
food as being cases of “reinforcement.” By means of these kinds of identifica-
tions Hull could apply his model to concrete situations.
Second, Hull was able to use the model for developing qualitative insight
into the consequences of the model in concrete situations. Undoubtedly, like
Tolman, Hull could derive useful hunches about this due to the above-men-
tioned skill of imagining. In addition, he could use the causal-mechanical sur-
plus meaning he had allocated to the theoretical terms in the behavioral model.
Because he interpreted “habit strength” as the outcome of a causal mechanism,
of which he had even built a simulation in the form of a mechanical apparatus,
Hull could use his skill of causal reasoning to develop qualitative insight in the
consequences of the behavioral model.
The intelligibility of the models induced by the surplus meaning of the theo-
retical terms is epistemologically relevant. Making the model unintelligible by
removing the surplus meaning (if that is at all possible) would have the conse-
quence that scientists lack the occasion of using their skills to apply the model
to concrete behavior. Except for the small domain in which operational defini-
tions of the theoretical terms are in effect (Hull’s examples), the connection
between model and phenomenon would be broken. Without this connection,
the theoretical principles in the model have no empirical content, which would
make them epistemologically insignificant. The case study about Hull there-
fore shows us that there are epistemological reasons for not removing surplus
meaning. Scientific models need to be intelligible.
8dcXajh^dc
The aim of this chapter is to demonstrate, from scientific practice, that in-
telligibility is one of the epistemic values that are constitutive of science; or to
put it in other words, that understanding is required for scientific knowledge.
The account of understanding used in this chapter is based on the idea that
merely possessing relevant theoretical knowledge is not enough for the scien-
tific understanding of phenomena. In addition, one should be able to use this
knowledge, which implies applying it to concrete cases by means of intelligible
models. A model is intelligible if it can be used to represent the phenomenon
and if its characteristic consequences can be seen intuitively. Whether a model
is intelligible for a user depends not only on the characteristics of the model but
also on the user’s skills.
The point of departure of the case study discussed in this chapter is Tolman’s
concrete example of rats in mazes, by which he exemplified his ideas of an ob-
jective psychological methodology. Because Tolman used his prescientific un-
derstanding of rats as a heuristic guide in formulating an abstract model about
u nd e r stan d i n g i n p s yc h ol o gy
their behavior, the meaning of the terms in the model was initially based on
this prescientific understanding. By applying his methodology, Tolman aimed
at fixing the meaning of these terms by objective definitions. But, as discussion
of his example shows, Tolman did not reach this goal. The terms in the model
could not be divested of all their surplus meaning originating from prescien-
tific understanding. In fact, a close look at Tolman’s example reveals that this
surplus meaning rendered the model intelligible to him. The right combina-
tion of Tolman’s skills (for example, his ability to imagine in the sense of to
empathize) and the virtues of the model originating from the surplus meaning
of the theoretical terms used in the model (such as its anthropomorphic inter-
pretability), enabled him to use the model for representing the phenomena of
rats in mazes and for developing qualitative insight into its consequences in
concrete situations.
The analysis of Tolman’s example illustrates the notion of the intelligibility
of a model. That this intelligibility is essential for scientific knowledge of phe-
nomena is demonstrated by an analysis of the models of behavior in the work of
Hull, who developed a very influential methodology for neobehaviorism, which
was an elaboration on Tolman’s work. The models of behavior in Hull’s work
contain theoretical terms that are defined by theoretical principles of behavior
(for example, by means of the learning curve). The epistemological value of
these models, and thereby of the principles of behavior, depends on the pos-
sibility of applying these models to concrete behavioral phenomena. Without
this possibility the theoretical principles lack empirical content. A close look at
Hull’s work reveals that establishing the connection with concrete phenomena
necessarily involves judgments that are not based on objective methodological
rules, for instance about the similarities of features of the models with features
of the phenomena and about the relevance of these similarities. Hull was able
to make these judgments (seemingly with ease) due to the intelligibility of the
models that, in a way similar to the intelligibility of Tolman’s model of the rat,
accrued through the surplus meaning of the theoretical terms in the models.
Only because the theoretical terms possessed surplus meaning was Hull able
to make the connection between model and phenomenon. This reveals that the
epistemological value of the behavioral models, and thereby of the theoreti-
cal principles of behavior, depends on the intelligibility of these models, which
supports the claim that intelligibility is one of the epistemic values that are
constitutive of science.
One cannot confirm, in a straightforward way, a philosophical claim about
the epistemic role of intelligibility and understanding in science. To elucidate
such a claim by means of empirical evidence from scientific practice requires
well-chosen case studies. The motivation for the choice of neobehaviorism as
a case study lies in the extraordinary positivistic inclination of the neobehav-
k ai eigner
iorists that originated from logical positivistic influences. The pursuit of the
neobehaviorists to set up an objective science of psychology, released from all
subjective elements—which in the end would also leave no room for intelligi-
bility and understanding—can be seen as a test case against the philosophical
idea that understanding plays a fundamental role in science. An analysis of the
scientific practice of the neobehaviorists, who did not reach the goal of an ob-
jective science, reveals that they instead came to realize that the use of intelli-
gible models is inevitable and even to be desired. Their embrace of intelligibility
as an epistemic value forms a corroboration of the claim that understanding is
more than a surplus epiphenomenon. It demonstrates that understanding is a
requirement for scientific knowledge.
CdiZh
I would like to thank Henk de Regt, Hans Radder, and Sabina Leonelli for their helpful
criticism and suggestions, and especially Hans Radder for his constructive ideas about the
nonlocal meaning of theoretical terms.
GZ[ZgZcXZh
Bailer-Jones, D. M. . Tracing the development of models in the philosophy of science.
In Model-based reasoning in scientific discovery, edited by L. Magnani, N. J. Nerses-
sian, and P. Thagard, –. New York: Kluwer Academic/Plenum.
———. . When scientific models represent. International Studies in the Philosophy of
Science :–.
Cartwright, N., T. Shomar, and M. Suárez. . The tool box of science: Tools for the
building of models with a superconductivity example. Poznań Studies in the Philoso-
phy of the Sciences and the Humanities :–.
de Regt, H. W. . Discussion note: Making sense of understanding. Philosophy of Sci-
ence :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
Giere, R. N. . Using models to represent reality. In Model-based reasoning in scientific
discovery, edited by L. Magnani, N. J. Nersessian, and P. Thagard, –. New York:
Kluwer Academic/Plenum.
———. . How models are used to represent reality. Philosophy of Science :–.
Hartmann, S. . Models and stories in Hadron physics. In Models as mediators:
Perspectives on natural and social science, edited by M. S. Morgan and M. Morrison,
–. Cambridge: Cambridge University Press.
Hergenhahn, B. R. . An introduction to the history of psychology. rd ed. Pacific Grove,
CA: Brooks/Cole.
Hull, C. L. . Logical positivism as a constructive methodology in the social sciences.
Einheitswissenschaft :–.
u nd e r stan d i n g i n p s yc h ol o gy
(,
Understanding in Political Science
The Plurality of Epistemic Interests
? : G D : C K6 C 7 D J L : A
u nd e r stan d i n g i n p ol i ti c a l s ci e nce
Jc^ÒXVi^dcVcYJcYZghiVcY^c\
In a traditional view, scientific understanding is linked to unification, or
reducing the number of theories (see Kitcher ). From this point of view,
the question of whether the unification of the different theoretical perspectives
in IR would increase understanding has to be answered positively: “Science
advances our understanding of nature by showing us how to derive descrip-
tions of many phenomena, using the same patterns of derivation again and
again, and, in demonstrating this, it teaches us how to reduce the number of
types of facts we have to accept as ultimate (or brute)” (Kitcher , ). In
contrast, I intend to show that an analysis of social scientific practice and the
reactions of scholars to unificationist attempts indicate the need for a more
nuanced answer to this question. This analysis can provide us with a deeper
insight into scientific understanding in social science, which is different from
the traditional view.
Several projects in the social sciences have attempted to unify the field. One
of the projects I discuss here—and the one that is considered the most success-
ful nowadays—is the incorporation of the social sciences within the framework
of rational choice theory and game theory. In order to test whether Kitcher’s
() view holds, then, we have to ask: do we achieve increased understand-
ing through such an incorporation within a single overarching framework? The
second project I discuss concerns the quest to reduce the plurality of perspec-
tives through a synthesis. Does such a unifying synthesis increase understand-
ing? Let us start with the promises of rational choice theory.
I]ZEgdb^hZhd[GVi^dcVa8]d^XZI]Zdgn
one that is closely linked to economics and the idea of the homo economicus.
“Economics imperialism is a matter of persistent pursuit to increase the de-
gree of unification provided by rational choice theory by way of applying it to
new types of explanandum phenomena that are located in territories that are
occupied by disciplines other than economics” (Mäki , ). As such, it
seems to be the best candidate in the social sciences with which to test Kitcher’s
() view.
Characteristic features of the economics takeover are the development of
formal models for the social sciences, whatever the subject matter; the inclusion
of the notion of maximizing agents, in line with rational choice theory (RCT);
and a general downgrading of the significance of history, culture, and dynamics
(in space and time). Current advocates of the unificationist project include, for
example, Herbert Gintis (see his work, in which he argues that—finally—
the conditions for unity in the behavioral sciences have been created, referring
to the theoretical tools of rational actor model and game theory); Robert Bates,
Rui de Figueiredo, and Barry Weingast, who “explore the possibilities for theo-
retical integration by suggesting how some of the fundamental concepts used
by interpretivists can be incorporated into rational choice theory” (Bates et al.
, ); Margaret Levi, Elinor Ostrom, and James Alt, who “expect the next
century to witness a major flowering of scientific achievement across the social
sciences similar to the neo-Darwinian synthesis of this past century in biology”
(Levi et al. , ).
Advocates of the economics takeover and the application of RCT in po-
litical science might claim that they provide explanatory mechanisms that are
transparent and coherent, and that their explanatory theory is eminently plau-
sible, and so should be confidently accepted, as a means of engendering under-
standing. However, it is questionable whether this unified explanatory theory,
and the class of understanding it purports to convey, are generally accepted as
providing (increased) understanding. Even if this unification induces a kind of
understanding (given some epistemic interests), certain critiques formulated
within political science point out that it does not provide the kind of under-
standing desired by some groups and communities in society.
Let us look at some of these critiques resisting unification. Some critics
start by noting that the economics takeover (that is, increasing the degree of
unification by applying rational choice theory in territories outside economics,
encapsulating political science, sociology, anthropology, history, and so on) has
indeed, especially in political science, led to a growing percentage of journal
articles using the RCT perspective. This has been documented for political sci-
ence in general by Donald Green and Ian Shapiro () and for international
relations studies by Stephen Walt (). These critics then go on to argue that
much of this growing body of formal RCT literature is irrelevant. According to
u nd e r stan d i n g i n p ol i ti c a l s ci e nce
them, it does not answer important societal questions and “real-world prob-
lems.” Concerning the discipline of international relations studies, Stephen
Walt (, ) has drawn the following conclusion:
In this sense, much of the recent formal work in security studies re-
flects the “cult of irrelevance” that pervades much of contemporary
social science. Instead of using their expertise to address important
real-world problems, academics often focus on narrow and trivial
problems that may impress their colleagues but are of little practical
value. If formal theory were to dominate security studies as it has oth-
er areas of political science, much of the scholarship in the field would
likely be produced by people with impressive technical skills but little
or no substantive knowledge of history, politics, or strategy.
today are confronted. But the teaching that is offered, that is to say for
the most part neoclassical theory or approaches derived from it, does
not generally answer this expectation. Indeed, even when the theory
legitimately detaches itself from contingencies in the first instance, it
rarely carries out the necessary return to the facts. . . . Furthermore,
this gap in the teaching, this disregard for concrete realities, poses an
enormous problem for those who would like to render themselves use-
ful to economic and social actors.5
I]Z>YZVd[VHnci]Zh^h
The thrust of the questions is that dialogue and synthesis are all of
one cloth, that everything can be debated out, and that some integral
new whole is likely to emerge to command our assent if we all do our
homework. But considering how seldom debates establish such con-
sensus, it might be useful to take this experiential datum as a start-
ing point and to inquire into the reasons why communication across,
and often even within, different theoretical perspectives is so difficult.
(Kratochwil , )
I]ZGdaZd[:e^hiZb^X>ciZgZhihVcY6YZfjVXn^cJcYZghiVcY^c\
Bouwel and Weber ; Weber and Van Bouwel ). There I emphasized
the importance of taking epistemic interests into account when making a
choice among different forms of explanation (and the plurality of theories
from which these explanations derive) and regarded explanations as answers
to explanation-seeking why-questions, the formulation of which helps to make
the explananda as explicit as possible and to draw attention to the underlying
epistemic interests of the explainee. From this viewpoint, a good explanation
should not be understood as merely the accurate (in relation to reality) ex-
planation of a social phenomenon, but also as an adequate (in relation to the
epistemic interests of the explainee) answer of an explanation-seeking ques-
tion concerning a social phenomenon.8 Usefulness depends on adequacy as
well as accuracy.
Analogous considerations have to be made when talking about understand-
ing. If we want to get a grip on what achieving understanding requires in social
science, attention should be paid to the (variety of ) interests of the understand-
ees.9 Understanding does not depend merely on the accuracy of the explana-
tions provided by a theory but also on its adequacy in relation to the interests
of the understandee.
In the objectivist view of understanding of Trout () this distinction
between accuracy and adequacy is missing. For Trout the accuracy suffices
to provide understanding. De Regt (, ) rightly criticizes Trout’s view
and pleads for the inclusion of pragmatic elements: “Understanding is not only
knowing the formula, but in addition being able to use the formula in the case
at hand.” Accuracy is not the only criterion: “Scientists prefer a more intelligible
theory over a less intelligible one, sometimes even at the cost of some accuracy,
not because it gives them a ‘feels right’ sense of understanding but rather be-
cause they have to be able to use the theory” ().
Following this approach, scientific theories provide understanding of the
social world if they offer the understandees an adequate answer to their prob-
lems, that is, one that that gives an idea of what to do, how to act to solve a
problem. To do this, they must fit with the interests of the understandees and
have a certain familiarity within their life-worlds and personal experiences.
This familiarity ensures that the adequate theory is accessible to the under-
standee, and usable given the questions to be answered or the problems to be
solved. It enables individuals and social groups to intervene in the social world
using the theory in order to reach their goals (empowerment, emancipation,
social benefits, and so on). The theories that help people to understand social
phenomena will thus provide more than merely theoretical knowledge.10
The overview of debates above provided us with several examples of how ac-
curate theories might be irrelevant for some groups, and how attention should
u nd e r stan d i n g i n p ol i ti c a l s ci e nce
;Zb^c^hbVcYGVi^dcVa8]d^XZI]Zdgn
These remarks are important in order to understand how a pluralist view of un-
derstanding has to be conceived so as to leave unificationist attempts behind.
A third group of feminists has developed a balanced view, which does not
reject RCT outright, but limits its usefulness to specific cases, arguing that in
other cases, alternative theories will be more adequate. The selection of the
most adequate theory is made by comparing theories and taking interests and
usefulness into account. A good example can be found in Anderson ().
Analyzing the feminist angle on health care policies taken by Kristin Luker’s
Taking Chances, Anderson reaches the conclusion that the answer to her title,
Should Feminists Reject Rational Choice Theory?, “depends on our purposes,
and on the aspect of RCT being used” (Anderson , ). She believes RCT
offers both resources for and obstacles to understanding the problems women
face and helping them overcome them.
Rational choice theory can be used to answer questions addressing feminist
interests in the following way:
vations that deter many sexually active women from using contracep-
tion, even when they have an interest in doing so. Here the function of
rational choice theory is to make women’s choices intelligible in terms
of their motivations, so the health-care system responds to women’s
needs as they see them. (Anderson , ; italics mine)
An example of the benefits of using RCT would be: “Luker suggests that once
health-care practitioners recognize that women have a stake in seeing their
sexual activity as spontaneous, and tend to stop contraception between rela-
tionships, they can respond by providing drop-in contraceptive services, rather
than requiring women to make appointments weeks in advance” (–).
On the other hand, rational choice theory also presents obstacles to under-
standing the problems women face. Only using RCT precludes the identifica-
tion of obstructions to women’s rationality and autonomy. Anderson does not
agree with Luker’s use of RCT to vindicate the rationality of women:
Anderson’s analysis shows us how, depending on the specific interests one has
(and the research questions these generate), different theories can be the most
adequate. Understanding that the existing plurality of theories in social sci-
ence can help to answer questions generated by a plurality of interests should
make us cherish theoretical pluralism rather than attempt unification. (This
does not imply that every existing theory is indispensable; comparing theories
when addressing specific questions will enable us to distinguish good from bad
theories. It does, however, entail that reducing the plurality of theories to one
unifying theory results in a loss of opportunities to achieve understanding.)
Anderson’s balanced analysis illustrates how the winner-takes-all approach can
be left behind; the proposed unification under RCT does not provide under-
standing for all at all times (independent of interests and context), but it does
serve some epistemic interests. Moreover, not rejecting RCT (and its ontologi-
cal commitments) outright solves the problem raised in relation to the first
feminist position, where questioning the rationality of the individual risks un-
dermining the emancipatory objectives of feminism. And it is precisely this
that feminism pursues: “Feminist theory . . . tries to understand the social world
so as to enable certain kinds of liberating changes in it” (Anderson , ;
italics mine).
6EajgVa^hi^XK^Zld[JcYZghiVcY^c\^ci]ZHdX^VaHX^ZcXZh
stand their problems. For instance, feminism theorizes in ways that women can
use to improve their lives; it implies that women should be able to recognize
themselves and their lives in feminist accounts of women’s predicaments (see
Anderson , ).
Feminism gives us a good illustration of the impact of interests on the se-
lection of theories to achieve understanding. Its relation to rational choice
theory—the best candidate in the social sciences to refute Kitcher ()—was
discussed, and I concluded that if we want to maximize understanding, the
ideal of unification (be it by RCT or an alternative theory) is not desirable be-
cause of the plurality of epistemic interests to be addressed. Even if unification
conveys understanding for some (relative to their interests at a given time), it
will always obstruct understanding for others (with different interests). Recog-
nizing the plurality of possible epistemic interests will help us to move on from
the winners-takes-all approach to social science (as was shown by Anderson’s
evaluation of RCT), and to stop regarding the plurality of theoretical perspec-
tives—present in international relations studies—as a problem. On the con-
trary, theoretical pluralism is a strength in achieving understanding.
CdiZh
. These data and the methodology used to obtain them can be found in Peterson,
Tierney, and Maliniak ().
. The term “imperialism” might sound too negative, but actually, one of the pioneers
of the approach, Gary Becker (), whose work is the locus classicus, agrees with this
label: “This definition of ‘economic imperialism’ is probably a good description of what I
do” (Swedberg , ).
. Robert D. Putnam warns us—in relation to irrelevance—of the dangers of policy re-
search migrating toward schools of public administration, just as happened with practical
economic studies, which changed from the economics department to business schools: “If
one compares the size of economics departments and business schools in today’s academy,
the cost of reducing a social science to sterile theoretical endeavors is obvious” (quoted on
www.paecon.net, consulted February ).
. The increasing “colonization” of political science by economists has led to so much
discontent that an anti-imperialist movement, named “Mr. Perestroika,” has been created.
The movement was initiated in the year , with a mass e-mailing by Mr. Perestroika.
Schram (, ) discusses the Mr. Perestroika movement in political science: “the ways
in which contemporary social science all too often fails to produce the kind of knowledge
that can meaningfully inform social life.” The Perestroikans’ main focus is that major
journals in the field have been preoccupied with publishing research that conforms to the
economics takeover features.
. Italics mine; this English version of the French petition can be found on
www.paecon.net, consulted February . Out of this petition arose a broad movement
called “post-autistic economics.” It has a lot in common with the Mr. Perestroika move-
jer o en van b o u wel
GZ[ZgZcXZh
Anderson, E. . Should feminists reject rational choice theory? In A mind of one’s
own: Feminist essays on reason and objectivity, edited by L. M. Antony and C. E. Witt,
–. Boulder: Westview Press.
———. . Feminist epistemology and philosophy of science. Stanford encyclopedia of
philosophy. Available at http://plato.stanford.edu/entries/feminism-epistemology.
Bates, R., R. de Figueiredo, and B. Weingast. . The politics of interpretation: Rational-
ity, culture, and translation. Politics and Society :–.
Becker, G. . The economic approach to human behavior. Chicago: University of Chi-
cago Press.
Bohman, J. . Theories, practices, and pluralism: A pragmatic interpretation of critical
social science. Philosophy of the Social Sciences :–.
Davis, J. . The theory of the individual in economics: Identity and value. London:
Routledge.
de Regt, H. W. . Discussion note: Making sense of understanding. Philosophy of Sci-
ence :–.
Giere, R. N. . Perspectival pluralism. In Scientific pluralism, edited by S. Kellert, H.
Longino, and K. Waters, –. Minneapolis: University of Minnesota Press.
u nd e r stan d i n g i n p ol i ti c a l s ci e nce
Gintis, H. . Towards the unity of human behavioral sciences. Politics, Philosophy and
Economics :–.
Green, D., and I. Shapiro. . Pathologies of rational choice theory. New Haven: Yale
University Press.
Hellmann, G. . In conclusion: Dialogue and synthesis in individual scholarship and
collective inquiry. International Studies Review :–.
Hempel, C. G. . Aspects of scientific explanation. In Aspects of scientific explanation
and other essays in the philosophy of science, –. New York: Free Press.
Kitcher, P. . Explanatory unification and the causal structure of the world. In Scientific
explanation, edited by P. Kitcher and W. C. Salmon, –. Minneapolis: University
of Minnesota Press.
Kratochwil, F. . The monologue of “science.” International Studies Review :–.
Levi, M., E. Ostrom, and J. Alt. . Conclusion. In Competition and cooperation: Con-
versations with Nobelists about economics and political sciences, edited by J. E. Alt, M.
Levi, and E. Ostrom. New York: Russell Sage.
Mäki, U. . Symposium on explanations and social ontology : Explanatory ecumenism
and economics imperialism. Economics and Philosophy :–.
Meyers, D. . Self, society, and personal choice. New York: Columbia University Press.
Moravcsik, A. . Theory synthesis in international relations: Real not metaphysical.
International Studies Review :–.
Nedelsky, J. . Reconceiving autonomy: Sources, thoughts, and possibilities. Yale Jour-
nal of Law and Feminism :–.
Peterson, S., M. J. Tierney, and D. Maliniak. . Teaching and research practices, views
on the discipline, and policy attitudes of international relations faculty at U.S. colleges
and universities. Available at http://mjtier.people.wm.edu/TRIP summary Aug .pdf.
Schram, S. . Return to politics: Perestroika and postparadigmatic political science.
Political Theory :–.
Shapiro, I. . Problems, methods, and theories in the study of politics, or what’s wrong
with political science and what to do about it. Political Theory :–.
Smith, S. . Dialogue and the reinforcement of orthodoxy in international relations.
International Studies Review :–.
Swedberg, R. . Economics and sociology: Redefining their boundaries: Conversations
with economists and sociologists. Princeton: Princeton University Press.
Trout, J. D. . Scientific explanation and the sense of understanding. Philosophy of Sci-
ence :–.
Van Bouwel, J. . Explanatory pluralism in economics: Against the mainstream? Philo-
sophical Explorations :–.
Van Bouwel, J., and E. Weber. . Remote causes, bad explanations? Journal for the
Theory of Social Behaviour :–.
Walt, S. . Rigor or rigor mortis? Rational choice and security studies. International
Security :–.
Weber, E., and J. Van Bouwel. . Can we dispense with structural explanations of social
facts? Economics and Philosophy :–.
pa rt title ver so
(-
Understanding in Historical Science
Intelligibility and Judgment
:9L>C@DHI:G
How historians relate to their own time, what are their innermost feelings
and experiences, what have been the decisive facts of their own lives—
these are all things that should not be distrusted and feared as threats to
so-called historical subjectivity but cherished as historians’ most crucial
asset in their effort to penetrate the mysteries of the past. ()
u n d e r sta n d i n g i n h i st o r i c a l s c i e n c e
Can the tension between these views be eliminated by stressing that the
first one is about the acceptance of knowledge in empirical sciences such as
physics and the second about the role of understanding in the study of his-
tory? Or is it better to stress that these different views are not typical of physics
and history, and that the choice of other representatives of the philosophy of
these disciplines would have given other positions? To answer these questions,
insight into the phenomenon of understanding in history in particular and in
science in general is necessary.
My aim in this chapter is twofold: I first want to explore how understanding
is obtained in the study of history and then, second, to specify the concept of
scientific understanding. I will begin by summarizing the conception of sci-
entific understanding in physics defended by de Regt and Dieks. After noting
some problems regarding de Regt and Dieks’s conception and their criteria for
achieving understanding, I will introduce the concept of historical understand-
ing. It will become clear that the tension between objectivity and subjectivity
is a general feature in the concepts of scientific and historical understanding.
By focusing on the notion of intelligibility, I will show that intentional explana-
tions, colligatory concepts, synoptic judgments, and historical narratives are
modes of understanding in history—modes in which the “nonobjective crite-
ria” (the “subjective dimension”) cannot be eliminated. I will explore the nature
of these nonobjective criteria at great length. These theoretical considerations
will be illustrated by examples taken from actual historiography of the First
World War. On the basis of discussions in this field, I will try to clarify the
notions of historical understanding and scientific understanding through the
concept of judgment.
JcYZghiVcY^c\^cE]nh^Xh
JcYZghiVcY^c\VhVc>cYZÒc^iZ8dcXZeiWZilZZcDW_ZXi^k^in
VcYHjW_ZXi^k^in
to their motives and ideas. On the other hand, Dilthey makes clear that the
understanding of historical agents also has to do with the verifiable inquiry of
documents in which their intentions are expressed (Plantinga , –; de
Mul , –, –, –, –).
Various scholars have proposed a way out of this ambiguity. Although they
are not univocal in their interpretations, they agree that Dilthey did not use
the term “objectivity” as synonymous with “universal validity.” According to
them, such a reading is impossible because of Dilthey’s emphasis on human
finitude and historicity. They further suggest that it would have been better if
Dilthey—speaking of the human sciences—would have used the concept of “in-
tersubjectivity” instead of “objectivity” because the former concept recognizes
the subjective nature of historical understanding without characterizing it as
arbitrary (Plantinga , –; de Mul , –).
With this reflection on the first phase of modern historiography, I have
gathered presumptive evidence for the claim that the tension between objec-
tivity and subjectivity is a general feature of the concept of understanding. Per-
haps the subjective dimension is to blame for the vagueness of this concept—a
vagueness to which de Regt and Dieks do not pay much attention. The plausi-
bility of this claim and the tenability of this suggestion will become clear in the
remainder of this chapter. But before gaining better insight into the concept
of understanding—much of the meaning of this concept is still implicit—we
first have to remove a source of misunderstanding. Understanding (in history)
is sometimes exclusively put on a par with our capacity to empathize with the
experiences of historical agents.
A classic expression of this identification is R. G. Collingwood’s (–)
term “reenactment,” one of the best-known aspects of his hermeneutics of his-
tory. According to Collingwood, to make history intelligible, historians reen-
act the experiences of the historical agents, that is, they empathize with their
thoughts and think through their actions (D’Oro , sec. .). Collingwood
(, ) even claims that in the act of reenactment the historian does not
think thoughts similar to those of the historical agent but the same thoughts:
“the process of argument which I go through is not a process resembling Plato’s,
it is actually Plato’s so far as I understand him correctly.” Sometimes reenact-
ment is incorrectly identified with (historical) understanding. We have already
noted that Dilthey has a broader concept of historical understanding because
he does not confine it to immediacy.
Hans-Georg Gadamer’s (, –) criticism of Collingwood’s theory
of reenactment is explicitly clear on this point. His ideas regarding the well-
known fusion of horizons and the Wirkungsgeschichte demonstrate that the gap
between the time of the historical agent and the time of the historian cannot be
bridged. This does not mean that, according to Gadamer, an understanding of
edwin koster
the past cannot be achieved: by virtue of the historian’s awareness of his or her
own situatedness, it is possible to have access to the individuality of the past.
However, Gadamer does not offer a specific method for achieving an under-
standing either of history or of (historical) texts. Truth and Method (Wahrheit
und Methode), Gadamer’s classic in the field of twentieth-century philosophi-
cal hermeneutics, can be considered a severe critique of the idea that truth and
knowledge are connected exclusively to an unambiguous method (Gadamer
, xi–xii). From the perspective of the adherents of such a method one
could even assert that Gadamer claims that the concept of understanding is an
intrinsically loose term. The plausibility of this view and, with this, the claim
that it is impossible to specify the slightly vague notions used by de Regt and
Dieks, will be tested below.
I]Z>ciZaa^\^W^a^ind[=^hidgn
Thus far, I have shown that there are similarities between the concept of un-
derstanding in physics and history: () the tension between objectivity and sub-
jectivity and () the vagueness of the concept. This supports de Regt and Dieks
because they claim to formulate a “generally applicable criterion for the attain-
ment of understanding.” However, one may doubt whether their conception
of scientific understanding can be applied to the practice of historians. They
describe the notion of intelligibility, for example, in connection with theories,
and they formulate a criterion for the intelligibility of theories in terms of the
absence of precise calculations. How are these ideas to be applied to the field of
history? Many scholars will agree that even if one assumes a strong connection
between history and the social sciences, the development of theories is rare in
the study of history. The same obtains for making exact calculations.
However, historians do refer to the notion of intelligibility. One of the ar-
ticles by the well-known philosopher of history W. H. Walsh is called “The In-
telligibility of History.” It is concerned with the question, among other things,
of whether it is possible to find intelligible connections between historical phe-
nomena, such as actions and events (Walsh , ). Other than in physics,
the intelligible connections in history are not meant to be extrinsic: according
to Walsh and many others, these connections are not causal. Walsh wants to
show that some sort of intrinsic relation exists between different historical phe-
nomena in a certain period of the human past. According to Walsh, this con-
nection is most obviously present in the case of actions: “two actions may be
connected as parts of the realization of a single consistent policy, and the ideas
behind them . . . may reciprocally determine each other. . . . the earlier actions
themselves . . . will take the form they do because the agent intended to carry
out certain other actions later” ().
u n d e r sta n d i n g i n h i st o r i c a l s c i e n c e
Actions are thus intrinsically related, rather than causally. Insight into
these connections makes history intelligible, writes Walsh, and can, I would
add, clarify the concept of historical understanding. Next to the reconstruc-
tion of a web of related actions, the historian tries to make a coherent whole of
the events of a selected period by putting them together under certain leading
ideas. To describe this activity of the historian, Walsh borrows a term made
famous by William Whewell and Mill: the historian “colligates” different events
according to “appropriate conceptions” such as “the Industrial Revolution” and
“the Enlightenment” (Walsh , ). An analysis of the role colligatory con-
cepts play in history throws light on the notion of historical understanding as
well. Walsh’s insights lead to at least two possible modes of understanding in
history: intentional understanding to gain insight into the actions of the histor-
ical agents and understanding by way of forming colligatory concepts to gain
knowledge of the way events are connected.
=^hidg^XVaJcYZghiVcY^c\VcY>ciZci^dcVa:meaVcVi^dc
Many people will agree that to understand, for instance, the events that
together constitute the First World War, it is necessary to acquire knowledge
about the political developments that gave rise to the outbreak of the Great
War, the interests of the superpowers at the time, the military strategies of the
armies from to , and the consequences of the Versailles Treaty in
. These macro-level developments form, of course, the indispensable in-
gredients of every historical attempt to understand this crucial period in Eu-
ropean history.
Although knowledge at this level is considered important, historians have
shown great interest in the documents that express the emotions, feelings, and
thoughts of the soldiers fighting in the trenches or recovering in field hospitals.
The insights offered by documents recalling the war, such as war poetry, novels,
letters, long-lost memoirs, diaries, and regimental histories written by partici-
pants—the micro-level of the First World War, so to speak—resulted in the last
decades of the twentieth century in a number of movies, the foundation of the
Great War Society (whose members try to reenact the war), and new histo-
ries of particular battles based on the reports of eyewitnesses (Stummer ).
An example of this last instance is the work of the historian Lyn Macdonald
on the British Tommies and German foot soldiers—who called themselves
Frontschweine—in the trenches of Flanders. By interviewing the survivors and
then making use of memoirs and studies of the terrain of the Western front,
she succeeds in sketching a lively picture of the war at trench level (Macdon-
ald , ). All this reveals the widespread belief that only the stories of
the historical agents can give insight into this traumatic event. In other words,
edwin koster
is indispensable for the explanation of individual agency, this does not entail
commitment to the claim that reenactment or empathy is the only method
historians should use. Stueber’s approach gives insight into the necessarily sub-
jective character of historical understanding.
It is sometimes suggested that our capacity to understand one’s own and
somebody else’s behavior is based on psychological theories. Therefore, the
concept of empathy would not be of any importance: to understand the mo-
tives of a historical agent, it would be enough to apply these general theories
to particular situations. However, how can one decide which theory is to be
applied in a certain situation? For that purpose, a general theory stating which
aspects of a specific situation are epistemically relevant seems to be necessary.
The existence of such a theory, Stueber writes, is not very likely. And even if
we had such a theory at our disposal, the right application of the theory in a
particular situation could not be guaranteed: there are no formal principles
that state how one is to use such a general theory. Therefore, one may look for
other theories that do state how the general theory is to be applied correctly
in different circumstances. But then we are threatened with an infinite regress
of theories and this, according to Stueber (, ), “can only be avoided by
appealing to a non-theoretical capacity.” This nontheoretical capacity is the hu-
man capacity for empathy: we can understand the intrinsic relation between
someone’s intentions and motives on the one hand and his or her actions on the
other only if we understand how these intentions and motives would lead us to
performs the same actions ().
It should be noted that our access to, for instance, complaints about trench-
feet or the short-arm inspection of soldiers serving during the First World War
is almost never immediate. This is not only due to the implications of herme-
neutical notions such as Wirkungsgeschichte—insofar as we believe that they
are tenable—but also because we know that the posthumous writings of the
men in the trenches are to a certain extent constructed: it has been shown
that popular war novels written by eyewitnesses functioned as templates for
soldiers and civilians to shape their own memories. The deliberate adaptation
of their own memories to the versions presented in these war fictions was so
natural that “the military doctor/novelist Georges Duhamel noted in that
‘if his former patients were to read today their own stories, they would rarely
recognize them’ ” (Shapiro, , –)—an indirect proof of Gadamer’s the-
sis, so to speak. Therefore, it is an illusion to think that empathy, important as
it is, takes us directly to the soldiers’ experiences.
To conclude, the understanding of intentions requires the act of empathy,
which cannot be eliminated in favor of the “objective application” of a set of
theories. However, the historian has to be aware that the experience of seeing
through the eyes of a historical agent might be illusionary. Here the historian
edwin koster
beyond the control of the human agents involved. One way to accomplish this
task is through the invention and application of colligatory concepts.
8daa^\Vidgn8dcXZeih!Hncdei^X?jY\bZcih!VcYCVggVi^kZ:medhjgZ
With colligatory concepts we make the transition from the first mode of
understanding—intentional understanding related to the actions of historical
agents—to the second mode, which is about the relation between historical
events. When historians consult their sources, they are confronted with a mass
of largely unrelated materials. It is their task to organize the data by showing
that they reveal certain patterns (themes or developments) that they judge to
be significant for the selected period. These patterns are seen as phases in a
continuous process and together they can be expressed by colligatory concepts
such as “the Renaissance” and “the First World War.”
According to Walsh (, –), colligation is primarily a stage in his-
torical interpretation because this operation is used principally to characterize
and analyze a certain period in history and, in doing so, colligatory concepts
produce a sense of understanding. However, as Walsh admits, the use of col-
ligatory concepts also has an explanatory function: by displaying how an event
fits into a development, the historian shows why it happened as it did (Walsh
, ; compare Von Wright , ). This again makes it clear that there
is no opposition between understanding and explanation in history.
Which conditions should govern the choice for certain colligatory con-
cepts? According to Walsh, two conditions must be fulfilled. In the first place,
a colligatory concept has to fit the facts: the statements that embody the con-
cept must be supported by evidence. To call a certain period in history “the
First World War,” for instance, it is thus necessary to show that our knowledge
of this period, the facts, can be organized into a picture of a unique and all-
embracing war. Second, it must illuminate the facts so that the use of this con-
cept increases our understanding of the selected period (Walsh , –).
Mink () explains what this last condition means. Mink makes a distinc-
tion between knowledge on the one hand and comprehension or understand-
ing on the other. We know many things as unrelated facts, such as the way
Franz Ferdinand died, the date of birth of the author of The Guns of August,
or the name of the capital of Germany. To know these facts is not to com-
prehend them. Comprehension, Mink writes, “is an individual act of seeing-
things-together, and only that. It . . . [is] the human activity by which elements
of knowledge are converted into understanding” (). This act of “seeing events
together” that are temporally and spatially separate is essentially the same as
the process of colligation. Mink calls it a “synoptic judgment” and, according
to him, this form of judgment is typical for the phenomenon of historical un-
edwin koster
But unlike these earlier contributions, De Mey and Weber emphasize that
thought experiments lose their explanatory power if they are not supported
by empirical data. They claim that premises such as “trench warfare is posi-
tively causally relevant for truces,” can be empirically supported (–). And,
indeed, they show by means of several examples that contrastive reasoning is
more than plain fantasy.
However, in my view, they overemphasize the role of generalizations that
can be tested empirically (but not experimentally). De Mey and Weber (,
–) present the determination of factors that are “more important” than
other factors as an almost entirely objective process. Only incidentally do they
acknowledge that contrastive reasoning is dependent on the knowledge of the
specificity of particular situations and on the possibility of accounting for rele-
vant values that are not always quantifiable (). I agree that the notion of “more
important than” is accessible to analysis and that empirical support has a vital
part to play here. But to determine the factors that, for instance, bring about
the truces, practical experience is needed as well. Thus, contra the arguments
of De Mey and Weber, and the suggestions of Ricoeur (, ), I maintain
that this is not an objective process and that subjective deliberations—though
not irrational or a matter of taste—are therefore inevitable. Thought experi-
ments are the tools (subjective, but not arbitrary) for determining the most
relevant factors.
It thus follows that a synoptic judgment, in which knowledge is converted
into understanding, must partly be characterized by subjective—but not arbi-
trary—considerations. Perhaps the best way to illustrate the transition from
historical knowledge to historical understanding—to borrow Mink’s terminol-
ogy once again—is to look at the process of constructing a historical narrative.
From the documentary phase, in which the historical archives serve as sources
for a set of unrelated facts, through the phase in which actions and events are
explained and understood by way of synoptic judgments, the historian arrives
at the representative phase: the writing of history through the production of a
narrative. Without claiming that these phases can be identified with distinct
chronological stages, the subdivision of the work of the historian in these inter-
woven methodological moments helps in recognizing the special features of the
concept of historical understanding (Ricoeur , –). Why, one can ask,
is the representative phase confined to narrative discourse? The answer is that
synoptic judgments have their written complement in the plot of a narrative.
As Ricoeur (, x) writes, “The plot of a narrative . . . ‘grasps together’ and
integrates into one whole and complete story multiple and scattered events.”
Therefore, the attainment of a synoptic judgment and the act of constructing a
plot are similar creative processes.
A beautiful example of the narrative exposure of a synoptic judgment is the
edwin koster
first chapter of Barbara Tuchman’s book on the beginning of the First World
War, The Guns of August. In this chapter she connects various developments
taking place at the royal courts during the first decade of the twentieth cen-
tury by describing in a grand style the funeral of Edward VII of England. She
makes a reasonable case for the important influence of the personal diplomacy
of Edward, made possible by his great gifts as a sociable king and by his fam-
ily ties: he was related to many members of Europe’s ruling houses, including
Kaiser Wilhelm of Germany and the Dowager Empress Marie of Russia. Tuch-
man writes that, when the procession of kings was passing by, the crowd “could
not keep back gasps of admiration” but also “felt a finality in the slow tread
of the marchers and in the solemn music played by the Royal Horse Guards’
band” (Tuchman , –). In this way, she renders a lively depiction of a
culture that would change enormously after August but was also largely
responsible for the outbreak of the First World War. That is why her historical
narrative can be seen as a synoptic judgment.
=^hidg^XVaJcYZghiVcY^c\VcYi]Z8dcXZeid[?jY\bZci
GZXdch^YZg^c\i]Z8dcXZeid[HX^Zci^ÒXJcYZghiVcY^c\
that the differences are negligible between the way in which the concept of
judgment is used in history and how it is used in the natural sciences, then we
can also indicate points of resemblance with the concept of understanding as
presented by de Regt and Dieks ()—for instance, the absence of algorithms
and the comparable role played by rules of thumb and skills. Regarding skills,
Brown () states, for example, “When we develop the ability to exercise
judgment in a particular field, we are developing a skill” (). And, as we have
seen, skills are also important to the approach of de Regt and Dieks.
The conditions to exercise judgment, such as being acquainted with com-
parable (for instance, historical or experimental) circumstances, come close
to their use of the notion of “familiarity.” However, it seems to me that a re-
description of the concept of scientific understanding in terms of “judgment”
is less vague than the definition of this concept in terms of “good sense” and
“intuition,” as de Regt and Dieks propose (compare de Regt, this volume, for an
attempt along these lines). I claim that, unlike the use of the latter terms, the
application of the notion of judgment gives insight into the way a scientist puts
together empirical evidence, (more or less) objective criteria, and subjective
considerations (even though based on sound reasons).
Of course, de Regt and Dieks have a great deal more to say about the con-
cept of understanding, so I am a bit unfair in my treatment of their position.
Moreover, the concept of understanding seems to possess a sense of elusiveness
that cannot be avoided. So, why does the vagueness of their concept constitute
a problem? Further, the imprecise character seems to be due to the subjective
dimension of the concept of understanding, and the presence of this dimension
in historiography has been sufficiently demonstrated in this chapter. Again, is it
not in de Regt and Dieks’s favor that they speak of nonobjective criteria and fo-
cus on the “not purely subjective” character of these criteria? It is and it is not.
It is in their favor because they rightly emphasize all this; but it is not enough,
for they do not provide sufficient insight into the subjective dimension of un-
derstanding and they do not discuss the possibility of eliminating distorted as-
pects of subjective preferences.
Another shortcoming in their approach is that they seem to confine the
concept of scientific understanding to the application of theories. They do not
say much—as far as I can see—about the possible role understanding plays in
the construction of theories and explanations. This may be due to the context
in which they raise their questions; that is, they start with the fact that we have
different kinds of explanations at our disposal and then ask how these expla-
nations provide understanding. They do not pose the question of how these
explanations are formed. However, de Regt and Dieks do not seem to be con-
cerned with the role of understanding in the formation of theories. In contrast,
in historiography the concept of understanding is central to its practice: it is
edwin koster
GZ[ZgZcXZh
Ankersmit, F. R. . Sublime historical experience. Stanford: Stanford University Press.
Bauer, H. H. . Scientific literacy and the myth of the scientific method. Urbana: Univer-
sity of Illinois Press.
Bernstein, R. J. . Beyond objectivism and relativism. Oxford: Blackwell.
Brown, H. I. . Rationality. London: Routledge.
———. . Judgment, role in science. In A companion to the philosophy of science, edited
by W. H. Newton-Smith, –. Oxford: Blackwell.
Cartwright, N. . How the laws of physics lie. Oxford: Clarendon Press.
Collingwood, R. G. . The idea of history. Edited by J. Van der Dussen. Oxford: Oxford
University Press.
De Mey, T., and E. Weber. . Explanation and thought experiments in history. History
and Theory :–.
de Mul, J. . The tragedy of finitude: Dilthey’s hermeneutics of life. New Haven: Yale
University Press.
de Regt, H. W. . Discussion note: Making sense of understanding. Philosophy of Sci-
ence :–.
de Regt, H. W., and D. Dieks. . A contextual approach to scientific understanding.
Synthese :–.
D’Oro, G. . Robin George Collingwood. Stanford encyclopedia of philosophy. Available
at http://plato.stanford.edu/entries/Collingwood.
Fromkin, D. . Europe’s last summer: Who started the Great War in ? New York:
Knopf.
Gadamer, H.-G. . Truth and method. New York: Seabury Press.
Hempel, C. G. . The philosophy of Carl G. Hempel: Studies in science, explanation, and
rationality. Edited by J. H. Fetzer. Oxford: Oxford University Press.
Joll, J. . The unspoken assumptions. In The origins of the First World War: Great power
rivalry and German war aims, edited by H. W. Koch, –. London: Macmillan.
Kuhn, T. S. . The essential tension: Selected studies in scientific tradition and change.
Chicago: University of Chicago Press.
Macdonald, L. . They called it Passchendaele: The story of the battle of Ypres and of the
men who fought in it. London: Penguin Books.
———. . To the last man: Spring . New York: Carroll and Graf.
u n d e r sta n d i n g i n h i st o r i c a l s c i e n c e
:FEKI@9LKFIJ
contribu tor s
inde x
; pairing with ontological principles, – explanation, contrastive, and principle of
, t, –; and scientific understanding, sufficient reason, t, , ,
–, , , –, explanation, mathematical, , –, , ,
epistemic humanism, –, , –. See also abstract
epistemic interests, and adequacy of mathematical models
understanding, , – explanation vs. description. See description, vs.
epistemic iteration, simulation-based explanation
understanding and, explanation in economics, –
epistemic skills: acquisition of, ; assessment explanation effect,
of, –; in biology, –; defined, , explanation in engineering sciences,
; types of, – explanation in historical science, as intrinsic to
epistemic value of feeling of understanding. understanding, –,
See feeling of understanding, reliability of explanation in physics: as pragmatic
epistemic value of models, –, –, (functional), , , –, . See also
, n constructive theories; principle theories
epistemic values constitutive of science, explanation vs. understanding, –; in
intelligibility and, , , – economic models, ; history of views on,
epistemology: feminist, ; and ontology, –, ; presence of subject and,
; and philosophy of science, -, ; explanatory claims, making explicit,
social, explanatory power: ambiguity of concept, ;
Evans, R., n vs. evidence, confusion of, –
event generators, – explanatory practices: improvement of, –
evidence vs. explanatory power, – ; lack of attention regarding, in sciences,
evolution: and psychology, , ; –, , –
understanding of, -, -, n, explanatory understanding, , , , ,
, –, ,
Ewens, W., n externality/objectivity, and observation,
exemplars, , –, ,
expectation, and understanding, , , –, familiarity, , , , , , –,
– ; feeling of understanding and, , –,
experience: cognitive, from performing
research, -, -; everyday, - feeling of knowing,
, ; personal, of historical agents, , feeling of understanding, –, , –, ,
, -; practical, -, ; from n; as ability or faculty, ; coherence
performing simulations, , , , . and, –, n, ; compared to
See also feeling of understanding other feelings, –; as motivating cause
explanandum: ambiguity in, ; contrastive of understanding, –, , –,
explanations of, ; in explanation, , –, –; necessity of relying on, ; as tacit
, , , , , –, guide to theory selection, –, ; vs.
explanans: ambiguity in, ; in explanation, , understanding, , ; understanding
-, , , , , without, , n
explanation: ambiguity of concept, –; feeling of understanding, reliability of, –,
cognitive benefits of, , –, –; , ; anecdotal evidence, –; clinical
defining of, , ; evaluating quality of, research on, –, –; conditional
–; history of philosophical treatment nature of, , –, , n; contextual
of concept, ; imprecision in explanandum factors, –, ; manipulation and,
or explanans, ; indeterminate end states ; methods of improving, ; and
of, , ; interpretive structures and, overconfidence bias, –; sources of
–, –; mechanism of providing unreliability, –. See also illusion of
understanding, –, , ; models and, depth of understanding
–, ; need for, abnormal events as Feigl, H.,
signal of, –; objectivism on, –; feminism, views on rational choice theory,
and objectivity, ; as pragmatic concept, –
, –; role of deduction in, –; Feynman, R. P., , –
simulacrum account of, –; as type of Fisher, R. A., –, , –, ,
knowledge, –; types of knowledge n
provided by, , ; understanding as Fox Keller, Evelyn, –, ,
ability to give, ; usefulness of, –; Frank, Philipp,
why-regress, Franklin, A., , n
ind e x
Knuuttila, Tarja, , , , , n, , , Longino, H., n, ,
n, –, , , n, Lorentz, H. A., , ,
Kochelmans, J. J., Lucas, R. E., –, ,
Kolmogorov-Chaitin complexity, – Lucas Critique, –
Koriat, A., Luker, Kristin, –
Koster, Edwin, , , , n
Kratochwil, F., MacCorquodale, K., ,
Krohs, U., Macdonald, Lyn,
Kuhn, Thomas S., –, n, –, , Mach, E, n
, – Machamer, P., ,
Kuorikoski, J., , , Mackay, H., n
Kvanvig, Jonathan, –, n Mäki, U., ,
Kydland, F. E., – Maliniak, D., n
manipulation, n; biological science and, ,
laboratory sciences, , –, –, , ; demonstration of understanding
n through, ; and reliability of feeling of
Lacey, H., n, understanding, ; simulations and, ;
Ladyman, J., as source of tacit causal information,
Lambert, K., , –, , ; understanding and, ,
Landman, U., , –, . See also understanding
language technology, –, through models
Latour, Bruno, n, n Mankiw, N. G., , n
Laub, J., Mattila, E., n
Laubichler, M., n Maxwell, James Clerk, , n, n, n
law(s): causal, , ; covering, , ; Mayo, D. G., n
empirical, ; fundamental, , , Mayr, Ernst, ,
, –, –, , ; vs. McAllister, J. W., n
generalizations, –, –; McMullin, Ernan, n,
general or universal, ; historical, mechanical explanation, ontological principles
, ; mathematical, –, ; underlying,
phenomenological, , , ; physical, mechanical models, –, , , –, ,
, , , , –, , –, , , , –, ,
, –; subsumption under, , ; mechanism(s): causal, , , , –, ,
theoretical, –, , –, , , ; exemplar, –; and extraction of
, . See also theoretical principles understanding, –, , , , –, ,
Leibniz, Gottfried Wilhelm: critique of , –, , ; models as, , ,
Descartes’ physics, –; principle of , ,
continuity, –, , ; and principle of mechanistic approach to understanding,
identity of indiscernibles, Meehl, P. E., ,
Lenhard, Johannes, , , , , , , Mendelism, synthesis with Darwinism, –
n , n
Leonelli, Sabina, , , , , n, n, mental models, understanding as possession
n, n, n, n of, –
Levi, Margaret, Merz, Martina, , , , , , ,
Lewis, C. I., n, n, –, , , n
Lewis, D., metaphysical belief(s), commitments(s),
Lewis, Gilbert, principle(s), truth(s), , , , , , ,
light principle, in theory of relativity, , ,
metaphysics as guide in science, and
Lin, L.-M., , ontological opinion vs. ontological principle,
Lipton, Peter, , , , –, , n, ,
n, , n, , , , meteorological models, accuracy vs.
local knowledge, –, , representativeness in,
local vs. nonlocal meaning, –, Meyers, D.,
logical empiricism: on intelligibility, ; and Mill, John Stuart, ,
neobehaviorism, –; on scientific Mills, C. M.,
understanding, , –, Mink, L. O., –,
logical positivism: on intelligibility, ; and Minkowski, Hermann, –, , n
neobehaviorism, –, , – Mitchell, S., , n, n
inde x
life, , ; natural, , , , , , , ; as model type, ; as scientific
, , , , –; normal, –; understanding, –; validity vs.
physical, , ; political, , –, representational accuracy in, –
n; revolutionary, , , –, ; simulations, –, –, , ; black-
role of scientific understanding in, ; box schemes in, –; representational
social. See social sciences; theoretical, ; accuracy of, ; undermining of link
unity of, ; values in, , , between theory and understanding in,
scientific communities: and epistemic –, –; uses of,
skills in biology, ; refinement of skill(s): as community asset, , ;
individual judgment by, –, ; scientific epistemic relevance of, –, –; in
understanding as social achievement, ; evaluation of scientific theories, ; in
skills as asset of, , ; and understanding historical studies, ; model use as, ;
in biology, – role in scientific practice, ; scientific
scientific method: as hardwired human understanding as, –, ; simulation-
function, ; internalized, as intellectual based understanding and, ; social skills
virtue, n as epistemic skill, –; and subjectivity,
scientific practice, , n, , , , , ; and surplus meaning in theoretical
, ; accounts of, , ; complexity terms, –, ; as tacit knowledge,
and, ; and explanation, ; and , ; vs. theoretical understanding, ;
intelligibility, , , , , ; models understanding as, , –, , , ,
and, , –, , , ; role of , –, –, ; understanding
skills in, , ; in social sciences, ; without explanation through, , , ,
tacit knowledge and, – (see also tacit . See also epistemic skills; judgment;
knowledge); theory and, –, , ; pragmatic understanding; tacit knowledge
and understanding, , , , , , , skill(s), development of: methods of, ;
, , . See also skill(s) through use and development of models,
scientific theories: development of, skills and, , –,
–; evaluation of, ; vs. model, n; Smith, L. D., ,
virtues of, , n Smith, S., , –
scientific understanding: establishing limits of, Sobel, D., n
; as explanatory understanding, , ; social commitments,
interpretive structures and, , –; social sciences, ; epistemic interests and,
judgment and, , –; as knowledge , –; vs. historical studies, ;
about how to perform an epistemic activity, policy research and, nn–; positivism
, –, ; objectivism on, –, , and, ; scientific practice in, ;
–; vs. other forms of understanding, understanding in, , –; usefulness
; as pragmatic, , , –, , –; of, n; vs.natural sciences, . See also
productive objectual models and, –; political science; unification in international
role in science, ; simulation-based relations studies; unification in social
understanding as, –; as skill, –, sciences
; as social achievement, ; unification social skills, –
and, , , –; variety of avenues to, sono-luminescence, –
. See also understanding Sosa, Ernest, n
scientific world view, comprehensibility as space-time, as context for intelligibility,
basic assumption of, special theory of relativity: constructive
Scriven, Michael, , , n theories of, , –, –, –;
semantic conception of models, –, debate over proper method of explanation,
Shanks, D. R., ; Minkowski’s explication of, –;
Shapin, S., n physical reality of relativistic effects, –;
Shapiro, Ian, , , as principle theory, –, –, ;
Shomar, T., relation between constructive and principle
Simon, H. A., –, n theories of, –
simplicity: of models, –; of theories, , specific heat ratio, –
, , , , , Spence, Kenneth W.,
simulacrum account of explanation, – Stachel, J., ,
simulation-based understanding: as ability, Stanford, Kyle, n
–; black-boxing of theory in, –, Star, S. L., n
–; Great Deluge Algorithm, –, stellar models,
inde x
intelligibility and, , , ; vs. knowledge, –; understanding of model as, –;
; as knowledge about coherence, ; understanding of theory as prerequisite for,
from mathematical explanations, –; , , , –
multiplicity of forms, ; prediction as feature understanding in physics, –, –,
of, n; psychological vs. cognitive, understanding in political science: epistemic
–; regulation of desire for, ; as spark interests and adequacy in, –; pluralistic
of insight, ; subjective dimensions of, view of, –; and rational choice theory,
–, ; types of, –; understanding –, –; synthesis and, –;
vs. explanation, –, , ; uses of term, unification and,
–; vagueness of concept, ; variety understanding in psychology: behaviorist
of avenues to, , ; without feeling methodology, –, –; models and,
of understanding, , n.. See also –
scientific understanding understanding in social sciences, , –
understanding, definition of: as ability to understanding of theory: confusing
coordinate theoretical and tacit knowledge, understanding of phenomenon with,
–; difficulty of formulating, , ; –; as prerequisite for understanding
as grasping of relations between facts, ; phenomenon, , , , –; simulation
historical variation in, ; impossibility modeling and, –
of, ; as skill, , –, , –, understanding without explanation, ; of
–, ; as skill to infer and intervene, causation, –; of necessity, –; of
, , possibility, –; of unification, –
understanding, demonstration of: unexpected events, –
through anticipation, –; through unification: in biological sciences, –;
communication, ; through successful and evaluation of theory, ; explanation
interaction, and, , ; models and, ; and scientific
understanding, simulation-based. See understanding, , , , , , , –,
simulation-based understanding , n, , , , –; skills
understanding in biology: as ability to and, ; and Supreme Court, ;
coordinate tacit and theoretical knowledge, understanding of, without explanation,
–, , –; scientific communities –
required for, –; types of, – unification in international relations studies:
understanding in economics: explanation advantages of, ; critique of, ; under
and, –; Lucas’s program of general- rational choice theory, –; and
equilibrium and, –; models and, scientific understanding, –; under
–, – theoretical synthesis, –
understanding in engineering sciences: unification in social sciences, –;
criteria for, –; defined, ; goals critique of, –, , –; goals of,
of, –; interpretive structures and, –
–; vs. intuitive skills, ; as pragmatic Universe, origin of, unintelligibility of,
(functional), –, , n unobservable: entities, n, ; world,
understanding in historical science: empathy , n
and, –, –; explanation as
intrinsic to, –, ; intelligibility validity of models in economics: assessment of,
and, , ; judgment and, –; –, –, –; defined,
reenactment and, , , , ; Van Bouwel, Jeroen, , –, , n
subjective dimensions of, –; through van Daalen, C. E.,
colligatory concepts, , –; through van Fraassen, Bas C., –, , , , , ,
intentionial explanation, – , n, n, n
understanding through models: with abstract Varičak, Vladimir,
mathematical models, , , , –; Verbraeck, A.,
in biological science, –, –; visualizability, , , , , , , ,
of complex systems, ; with economic –,
models, –, –; with gray-box visualization: explanation and, , ,
models, , –; by model building, , , –, –, , –;
, –, , –, ; and models and, , ; ontological principles
theoretical understanding, – underlying,
understanding of phenomenon: confusing voluntary action, and principle of agency,
understanding of theory with, –; von Ignatowski, W.,
as pragmatic, ; scientific skill and, von Neumann, J.,
inde x