Professional Documents
Culture Documents
Rastefer
Rastefer
converging paradigm for intelligence dencies among variables capturing the likelihoods
of relevant states in the world. In light of incom-
ing streams of perceptual data, Bayesian updating
in brains, minds, and machines procedures or approximations are used to prop-
agate information and to compute and revise
Samuel J. Gershman,1* Eric J. Horvitz,2* Joshua B. Tenenbaum3* probability distributions over states of variables.
Beyond base processes for evaluating probabil-
After growing up together, and mostly growing apart in the second half of the 20th century, ities, models of computational rationality require
the fields of artificial intelligence (AI), cognitive science, and neuroscience are mechanisms for reasoning about the feasibility
reconverging on a shared view of the computational foundations of intelligence that and implications of actions. Deliberation about
promotes valuable cross-disciplinary exchanges on questions, methods, and results. the best action to take hinges on an ability to make
We chart advances over the past several decades that address challenges of perception predictions about how different actions will in-
and action under uncertainty through the lens of computation. Advances include the fluence likelihoods of outcomes and a considera-
I
Thus, models of computational rationality may
magine driving down the highway on your purpose ideal for decision-making under uncer- include policies or deliberative machinery that
way to give an important presentation, when tainty. Second, maximizing expected utility is make inferences and decisions at the “metalevel”
suddenly you see a traffic jam looming ahead. nontrivial for most real-world problems, necessi- in order to regulate base-level inferences. These
In the next few seconds, you have to decide tating the use of approximations. Third, the choice decisions rely on reflection about computational
whether to stay on your current route or take of how best to approximate may itself be a decision effort, accuracy, and delay associated with the
the upcoming exit—the last one for several miles— subject to the expected utility calculus—thinking invocation of different base-level algorithms in dif-
all while your head is swimming with thoughts is costly in time and other resources, and some- ferent settings. Such metalevel decision-making,
about your forthcoming event. In one sense, this times intelligence comes most in knowing how or “metareasoning,” can be performed via real-
problem is simple: Choose the path with the best to allocate these scarce resources. time reflection or as policies computed during
highest probability of getting you to your event The broad acceptance of guiding action with offline optimizations. Either way, the goal is to
on time. However, at best you can implement expected utility, the complexity of formulating identify configurations and uses of base-level pro-
this solution only approximately: Evaluating the and solving decision problems, and the rise of cesses with the goal of maximizing the expected
full branching tree of possible futures with high approximate methods for multiple aspects of value of actions taken in the world. These com-
uncertainty about what lies ahead is likely to be decision-making under uncertainty has motivated putational considerations become increasingly
infeasible, and you may consider only a few of artificial intelligence (AI) researchers to take a important when we consider richer representa-
the vast space of possibilities, given the urgency fresh look at probability through the lens of com- tions (graphs, grammars, and programs) that sup-
of the decision and your divided attention. How putation. This examination has led to the devel- port signature features of human intelligence, such
best to make this calculation? Should you make a opment of computational representations and as recursion and compositionality (5).
snap decision on the basis of what you see right procedures for performing large-scale probabil- Key advances in AI on computational machin-
now, or explicitly try to imagine the next several istic inference; methods for identifying best ac- ery for performing inference, identifying ideal
miles of each route? Perhaps you should stop tions, given inferred probabilities; and machinery actions, and deliberating about the end-to-end
thinking about your presentation to focus more for enabling reflection and decision-making about operation of systems have synergies and reso-
on this choice, or maybe even pull over so you tradeoffs in effort, precision, and timeliness of nances with human cognition. After a brief his-
can think without having to worry about your computations under bounded resources. Analo- tory of developments in AI, we consider links
driving? The decision about whether to exit has gous ideas have come to be increasingly important between computational rationality and findings
spawned a set of internal decision problems: how in how cognitive scientists and neuroscientists in cognitive psychology and neuroscience.
much to think, how far should you plan ahead, think about intelligence in human minds and
and even what to think about. brains, often being explicitly influenced by AI re- Foundations and early history
This example highlights several central themes searchers and sometimes influencing them back. AI research has its roots in the theory of compu-
in the study of intelligence. First, maximizing some In this Review, we chart this convergence of ideas tability developed in the 1930s. Efforts then high-
measure of expected utility provides a general- around the view of intelligence as computational lighted the power of a basic computing system
rationality: computing with representations, algo- (the Turing Machine) to support the real-world
rithms, and architectures designed to approximate mechanization of any feasible computation (6).
1
Department of Psychology and Center for Brain Science, decisions with the highest expected utility, while The promise of such general computation and the
Harvard University, Cambridge, MA 02138, USA. 2Microsoft taking into account the costs of computation. We fast-paced rise of electronic computers fueled the
Research, Redmond, WA 98052, USA. 3Department of Brain share our reflections about this perspective on imagination of early computer scientists about
and Cognitive Sciences, Massachusetts Institute of
Technology, Cambridge, MA 02139, USA.
intelligence, how it encompasses interdiscipli- the prospect of developing computing systems
*Corresponding author. E-mail: gershman@fas.harvard.edu nary goals and insights, and why we think it will that might one day both explain and replicate
(S.J.G.); horvitz@microsoft.com (E.J.H.); jbt@mit.edu (J.B.T.) be increasingly useful as a shared perspective. aspects of human intelligence (7, 8).
A S
NP VP
S
S
DT NN VBD NP PP
NP VP
NP VP
DT NN IN NP
DT NN VBD
NP VP VBD
main verb DT NN
1
reduced relative DT NN VBN NP PP
DT NN IN NP
0.8
S DT NN
NP
0.6
Probability
S
NP VP
NP
0.4 DT NN VBN NP PP
NP VP
DT NN IN NP
0.2 DT NN VBN
DT NN
---
h ki
0
B Time t Time t + C
Expected utility Decision Expected utility
x =
Low velocity
Action/sample cost
0.75 1 0.7 1000
0.6
0.5 100
0.5 0.4
0.3 10
0.2
0.1 1
0.5 0 0
0 1 10 100 1000 0 1 10 100 1000 0 1 10 100 1000
High velocity
Fig. 3. Resource-constrained sampling in human cognition. (A) Incre- indicates the strength of belief that a dot is in the tracked set (More red =
mental parsing of a garden-path sentence. [Adapted from (36)] (Top) Evo- higher confidence) after some time interval D, for a particle-filter object
lution of the posterior probability for two different syntactic parses (shown tracker. Uncertainty scales with velocity, explaining why high-velocity ob-
in boxes). The initially favored parse is disfavored by the end of the sen- jects are harder to track (32). (C) For a sampling-based approximate Bayesian
tence. (Bottom) A resource-constrained probabilistic parser (based on a decision-maker, facing a sequence of binary choices, expected utility per
particle filter with few particles) may eliminate the initially unlikely parse decision, and number of decisions per unit time can be combined to
and therefore fail to correctly parse the sentence by the end, as shown by compute the expected utility per unit time as a function of the number of
the proportion of particle filters with 20 particles that successfully parse posterior samples and action/sample cost ratios. Circles in the rightmost
the sentence up to each word. (B) Sample-based inference for multiple- graph indicate the optimal number of samples at a particular action/sample
object tracking. In this task, subjects are asked to track a subset of dots cost ratio. For many decisions, the optimal choice tradeoff between accuracy
over time, marked initially in red. Lines denote velocity, and circles denote and computation time suggests deciding after only one sample. [Reprinted
uncertainty about spatial transitions. In the second frame, red shading from (38) with permission]
sampled (1). Evidence suggests that humans for tracking multiple objects moving in a dynamic quent data, can be explained by particle filters
use this strategy across several domains, includ- uncertain environment (36)—are at the heart of for approximate online parsing in probabilistic
ing causal reasoning (31), perception (32, 33), the Google self-driving car’s picture of its sur- grammars (Fig. 3A) (37). These biases may in
and category learning (34). Sampling algorithms roundings (Fig. 1B) and also may describe how fact be rational under the assumption that sam-
can also be implemented in biologically plausible humans track multiple objects (Fig. 3B) (32). pling is costly and most gains or losses are small,
neural circuits (35), providing a rational expla- When only a small number of hypotheses are as in many everyday tasks; then, utility can be
nation for the intrinsic stochasticity of neurons. sampled, various biases emerge that are con- maximized by sampling as few as one or a few
We see a correspondence between the sam- sistent with human behavior. For instance, “gar- high-posterior probability hypotheses for each
pling algorithms humans appear to use and those den path” effects in sentence processing, in decision (Fig. 3C) (38).
used in state-of-the-art AI systems. For example, which humans perseverate on initially promising This argument rests crucially on the assertion
particle filters—sequential sampling algorithms hypotheses that are disconfirmed by subse- that the brain is equipped with metareasoning
Computational tradeoffs in
sequential decision-making 100 ms 125 ms 150 ms 175 ms fully- fully-
connected- connected-
Computational rationality has played an impor- conv-layer conv-layer layer layer
tant role in linking models of biological intelli- C (tanh) (tanh) [max(0x)] (linear)
example, reliance on the model-based system support of maximizing expected utility; that 24. A. Tversky, D. Kahneman, Science 185, 1124–1131 (1974).
decreases when the availability of cognitive ideal MEU calculations may be intractable for 25. G. Gigerenzer, Rationality for Mortals: How People Cope with
Uncertainty (Oxford Univ. Press, Oxford, 2008).
resources are transiently disrupted (53). Recent real-world problems, but can be effectively ap- 26. J. R. Anderson, The Adaptive Character of Thought (Lawrence
data show that the arbitration mechanism may proximated by rational algorithms that maximize Erlbaum, Hillsdale, NJ, 1990).
be supported by the lateral prefrontal cortex (54), a more general expected utility incorporating the 27. M. Oaksford, N. Chater, Bayesian Rationality (Oxford Univ.
the same region involved in the registration of costs of computation; and that these algorithms Press, Oxford, 2007).
28. T. L. Griffiths, J. B. Tenenbaum, Cognit. Psychol. 51, 334–384
cognitive demand. can be rationally adapted to the organism’s spe-
(2005).
Finer-grained metareasoning may play a role cific needs, either offline through engineering 29. K. Doya, S. Ishii, A. Pouget, R. P. N. Rao, Eds. The Bayesian
within the richer model-based systems them- or evolutionary design, or online through meta- Brain: Probabilistic Approaches to Neural Coding (MIT Press,
selves. One way to approximate values is to adapt reasoning mechanisms for selecting the best Cambridge, MA, 2007).
the sampling hypothesis to the sequential deci- approximation strategy in a given situation. 30. T. L. Griffiths, F. Lieder, N. D. Goodman, Top. Cogn. Sci. 7,
217–229 (2015).
sion setting, stochastically exploring trajectories We discussed case studies in which these ideas 31. S. Denison, E. Bonawitz, A. Gopnik, T. L. Griffiths, Cognition
through the state space and using these sample are being fruitfully applied across the disci- 126, 285–300 (2013).
paths to construct a Monte Carlo estimator. Re- plines of intelligence, but we admit that a ge- 32. E. Vul, M. Frank, G. Alvarez, J. B. Tenenbaum, Adv. Neural Inf.
cently, a class of sampling algorithms known as nuine unifying theory remains mostly a promise Process. Syst. 29, 1955–1963 (2009).
Monte Carlo Tree Search (MCTS) has gained con- for the future. We see great value in pursuing 33. S. J. Gershman, E. Vul, J. B. Tenenbaum, Neural Comput. 24,
1–24 (2012).
siderable traction on complex problems by bal- new studies that seek additional confirmation 34. A. N. Sanborn, T. L. Griffiths, D. J. Navarro, Psychol. Rev. 117,
ancing exploration and exploitation to determine (or disconfirmation) of the roles of machinery for 1144–1167 (2010).
Editor's Summary
Article Tools Visit the online version of this article to access the personalization and
article tools:
http://science.sciencemag.org/content/349/6245/273
Science (print ISSN 0036-8075; online ISSN 1095-9203) is published weekly, except the last week
in December, by the American Association for the Advancement of Science, 1200 New York
Avenue NW, Washington, DC 20005. Copyright 2016 by the American Association for the
Advancement of Science; all rights reserved. The title Science is a registered trademark of AAAS.