Professional Documents
Culture Documents
Murphy2017 Parameterisasi
Murphy2017 Parameterisasi
Murphy2017 Parameterisasi
Management Science
Publication details, including instructions for authors and subscription information:
http://pubsonline.informs.org
. http://dx.doi.org/10.1287/mnsc.2016.2591
This article may be used only for the purposes of research, teaching, and/or private study. Commercial use
or systematic downloading (by robots or other automatic processes) is prohibited without explicit Publisher
approval, unless otherwise noted. For more information, contact permissions@informs.org.
The Publisher does not warrant or guarantee the article’s accuracy, completeness, merchantability, fitness
for a particular purpose, or non-infringement. Descriptions of, or references to, products or publications, or
inclusion of an advertisement in this article, neither constitutes nor implies a guarantee, endorsement, or
support of claims made of that product, publication, or service.
INFORMS is the largest professional society in the world for professionals in the fields of operations research, management
science, and analytics.
For more information on INFORMS, its publications, membership, or meetings visit http://www.informs.org
MANAGEMENT SCIENCE
Articles in Advance, pp. 1–19
http://pubsonline.informs.org/journal/mnsc/ ISSN 0025-1909 (print), ISSN 1526-5501 (online)
Received: January 28, 2015 Abstract. An individual’s tolerance of risk can be quantified by using decision models
Accepted: June 20, 2016 with tuned parameters that maximally fit a set of risky choices the individual has made.
Published Online in Articles in Advance: A goal of this model fitting procedure is to identify parameters that correspond to stable
January 6, 2017 underlying risk preferences. These preferences can be modeled as an individual differ-
https://doi.org/10.1287/mnsc.2016.2591 ence, indicating a particular decision maker’s tastes and willingness to accept risk. Using
hierarchical statistical methods, we show significant improvements in the reliability of
Copyright: © 2017 INFORMS individual risk preference parameter estimates over other common methods for cumu-
lative prospect theory. This hierarchical procedure uses population-level information (in
addition to an individual’s choices) to break “ties” (or near ties) in the fit quality for sets
of possible risk preference parameters. By breaking these statistical ties in a sensible way,
researchers can avoid overfitting choice data and thus more resiliently measure individual
differences in people’s risk preferences.
Keywords: measuring risk preferences • cumulative prospect theory • decision making under risk • hierarchical parameter estimation • individual
differences • risk profile • reliability
risky choices and thus elicit risk preferences. One rea- (e.g., correlating risk preferences with other variables,
son to use binary lotteries is because people appear including measures of physiological processes such as
to have problems with assessing a lottery’s certainty Huettel et al. 2006, Engelmann and Tamir 2009, Figner
equivalent (Lichtenstein and Slovic 1971, Harrison and and Murphy 2010, or process tracing approaches that
Rutström 2008), although this simplicity comes at the track attention as in Schulte-Mecklenbeck et al. 2016)
cost of requiring a DM to make many binary choices as well as tuning a model to make predictions about
to achieve the same degree of estimation fidelity that what a particular DM will choose when presented
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
other methods purport to have (e.g., the three decision with different options. Here we focus on the nonex-
risk measure from Tanaka et al. 2010). pected utility model cumulative prospect theory (Kah-
The approach we cover here employs a set of static neman and Tversky 1979, Tversky and Kahneman 1992)
lotteries; this fixed approach has some advantages in as our decision model, using functional specifications
its ease of implementation and the comparability of explained below. Prospect theory is arguably the most
results between subjects. Other approaches have used important and influential descriptive model of risky
adaptive lotteries that require sophisticated algorithms
choice to date (Starmer 2000, Wakker 2010, Barberis
to dynamically design risk elicitation items contin-
2013, Fox et al. 2015), and it has been used extensively
gent upon what choices DMs have made so far. Cav-
in research related to model fitting and risky decision
agnaro et al. (2013) develop an adaptive approach for
making.
model discrimination. This is a useful tool when the
research goal is to determine if a DM’s choices are 1.1.3. Parameter Estimation Procedure. Given a set
more consistent with one particular choice model or of lotteries, risky choices, and a particular decision
another. Toubia et al. (2013) have developed a dif- model, the best fitting parameter(s) can be identified
ferent adaptive approach that aims to maximize the via a statistical estimation procedure. A perfectly fit-
information obtained in each step of a choice experi- ting parameterized model would exactly reproduce
ment involving risk and time preferences. Such adap- all of a DM’s choices. Maximum likelihood estimation
tive methods appear to be more efficient (ultimately methods have been developed that allow for a more
requiring fewer choices from DMs for estimating pref- nuanced approach (i.e., not just counting the number
erences and thus demanding less time and effort) but
of correct predictions) in evaluating the fit of a choice
are also more challenging to implement, which may
model (see, e.g., Harless and Camerer 1994, Hey and
act as a barrier to their widespread adoption. Adap-
Orme 1994, Regenwetter and Robinson 2016) to behav-
tive methods are also predicated on assumptions that
ioral data. Other methods build on this and use mix-
preferences are stable and independent of the elicita-
ture models, in which DMs’ choices are fit to several
tion process. Other approaches (Payne et al. 1992; Par-
ducci 1995; Stewart et al. 2003, 2006; Lichtenstein and models simultaneously rather than only one (see Har-
Slovic 2006) posit that often DMs do not in many sit- rison and Rutström 2009, Conte et al. 2011). Mixture
uations have stable long-term preferences but rather models can also be used to identify the most com-
dynamically sample information, make comparative mon preference types by specifying a flexible deci-
judgments from this particular local context, and then sion model (Bruhin et al. 2010). This approach can
construct preferences in real time. Such a view would be powerful because clusters of types of DMs can
also favor using a static set of lotteries over adaptive emerge endogenously, yielding a structure to orga-
methods. nize the considerable heterogeneity observed in peo-
ple’s risk taking behavior. Regardless of the model-
1.1.2. Decision Models (Nonexpected Utility Theory).
ing approach, all of these evaluation methods can
Models may reflect the general structure (e.g., styl-
include in-sample and out-of-sample tests, the lat-
ized facts) of DMs’ aggregate preferences by invok-
ing a latent construct like utility. These models typi- ter of which are especially useful because they can
cally also have free parameters that can be tuned to diagnose and mitigate the overfitting of a model to
improve the accuracy of how they represent differ- choice data.
ent choice patterns. An individual’s choices from the The interrelationship between lotteries, a decision
lotteries are fit to a decision model by tuning these model, and an estimation procedure is shown in
parameters, thus identifying individual differences in Figure 1. Lotteries provide stimuli and the resulting
revealed preferences and summarizing the pattern of choices are the behavioral input for the model; a deci-
choices by using particular combinations of parame- sion model and an estimation procedure tune param-
ter values. These parameters can, for example, not just eters to maximize correspondence between the model
establish the existence of risk aversion but also quan- and the actual choices from the DM facing the lotter-
tify the degree of this particular preference. This is ies. This process of eliciting choices can be repeated
useful in characterizing an individual DM and isolat- again using the same lotteries and the same experimen-
ing the cognitive mechanisms that underlie behavior tal subjects. This test-retest design allows for both the
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS 3
Input
Choices
Model
Time 1 Time 2
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
Out-of-sample
(prediction)
In-sample (summarization)
Notes. On the left are choices from binary lotteries that serve as input for the model. Using an estimation method, we fit the model’s risk
preference parameters (of cumulative prospect theory) to the individual risky choices at time 1. These parameters, to some extent, summarize
the choices made in time 1 by using the parameters in the model and reproduce the choices (depending on the quality of the fit). The out-
of-sample predictive capacity of the model can be evaluated by comparing the predictions against actual choices at time 2. The reliability
of the estimates is evaluated by comparing the time 1 parameter estimates to estimates obtained using the same estimation method on the
time 2 data. This experimental design holds the lotteries and choices constant but varies the estimation procedures; moreover, this is done in
a test-retest design that allows for the computation and comparison of both in-sample and out-of-sample statistics.
development of in-sample parameter fitting (i.e., statis- extreme parameter values and thus can reduce over-
tical “explanation” or summarization) as well as out- fitting, but on the downside they negate the possibil-
of-sample parameter fitting (i.e., prediction). Moreover, ity of detecting extreme preferences altogether, even
the test-retest reliability of the different parameter esti- though these preferences may be real but unusual.
mates can be computed as a correlation coefficient Boundaries are also defined arbitrarily and may create
because two sets of parameters exist for each DM. serious estimation problems due to parameter interde-
This is a powerful experimental design because it can pendence. For example, setting a different boundary
accommodate both in-sample fitting and also predic- on one parameter may radically change the estimate of
tion and further can diagnose instances of overfitting another parameter (e.g., restricting the range of proba-
which can undermine reliability as well as meaning- bility distortion may unduly influence estimates of loss
ful psychological interpretations (see Pitt and Myung aversion) for one particular subject. The interdepen-
2002, Lewandowsky and Farrell 2010). dence of parameter estimates has been noted (Nilsson
et al. 2011, Zeisberger et al. 2012). To circumvent the
1.2. Overfitting and Bounding Parameters pitfalls of arbitrary parameter boundaries, we use a
Multiparameter models’ estimation methods may be hierarchical estimation method based on Farrell and
prone to overfitting and in doing so adjust to noise Ludwig (2008) without such boundaries. At its core,
instead of real risk preferences (Roberts and Pash- this hierarchical method uses estimates of risk prefer-
ler 2000). This can sometimes be observed when pa- ences of the whole sample to inform estimates of the
rameter values emerge that are highly atypical and risk preferences at the individual level. We therefore
extreme. A common solution to this problem is to address to what degree an estimation method com-
set boundaries and limit the range of parameter val- bining group-level information with individual-level
ues that are potentially estimated. Boundaries prevent information can more reliably represent individual risk
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
4 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
preferences compared with using either individual or that individual risk preferences are generally stable
aggregate information exclusively. and that the individual parameter values outperform
aggregate values in terms of prediction. The authors
1.3. Justification for Using Hierarchical concluded that the reliability of parameters suffered
Estimation Methods when extra model parameters were added. This result
The ultimate goal of the hierarchical estimation proce- is contrasted against the work of Fehr-Duda and Epper
dure here is to obtain improvements in the reliability (2012), who conclude, based on experimental data
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
of estimates for individual risk preferences that can and a literature review, that those additional param-
be used to make better predictions about risky choice eters (related to the probability weighting function)
behavior, contrasted to other estimation methods. This are necessary to capture individual risk preferences.
is not modeling as a means only to maximize in-sample Zeisberger et al. (2012) also tested individual param-
fit but rather to maximize out-of-sample correspon- eter reliability using a different type of risky choice
dence; moreover, these methods provide a way to gain (certainty equivalent rather than binary choice). They
insights into what people are actually motivated by found significant differences in parameter value esti-
as they cogitate about risky options and make choices mates over time but did not use a hierarchical estima-
when confronting irreducible but quantified uncer- tion procedure.
tainty. Ideally, the parameters should not only be about
merely summarizing choices but also reflect some psy- 1.6. Major Contributions
chological mechanisms that underlie risky decision- In this paper we show concretely how the use of a
making behavior. hierarchical estimation method is able to circumvent
many of the problems that occur with maximum like-
1.4. Other Applications of Hierarchical lihood procedures when estimating individual param-
eters in cumulative prospect theory. The hierarchical
Estimation Methods
method outperforms maximum likelihood estimation
Other researchers have applied hierarchical estimation
with regard to out-of-sample fit and parameter relia-
methods to risky choice modeling at the individual
bility and does not resort to arbitrarily chosen param-
level. Nilsson et al. (2011) have applied a Bayesian
eter bounds or to the elimination of subjects to accom-
hierarchical parameter estimation model to simple
plish this efficiency gain. As a statistical method it is
risky choice data. The hierarchical procedure outper-
simpler, more transparent, and computationally less
formed maximum likelihood estimation in a param-
demanding than other approaches. We explore how
eter recovery test. The authors, however, did not test
this method is a more nuanced way of discerning
out-of-sample predictions nor the test-retest reliabil-
parameters through the use of a hierarchical step,
ity of their parameter estimates, a gap we address
which produces a kind of plausibility filter for sets
in this paper. Wetzels et al. (2010) applied Bayesian
of individual-level parameters. This helps develop an
hierarchical parameter estimation in a learning model
explanation and intuition for how hierarchical meth-
and found it was more robust with regard to extreme ods achieve these superior results. We conclude that
estimates and misunderstandings of the nature of the hierarchical maximum likelihood (HML) method
the decision-making task. Scheibehenne and Pachur is readily applicable “off the shelf,” making it a pow-
(2013) applied Bayesian hierarchical parameter estima- erful approach for both researchers and practitioners
tion to risky choice data using a Transfer of Atten- interested in measuring DMs’ risk preferences.
tion eXchange (TAX) model (Birnbaum and Chavez
1997, Birnbaum 1999) but reported no improvement 1.7. Structure of the Rest of This Paper
in parameter stability. In another study, Scheibehenne In this paper we focus on evaluating different statisti-
and Pachur (2015) also find no improvement in param- cal estimation procedures for fitting individual choice
eter stability for prospect theory. The authors offer the data to a risky choice model. To this end, we hold
explanation that hierarchical estimation at times over- constant the set of lotteries DMs made choices with
corrects for extreme but correct parameter values and and the functional form of the risky choice model. We
that the benefits of hierarchical estimation may be lim- then fit the same set of choice data using two estima-
ited to cases where data are sparse. tion methods and contrast the results to differentiate
the quality of the different statistical methods. We also
1.5. Other Research Examining Parameter compute the test-retest reliability of individual-level
Stability parameters that resulted from different estimation pro-
Besides results from the hierarchical approaches out- cedures. This broad approach allows us to evaluate
lined above, there exists other research on estimated different estimation methods and make substantiated
parameter stability. Glöckner and Pachur (2012) tested conclusions about the fit quality as well as diagnose
the reliability of parameter estimates using a non- overfitting. We conclude with recommendations for
hierarchical estimation procedure. The results show estimating risk preferences at the individual level.
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS 5
32 0.78 39 0.32
1 with probability with probability Gain
99 0.22 56 0.68
−24 0.99 −15 0.44
2 with probability with probability Loss
−13 0.01 −62 0.56
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
58 0.48 96 0.60
3 with probability with probability Mixed
−5 0.52 −40 0.40
−30 0.50
4 with probability 0 for sure Mixed-zero
60 0.50
Note. These lotteries were part of the set of 91 lotteries used in the experiment in this paper.
2. Lotteries and Experimental Design based on their choices. Participants worked through
2.1. Participants several examples of risky choices to familiarize them-
One hundred eighty-five participants from the subject selves with the interface of the MouselabWeb software
pool at the Max Planck Institute for Human Devel- (Willemsen and Johnson 2011, 2014) that was used to
opment in Berlin volunteered to participate in two administer the study. Although the same 91 lotteries
research sessions (referred to hereafter as time 1 and were used for both experimental sessions, the item
time 2) that were administered approximately two order was fully randomized. Additionally, the order of
weeks apart. After both sessions were conducted, the outcomes (top-bottom) and option order (A or B)
complete data from both sessions were available for and the orientation (A above B, A left of B) were ran-
142 participants; these subjects with complete data sets domized and stored during the first session. The exact
are retained for subsequent analysis.1 The experiment opposite spatial representation of all of this was used
was incentive compatible and was conducted without in the second session to mitigate potential order or pre-
deception. The data used here are the resulting choices sentation effects.
from the Schulte-Mecklenbeck et al. (2016) experiment, Incentive compatibility was implemented by ran-
and we are indebted to these authors for their generos- domly selecting one lottery at the end of each exper-
ity in sharing their raw data. imental session, playing it out with the stated prob-
abilities, and paying the participant according to her
2.2. Lotteries choice and the realized outcome on the selected item.
In each session of the experiment, individuals made An exchange rate of 10:1 was used between experimen-
choices from a set of 91 simple binary lotteries. Each tal values and payments for choices. Thus, all partici-
option has two possible outcomes between −100 and pants earned their fixed payment plus one tenth of the
100 that occur with known probabilities that sum to outcome of one randomly selected lottery for each com-
one. There were four types of lotteries for which exam- pleted experimental session. Subjects knew all of this
ples are shown in Table 1: gains only, losses only, mixed information. Participants earned in total about EUR 30
lotteries with both gains and losses, and mixed-zero (approximately USD 40) on average for participating in
lotteries with one gain and one loss and zero (status both sessions.
quo) as the alternative outcome. The first three types
were included to cover the spectrum of risky deci-
sions and the mixed-zero type allows for measuring 3. Model Specification
loss aversion separately from risk aversion (Rabin 2000, We use cumulative prospect theory (Tversky and Kah-
Wakker 2005). The same 91 lotteries were used in both neman 1992) to model risk preferences. A two-outcome
test-retest sessions of this research. The set of lotter- lottery L is valued in utility u( · ) as the sum of its com-
ies was compiled of existing items used by Rieskamp ponents in which the monetary reward x i is weighted
(2008), Gächter et al. (2007), and Holt and Laury (2002). by a value function v( · ) and the associated probabil-
In total, 35 lotteries are gain only, 25 are loss only, 25 ity p i , which is transformed by a probability weight-
are mixed, and 6 are mixed-zero. All of the lotteries are ing function w( · ). This is shown in the following
listed in Appendix E. equation:2
2.3. Procedure
v(x 1 )w(p1 ) + v(x2 )(1 − w(p 1 ))
Participants received extensive instructions regarding
if both positive/both negative,
the experiment at the beginning of the first session. All u(L) |x1 | > |x2 |; (1)
participants received EUR 10 as a guaranteed payment
v(x1 )w(p1 ) + v(x2 )w(p2 ) if mixed.
for participation in the research and could earn more
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
6 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
Cumulative prospect theory has many possible mathe- The γ parameter controls the curvature of the proba-
matical specifications. These different functional spec- bility weighting function. The psychological interpre-
ifications have been tested and reported in the litera- tation of the curvature of the function is a diminishing
ture (see, e.g., Stott 2006) and the functional forms and sensitivity away from the end points: both zero and
parameterizations outlined below are justifiable given one serve as necessary boundaries and the further
the preponderance of previous findings. from these edges, the less sensitive individuals are to
changes in probability (Tversky and Kahneman 1992).
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
3.1. Value Function The δ parameter controls the general elevation of the
In general, power functions have been shown to fit probability weighting function. It is an index of how
behavioral choice data better than many other func- attractive lotteries are in general (Gonzalez and Wu
tional forms at the individual level (Stott 2006). Stevens 1999) and it corresponds to how optimistic or pes-
(1957) also cites experimental evidence showing the simistic an individual DM is.
merit of power functions in general for modeling psy- The use of Prelec’s two-parameter weighting func-
chological processes. Here we use a power value func- tion rather than the original specification used in
tion as displayed in the following equation: cumulative prospect theory (Tversky and Kahneman
( 1992) requires additional explanation. In the original
xα if x ≥ 0, α > 0; specification, the point of intersection with the diag-
v(x) α
(2)
−λ(−x) if x < 0, α > 0; λ > 0. onal changes simultaneously with the shape of the
weighting function, whereas in Prelec’s specification,
The α parameter controls the curvature of the value the weighting line always intersects at the same point
function. If the value of α is below one, there are dimin- if δ is kept constant.5 However, changing the point of
ishing marginal returns in the domain of gains. If the intersection and curvature simultaneously induces a
value of α is one, the function is linear and consistent negative correlation with the value-function parameter
with risk neutrality (i.e., EV maximizing) in decision α because both parameters capture similar character-
making. For α values above one, there are increasing istics; moreover, this specification does not allow for a
marginal returns for positive values of x. This pattern wide variety of individual preferences (see Fehr-Duda
reverses in the domain of losses (values of x less than and Epper 2012). Furthermore, the original specifica-
zero), which is referred to as the reflection effect. tion of the weighting function is nonmonotonic for γ <
For this paper we use the same value-function pa- 0.279 (Ingersoll 2008). Although that low value is gen-
rameter α for both gains and losses, and our reasoning erally not in the range of reported aggregate param-
is as follows. First, this is parsimonious. Second, when eters (Camerer and Ho 1994), this nonmonotonicity
using a power function with different parameters for may become relevant and problematic when estimat-
gains and losses, loss aversion cannot be defined any- ing individual preferences, where there is considerably
more as the magnitude of the kink at the reference more heterogeneity and noise in the choice data.
point (Köbberling and Wakker 2005) but would instead
need to be defined contingently over the whole curve.3
This complicates interpretation and undermines clear 4. Estimation Methods
separability between utility curvature and loss aver- Individual risk preferences are captured by obtaining
sion; further, it may cause modeling problems for a set of parameters that best fit the observed choices
mixed lotteries that include an outcome of zero. Prob- implemented via a particular choice model. In this sec-
lems of induced correlation between the loss aversion tion we explain the workings of standard maximum
parameter and the curvature of the value function in likelihood estimation (MLE) and of hierarchical maxi-
the negative domain have furthermore been reported mum likelihood estimation (HML).
when using αgain , αloss in a power utility model (Nils-
4.1. Maximum Likelihood Estimation
son et al. 2011).
A narrow interpretation of cumulative prospect theory
3.2. Probability Weighting Function dictates that even a trivial difference in utility leads the
To capture subjective probability distortion, we use DM to always choose the option with the highest utility.
Prelec’s functional specification (Prelec 1998; see Fig- However, even in early experiments with repeated lot-
ure B.1 in Appendix B). Prelec’s two-parameter prob- teries, Mosteller and Nogee (1951) found that this is not
ability weighting function can accommodate a wide the case with real DMs. Choices were instead partially
range of curves and it has been shown to fit individ- stochastic, with the probability of a DM choosing the
ual data well (Gonzalez and Wu 1999, Fehr-Duda and generally more favored option increasing as the util-
Epper 2012). Its specification4 can be found in the fol- ity difference between the options increased. This idea
lowing equation: of random utility maximization has been developed by
Luce (1959) and others (e.g., Luce and Suppes 1965,
w(p) exp (−δ(− ln(p))γ ), δ > 0; γ > 0. (3) McFadden 1980, Harless and Camerer 1994, Hey and
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS 7
Orme 1994, Loomes and Sugden 1995). In this tradi- tolerates several smaller prediction mistakes instead of
tion, we use a generalized logistic function that spec- fewer, but very large, mistakes. The maximum like-
ifies the probability of picking one option, depending lihood quality of the fit drives the parameter esti-
on each option’s utility (i.e., softmax). This formulation mates, and this improves out-of-sample performance
is displayed in Equation (4). There are other stochastic over methods that aim to maximize the fraction of
choice functions that have been used in the literature explained choices in-sample.
(see Stott 2006, Table 4). A logistic function has a long On the downside, MLE is not robust to aberrant
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
history, being used to fit choice data from Mosteller choices. If by confusion or accident a DM chooses an
and Nogee (1951), and has been shown generally to atypical option for a lottery that is greatly out of line
perform well with behavioral data (Stott 2006): with her other choices, the resulting parameters may
be disproportionately affected by this single anoma-
1
p(A B) , ϕ ≥ 0. (4) lous choice. Although MLE does take into account the
1+ e ϕ(u(B)−u(A)) quality of the fit with respect to utility discrepancies,
The parameter ϕ is an index of the sensitivity to the fitting procedure has the simple goal of finding the
differences in utility. Lower values for ϕ diminish the parameter set that generates the best overall fit (even
importance of the difference in utility. It is important to if the differences in fit among parameter sets are neg-
note that this parameter operates on the absolute dif- ligible). This approach ignores the reality that there
ference in utility assigned to both options. Two individ- may be different parameter sets that fit choice sets vir-
uals with different risk attitudes, but equal sensitivity, tually the same. Consider estimations for subject A in
will almost certainly have different ϕ parameters sim- Figure 4 (see Section 6.3), which shows the resulting
ply because the differences in the options’ utilities are distribution of parameter estimates when one single
also determined by their risk attitude. Although the choice is changed and then the MLE parameters are
parameter is useful to help fit an individual’s choice reestimated. Especially for loss aversion and the eleva-
data, one cannot compare the values of ϕ across indi- tion of the probability weighting function, these sensi-
viduals unless both the lotteries and the individuals’ tivity analysis results show that a different response on
risk attitudes are held constant. The interpretation of a single choice can make a large difference among the
the sensitivity parameter is therefore ambiguous and resulting best fitting parameter values.
it should be here considered simply as an aid in the The MLE method solves parameter tie-breaking
estimation method. problems but does so by ascribing a great deal of im-
In maximum likelihood estimation the goal function portance to potentially trivial differences in fit quality.
is to maximize the likelihood of observing the out- Radically different combinations of parameters may
come, which consists of the observed choices, given a be identified by the procedure as nearly equivalently
set of parameters. The likelihood is expressed using the good, and the winning parameter set may emerge but
choice function above, yielding a stochastic specifica- only differ from the other “almost-as-good” sets by the
tion. We use the notation M {α, λ, δ, γ, ϕ}, where yi slightest of margins in fit quality. This process of select-
denotes the choice for the ith lottery: ing the winning parameter set from among the space of
possible combinations ignores how plausible the sets
N
Y of parameters may be in capturing the DM’s actual
M̂i arg max c(yi | M). (5)
M i1
preferences. Moreover, the ill-behaved structure of the
parameter space makes it a challenge to find reliable
By c( · ), we denote the choice itself, where p(A B) is results because different solutions can be nearly identi-
given by the logistic choice function in Equation (4): cal in fit quality but be located in far distant regimes of
( the ill-behaved parameter space. A minor change in the
p(A B) if A is chosen (yi is 0), lotteries, or just one different choice, may bounce the
c(yi | M) (6)
1 − p(A B) if B is chosen (yi is 1). resulting parameter estimations around substantially.
Furthermore, the MLE procedure does not consider
MLE takes the utility difference between the two the psychological interpretability or plausibility of the
options into account (see Equation (4)) and there- parameters and therefore may yield a parameter set
fore does not only fit the number of choices correctly that is grossly atypical, doing so with only very weak
predicted but also is sensitive to the magnitude of evidence to justify the extreme conclusion.
the difference between the options’ utilities. Ironically, Results for MLE using risky binary choice problems
because it is the utility difference that drives the fit (and may lead to particularly lumpy likelihood surfaces. To
not the frequency of correct predictions), the resulting prevent the evaluation algorithm from becoming stuck
best-fitting parameters may explain fewer choices than in local minima, a large number of starting solutions
can be obtained by finding parameters that maximize was used. For each subject, the likelihood of a grid of
the fraction of explained choices. This is because MLE starting values was evaluated. The resolution in the
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
8 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
range of commonly observed values was chosen to be This first step can be explained using a simplified
higher than those outside of it. The 25 solutions for example that is a discrete version of Equation (7). For
which the likelihood was highest for that particular each of the four risk preference parameters let us pick
subject were used as starting points for the Nelder- a set of values that covers a sufficiently large interval,
Mead algorithm. Additionally, a wider grid of 216 for example {0.5, 0.6, 0.7, . . . , 3.0}. We then take all pos-
starting solutions that covered a large range of parame- sible parameter combinations (i.e., the Cartesian prod-
ter values was used for all subjects. Thus, a total of 241 uct) of these four sets, i.e., α 0.5, λ 0.5, δ 0.5,
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
starting points was used to estimate MLE parameters.6 γ 0.5, then α 0.5, λ 0.5, δ 0.5, γ 0.6, and so
forth. Each of these combinations has a certain likeli-
4.2. Bounded MLE hood of explaining a subject’s observed choices, dis-
A potential weakness of MLE is that only the likelihood played within the block parentheses in Equation (7).
drives the parameter estimations; thus, the method However, not all of the parameter values in these com-
sometimes returns parameter estimates that are con- binations are equally supported by the data. It could
sidered outliers. To circumvent this while avoiding be, for example, that the combination above that con-
the complete removal of subjects from the analysis, tains γ 0.5 is 10 times more likely to explain the
we applied an established practice of using (some- observed choices than γ 1.0. How likely each parame-
what arbitrarily chosen) bounds to the MLE estimation ter value is in relation to another value is precisely what
method. These bounds are 0.2 ≤ α ≤ 2, 0.2 ≤ λ ≤ 5, we intend to quantify. The fact that one set of parame-
0.2 ≤ δ ≤ 3, and 0.2 ≤ γ ≤ 3. They have been chosen such ters is associated with a higher likelihood of explaining
that extreme values are excluded while still allowing observed choices (and therefore is more strongly sup-
for a wide range of risk preferences. This is impor- ported by the data) can be reflected by changing the
tant because using too narrow of bounds can adversely four log-normal density functions for each of the model
affect the validity of the estimation procedure. parameters. We multiply the likelihood that a set of val-
ues explains the observed choices by the weight put on
4.3. Hierarchical Maximum Likelihood this set of parameters through density functions; Equa-
Estimation tion (7) achieves this. The density functions, and the
The HML estimation procedure developed here is product thereof, denoted by LN( · ) within the integrals,
based on the work of Farrell and Ludwig (2008). It is a are therefore ways to assess the relative plausibility of
two-step procedure that is explained below. particular parameter combinations. The maximization
Step 1. First, the likelihood of occurrence for each is done over all participants simultaneously under the
of the four model parameters in the population as a condition that it is a probability density distribution.
whole is estimated. These likelihoods of occurrence are Applying this step to our data leads to the density dis-
captured using probability density distributions and tributions displayed in Figure 2.
reflect how likely a particular parameter value is in the Step 2. In this step, we estimate individual param-
population, given everyone’s choices. These density eters as carried out in standard maximum likelihood
distributions are found by solving the integral in Equa- estimation but now weigh these parameters by the like-
tion (7). For each of the four cumulative prospect the- lihood of their co-occurrence as given by the density
ory parameters, we use a log-normal (denoted LN( · )) functions obtained in Step 1. Those distributions and
density distribution since this distribution has only both steps are illustrated in Figure 2. This second step
positive values, is positively skewed, has only two dis- of the procedure is described in Equation (7). With Mi
tribution parameters, and is not too computationally we denote M as above for subject i. The Ŝ values used
demanding. Because the sensitivity parameter ϕ is not in Equation (8) are from the output of Equation (7). The
independent of the other parameters and because it resulting parameters are driven not only by individual
has no clear psychological interpretation, we do not decision data but also by how likely it is that such a
estimate a distribution of occurrence for ϕ but instead parameter combination for an individual occurs in the
take the value we find for the aggregate data with population:
classical maximum likelihood estimation for this step.7
We use the notation M {α, λ, δ, γ, ϕ}, Sα {µ α , σα }, N
Y
Sλ {µ λ , σλ }, Sδ {µ δ , σδ }, Sγ {µ γ , σγ }, and S M̂i arg max c(yi | Mi ) LN(α | Ŝα )LN(λ | Ŝλ )
Mi i1
{Sα , Sλ , Sδ , Sγ }; S and N denote the number of partic-
ipants and the number of lotteries, respectively: · LN(δ | Ŝδ )LN(γ | Ŝγ ). (8)
Figure 2. A Simplified Example of the Two Steps of the Hierarchical Maximum Likelihood Estimation Procedure
Step 1 Step 2
3 3
← (curvature) ←
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
Density
Density
2 2
1 1
← (loss aversion) ←
0 0
0 1 2 3 0 1 2 3
Parameter value Parameter value
3 3
Density
Density
2 2
(shape)
→ →
1 1
← (elevation) ←
0 0
0 1 2 3 0 1 2 3
Parameter value Parameter value
Notes. In the first step, we fit the data to four probability density distributions. In the second step, one obtains the individual estimates. In
the absence of individual choice data, the small square has the highest likelihood for any individual, which is at the modal value of each
population-level-parameter distribution. The individual-level parameters are then considered in the context of their likelihood of occurrence
given Step 1 results, and the emerging best-fitting individual parameter is depicted by the big squares. It is worth noting that this visual
representation is a simplification, as the HML procedure is multidimensional in practice (see Equations (7) and (8)).
the second step, the likelihood of a particular parame- the best parameter estimates for the DM will simply be
ter value is determined by how well it fits the observed those for the population average.
choices from an individual, and at the same time, it is The estimation method has been implemented sep-
weighted by the likelihood of observing that particular arately in both MATLAB and C. Parameter estimates
parameter set in the population distribution (defined by in the second step were found using the simplex
the density function obtained in Step 1). The param- algorithm by Nelder and Mead (1965) and initialized
eter set with the most likely combination of both of with multiple starting points to avoid local minima.
those steps is selected. The weaker the evidence is, Negative log-likelihood was used for computational
the more extreme parameter estimates are “filtered,” stability when possible. For the first step of the hier-
archical method, negative log-likelihood was not used
resulting in more plausible parameter sets nearer the
for the whole equation because of the presence of the
center of the density distribution being favored. HML
integral, so negative log-likelihood was only used for
therefore counters maximum likelihood estimation’s the multiplication across participants. Solving the vol-
tendency to fit parameters to extreme choices by requir- ume integral is computationally intense because of the
ing stronger evidence to establish extreme parameter number of dimensions and the low probability val-
estimates. Thus, if a DM makes very consistent choices, ues therein. Quadrature methods were found to be
the population parameter densities will have very little prohibitively slow. Because the integral-maximizing
effect on the estimates. Conversely, if a DM has greater parameters are what is of central importance, and not
inconsistency in his choices, the population parameters the precise value of the integral, we used Monte Carlo
will have more influence on the individual estimates. integration with 2,000 (uniform random) values for
In the extreme, if a DM has wildly inconsistent choices, each dimension as a proxy. We repeated this 50 times
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
10 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
and used the average of these estimates as the final This shows that despite values being drawn from a uni-
distribution parameters.8 Another possible benefit of form distribution, HML is flexible and robust enough
HML is that the Step 1 distributions appear to be very to deal with it. These parameter recovery results are
stable. With different populations it may be possible to consistent with the empirical results, which we turn to
simply take the distributions found in previous stud- next.
ies, whereas other populations may require their own
aggregate estimations.9 6. Empirical Results
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
to perform well. The correlation between the predicted Although this seems counterintuitive, it is the fit (like-
and observed probability of picking A over B is 0.93. lihood) that drives the MLE parameter estimates and
Details are shown in Appendix D. not the fraction of explained choices. For most partici-
pants, the HML parameter estimates are different from
6.2. Individual Risk Preference Estimates the MLE estimates, but the different estimates do not
When benchmarking the explained fraction of choices result in a different percentage of explained choices.
and the fit of individual parameters compared to their So far we have established that the HML estimates
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
estimated aggregate counterparts (complete pooling of do not hurt the explained fraction of choices accounted
choices, where each individual is assigned the aggre- for, even if it has a somewhat lower fit in-sample. The
gate parameters), we find that individual parameters next and arguably more important question is whether
provide clear benefits because both the explained frac- HML is worth the effort in terms of prediction (i.e.,
tion of choices and the fit are substantially improved does it perform better out-of-sample). The estimates of
by using individual-level parameters. After having individual parameters of the first session were used
established that individual parameters have clear to predict choices in the second session. In Table 2
value in terms of model fit, a more direct compar- we have listed simple statistics for the fit (again com-
ison between MLE and HML is warranted. Particu- parable by excluding the hierarchical component) and
larly for HML, it is important to determine to what the explained fractions of all three estimation methods
degree its parameter estimates sacrifice in-sample per- for the second session. Both methods explain approxi-
formance, which is the performance of the individ- mately the same fraction of choices that people make.
ual parameters of the first session in terms of how It is interesting to compare this with the baseline pre-
well the parameters summarize the actual choice data. diction that participants make the same choices in both
Individual risk preference parameters were estimated sessions. That approach correctly predicts 71% on aver-
for each subject using both estimation methods. Fits age. Judging by the fraction of lotteries explained, it
are expressed in deviance11 (lower is better because is not immediately obvious which method is better.
this value is twice the negative log-likelihood) and are A more useful and sensitive measure to compare the
directly comparable because the hierarchical compo- estimation methods is the goodness-of-fit statistic. Fig-
nent is excluded from the deviance values for evalu- ure 3 compares the fit for MLE and HML parameters.
ation purposes. The mean and standard deviation of When comparing HML to MLE, we find that HML’s
the subject’s fits are displayed in Table 2. As to be out-of-sample fit is better on average and its deviance is
expected, the log-likelihood is lowest for MLE because lower for 92 out of 142 participants. Using a paired sign
that method attempts to find the set of parameters rank test for medians, we find that the fit for HML is
with the best possible fit in-sample. However, HML is significantly lower (better) than for MLE and bounded
very close behind, and more importantly it has sub- MLE (p < 0.001). The results are perhaps clearer when
stantial advantages in terms of reliability, which we examining the fit in the figure, which shows many
explore in the next section. points around the diagonal and the majority below it
We also determined the explained fraction of choices (lower deviance is better). This shows that HML sacri-
for each individual. The means and standard devi- ficed only a small part of its (in-sample) fit at time 1,
ations are displayed in Table 2. Both methods explain for an improved (out-of-sample) fit at time 2.
76% of choices. For some participants the HML esti- Examination of the differences in parameter esti-
mates explain a larger proportion of choices at time 1. mates for those participants for which HML outper-
Table 2. Mean and Standard Deviation (in Parentheses) for forms MLE reveals higher MLE mean parameter esti-
the Three Estimation Methods mates for the individual probability weighting func-
tion parameters. The use of bounded MLE, where
Explained (time 1) Predicted (time 2) this issue is reduced, did not change the results be-
Fraction cause HML still outperformed MLE with boundaries.
MLE 0.76 (0.09) 0.73 (0.09) In many instances, HML offers an improvement over
MLE (bounded) 0.76 (0.09) 0.73 (0.09) MLE for seemingly plausible MLE estimates, so HML’s
HML 0.76 (0.08) 0.73 (0.09) effect is not limited to outliers or only extreme param-
Fit eter values. Interestingly, some of the cases in which
MLE 83.96 (21.13) 103.11 (27.42)
MLE outperforms HML are those in which extreme
MLE (bounded) 85.00 (20.92) 101.77 (26.45)
HML 86.64 (21.66) 99.31 (24.99) low values emerge that are consistent with choices in
the second experimental session, such as loss seeking
Notes. Explained refers to the time 1 statistics given that time 1 esti-
(less weight on losses relative to gains) or weighting
mates are used, predicted refers to time 2 statistics using parameters
from time 1. The explained fraction of choices is approximately the functions that are nearly binary step functions. This
same. The fit at time 2 is significantly better for the HML method lack of sensitivity highlights one of the potential down-
and indicated by using signed rank tests. sides of HML because it may pull its estimates toward
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
12 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
Figure 3. The Goodness-of-Fit at Time 2 When Using (Bounded) MLE Parameters from Time 1 (x axis) and HML Parameters
from Time 1 (y axis)
200
170
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
140
Deviance HML
110
80
50
the peak of the aggregate distribution too much in The highest test-retest correlations are obtained using
some cases. HML, which yields the most reliable parameter esti-
mates for all parameters. Additionally, bootstrapping
6.3. Parameter Reliability on the differences in correlation between HML and the
A key issue in measuring preferences is the reliabil- other two methods reveals HML’s estimates are statisti-
ity of the parameter estimates. This viewpoint is con- cally more reliable for the α, δ, and γ parameters (with
sistent with considering risk preferences as a trait, a 95% intervals).
stable characteristic of an individual DM, and amena- Another form of parameter reliability is the change
ble to psychological interpretation. Reliability can be in parameter estimates resulting from small changes
assessed by applying the different estimation proce- in choice input. One way of establishing this form of
dures independently on the same data from the two reliability is by temporarily changing a choice on one
experimental sessions. Large changes in parameter val- of a subject’s lotteries, after which the parameters are
ues between time 1 and 2 are a problem and indicate reestimated. This is repeated for each of the 91 lot-
that we are mostly measuring/modeling noise instead teries separately and one at a time. HML’s estimates
of real individual risk preferences. At the individual appear to be fairly robust to this procedure, whereas
level we can examine the test-retest correlations of the MLE’s estimates are less so. In Figure 4 the results of
parameters from the estimation methods, displayed reestimation after perturbing one choice is plotted for
in Table 3. MLE provides rather unreliable estimates. two representative subjects using both MLE and HML.
The test-retest reliability for MLE can be improved A single different choice can yield large changes in
by applying admittedly somewhat arbitrary bounds. MLE parameter estimates and this contributes to unre-
liable parameter estimates. This fragility is an unde-
Table 3. Test-Retest Correlation Coefficients (r Values) for sirable property because minor inattention, confusion,
Parameter Values Obtained Using the Listed Estimation or a mistake from a DM would lead to very differ-
Procedures for Both Sessions Independently ent conclusions about that person’s risk preferences.
The robustness of HML estimates is a general finding
rα rλ rδ rγ
among subjects in our sample. Moreover, the delicate-
Correlations ness of the MLE estimates quickly becomes more prob-
MLE 0.25 0.37 0.12 0.19 lematic as the number of aberrant choices increases by
MLE (bounded) 0.29 0.47 0.12 0.43 only a few more. In these noisier instances, HML esti-
HML 0.46 0.56 0.49 0.67
mates continue to be more resilient.
Notes. The use of parameter bounds 0.2 ≤ α ≤ 2, 0.5 ≤ λ ≤ 5, 0.2 ≤
δ ≤ 3, 0.2 ≤ γ ≤ 3 improves the test-retest correlation for MLE. The
highest test-retest correlations are obtained with HML. Bootstrap-
7. Discussion
ping was used to test the significance of the increases in correlation This paper illustrates the merits of a hierarchical max-
that result from using HML. Significant improvements occur for α, imum likelihood (HML) procedure in estimating indi-
γ, and δ over both MLE methods. vidual risk parameters. The HML method does reduce
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS 13
Figure 4. Parameter Outcomes of MLE Parameter Estimates for Two Typical Subjects If There Is a Change to a Single Choice
for Each of the 91 Items
Subject A—MLE
90
← MLE
Frequency
estimate
60
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
30
0
0 1 2 0 1 2 0 2 4 0 1 2
Subject A—HML
90
Frequency
60
30
0
0 1 2 0 1 2 0 2 4 0 1 2
Subject B—MLE
90
Frequency
60
30
0
0 1 2 0 1 2 0 2 4 0 1 2
Subject B—HML
90
Frequency
60
30
0
0 1 2 0 1 2 0 2 4 0 1 2
Notes. The black line corresponds to the best MLE estimate of the original choices. The light gray horizontal line shows the range of estimations
that could emerge for each parameter given one different choice. The multimodal pattern we observe for loss aversion and the weighting
function for subject A is not a desirable feature. A momentary lapse of attention or a minor mistake from a DM may result in very different
estimated parameter combinations. Compared to MLE, HML produces more reliable estimates by being less sensitive to one aberrant choice.
the in-sample performance, albeit trivially, while at the the importance of fit and parameter reliability would
same time significantly improving the out-of-sample be necessary to evaluate the cost/benefit of HML. No
fit compared to MLE. Of central importance, HML esti- such trade-off is required here. The use of population-
mation yields parameter estimates with higher relia- level information in establishing individual parame-
bility than MLE. In terms of the bias-variance trade-off ter estimates mitigates the estimation difficulties that
often discussed in statistical model evaluation, HML ensnare MLE and bounded MLE, and consequently
dominates HML by having equivalent in-sample per- HML yields a better measure of individual differences.
formance, better out-of-sample performance, and bet- Some caveats are in order, however. The benefits
ter parameter reliability, all while using the exact of hierarchical modeling may, for example, diminish
same data and choice model. If improvements had when more choice data are available. In cases where
occurred in either out-of-sample fit or parameter relia- many choices have been observed, the signal can more
bility, but not both simultaneously, a trade-off between easily be teased apart from the noise and hierarchical
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
14 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
methods may be less fruitful. Its benefits may also to verify whether or not the calibrated model outper-
depend on the set of lotteries used. For lotteries specifi- forms other methods. In this multitest setup, the corre-
cally designed to test between models, the consequence lation structure among parameters could also be esti-
being a narrow range of potential parameter values, mated and used as an additional source of information
the choices alone may be insufficient to reliably esti- in breaking “ties” in goodness of fit. As correlations
mate individual risk preference parameters. Further- among parameters exist, the use of partially dependent
more, cumulative prospect theory and the instance multivariate distributions may be useful. These exten-
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
implemented here are but one particular choice model. sions may be of both theoretical and practical interest,
The possibility exists that other models fit choice data but are beyond the scope of the paper.
better and that the resulting parameter estimates are A major conclusion from the current paper is that
even more reliable over time; in this case there may not the estimation method can greatly affect the inferences
be sufficient added value from hierarchical parameter one makes about individual risk preferences, particu-
larly in multiparameter risky choice models. A number
estimation.
of different parameter combinations may produce vir-
In this paper we have not discussed some alterna-
tually the same fit result in-sample. When the interpre-
tive hierarchical estimation methods. Bayesian models
tation of individual parameter values is used to make
have gained attention in the literature recently (Zyphur
predictions about what an individual is like (e.g., psy-
and Oswald 2013). A frequentist model is used here in chological traits) or what she will do in the future (out-
part for its transparency and in part because it requires of-sample prediction), it is important to realize that
the addition of only one term in the likelihood func- other constellations of parameters are virtually equally
tion, given that the hierarchical distribution is known plausible in terms of fit but may lead to vastly differ-
or can be reasonably approximated. Our data indi- ent conclusions about individual differences and pre-
cate that this distribution is fairly stable in a popula- dictions about behavior. This paper shows that using
tion and can be well represented with the log-normal HML estimation methods can help us extract more
function. Future results may indicate that this distri- “signal” from a set of noisy choices and thus yield a
bution can be applied “out of the box” to different better measure of people’s innate risk preferences.
contexts, which would make the HML method poten-
tially attractive as a general tool for measuring risk Acknowledgments
preferences. The simple distributions used here stand The authors thank Michael Schulte-Mecklenbeck, Thorsten
in sharp contrast with the substantial flexibility offered Pachur, and Ralph Hertwig for the use of the experimental
data analyzed in this paper. These data are a subset from a
by Bayesian methods. However, the “rigidity” of HML
large decision-making study conducted at the Max Planck
is (perhaps unexpectedly) valuable because it produces Institute for Human Development in Berlin. The authors also
better out-of-sample performance and improved reli- thank Helga Fehr-Duda, Thomas Epper, and Didier Sornette
ability. The simplicity of the hierarchical component for helpful comments and suggestions on this project. This
of the HML method appears to prevent overfitting work was supported in part by the ETH Zürich Risk Center
in-sample, which can undermine out-of-sample perfor- and is related to Future Resilient Systems (FRS) Module 3.1
mance. Along these lines there is evidence of exces- from the Singapore ETH Centre.
sive shrinkage (Scheibehenne and Pachur 2013, 2015)
with existing Bayesian hierarchical estimation methods Appendix A. Example of a Binary Risky Choice
when applied to cumulative prospect theory. A one-shot binary lottery (also referred to as a gamble or a
The hierarchical component that is used here is rel- prospect) is a common tool for studying risky decision mak-
atively simple. It is possible to modify both the hier- ing.12 It is so common that it has been called the fruit fly of
decision research (Lopes 1983). In these simple risky choices,
archical component as well as the extent to which
the hierarchical method modulates the individual risk Table A.1. This Is an Example of a Simple, One-Shot Risky
preference estimates, and there are certainly more com- Decision Where the Choice Is Simply Whether to Select
plicated ways to implement the two-step procedure. Option A or Option B
One way to extend the HML method is to fit a new
Option A Option B
parameter that influences the weight of the hierarchi-
cal component by maximizing the fit in a retest, for $1 0.62 $37 0.41
with probability with probability
example by truncating the probability of the hierarchi- $83 0.38 $24 0.59
cal component to a lower bound and scaling it accord- Notes. This is one of the actual items used in the research, and the
ingly. Determining the advantages of such a method full list of items is included in Appendix E. The expected value of
option A is higher ($32.16) than that of option B ($29.33), but option
would require an experiment with three sessions: one
A contains the possibility of an outcome with a relatively small value
to obtain parameter estimates, another to calibrate the ($1). As it turns out, the majority of DMs prefer option B, even though
hierarchical weighting parameters, and a third session it has a smaller expected value.
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS 15
15 0.8
5 0.6
–5 0.4
–15 0.2
– 25 0
– 25 –15 –5 5 15 25 0 0.2 0.4 0.6 0.8 1.0
x (objective value) p (objective probability)
Notes. The parameters used to plot the solid lines are from Tversky and Kahneman (1992) and reflect the empirical findings at the aggregate
level. The dashed lines represent risk-neutral preferences and veridical probability weighting. The solid line for the value function is concave
over gains and convex over losses and further exhibits a “kink” at the reference point (in this case the origin) consistent with loss aversion. The
probability weighting function overweights small probabilities and underweights large probabilities. It is worth noting that at the individual
level, the plots can differ significantly from these aggregate-level plots.
DMs are presented with monetary outcomes x i , each asso- make many independent choices; from the overall pattern
ciated with an explicit probability p i . Consider, for example, of behavior, researchers can then draw inferences about the
the lottery in Table A.1, offering a DM the choice between DM’s underlying risk preferences. Preferences are revealed
option A, a payoff of $1 with probability 0.62 or $83 with in this way, although the mapping from choices to model
probability 0.38, or option B, a payoff of $37 with probability parameters is not so straightforward.
0.41 or $24 with probability 0.59. The DM is called upon to
choose either option A or option B and then the lottery can
Appendix C. Interparameter Correlation
be played out via a random process for real consequences,
Table C.1 reports the interparameter correlations. The corre-
thus ensuring incentive compatibility.
lation between α (utility curvature) and ϕ (the noise param-
The expected value maximizing solution to the decision
eter) is particularly low. This is not surprising because the
problem in Table A.1 is straightforward. Although this deci-
value of α primarily affects utility (and scales its magnitude),
sion policy has desirable properties, it is often not an accu-
whereas ϕ operates directly on the resulting utility.
rate description of what people prefer nor what they choose.
Similar degrees of noise for lower values of α therefore lead
Different people have different tastes of course (de gustibus
to lower values of ϕ.13 Other correlations are low or at most
non est disputandum), and this heterogeneity includes pref-
moderate. Note that it is not implausible that such correlations
erences for risk. For example, the majority of incentivized
are real (i.e., not a measurement artifact), which may suggest
DMs select option B from the lottery shown in Table A.1,
some as yet undetermined (partial) common cause.
indicating, perhaps, that these DMs have some degree of
risk aversion. However, the magnitude of this risk aversion
is still unknown and it cannot be estimated from only one Appendix D. Aggregate Stochastic Predictions
choice resulting from a binary lottery. This limitation has led The workings of the stochastic choice function, shown in
researchers to use larger sets of binary lotteries where DMs Equation (4), can be verified by comparing its predicted
Table C.1. Interparameter Correlation for MLE, Bounded MLE, and HML
α λ δ γ ϕ α λ δ γ ϕ α λ δ γ ϕ
α — −0.10 0.07 −0.05 −0.62 — −0.06 0.22 −0.04 −0.58 — −0.21 −0.25 0.14 −0.44
λ — 0.05 0.03 −0.10 — −0.14 0.01 −0.15 — −0.15 −0.06 −0.18
δ — 0.27 −0.07 — 0.07 −0.23 — 0.11 0.12
γ — −0.11 — −0.13 — 0.05
ϕ — — —
Note. The correlation between α (utility curvature) and ϕ (noise parameters) for both MLE methods stands out.
Murphy and ten Brincke: Hierarchical Parameter Estimation for Prospect Theory
16 Management Science, Articles in Advance, pp. 1–19, © 2017 INFORMS
Figure D.1. The Predicted Fraction of Subjects Choosing A over B vs. the Actual Fraction of Subjects Choosing A over B
Time 1 data Time 2 data
1.0
0.8
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
Data, A over B
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1.0 0 0.2 0.4 0.6 0.8 1.0
Modeled fraction, A over B Modeled fraction, A over B
Notes. Each dot represents one lottery. The correlation for both time 1 data (explanation) and time 2 data (prediction) is 0.93.
proportion of DMs selecting an option to the actu- well as time 2 data, the correlation between the mod-
ally observed proportion of DMs choosing that option. eled fractions and the actual fractions is very high
This is shown in Figure D.1. For both time 1 as at 0.93.
Appendix E. Lotteries
Table E.1. Lotteries Used in the Experiment
% that chose A
% that chose A
Huettel SA, Stowe CJ, Gordon EM, Warner BT, Platt ML (2006) Neu- Rieskamp J (2008) The probabilistic nature of preferential choice.
ral signatures of economic preferences for risk and ambiguity. J. Experiment. Psych. Learn. Memory Cognition 34(6):1446–1465.
Neuron 49(5):765–775. Roberts S, Pashler H (2000) How persuasive is a good fit? A comment
Ingersoll J (2008) Non-monotonicity of the Tversky-Kahneman on theory testing. Psych. Rev. 107(2):358–367.
probability-weighting function: A cautionary note. Eur. Financial Samuelson PA (1938) A note on the pure theory of consumers’
Management 14(3):385–390. behaviour. Economica 5(17):61–71.
Kahneman D, Tversky A (1979) Prospect theory: An analysis of deci- Scheibehenne B, Pachur T (2013) Hierarchical Bayesian modeling:
sion under risk. Econometrica 47(2):263–291. Does it improve parameter stability? Proc. 35th Annual Conf. Cog-
Kahneman D, Tversky A (2000) Choices, values and frames. Kahne- nitive Sci. Soc., Berlin, 1277–1282.
Downloaded from informs.org by [134.148.10.13] on 13 January 2017, at 05:14 . For personal use only, all rights reserved.
man D, Tversky A, eds. Choices, Values and Frames (Cambridge Scheibehenne B, Pachur T (2015) Using Bayesian hierarchical param-
University Press, Cambridge, UK), 1–16. eter estimation to assess the generalizability of cognitive models
Karmarkar US (1979) Subjectively weighted utility and the Allais of choice. Psychonomic Bull. Rev. 22(2):391–407.
paradox. Organ. Behav. Human Performance 24(1):67–72. Schulte-Mecklenbeck M, Pachur T, Murphy RO, Hertwig R (2016)
Knight FH (1921) Risk, Uncertainty, and Profit (Houghton Mifflin Prospect theory tracks selective allocation of attention. Working
Company, Chicago). paper, Max Planck Institute, Berlin.
Köbberling V, Wakker PP (2005) An index of loss aversion. J. Econom. Starmer C (2000) Developments in non-expected utility theory: The
Theory 122(1):119–131. hunt for a descriptive theory of choice under risk. J. Econom.
Lewandowsky S, Farrell S (2010) Computational Modeling in Cognition: Literature 38(2):332–382.
Principles and Practice (Sage, Los Angeles). Stevens SS (1957) On the psychophysical law. Psych. Rev. 64(3):
Lichtenstein S, Slovic P (1971) Reversals of preference between bids 153–181.
and choices in gambling decisions. J. Experiment. Psych. 89(1): Stewart N, Chater N, Brown GDA (2006) Decision by sampling. Cog-
46–55. nitive Psych. 53(1):1–26.
Lichtenstein S, Slovic P (2006) The Construction of Preference (Cam- Stewart N, Chater N, Stott HP, Reimers S (2003) Prospect relativity:
bridge University Press, Cambridge, UK). How choice options influence decision under risk. J. Experiment.
Loomes G, Sugden R (1995) Incorporating a stochastic element into Psych.: General 132(1):23–46.
decision theories. Eur. Econom. Rev. 39(3):641–648. Stott HP (2006) Cumulative prospect theory’s functional menagerie.
Lopes LL (1983) Some thoughts on the psychological concept of risk. J. Risk Uncertainty 32(2):101–130.
J. Experiment. Psych. Human Perception Performance 9(1):137–144. Tanaka T, Camerer CF, Nguyen Q (2010) Risk and time preferences:
Luce RD (1959) Individual Choice Behavior: A Theoretical Analysis (John Linking experimental and household survey data from Vietnam.
Wiley & Sons, New York). Amer. Econom. Rev. 100(1):557–571.
Luce RD, Raiffa H (1957) Games and Decisions: Introduction and Critical Toubia O, Johnson EJ, Evgeniou T, Delquié P (2013) Dynamic experi-
Survey (John Wiley & Sons, New York). ments for estimating preferences: An adaptive method of elicit-
Luce RD, Suppes P (1965) Preference, utility, and subjective probabil- ing time and risk parameters. Management Sci. 59(3):613–640.
ity. Luce RD, Bush RR, Galante E, eds. Handbook of Mathematical Tversky A, Kahneman D (1992) Advances in prospect theory: Cumu-
Psychology (John Wiley & Sons, New York), 249–410. lative representations of uncertainty. J. Risk Uncertainty 5(4):
McFadden D (1980) Econometric models for probabilistic choice 297–323.
among products. J. Bus. 53(3):S13–S29. Varian HR (2006) Revealed preference. Szenberg M, Ramrattan L,
Mosteller F, Nogee P (1951) An experimental measurement of utility. Gottesman AA, eds. Samuelsonian Economics and the Twenty-First
J. Political Econom. 59(5):371–404. Century (Oxford University Press, Oxford, UK), 99–115.
Nelder JA, Mead R (1965) A simplex method for function minimiza- Wakker PP (2005) Formalizing reference dependence and initial
tion. Comput. J. 7(4):308–313. wealth in Rabin’s calibration theorem. Working paper, Erasmus
Nilsson H, Rieskamp J, Wagenmakers EJ (2011) Hierarchical University, Rotterdam, Netherlands.
Bayesian parameter estimation for cumulative prospect theory. Wakker PP (2010) Prospect Theory: For Risk and Ambiguity (Cambridge
J. Math. Psych. 55(1):84–93. University Press, Cambridge, UK).
Parducci A (1995) Happiness, Pleasure, and Judgment: The Contextual Wetzels R, Vandekerckhove J, Tuerlinckx F, Wagenmakers EJ (2010)
Theory and Its Applications (Psychology Press, New York). Bayesian parameter estimation in the expectancy valence model
Payne JW, Bettman JR, Johnson EJ (1992) Behavioral decision of the Iowa gambling task. J. Math. Psych. 54(1):14–27.
research: A constructive processing perspective. Ann. Rev. Psych. Willemsen MC, Johnson EJ (2011) Visiting the decision factory:
43(1):87–131. Observing cognition with MouselabWEB and other informa-
Pitt MA, Myung IJ (2002) When a good fit can be bad. Trends in tion acquisition methods. Schulte-Mecklenbeck M, Kühberger
Cognitive Sci. 6(10):421–425. A, Ranyard R, eds. A Handbook of Process Tracing Methods for
Prelec D (1998) The probability weighting function. Econometrica Decision Research: A Critical Review and User’s Guide (Taylor and
66(3):497–527. Francis, New York), 21–42.
Rabin M (2000) Risk aversion and expected-utility theory: A calibra- Willemsen MC, Johnson EJ (2014) Mouselab software. Last accessed
tion theorem. Econometrica 68(5):1281–1292. November 17, 2016, http://www.mouselabweb.org.
Regenwetter M, Robinson M (2016) The construct-behavior gap in Zeisberger S, Vrecko D, Langer T (2012) Measuring the time stability
behavioral decision research: A challenge beyond replicability. of prospect theory preferences. Theory Decision 72(3):359–386.
Working paper, University of Illinois at Urbana–Champaign, Zyphur MJ, Oswald FL (2013) Bayesian probability and statistics in
Champaign. management research: A new horizon. J. Management 39(1):5–13.