Download C:\papers\ee\loss\papers\Expected Utility Theory and Prospect

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Ars Conjectandi wikipedia , lookup

Probability interpretations wikipedia , lookup

Inductive probability wikipedia , lookup

Mixture model wikipedia , lookup

Expected utility hypothesis wikipedia , lookup

Risk aversion (psychology) wikipedia , lookup

Transcript
Expected Utility Theory and Prospect Theory:
One Wedding and A Decent Funeral
by
Glenn W. Harrison and E. Elisabet Rutström*
June 2006
Abstract. Choice behavior is typically evaluated by assuming that the data is
generated by one latent decision-making process or another. What if there are two
latent decision-making processes generating the observed choices? Some choices
might then be better characterized as being generated by one process, and other
choices by the other process. A finite mixture model can be used to estimate the
parameters of each decision process at the same time as one estimates the probability
that each process applies to the sample. We consider the canonical case of lottery
choices and assume that the data is generated by expected utility theory and prospect
theory decision rules. We jointly estimate the parameters of each theory as well as the
fraction of choices characterized by each. The methodology provides the wedding
invitation, and the data consummates the ceremony with a decent funeral for the
representative agent model that assumes only one type of decision process and
homogenous preferences. The evidence suggests support for each theory, and goes
further to identify under what demographic domains one can expect to see one
theory perform better than the other. We therefore propose a statistical
reconciliation of the debate over two of the dominant theories of choice under risk,
at least for the tasks and samples we consider. The methodology is broadly applicable
to a range of debates over competing theories generated by experimental and nonexperimental data.
*
Department of Economics, College of Business Administration, University of Central Florida,
USA. E-mail contacts: [email protected] and [email protected].
Rutström thanks the U.S. National Science Foundation for research support under grant NSF/IIS
9817518, and Harrison and Rutström are grateful for support under grant NSF/HSD 0527675. We
thank Ryan Brosette, Harut Hovsepyan, David Millsom and Bob Potter for research assistance, and
Steffen Andersen, Vince Crawford, Curt Eaton, John Hey, Peter Kennedy, Jan Kmenta, Peter
Wakker and seminar participants for helpful comments. Supporting data and instructions are stored
at the ExLab Digital Library at http://exlab.bus.ucf.edu.
One of the enduring contributions of behavioral economics is that we now have a rich set of
competing models of behavior in many settings. Debates over the validity of these models have
often been framed as a horse race, with the winning theory being declared on the basis of some
statistical test where the theory is represented as a latent process explaining the data. In other words,
we seem to pick the best theory by “majority rule.” If one theory explains more of the data than
another theory, we declare it the better theory and discard the other one. The problem with this
approach is that it does not recognize the possibility that several behavioral latent processes may
coexist in a population.
Ignoring this possibility can lead to erroneous conclusions about the domain of applicability
of each theory, and is likely an important reason for why the horse races appear to pick different
winners in different domains. Heterogeneity in responses is well recognized as causing statistical
problems in experimental data. However, allowing for heterogeneity in responses through standard
methods, such as fixed or random effects, is not helpful when we want to identify which people
behave according to what theory, and when. For purely statistical reasons, if we have a belief that
there are two or more latent population processes generating the observed sample, one can make
more appropriate inferences if the data are not forced to fit a specification that assumes one latent
population process.
One way to allow explicitly for alternative theories is to simply collect sufficient data at the
individual level, and test the different theories for that individual. One classic example of this
approach is the experimental study of expected utility theory by Hey and Orme [1994]. In one
experimental treatment they collected 200 binary choice responses from each of 80 subjects, and
were able to test alternative estimated models for each individual. Thus, at the beginning of their
discussion of results (p.1300), they state simply, “We assume that all subjects are different. We
-1-
therefore fit each of the 11 preference functionals discussed above to the subject’s stated preferences
for each of the 80 subjects indvidually.” Their approach would not be so remarkable if it were not for
the prevailing popularity of the “representative agent” assumption in previous tests of expected
utility theory.
We consider the more familiar case in which one cannot collect considerable data at the level
of an individual. How is one to account for heterogeneous population processes in such settings? One
solution is to use statistical mixture models, in which the “grand likelihood” of the data is specified
as a probability-weighted average of the likelihood of the data from each of K>1 models, where
each model implies a distinct and latent data-generating process. With two modeling alternatives the
overall likelihood of a given observation is just the probability of model A being true times the
likelihood conditional on model A being true, plus the probability of model B being true times the
likelihood conditional on model B being true.
One obvious application of this approach is to offer a statistical reconciliation of the debate
over alternative theories of choice under risk. A canonical setting for these theories has been the
choice over pairs of monetary lotteries, as reviewed by Camerer [1995]. To keep things sharply
focused, two alternative models are posited here. One is a simple expected utility theory (EUT)
specification, assuming a CRRA utility function defined over the “final monetary prize” that the
subject would receive if the lottery were played out. The other model is a popular specification of
prospect theory (PT) due to Kahneman and Tversky [1979], in which the utility function is defined
over gains and losses separately, and there is a probability weighting function that converts the
underlying probabilities of the lottery into subjective probabilities.1
1
We use the language of EUT, but prospect theorists would instead refer to the utility function as a
“value function,” and to the transformed probabilities as “decision weights.” Our implementation of PT uses
the original version in which there are no editing processes assumed. One could easily extend our approach to
-2-
Our primary methodological contribution is to illustrate how one can move from results that
rely on representative agent assumptions about the decision-making model to a richer characterization
that allows for the potential coexistence of multiple decision-making models. Our focus is on the coexistence of latent EUT and PT data-generating processes, and not just the heterogeneity of behavior
given EUT or PT.
Section 1 discusses our approach in more detail. Section 2 reviews the experimental data,
section 3 the competing models, section 4 presents the results of estimating a mixture model for
both models, and section 5 draws conclusions.
1. Weddings, Funerals and Heterogeneity
One way that heterogeneity can be recognized is by collecting information on observable
characteristics and controlling for them in the statistical analysis. For example, a given theory might
allow some individuals to be more risk averse than others as a matter of personal preference. To
allow for this when we measure risk attitudes we can condition on observable individual
characteristics and evaluate the presence of heterogeneity. Additionally, one can allow for
unobserved individual characteristics when appropriate and the data is in a panel. But this approach
only recognizes heterogeneity within a given theory. This may be important for valid inferences about the
ability of the theory to explain the data, but it does not allow for heterogeneous theories to co-exist in
the same sample.
We seek a statistical reconciliation of the debate over two dominant theories of choice under
risk, EUT and PT. We collect data in a setting in which subjects make risky decisions in a gain
frame, a loss frame, or a mixed frame, explained in detail below. Such settings allow one to test the
allow editing processes or Cumulative PT formulations.
-3-
competing models on the richest domain. All data is collected using standard procedures in
experimental economics in the laboratory: no deception, no field referents, fully salient choices, and
with information on individual characteristics of the subjects. The experimental procedures are
described in section 1.
Our experiments represent a replication and extension of the procedures of Hey and Orme
[1994]. The major extension is to consider lotteries in which some or all outcomes are framed as
losses, as well as the usual case in which all outcomes are frames as gains. The subject received an
initial endowment that resulted in their final outcomes being the same as those in the other frames.
Thus the final income level of the subjects is essentially equalized across all frames.
Our EUT specification is defined over the “final monetary prize” that the subject would
receive if the lottery were played out. That is, the argument of the utility function is the prize plus
the initial endowment, which are always jointly non-negative. Our PT specification defines the utility
function over gains and losses separately, and there is a probability weighting function that converts
the underlying probabilities of the lottery into subjective probabilities. The three critical features of
the PT model are (i) that the arguments of the utility function be gains and losses relative to some
reference point, taken here to be the endowment; (ii) that losses loom larger than gains in the utility
function; and (iii) that there be a nonlinearity in the transformed probabilities that could account for
apparently different risk attitudes for different lottery probabilities. There may be some debate over
whether the endowment serves as a “reference point” to define losses, but that is one thing that is
being tested with these different models.2 We specify the exact forms of the models tested in
2
Some might argue that negative lottery prizes would count as losses only when the subject “earns”
the initial stake. This is a fair point, but testable by simple modification of the design. There is evidence from
related settings that such changes in how subjects receive their initial endowment can significantly change
behaviour: see Cherry, Frykblom and Shogren [2002], Johnson, Rutström and George [2004] and George,
Johnson, and Rutström [2006]. Moreover, one can extend the PT specification to estimate the endogenous
-4-
section 2.
It is apparent that there are many variants of these two models, and many others that deserve
to be considered. Wonderful expositions of the major alternatives can be found in Camerer
[1995] and Hey and Orme [1994], among others. But this is not the place to array every feasible
alternative. Instead, the initial exploration of this approach should consider two major alternatives
and the manner in which they are characterized. Many of the variants involve nuances that would be
hard to empirically evaluate with the data available here, rich as it is for the task at hand.3 But the
methodological point illustrated here is completely general.
The “wedding” arises from the specification and estimation of a grand likelihood function
that allows each model to co-exist and have different weights. The data can then identify what
support each model has. The most intuitive interpretation of these estimated probabilities is that the
sample consists of two types of decisions or decison-makers: those that are relatively well
characterized by EUT, and those that are relatively well characterized by PT. The wedding is
consummated by the maximum likelihood estimates converging on probabilities that apportion nontrivial weight to both models. Thus the data, when allowed, indicates that one should not think
exclusively of EUT or PT as the correct model, but as models that are correct for distinct parts of
the sample or the decisions.
Kirman [1992; p.119] considers the dangers of using representative agent characterizations
reference point that subjects employ: see Harrison, List and Towe [2005] and Andersen, Harrison and
Rutström [2006] for examples.
3
For example, on the EUT side one would prefer to consider specifications that did not assume
constant relative risk aversion after the empirical “wake up call” provided by Holt and Laury [2002]. But the
reliable estimation of the parameters of such specifications likely requires a much wider range of prizes than
considered here: hence the design of Holt and Laury [2002], where prizes were scaled by a factor of 20 for
most of their sample in order to estimate such a flexible specification. For an example on the PT side, there
are a plethora of specifications available for the probability weighting function (e.g., Wu and Gonzalez [1996]
and Prelec [2002]).
-5-
in macroeconomics, and argues that “... it is clear that the ‘representative’ agent deserves a decent
burial, as an approach to economics analysis that is not only primitive, but fundamentally
erroneous.” The “funeral” we offer arises from the ability to identify which individual or task
characteristics are associated with subjects being better characterized as EUT decision-makers rather
than as PT decision-makers, or vice versa. We can “bury” a theory for a specific domain. The
funeral we provide is “decent” because the ability to identify empirically valid domains for each
theory is based on the recognition of the dangers of statistical inference using representative agent
models with homogenous preferences.4 Moreover, within each type, we can identify those
individual characteristics associated with different preferences. For example, we can identify which
EUT decision-makers are more or less risk averse, or which PT decision-makers are more or less
averse to losses.
Mixture models have an astonishing pedigree in statistics: Pearson [1894] examined data on
the ratio of forehead to body length of 1000 crabs to illustrate “the dissection of abnormal
frequency curves into normal curves.” In modern parlance he was allowing the observed data to be
generated by two distinct Gaussian processes, and estimated the two means and two standard
deviations. Modern surveys of the evolution of mixture models are provided by Titterington, Smith
and Makov [1985], Everitt [1996] and McLachlan and Peel [2000]. Mixture models are also virtually
identical to “latent class models” used in many areas of statistics, marketing and econometrics, even
though the applications often make them seem quite different (e.g., Goodman [1974a][1974b] and
Vermunt and Magidson [2003]). In experimental economics, El-Gamal and Grether [1995; §6]
estimate a finite mixture model of Bayesian updating behavior, and contrast it to a related approach
4
Camerer and Ho [1994] remains the most impressive work of this genre, and contains a very clear
statement of the strengths and weaknesses of the approach (p.186).
-6-
in which individual subject behavior is classified completely as one type of the other. Stahl and
Wilson [1995] develop a finite mixture model to explain behavior in a normal form game,
differentiating between five types of boundedly rational players.5
One challenge with mixture models of this kind is the joint estimation of the probabilities
and the parameters of the conditional likelihood functions. If these are conditional models that each
have some chance of explaining the data, then mixture models will be characterized numerically by
relatively flat likelihood functions. On the other hand, if there are indeed K distinct latent processes
generating the overall sample, allowing for this richer structure is inferentially useful. Most of the
applications of mixture modeling in economics have been concerned with their use in better
characterizing unobserved individual heterogeneity in the context of a given theory about behavior,6
although there have been some applications to hypothetical survey data that consider individual
heterogeneity over alternative behavioral theories.7
5
Additional applications of mixture models to experimental data include Stahl [1996][1998], Haruvy,
Stahl and Wilson [2001], Coller et al. [2003], Bardsley and Moffatt [2005], Harrison, Humphrey and
Verschoor [2005] and Andersen et al. [2005][2006].
6
For example, Heckman and Singer [1984], Geweke and Keane [1999] and Araña and León [2005].
The idea here is that the disturbance term of a given specification is treated as a mixture of processes. Such
specifications have been used in many settings that may be familiar to economists under other names. For
example, stochastic frontier models rely on (unweighted) mixture specifications of a symmetric “technical
efficiency” disturbance term and an asymmetric “idiosyncratic” disturbance term; see Kumbhakar and Lovell
[2000] for an extensive review.
7
For example, Werner [1999] uses mixture models to characterize the “spike at zero” and “nonspike” responses common in contingent valuation surveys. Wang and Fischbeck [2004] provide an
application to framing effects within prospect theory, using hypothetical field survey data on health insurance
choices. Their approach is to view the frames as different data generation processes for responses.
-7-
2. Experimental Design
Subjects were presented with 60 lottery pairs, each represented as a “pie” showing the
probability of each prize. The subject could choose the lottery on the left or the right, or explicitly
express indifference (in which case the experimenter would flip a coin on the subject’s behalf). After
all 60 lottery pairs were evaluated, three were selected at random for payment. The lotteries were
presented to the subjects in color on a private computer screen,8 and all choices recorded by the
computer program. This program also recorded the time taken to make each choice. An appendix
presents the instructions given to our subjects, as well as an example of the “screen shots” they saw.
In the gain frame experiments the prizes in each lottery were $0, $5, $10 and $15, and the
probabilities of each prize varied from choice to choice, and lottery to lottery. In the loss frame
experiments subjects were given an initial endowment of $15, and the corresponding prizes from the
gain frame lotteries were transformed to be -$15, -$10, -$5 and $0. Hence the final outcomes,
inclusive of the endowment, were the same in the gain frame and loss frame. In the mixed frame
experiments subjects were given an initial endowment of $8, and the prizes were transformed to be $8, -$3, $3 and $8, generating final outcomes inclusive of the endowment of $0, $5, $11 and $16.9
In addition to the fixed endowment, each subject received a random endowment between $1
and $10. This endowment was generated using a uniform distribution defined over whole dollar
amounts, operationalized by a 10-sided die.
The probabilities used in each lottery ranged roughly evenly over the unit interval. Values of
8
The computer laboratory used for these experiments has 24 subject stations. Each screen is
“sunken” into the desk, and subjects were typically separated by several empty stations due to the staggered
recruitment procedures. No subject could see what the other subjects were doing, let alone mimic what they
were doing since each subject was started individually at different times.
9
These final outcomes differ by $1 from the two highest outcomes for the gain frame and mixed
frame, because we did not want to offer prizes in fractions of dollars.
-8-
0, 0.13, 0.25, 0.37, 0.5, 0.62, 0.75 and 0.87 were used.10
Subjects were recruited at the University of Central Florida, primarily from the College of
Business Administration. Each subject received a $5 fee for showing up to the experiments, and
completed an informed consent form. Subjects were deliberately recruited for “staggered” starting
times, so that the subject would not pace their responses by any other subject. Each subject was
presented with the instructions individually, and taken through the practice sessions at an individual
pace. Since the rolls of die were important to the implementation of the objects of choice, the
experimenters took some time to give each subject “hands-on” experience with the (10-sided, 20sided and 100-sided) die being used. Subjects were free to make their choices as quickly or as slowly
as they wanted.
The presentation of a given lottery on the left or the right was determined at random, so that
the “left” or “right” lotteries did not systematically reflect greater risk or greater prize range than the
other.
Our data consists of responses from 158 subjects making 9311 choices that do not involve
indifference. Only 1.7% of the choices involved explicit choice of indifference, and to simplify we
drop those.11 Of these 158 subjects, 63 participated in gain frame tasks, 37 participated in mixed
frame tasks, and 58 participated in loss frame tasks.
10
To ensure that probabilities summed to one, we also used probabilities of 0.26 instead of 0.25, 0.38
instead of 0.37, 0.49 instead of 0.50 or 0.74 instead of 0.75.
11
For the specification of likelihoods of strictly binary responses, such observations add no
information. However, one could augment the likelihood for the strict binary responses with a likelihood
defined over “fractional responses” and assume that the fraction for these indifferent responses was exactly
½. Such specifications are provided by Papke and Wooldridge [1996], but add needless complexity for present
purposes given the small number of responses involved.
-9-
3. Meet the Bride and Groom
A. Expected Utility Specification
We assume that utility of income is defined by U(s, x) = (s+x)r where s is the fixed
endowment provided at the beginning of the experiment, x is the lottery prize, and r is a parameter
to be estimated. With this specification we assume “perfect asset integration” between the
endowment and lottery prize.12 Probabilities for each outcome k, p(k), are those that are induced by
the experimenter, so expected utility is simply the probability weighted utility of each outcome in
each lottery. Since there were up to 4 implicit outcomes in each lottery i, EUi = 3k [ p(k) × U(k) ]
for k = 1, 2, 3, 4.
A simple stochastic specification was used to specify likelihoods conditional on the model.
The EU for each lottery pair was calculated for a candidate estimate of r, and the difference LEU =
EUR - EUL calculated, where EUL is the left lottery in the display and EUR is the right lottery.13 The
index LEU is then used to define the cumulative probability of the observed choice using the
logistic function: G(LEU) = Λ(LEU) = exp(LEU)/[1+exp(LEU)].
Thus the likelihood, conditional on the EUT model being true, depends on the estimates of
r given the above specification and the observed choices. The conditional log-likelihood is
12
A valuable extension, inspired by Cox and Sadiraj [2005], would be to allow s and x to be
combined in some linear manner, with weights to be estimated.
13
This may be viewed as a deterministic theory of EUT, in the sense that any difference in the EU of
the two lotteries implies that the subject will make the implied choice with certainty. It is possible to extend
this specification to include a “structural” stochastic story. Harless and Camerer [1994], Hey and Orme [1994]
and Loomes and Sugden [1995] provided the first wave of empirical studies including some formal stochastic
specification in the version of EUT tested. There are several species of “errors” in use, reviewed by Hey
[1995][2002], Loomes and Sugden [1995], Ballinger and Wilcox [1997], and Loomes, Moffatt and Sugden
[2002]. Some place the error at the final choice between one lottery or the other after the subject has decided
deterministically which one has the higher expected utility; some place the error earlier, on the comparison of
preferences leading to the choice; and some place the error even earlier, on the determination of the expected
utility of each lottery.
-10-
ln LEUT(r; y,X) = 3i [ (ln Λ(LEU) * yi=1) + (ln (1-Λ(LEU)) * yi=0) ]
where yi =1(0) denotes the choice of the right (left) lottery in task i, and X is a vector of individual
characteristics.
B. Prospect Theory Specification
Tversky and Kahneman [1992] propose a popular specification, which is employed here.
There are two components, the utility function and the probability weighting function.
Tversky and Kahneman [1992] assume a power utility function defined separately over gains
and losses: U(x) = xα if x $ 0, and U(x) = -λ(-x)β for x < 0. So α and β are the risk aversion
parameters, and λ is the coefficient of loss aversion.
There are two variants of prospect theory, depending on the manner in which the probability
weighting function is combined with utilities. The original version proposed by Kahneman and
Tversky [1979] posits some weighting function which is separable in outcomes, and has been
usefully termed Separable Prospect Theory (SPT) by Camerer and Ho [1994; p. 185]. The alternative
version, proposed by Tversky and Kahneman [1992], posits a weighting function defined over the
cumulative probability distributions. In either case, the weighting function proposed by Tversky and
Kahneman [1992] has been widely used: it assumes weights w(p) = pγ/[ pγ + (1-p)γ ]1/γ.
Assuming that SPT is the true model, prospective utility PU is defined in much the same
manner as when EUT is assumed to be the true model. The PT utility function is used instead of the
EUT utility function, and w(p) is used instead of p, but the steps are otherwise identical.14 The same
error process is assumed to apply when the subject forms the preference for one lottery over the
14
We ignore the editing processes that were discussed by Kahneman and Tversky [1979]. They have
not been used in the empirical implementations of SPT by Camerer and Ho [1994] or Hey and Orme [1994],
to take two prominent examples.
-11-
other. Thus the difference in prospective utilities is defined similarly as LPU = PUR - PUL.
Thus the likelihood, conditional on the SPT model being true, depends on the estimates of
α, β, λ and γ given the above specification and the observed choices.15 The conditional log-likelihood
is
ln LPT(α, β, λ, γ; y,X) = 3i l iPT = 3i [ (ln Λ(LPU) * yi=1) + (ln (1-Λ(LPU)) * yi=0) ].
C. The Nuptial
If we let πEUT denote the probability that the EUT model is correct, and πPT = (1-πEUT)
denote the probability that the PT model is correct, the grand likelihood can be written as the
probability weighted average of the conditional likelihoods. Thus the likelihood for the overall
model estimated is defined by
ln L(r, α, β, λ, γ, πEUT; y,X) = 3i ln [ (πEUT × l iEUT ) + (πPT × l iPT ) ].
This log-likelihood can be maximized to find estimates of the parameters.16
We allow each parameter to be a linear function of the observed individual characteristics of
the subject. This is the X vector referred to above. We consider six characteristics: binary variables
to identify Females, subjects that self-reported their ethnicity as Black, those that reported being
Hispanic, those that had a Business major, and those that reported a low cumulative GPA (below
3¼). We also included Age in years. The estimates of each parameter in the above likelihood function
actually entails estimation of the coefficients of a linear function of these characteristics. So the
15
The fact that the SPT specification has three more parameters than the EUT specification is
meaningless. One should not pick models on some arbitrary parameter-counting basis. We refer to SPT or PT
interchangeably.
16
This approach assumes that any one observation can be generated by both models, although it
admits of extremes in which one or other model wholly generates the observation. One could alternatively
define a grand likelihood in which observations or subjects are classified as following one model or the other
on the basis of the latent probabilities πEUT and πPT. El-Gamal and Grether [1995] illustrate this approach in
the context of identifying behavioral strategies in Bayesian updating experiments.
-12-
estimate of r, ^r , would actually be
^r = ^r 0 + (r^FEMALE × FEMALE) + (r^BLACK × BLACK) + (r^HISPANIC × HISPANIC) +
(r^BUSINESS × BUSINESS) + (r^GPAlow × GPAlow) + (r^AGE × AGE)
where ^r 0 is the estimate of the constant. If we collapse this specification by dropping all individual
characteristics, we would simply be estimating the constant terms for each of r, α, β, λ, µ. Obviously
the X vector could include treatment effects as well as individual effects, or interaction effects.17
The estimates allow for the possibility of correlation between responses by the same subject,
so the standard errors on estimates are corrected for the possibility that the 60 responses are
clustered for the same subject. The use of clustering to allow for “panel effects” from unobserved
individual effects is common in the statistical survey literature.18
Flexible as this specification is, it rests on some assumptions, just as many marriages start
with a pre-nuptial agreement. First, we only consider two data generating processes, despite the fact
that there are many alternatives to EUT. Still, two is double the one that is customarily assumed, and
that is the main point of our analysis: to assume more than one data generating process.
Furthermore, there is a natural “family similarity” between many of the alternatives that might make
them hard to identify. This similarity is to be expected, given the common core of “empirical facts”
about violations of EUT that spawned them. But it implies low power for any statistical specification
17
Haruvy, Stahl and Wilson [2001] also explicitly consider within-type diversity in the context of a
between-type mixture model. They find that a measure of the intensity of individual calculator use helps
explain the allocation to types.
18
Clustering commonly arises in national field surveys from the fact that physically proximate
households are often sampled to save time and money, but it can also arise from more homely sampling
procedures. For example, Williams [2000; p.645] notes that it could arise from dental studies that “collect data
on each tooth surface for each of several teeth from a set of patients” or “repeated measurements or recurrent
events observed on the same person.” The procedures for allowing for clustering allow heteroskedasticity
between and within clusters, as well as autocorrelation within clusters. They are closely related to the
“generalized estimating equations” approach to panel estimation in epidemiology (see Liang and Zeger
[1986]), and generalize the “robust standard errors” approach popular in econometrics (see Rogers [1993]).
Wooldridge [2003] reviews some issues in the use of clustering for panel effects, in particular noting that
significant inferential problems may arise with small numbers of panels.
-13-
that seeks to identify their individual contribution to explaining behavior, unless one focusses on
“trip wire” tasks that are designed just to identify these differences by setting up one or other
models for a fall if choices follow certain patterns. Second, we adopt specific parameterizations of
the EUT and PT models, recognizing that there exist many variants. On the other hand, our
parameterizations are not obviously restrictive compared to those found in the literature. Third, we
employ specific parameterizations of the link between latent preferences and observed choices: we
do not allow a “structural stochastic error” story, and we use a logistic cumulative distribution
function. Fourth, we assume that observable characteristics of the individual have a linear effect on
the parameter, and that there are no interactions. Since these characteristics will provide some of the
most entertaining moments of the marriage, as discussed below, this assumption would be worth
examining in future work (with even larger data sets). Finally, we do not consider much variation in
task domain, other than the obvious use of gain, loss or mixed frames. Our priors are that the
relative explanatory power of EUT and PT will vary with demographics and task domain, and
possibly some interaction of the two. To provide sharp results we therefore controlled the variability
of the task domain, and focus on the possible effect of demographics.19 We suspect that significantly
larger samples will be needed to fully examine the effects of demographics and task domain, even in
a simple task such as ours.
19
These priors also imply that we prefer not to use mixture specifications in which subjects are
categorized as completely EUT or PT. It is possible to rewrite the grand likelihood (CC) such that π iEUT = 1
and π iPT = 0 if l iEUT > l iPT, and π iEUT = 0 and π iPT = 1 if l iEUT < l iPT, where the subscript i now refers to the
individual subject. The problem with this specification is that it assumes that there is no effect on the
probability of EUT and PT from task domain. We do not want to impose that assumption, even for a
relatively homogenous task design such as ours. One could use a categorization mixture specification in which
each binary choice was classified as wholly EUT or PT without assuming that task domain has no effect, but
we see no numerical or inferential advantage in doing so compared to our specification.
-14-
4. Results
A. Heterogeneity of Process
Table 1 presents the maximum likelihood estimates of the conditional models and the
mixture model when we assume no individual covariates. The estimates for the conditional models
therefore represent the traditional representative agent assumption, that every subject has the same
preferences and behaves in accord with one theory or the other. The estimates for the mixture model
allow each theory to have positive probability, and therefore makes one major step towards
recognizing heterogeneity of preferences.
The first major result is that the estimates for the probability of the EUT and PT specifications indicate
that each is equally likely. Each estimated probability is significantly different from zero.20 EUT wins by
a (quantum) nose, with an estimated probability of 0.55, but the whole point of this approach is to
avoid such rash declarations of victory. The data suggest that the sample is roughly evenly split
between those subjects that are best characterized by EUT and those subjects that are best
characterized by PT. In fact, one cannot reject the formal hypothesis that the probabilities of each
theory are identical and equal to ½ (p-value = 0.490).
The second major result is that the estimates for the PT specification are very weakly
consistent with the a priori predictions of that theory when the specification is assumed to fit every
subject, but strikingly consistent with the predictions of the theory when it is only assumed to fit some of the subjects.
When the conditional PT model is assumed for all subjects, the loss aversion parameter is greater
than 1, but only equals 1.381. This is significantly different from 1 at the 5% significance level, since
20
The likelihood function actually estimates the log odds in favor of one model or the other. If we
denote the log odds as κ, one can recover the probability for the EUT model as πEUT = 1/(1+exp(κ)). This
non-linear function of κ is calculated using the “delta method,” which also provides estimates of the standard
errors and p-values. Since πEUT = ½ when κ=0, the standard p-value on the estimate of κ provides the estimate
for the null hypothesis H0: πEUT=πEUT listed in Table 1.
-15-
the p-value is 0.088 and a one-sided test is appropriate here given the priors from PT that λ>1. But
the size of the effect is much smaller than the 2.25 estimated by Tversky and Kahneman [1992] and
almost universally employed.21 However, when the same PT specification is estimated in the mixture
model, where it is only assumed to account for the behavior of some of the subjects, the estimated
value for λ jumps to 5.781 and is even more clearly greater than 1 (although the estimated standard
error also increases, from 0.223 to 1.612). Similarly, the γ parameter is estimated to be 0.942 with the
conditional PT model, and is not statistically different from 1 (p-value = 0.266), which is the special
EUT case of the probability weighting function where w(p)=p for all p. But when the mixture model
is estimated, the value of γ drops to 0.681 and is significantly smaller than 1, against consistent with
many of the priors from PT.22 Finally, the risk aversion coefficients α and β are not significantly
different from each other under the conditional model, but are different from each other under the
mixture model. This difference is not critical to PT, and is often assumed away in applied work with
PT, but it is worth noting since it points to another sense in which gains and losses are viewed
differently by subjects, which is a primary tenet of PT.
B. Heterogeneity of Process and Parameters
Table 2 presents estimates from the mixture model with all individual characteristics
included. In general we see many individual coefficients that are statistically significant, and an
overall F-test that all coefficients are jointly zero can be easily rejected with a p-value of 0.0003.
Apart from indicating marginal statistical significance, the detailed coefficient estimates of Table 2
21
For example, Benartzi and Thaler [1995; p.79] assume this value in their evaluation of the “equity
premium puzzle,” and note (p.83) that it is the assumed value of the loss aversion parameter that drives their
main result.
22
This prior is not universally held: see Humphrey and Verschoor [2004] for a review of the literature
and some contrary evidence that γ>1 in some settings.
-16-
are not particularly informative. It is more instructive to examine the distribution of coefficients
across the sample, generated by predicting each parameter for each subject by evaluating the linear
function with the characteristics of that subject.23 Such predictions are accompanied by estimated
standard errors, just as the coefficient estimates in Table 1 are, so we also take the uncertainty of the
predicted parameter value into account.
Figure 1 displays the main result, a distribution of predicted probabilities for the competing
models. The two panels are, by construction, mirror images of each other since we only have two
competing models, but there is pedagogic value in displaying both. The support for the EUT
specification is extremely high for some people, but once that support wanes the support for PT
picks up. It is not the case that there are two sharp modes in this distribution, which is what one
might have expected with priors that there are two types of subjects. Instead, subjects are better
characterized as either “clearly EUT” or “probably PT.” It is as if EUT is fine for city folk, but PT
rules the hinterland.24 No doubt this is due to many subjects picking lotteries according to their
estimate of the expected value of the lotteries, which would make them EUT-consistent subjects
that are risk neutral (r=1). The average probability for the EUT model is 0.51, as is the median,
despite the apparent skewness in the distributions in Figure 1.
Figure 2 indicates the uncertainty of these predicted probabilities. Consistent with the use of
discrimination functions such as the logistic, uncertainty is smallest at the end-points and greatest at
the mid-point. Since EUT has more of it’s support from subjects that are at one extreme probability,
23
For example, a black 21-year-old female that was not hispanic, did not have a business major, and
had an average GPA, would have an estimate of r equal to ^r 0 + ^r FEMALE + ^r BLACK + ^r AGE × 21. She would
have the same estimate of r as anyone else that had exactly the same set of characteristics.
24
Andersen et al. [2006] report the reverse qualitative pattern when considering dynamic lottery choices
in which subjects could accumulate income. We again caution that the relative explanatory power of EUT and
PT likely interacts with task domain.
-17-
this implies that one can be more confident about declaring a subject as better characterized by EUT
than one can about the alternative model.
Figures 3 through 7 display stratifications of the probability of the EUT model being correct
in terms of individual characteristics. The results are striking. Figure 3 shows that men have a much
stronger probability of behaving in accord with the EUT model than women. Figure 4 shows that
“whites” have a dramatically higher probability of being EUT decision makers than “blacks” or
“hispanics.” Consider the implications of these two Figures for the dramatic differences that List
[2002][2003][2004] finds between subjects in field experiments on the floor of sportscard shows and
conventional lab experiments with college students. The vast majority of participants on those shows
are white males. No doubt there is much more going on in these field experiments than simply
demographic mix, as stressed by Harrison and List [2004], but one might be able to go a long way to
explaining the differences just on these two demographics alone.25
Figure 5 shows that business majors exhibit behavior that is much more consistent with
EUT, helping us track homo economicus down among, thankfully, the Econ. Figure 6 expands on this
characteristic by showing that EUT is more common for those subjects with a lower GPA as
compared to those with a higher GPA; the closing credits of Animal House remind us, of course, that
those with lower GPA might become the political and economic leaders of the future. Figure 7
shows some modest effects of age on support for EUT. Our sample obviously has a limited range in
terms of age, with these ages also confounded with “standing” within the undergraduate program.
25
Haigh and List [2005] report contrary evidence, using tasks involving risky lotteries and comparing
field traders from the Chicago Board of Trade and students. However, the behavior they observe can be easily
explained under EUT using a specification of the utility function that allows non-constant relative risk
aversion, such as the expo-power specification favored by Holt and Laury [2002] in their experiments. Thus
one should not count this as contrary evidence to the claim in the text unless one constrains EUT to the
special case of CRRA.
-18-
But there does appear to be a marked reduction in support for EUT as subjects age or get more
university education.
Finally, Figures 8 and 9 display the distributions of the parameters of each model. The raw
predictions for all 158 subjects include some wild estimates, but for completely sensible reasons. If
some subject has a probability of behavior in accord with PT of less than 0.0001, then one should
not expect the estimates of α, β, λ or γ to be precise for that subject. In fact, most of the extreme
estimates are associated with parameters that have very low support for that subject, or that have
equally high standard errors. Thus, we restrict the displays in Figures 8 and 9 to those subjects
whose probability of behaving in accord with EUT or PT, respectively, is at least ¼.
Figure 7 displays results for the EUT subjects that are slightly lower than those reported in
Table 1. The average EUT subject has a CRRA coefficient r equal to 0.70, with the same median,
whereas the estimate from Table 1 was 0.85 when all EUT subjects were assumed to have the same
risk attitude. In Figure 7 we observe two modes close to the risk-neutral case, r=1, reflecting the
ability of the mixture model that includes individual characteristics to allow for heterogeneity of risk
attitudes within the subset of the sample that behave according to EUT. Some subjects exhibit
considerable risk aversion (r<1), and some even exhibit mild risk-loving behavior (r>1). From Table
2 we see that the characteristics that drive this heterogeneity include sex, ethnicity, and age.
Figure 8 displays results for the PT parameters that are generally consistent with the
estimates in Table 1. The parameters α and β are generally less than 1, consistent with mild risk
aversion in the usual sense. The average estimate for α is 0.59, with a median of 0.50. The loss
aversion coefficient λ is generally much greater than 1, consistent with the subjects that behave in
accord with PT having significant loss aversion. The average estimate for λ is 5.9, with a median of
6.8. Of course, this applies only to the fraction of the sample deemed to have certain level of
-19-
support for PT; the value of λ for EUT subjects is of course 1, so the weighted average for the
whole sample would be much smaller (as Table 1 indicates, it is 1.381 when estimated over all
subjects). The probability weighting coefficient γ is generally less than 1, which is the value at which
it collapses back to the standard EUT assumption that w(p)=p for all p. The average estimate is
0.89, and the median is 0.77.
5. Conclusion
Characterizing behavior in lottery choice tasks as generated by potentially heterogenous
processes generates several important insights.
First, it provides a framework for systematically resolving debates over competing models.26
This method does not rest on exploiting structural aspects of one type of task or another, such as
have been a staple in tests of EUT. Such extreme domains are still interesting of course, just as the
experiments employed here deliberately generated loss frames and mixed frames. But the test of the
competing models should not be restricted to such trip-wire tests, which might not characterize
behavior in a broader domain.
Second, it provides a more balanced metric for deciding which theory does better in a given
task, rather than extreme declarations of winners and losers. If some theory has little support, this
approach will still show that.
Third, the approach can provide some insight into when one theory does better than another,
through the effects of individual characteristics and treatments on estimated probabilities of support.
This insight might be fruitfully applied to other settings in which apparently conflicting findings
26
The approach readily generalizes to other contexts, such as competing models of intertemporal
discounting behavior. See Andersen et al. [2005] and Coller et al. [2003] for applications.
-20-
have been reported. We fully expect that the relative explanatory power of EUT and PT will vary
with task domain as well as demographics. So the constructive challenge is to characterize those
domains so that we know better when to use one model or the other.
Fourth, the approach readily generalizes to include more than two models, although
likelihoods are bound to become flatter as one mixes more and more models that have any support
from the data. The obvious antidote for those problems is to generate larger samples, to pool data
across comparable experiments, and to marshall more exhaustive numerical methods.
-21-
Table 1: Estimates of Parameters of Models With No Individual Covariates
Estimates from Conditional Models
Parameter
or Test
Estimate Standard Error
Estimates from Mixture Model
p-value
Lower 95%
Confidence
Interval
Upper 95%
Confidence
Interval
Estimate
Standard Error
p-value
Lower 95%
Confidence
Interval
Upper 95%
Confidence
Interval
r
0.837
0.025
0.000
0.788
0.886
0.846
0.044
0.000
0.759
0.933
α
β
λ
γ
0.730
0.733
1.381
0.942
0.033
0.062
0.223
0.052
0.000
0.000
0.000
0.000
0.665
0.611
0.943
0.841
0.796
0.854
1.820
1.044
0.614
0.312
5.781
0.681
0.057
0.132
1.612
0.047
0.000
0.019
0.000
0.000
0.501
0.052
2.598
0.587
0.727
0.572
8.965
0.774
0.550
0.450
0.072
0.072
0.000
0.000
0.407
0.308
0.692
0.592
πEUT
πPT
H0: πEUT=πPT
H0: α=β
H0: λ=1
H0: γ=1
0.490
0.973
0.088
0.266
0.046
0.003
0.000
-22-
Table 2: Estimates of Parameters of Mixture Model With Individual Covariates
0.509
0.006
0.827
0.002
0.001
0.266
0.560
Lower 95%
Confidence
Interval
-0.931
-0.473
-0.408
-0.965
0.028
-0.305
-0.113
Upper 95%
Confidence
Interval
0.464
-0.079
0.327
-0.222
0.100
0.085
0.208
0.370
0.647
0.550
0.054
0.957
0.839
0.798
0.485
0.260
0.735
0.324
0.754
0.582
0.817
0.685
0.038
0.014
0.354
0.160
0.860
0.029
0.000
0.976
0.100
0.147
0.104
0.447
0.739
0.002
0.215
0.661
0.105
0.223
0.429
-0.482
-0.519
-0.743
0.000
-0.457
-0.366
-2.348
-0.811
-0.224
-0.818
-0.014
-0.673
-0.232
-13.726
-24.411
-8.969
-9.253
-0.593
-7.287
-3.918
0.059
0.320
-0.254
-0.190
-0.037
-0.069
-0.475
-2.473
0.610
-1.374
-5.204
-0.260
-0.478
-0.721
0.181
0.323
0.397
0.045
0.483
0.449
1.809
1.704
0.824
1.157
0.042
0.928
0.411
17.366
16.068
-0.264
-1.062
1.647
1.213
3.275
1.067
0.661
0.262
2.142
0.006
0.727
0.211
3.479
2.696
6.062
8.181
0.025
2.034
1.688
Parameter
Variable
Estimate
Standard Error
p-value
r
Constant
Female
Black
Hispanic
Age
Business
GPAlow
Constant
Female
Black
Hispanic
Age
Business
GPAlow
Constant
Female
Black
Hispanic
Age
Business
GPAlow
Constant
Female
Black
Hispanic
Age
Business
GPAlow
Constant
Female
Black
Hispanic
Age
Business
GPAlow
Constant
Female
Black
Hispanic
Age
Business
GPAlow
-0.234
-0.276
-0.041
-0.594
0.064
-0.110
0.047
0.119
-0.151
-0.098
-0.173
0.022
0.013
0.042
-0.269
0.446
0.300
0.170
0.014
0.127
0.090
1.820
-4.171
-4.616
-5.157
0.527
-3.037
-0.322
0.563
0.490
0.004
0.976
-0.016
0.329
-0.132
0.503
1.653
2.344
1.489
-0.118
0.778
0.483
0.353
0.100
0.186
0.188
0.018
0.099
0.081
†
0.168
0.213
0.288
0.012
0.238
0.206
1.052
0.637
0.265
0.500
0.014
0.405
0.163
7.871
10.247
2.204
2.073
0.567
2.152
1.821
0.255
0.086
0.130
0.590
0.011
0.201
0.174
1.507
0.528
1.882
3.388
0.072
0.636
0.610
α
β
λ
γ
κ‡
Notes: † The standard error for the constant is not reliably estimated.
‡ κ is the log odds of the probabilities of each model, and πEUT = 1/(1+exp(κ)) .
-23-
Figure 1: Probability of Competing Models
EUT Model
15
PT Model
15
10
Percent
Percent
10
5
0
5
0
.1
.2
.3
.4 .5 .6
Probability
.7
.8
.9
1
0
0
.1
.2
.3
.4 .5 .6
Probability
.7
.8
.9
1
Figure 2: Uncertainty In Predicted Probability of Models
Predicted probabilities and standard errors from estimates in Table 2
Quadratic regression line added
1
.8
Predicted
Standard
Error of
Probability
.6
.4
.2
0
0
.2
.4
.6
Predicted Probability of EUT Model
-24-
.8
1
Figure 3: Sex, EUT and Prospect Theory
Males
40
30
Percent
Percent
30
20
10
0
Females
40
20
10
0
.1
.2 .3 .4 .5 .6 .7 .8
Probability of EUT Model
.9
1
0
0
.1
.2
.3 .4 .5 .6 .7 .8
Probability of EUT Model
.9
1
Figure 4: Race, EUT and Prospect Theory
Black (N=16)
80
40
20
0
80
60
Percent
Percent
60
Hispanic (N=19)
40
20
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
0
Others (N=123)
60
Percent
80
40
20
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
-25-
0
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
Figure 5: Business Majors, EUT and Prospect Theory
Non-Business Majors
25
20
20
15
15
Percent
Percent
Business Majors
25
10
10
5
5
0
0
.1
.2 .3 .4 .5 .6 .7 .8
Probability of EUT Model
.9
0
1
0
.1
.2
.3 .4 .5 .6 .7 .8
Probability of EUT Model
.9
1
.9
1
Figure 6: GPA, EUT and Prospect Theory
Low GPA
30
Higher GPA
30
20
Percent
Percent
20
10
0
10
0
.1
.2 .3 .4 .5 .6 .7 .8
Probability of EUT Model
.9
1
-26-
0
0
.1
.2
.3 .4 .5 .6 .7 .8
Probability of EUT Model
Figure 7: Age, EUT and Prospect Theory
17 or 18
50
19
50
30
30
30
Percent
40
Percent
40
Percent
40
20
20
20
10
10
10
0
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
20 and Over
50
0
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
0
0 .1 .2 .3 .4 .5 .6 .7 .8 .9 1
Probability of EUT Model
Figure 8: CRRA Parameter of the EUT Model
15
Percent
10
5
0
0
.25
.5
.75
-27-
1
1.25
Figure 9: Parameters of the PT Model
ë
Percent
Percent
á
0
.25
.5
.75
1
1.25
0
1
2
3
4
5
6
7
8
9
10 11
ã
Percent
Percent
â
0
.25
.5
.75
1
1.25
-28-
0
.5
1
1.5
2
References
Andersen, Steffen; Harrison, Glenn W.; Lau, Morten Igel, and Rutström, E. Elisabet, “Eliciting Risk
and Time Preferences,” Working Paper 05-24, Department of Economics, College of Business
Administration, University of Central Florida, 2005.
Andersen, Steffen; Harrison, Glenn W., and Rutström, E. Elisabet, “Choice Behavior, Asset
Integration, and Natural Reference Points,” Working Paper 06-07, Department of Economics,
College of Business Administration, University of Central Florida, 2006.
Araña, Jorge E., and León, Carmelo J., “Flexible mixture distribution modeling of dichotomous
choice contingent valuation with heterogeneity,” Journal of Environmental Economics &
Management, 49, 2005, forthcoming.
Ballinger, T. Parker, and Wilcox, Nathaniel T., “Decisions, Error and Heterogeneity,” Economic
Journal, 107, July 1997, 1090-1105.
Bardsley, Nicholas, and Moffatt, Peter G., “The Experimetrics of Public Goods: Inferring
Motivations from Contributions,” CEDEX Discussion Paper 2005-09, School of Economics,
University of East Anglia, June 2005.
Bolton, Gary E., and Ockenfels, Axel, “ERC: A Theory of Equity, Reciprocity, and Competition,”
American Economic Review, 90(1), March 2000, 166-193.
Bryant, Peter, and Williamson, John A., “Asymptotic Behavior of Classification Maximum
Likelihood Estimates,” Biometrika, 65(2), 1978, 273-281.
Camerer, Colin F., “Individual Decision Making,” in J. H. Kagel and A.E. Roth (eds.), The Handbook
of Experimental Economics (Princeton: Princeton University Press, 1995)
Camerer, Colin F., and Ho, Teck-Hua, “Violations of the Betweeness Axiom and Nonlinearity in
Probability,” Journal of Risk and Uncertainty, 8, 1994, 167-196.
Cherry, Todd L.; Frykblom, Peter, and Shogren, Jason F., “Hardnose the Dictator,” American
Economic Review, 92(4), September 2002, 1218-1221.
Coller, Maribeth; Harrison, Glenn W., and Rutström, E. Elisabet, “Are Discount Rates Constant?
Reconciling Theory and Observation,” Working Paper 3-31, Department of Economics,
College of Business Administration, University of Central Florida, 2003.
Cox, James C., and Sadiraj, Vjollca, “Small- and Large-Stakes Risk Aversion: Implications of
Concavity Calibration for Decision Theory,” Unpublished Manuscript, Department of
Economics, University of Arizona, May 2005; forthcoming, Games & Economic Behavior.
El-Gamal, Mahmoud A., and Grether, David M., “Are People Bayesian? Uncovering Behavioral
Strategies,” Journal of the American Statistical Association, 90, 432, December 1995, 1137-1145.
Everitt, Brian S., “An Introduction to Finite Mixture Distributions,” Statistical Methods in Medical
Research, 5, 1996, 107-127.
-29-
George, J. Greg; Johnson, Laurie T., and Rutström, E. Elisabet, “Social Preferences in the Face of
Regulatory Change,” in T. Cherry, S. Kroll and J.F. Shogren (eds.) Experimental Methods,
Environmental Economics (Oxford, UK: Routledge, 2006).
Geweke, John, and Keane, Michael, “Mixture of Normals Probit Models,” in C. Hsio, K. Lahiri, LF. Lee and M.H. Pesaran (eds.), Analysis of Panel and Limited Dependent Variables: A Volume in
Honor of G.S. Maddala (New York: Cambridge University Press, 1999).
Goodman, Leo A., “The analysis of systems of qualitative variables when some of the variables are
unobservable: Part I. A modified latent structure approach,” American Journal of Sociology, 79,
1974a, 1179–1259.
Goodman, Leo A., “Exploratory latent structure analysis using both identifiable and unidentifiable
models,” Biometrika, 61, 1974b, 215–231.
Haigh, Michael, and List, John A., “Do Professional Traders Exhibit Myopic Loss Aversion? An
Experimental Analysis,” Journal of Finance, 60(1), February 2005, 523-534.
Harless, David W., and Camerer, Colin F., “The Predictive Utility of Generalized Expected Utility
Theories,” Econometrica, 62, 1994, 1251-1289.
Harrison, Glenn W., and List, John A., “Field Experiments,” Journal of Economic Literature, 42(4),
December 2004, 1013-1059.
Harrison, Glenn W.; Humphrey, Steven J., and Verschoor, Arjan, “Choice Under Uncertainty in
Developing Countries,” Working Paper 05-29, Department of Economics, College of
Business Administration, University of Central Florida, 2005.
Harrison, Glenn W.; List, John A., and Towe, Charles, “Naturally Occurring Preferences and
Exogenous Laboratory Experiments: A Case Study of Risk Aversion,” Working Paper 04-22,
Department of Economics, College of Business Administration, University of Central
Florida, 2004.
Haruvy, Ernan; Stahl, Dale O., and Wilson, Paul, W., “Modeling and Testing for Heterogeneity in
Observed Strategic Behavior,” Review of Economics and Statistics, 83(1), February 2001, 146-157.
Heckman, James, and Singer, Brian, “A Method for Minimizing the Impact of Distributional
Assumptions in Econometric Models for Duration Data,” Econometrica, 52(2), 1984, 271-320.
Hey, John, “Experimental Investigations of Errors in Decision Making Under Risk,” European
Economic Review, 39, 1995, 633-640.
Hey, John D., “Experimental Economics and the Theory of Decision Making Under Uncertainty,”
Geneva Papers on Risk and Insurance Theory, 27(1), June 2002, 5-21.
Hey, John D., and Orme, Chris, “Investigating Generalizations of Expected Utility Theory Using
Experimental Data,” Econometrica, 62(6), November 1994, 1291-1326.
Holt, Charles A., and Laury, Susan K., “Risk Aversion and Incentive Effects,” American Economic
-30-
Review, 92(5), December 2002, 1644-1655.
Humphrey, Steven J., and Verschoor, Arjan, “The Probability Weighting Function: Experimental
Evidence from Uganda, India and Ethiopia,” Economics Letters, 84, 2004, 419-425.
Johnson, Laurie T.; Rutström, E. Elisaet, and George, J. Gregory, “Income Distribution
Preferences: The Role of History,” Working Paper 04-11, Department of Economics, College
of Business Administration, University of Central Florida, 2004; forthcoming, Journal of
Economic Behavior & Organization.
Kahneman, Daniel, and Tversky, Amos, “Prospect Theory: An Analysis of Decision under Risk,”
Econometrica, 1979, 47(2), 263-291.
Kirman, Alan P., “Whom or What Does the Representative Individual Represent?” Journal of
Economic Perspectives, 6(2), Spring 1992, 117-136.
Kumbhakar, Subal C., and Lovell, C.A. Knox, Stochastic Frontier Analysis (New York: Cambridge
University Press, 2000).
Liang, K-Y., and Zeger, S.L., “Longitudinal Data Analysis Using Generalized Linear Models,”
Biometrika, 73, 1986, 13-22.
List, John A., “Preference Reversals of a Different Kind: The More Is Less Phenomenon,” American
Economic Review, 92, 2002, 1636-1643.
List, John A., “Does Market Experience Eliminate Market Anomalies,” Quarterly Journal of Economics,
118, 2003, 41-71.
List, John A., “Neoclassical Theory Versus Prospect Theory: Evidence from the Marketplace,”
Econometrica, 72(2), March 2004, 615-625.
Loomes, Graham; Moffatt, Peter G., and Sugden, Robert, “A Microeconometric Test of Alternative
Stochastic Theories of Risky Choice,” Journal of Risk and Uncertainty, 24(2), 2002, 103-130.
Loomes, Graham, and Sugden, Robert, “Incorporating a Stochastic Element Into Decision
Theories,” European Economic Review, 39, 1995, 641-648.
McLachlan, Geoffrey, and Peel, David, Finite Mixture Models (New York: Wiley, 2000).
Papke, Leslie E., and Wooldridge, Jeffrey M., “Econometric Methods for Fractional Response
Variables with an Application to 401(K) Plan Participation Rates,” Journal of Applied
Econometrics, 11, 1996, 619-632.
Pearson, Karl, “Contribution to the mathematical theory of evolution,” Philosophical Transactions A,
185, 1894, 71-110.
Prelec, Drazen, “The Probability Weighting Function,” Econometrica, 66(3), May 1998, 497-527.
Rogers, W. H., “Regression standard errors in clustered samples,” Stata Technical Bulletin, 13, 1993,
-31-
19-23.
Rutström, E. Elisabet, and Williams, Melonie, “Entitlements and Fairness: An Experimental Study
of Distributive Preferences,” Journal of Economic Behavior & Organization, 43(1), 2000, 75-89.
Stahl, Dale O., “Boundedly Rational Rule Learning in a Guessing Game,” Games and Economic
Behavior, 16, 1996, 303-330.
Stahl, Dale O., “Is step-j thinking an arbitrary modelling restriction or a fact of human nature?”
Journal of Economic Behavior & Organization, 37, 1998, 33-51.
Stahl, Dale O., and Wilson, Paul W., “On Players’ Models of Other Players: Theory and
Experimental Evidence,” Games and Economic Behavior, 10, 1995, 218-254.
Titterington, D.M., Smith, A.F.M., and Makov, U.E., Statistical Analysis of Finite Mixture Distributions
(New York: Wiley, 1985).
Tversky, Amos, and Kahneman, Daniel, “Advances in Prospect Theory: Cumulative
Representations of Uncertainty,” Journal of Risk & Uncertainty, 5, 1992, 297-323; references to
reprint in D. Kahneman and A. Tversky (eds.), Choices, Values, and Frames (New York:
Cambridge University Press, 2000).
Vermunt, Jeroen K., and Magidson, Jay, “Latent Class Models for Classification,” Computational
Statistics & Data Analysis, 41, 2003, 531-537.
Wang, Mei, and Fischbeck, Paul S., “Incorporating Framing into Prospect Theory Modeling: A
Mixture-Model Approach,” Journal of Risk & Uncertainty, 29(2), 2004, 181-197.
Watt, Richard, “Defending Expected Utility Theory,” Journal of Economic Perspectives, 16(2), Spring
2002, 227-228.
Williams, Rick L., “A Note on Robust Variance Estimation for Cluster-Correlated Data,” Biometrics,
56, June 2000, 645-646.
Wooldridge, Jeffrey, “Cluster-Sample Methods in Applied Econometrics,” American Economic Review
(Papers & Proceedings), 93, May 1993, 133-138.
Wu, George, and Gonzalez, R., “Curvature of the Probability Weighting Function,” Management
Science, 42, 1996, 1676-1690.
-32-
Appendix A: Experimental Instructions (NOT FOR PUBLICATION)
This is an experiment about choosing between lotteries with varying prizes and chances of
winning. You will be presented with a series of lotteries where you will make choices between pairs
of them. There are 60 pairs in the series. For each pair of lotteries, you should indicate which of the
two lotteries you prefer to play. You will actually get the chance to play three of the lotteries you
choose, and will be paid according to the outcome of those lotteries, so you should think carefully
about which lotteries you prefer.
Here is an example of what the computer display of such a pair of lotteries will look like.
The display on your screen will be bigger and easier to read.
-33-
The outcome of the lotteries will be determined by the draw of a random number between 1
and 100. Each number between (and including) 1 and 100 is equally likely to occur. In fact, you will
be able to draw the number yourself using a special die that has all the numbers 1 through 100 on it.
In the above example the left lottery pays five dollars ($5) if the number on the ball drawn is
between 1 and 40, and pays fifteen dollars ($15) if the number is between 41 and 100. The yellow
color in the pie chart corresponds to 40% of the area and illustrates the chances that the ball drawn
will be between 1 and 40 and your prize will be $5. The black area in the pie chart corresponds to
60% of the area and illustrates the chances that the ball drawn will be between 41 and 100 and your
prize will be $15.
We have selected colors for the pie charts such that a darker color indicates a higher prize.
White will be used when the prize is zero dollars ($0).
Now look at the pie in the chart on the right. It pays five dollars ($5) if the number drawn is
between 1 and 50, ten dollars ($10) if the number is between 51 and 90, and fifteen dollars ($15) if
the number is between 91 and 100. As with the lottery on the left, the pie slices represent the
fraction of the possible numbers which yield each payoff. For example, the size of the $15 pie slice
is 10% of the total pie.
Each pair of lotteries is shown on a separate screen on the computer. On each screen, you
should indicate which of the lotteries you prefer to play by clicking on one of the three boxes
beneath the lotteries. You should click the LEFT box if you prefer the lottery on the left, the
RIGHT box if you prefer the lottery on the right, and the DON’T CARE box if you do not prefer
one or the other.
You should approach each pair of lotteries as if it is one of the three that you will play out. If
you chose DON’T CARE in the lottery pair that we play out, you will pick one using a standard six
-34-
sided die, where the numbers 1-3 correspond to the left lottery and the numbers 4-6 to the right
lottery.
After you have worked through all of the pairs of lotteries, raise your hand and an
experimenter will come over. You will then roll a 20 sided die three times to determine which pairs
of lotteries that will be played out. One lottery pair from the first 20 pairs will be selected, one from
the next 20 pairs, and finally one from the last 20 pairs. If you picked DON’T CARE for one of
those pairs, you will use the six sided die to decide which one you will play. Finally, you will roll the
100 sided die to determine the outcome of the lottery you chose.
For instance, suppose you picked the lottery on the left in the above example. If the random
number was 37, you would win $5; if it was 93, you would get $15. If you picked the lottery on the
right and drew the number 37, you would get $5; if it was 93, you would get $15.
Therefore, your payoff is determined by three things:
•
by which three lottery pairs that are chosen to be played out in the series of 60 such pairs
using the 20 sided die;
•
by which lottery you selected, the left or the right, for each of these three pairs; and
•
by the outcome of that lottery when you roll the 100 sided die.
This is not a test of whether you can pick the best lottery in each pair, because none of the
lotteries are necessarily better than the others. Which lotteries you prefer is a matter of personal
taste. The people next to you will have difference lotteries, and may have different tastes, so their
responses should not matter to you. Please work silently, and make your choices by thinking
carefully about each lottery.
All payoffs are in cash, and are in addition to the $5 show-up fee that you receive just for
being here. In addition to these earnings you will also receive a randomly determined amount of
-35-
money before we start. This can be any amount between $1 and $10 and will be determined by
rolling the 10 sided die.
When you have finished reading these instructions, please raise your hand. We will then
come to you, answer any questions you may have, let you roll the 10 sided die for the additional
money, and start you on a short series of practice choices. The practice series will not be for
payment, and consists of only 6 lottery pairs. After you complete the practice we will start you on
the actual series for which you will be paid.
As soon as you have finished the actual series, and after you have rolled the necessary dice,
you will be paid in cash and are free to leave.
Practice Session
In this practice session you will be asked to make 6 choices, just like the ones that you will be
asked to make after the practice. You may take as long as you like to make your choice on each
lottery.
You will not be paid your earnings from the practice session, so feel free to use this practice
to become familiar with the task. When you have finished the practice you should raise your hand so
that we can start you on the tasks for which you will be paid.
-36-
B. Individual Characteristics Questionnaire27
Stage 2: Some Questions About You
In this survey most of the questions asked are descriptive. We will not be grading your
answers and your responses are completely confidential. Please think carefully about each question
and give your best answers.
1.
What is your AGE? ____________ years
2.
What is your sex? (Circle one number.)
01
3.
Male
02
Female
Which of the following categories best describes you? (Circle one number.)
01
02
03
04
05
White
African-American
African
Asian-American
Asian
06
07
08
09
Hispanic-American
Hispanic
Mixed Race
Other
4. What is your major? (Circle one number.)
01
02
03
04
05
06
07
08
09
10
11
12
13
14
5.
What is your class standing? (Circle one number.)
01
02
03
6.
Accounting
Economics
Finance
Business Administration, other than Accounting, Economics, or Finance
Education
Engineering
Health Professions
Public Affairs or Social Services
Biological Sciences
Math, Computer Sciences, or Physical Sciences
Social Sciences or History
Humanities
Psychology
Other Fields
Freshman
Sophomore
Junior
04
05
06
Senior
Masters
Doctoral
What is the highest level of education you expect to complete? (Circle one number)
01
Bachelor’s degree
27
In the original experiments question #35 was inadvertently a repeat of question #31. We show
here the correct question #35, but the responses for these experiments for question #35 are invalid. This has
no effect on any of the results in the text, since these questions were not employed in the statistical analysis
reported here.
-37-
02
03
04
7.
What was the highest level of education that your father (or male guardian) completed?
(Circle one number)
01
02
03
04
05
8.
Single and never married?
Married?
Separated, divorced or widowed?
On a 4-point scale, what is your current GPA if you are doing a Bachelor’s degree, or what
was it when you did a Bachelor’s degree? This GPA should refer to all of your coursework,
not just the current year. Please pick one:
01
02
03
04
05
06
07
08
13.
Yes
No
Are you currently...
01
02
03
12.
U.S. Citizen
Resident Alien
Non-Resident Alien
Other Status
Are you a foreign student on a Student Visa?
01
02
11.
Less than high school
GED or High School Equivalency
High School
Vocational or trade school
College or university
What is your citizenship status in the United States?
01
02
03
04
10.
Less than high school
GED or High School Equivalency
High school
Vocational or trade school
College or university
What was the highest level of education that your mother (or female guardian) completed?
(Circle one number)
01
02
03
04
05
9.
Master’s degree
Doctoral degree
First professional degree
Between 3.75 and 4.0 GPA (mostly A’s)
Between 3.25 and 3.74 GPA (about half A’s and half B’s)
Between 2.75 and 3.24 GPA (mostly B’s)
Between 2.25 and 2.74 GPA (about half B’s and half C’s)
Between 1.75 and 2.24 GPA (mostly C’s)
Between 1.25 and 1.74 GPA (about half C’s and half D’s)
Less than 1.25 (mostly D’s or below)
Have not taken courses for which grades are given.
How many people live in your household? Include yourself, your spouse and any
dependents. Do not include your parents or roommates unless you claim them as
dependents.
___________________
-38-
14.
Please circle the category below that describes the total amount of INCOME earned in 2002
by the people in your household (as “household” is defined in question 13).
[Consider all forms of income, including salaries, tips, interest and dividend payments,
scholarship support, student loans, parental support, social security, alimony, and child
support, and others.]
01
02
03
04
05
06
07
08
15.
Please circle the category below that describes the total amount of INCOME earned in 2002
by your parents. [Consider all forms of income, including salaries, tips, interest and dividend
payments, social security, alimony, and child support, and others.]
01
02
03
04
05
06
07
08
09
16.
Part-time
Full-time
Neither
Before taxes, what do you get paid? (fill in only one)
01
02
03
04
18.
$15,000 or under
$15,001 - $25,000
$25,001 - $35,000
$35,001 - $50,000
$50,001 - $65,000
$65,001 - $80,000
$80,001 - $100,000
over $100,000
Don’t Know
Do you work part-time, full-time, or neither? (Circle one number.)
01
02
03
17.
$15,000 or under
$15,001 - $25,000
$25,001 - $35,000
$35,001 - $50,000
$50,001 - $65,000
$65,001 - $80,000
$80,001 - $100,000
over $100,000
_____ per hour before taxes
_____ per week before taxes
_____ per month before taxes
_____ per year before taxes
Do you currently smoke cigarettes? (Circle one number.)
00
01
No
Yes
If yes, approximately how much do you smoke in one day? _______ packs
For the remaining questions, please select the one response that best matches your reaction
to the statement.
19.
“I believe that fate will mostly control what happens to me in the years ahead.”
Strongly
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
-39-
Agree
Strongly
Agree
20.
“I am usually able to protect my personal interests.”
Strongly
Disagree
21.
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
Agree
Strongly
Agree
“When I get what I want, it's usually because I'm lucky.”
Strongly
Disagree
22.
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
“In order to have my plans work, I make sure that they fit in with the desires of people who
have power over me.”
Strongly
Disagree
23.
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
“When I make plans, I am almost certain to make them work.”
Strongly
Disagree
28.
Strongly
Agree
“I feel like other people will mostly determine what happens to me in the future.”
Strongly
Disagree
27.
Agree
“Chance occurrences determined most of the important events in my past.”
Strongly
Disagree
26.
Slightly
Agree
“Whether or not I get into a car accident depends mostly on the other drivers.”
Strongly
Disagree
25.
Slightly
Disagree
“I have mostly determined what has happened to me in my life so far.”
Strongly
Disagree
24.
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
“Getting what I want requires pleasing those people above me.”
Strongly
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
-40-
Agree
Strongly
Agree
29.
“Whether or not I get into a car accident depends mostly on how good a driver I am.”
Strongly
Disagree
30.
Disagree
Disagree
Disagree
Slightly
Disagree
Agree
Agree
Strongly
Agree
Disagree
Slightly S
Disagree
lightly
Agree
Agree
Strongly
Agree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
“I think that I will mostly control what happens to me in future years.”
Strongly
Disagree
35.
Slightly
“Whether or not I get into a car accident is mostly a matter of luck.”
Strongly
Disagree
34.
Strongly
Agree
“Most of my personal history was controlled by other people who had power over me.”
Strongly
Disagree
33.
Agree
“When I get what I want, it's usually because I worked hard for it.”
Strongly
Disagree
32.
Slightly
Agree
“Often there is no chance of protecting my personal interests from bad luck.”
Strongly
31.
Slightly
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
“People like myself have very little chance of protecting our personal interests when they
conflict with those of strong pressure groups.”
Strongly
Disagree
36.
Disagree
Slightly
Disagree
Slightly
Agree
Agree
Strongly
Agree
“It's not always wise for me to plan too far ahead because many things turn out to be a
matter of good or bad fortune.”
Strongly
Disagree
Disagree
Slightly
Disagree
Slightly
Agree
-41-
Agree
Strongly
Agree