Download Expected Utility Theory in Medical Decision Making

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Neuroeconomics wikipedia , lookup

Heuristics in judgment and decision-making wikipedia , lookup

Behavioral economics wikipedia , lookup

Risk aversion (psychology) wikipedia , lookup

Expected utility hypothesis wikipedia , lookup

Transcript
Expected Utility Theory and Deviations from it in Medical Decision Making
Margaret Sala
Introduction
The topic of the paper is on expected utility theory and deviations from it in medical decisionmaking. Expected utility theory is useful in medical decisions and gives patients and physicians the tools
to make optimal decisions regarding tests and treatments when risk is involved. However, deviance from
the expected utility theory is very common. Prospect theory, framing effects, and heuristics can account
for deviation from the expected utility theory model in medical decision-making.
Literature Review/Relevance to Other Topics
Expected utility theory is used to assess medical risk. Expected utility is the utility that is expected on
the average. In order to use expected utility theory, we must understand probability and it’s tradeoffs with
the disutility of possible outcomes. (Gurmankin, 2005) Utility in medical conditions is often converted to
Quality Adjusted Life Years (QALY). (Baron, 2007) A QALY is the total utility of a year in normal
health. Medicine shows us how expected utility is better in the long run. For example, if many people
choose to have a surgery with .00000001 probability of death, a few people will die, but if the utility of
the medical benefit is high enough, then it’s worth it. (Baron, 2007) The loss to some is more than
compensated by a gain to many others. (Baron, 2007)
There are many strengths to the use of expected utility theory in medicine. Expected utility theory
allows for the integration of patient values with medical facts. (Ubel, 1997) That is, it uses both
information that only a patient posses and probabilistic information. (Ubel, 1997) By using the expected
utility theory in medical decision-making, all the expected utility theory axioms (dominance, transitivity,
independence, etc) are satisfied. (Ubel, 1997) Furthermore, expected utility models closely resemble
linear models, which have proven to be successful in judging and prediction. (Ubel, 1997) By using
expected utility theory, we can accommodate two conflicting probabilities by using one piece of
information and then the other. (Ubel, 1997)
Although expected utility theory provides us with a normative model for medical decision-making, it
is important to note that deviance from the normative model is very common in medical decision-making.
Deviance from the normative model can be explained by prospect theory, framing effects, a difference
between experienced, predicted, and decision utilities, and several biases and heuristics.
Prospect theory is a descriptive theory of decision utility. (Baron, 2007) It predicts deviations from
expected utility theory. (Baron, 2007) Prospect theory explains why subjects are more affected by
differences among high probabilities than small ones when it comes to medical risk. (Gurmankin, 2005)
Framing effects are also very common in medical decision-making. Gigerenzer found that physicians
are better able to judge probability of a woman having breast cancer if data is presented in frequency
format rather than in probability format. (Galanter, 2005)
Medicine shows us that experienced, predicted, and decision utilities differ. Sieff, Dawes, and
Lowenstein (1999) asked people who were tested for HIV to predict how they would feel five weeks after
the test if the results were positive and if they were negative. Those with positive results thought they
Sala 2
would be happier than they ended up being, and those with negative results thought they would be
unhappier than they ended up being. Read and Lowenstein (1999) assessed whether people’s willingness
to accept pain (WTAP) depended on whether they felt pain either a moment or a week earlier and whether
they were focused or distracted from the pain. Read et. al also looked at people’s WTAP when they
hadn’t experienced the pain. Read et. al found that the distraction group displayed less WTAP than the
sensation focused group immediately after experiencing pain, but greater WTAP a week after accepting
the pain. Read et. al argued that sensation-focus reduces the experience of pain, but distraction reduces
the unpleasantness of remembered pain. He argued that remembering pain might sometimes be beneficial
for future decision-making. Also, people who never experienced pain experienced a greater WTAP. Read
and Lowenstein’s experiment shows us how the predicted disutility of pain is influenced by several
factors, and how it is thus different the experienced disutility of pain.
Doctors often fall into the availability heuristic, where they overestimate the frequency of an easily
recalled event and underestimate the frequency of a difficult to recall event. (Galanter, 2005) For
example, psychological illnesses such as bipolar disorder have recently been over diagnosed because they
have been talked about a lot over the last few years. (Galanter, 2005)
Doctors also tend to fall into the representativeness bias, where they diagnose a disorder in a person
based on how similar it is to the description of the disorder, rather than how likely it is for a person to
have the disorder based on it’s frequency of occurrence in the population. (Galanter, 2005) In other
words, physicians ignore base rates when making diagnostic predictions. (Galanter, 2005) For example,
Eddy (1982) asked physicians to estimate a young woman’s risk of having breast cancer given that she
had a positive mammogram. Physicians were provided with the probability of a woman having breast
cancer and the false positive rate. Physicians failed to use Bayes Theorem (the normative way of solving
this problem) and thus overestimated the probability of this woman having breast cancer. (Galanter, 2005)
The confirmation bias occurs when a doctor generates a hypothesis and then seeks evidence that
confirms the hypothesis. (Galanter, 2005)
The individual versus statistical heuristic is also very common. This heuristic is the idea that saving a
single life is seen as correcting 100% of a problem, while a small reduction in the death rate is seen as a
small correction of a problem. (Baron, 2007) Redelmeir et. al (1990) invited physicians to participate in a
medical decision making study, composed of three experiments. In the first experiment, physicians were
presented problems concerning either one patient, or concerning a group of patients. In the second
experiments, doctors were asked to analogously compare problems concerning either one patient or
concerning a group of patients. In the third experiment, undergraduate students were asked to (randomly)
interpret either an individual or aggregate version problem. It was found that both physicians and
laypeople give more concern to individual versus group patient problems.
The omission versus commission bias favors harms of omission over equivalent harms from action.
(Baron, 2007) For example people don’t like the idea of causing death by a vaccine. (Ritov and Baron,
1990) This is very relevant to medicine, as somebody would probably feel worse if they lost sight due to
having a surgery than if they lost sight due to not having a surgery.
Conclusion
We have seen that many topics from this course are relevant to medical decision making,
including:

Expected Utility Theory
Sala 3

Prospect Theory

Framing Effects

Difference between
 Biases and Heuristics: Availability heuristic, representativeness bias, confirmation bias,
individual versus statistical heuristic, and omission versus commission bias
References
Baron, J. (1988, 1994, 2000, 2008). Thinking and Deciding. New York: Cambridge University Press.
Chapman, G. B., & Sonnenberg, F. A. (Eds) (2000) Decision making in health care. New York:
Cambridge University Press.
Eddy, D.M. (1982) Probabilistic reasoning in clinical medicine: Problems and opportunities. In D.
Kahneman, P. Slovic, & A. Traversky (Eds.), Judgment under unceirtanty: Heuristics and biases (p.
249-267). Cambridge University Press.
Galanter, C.A., & Patel, V.L. (2005). Medical decision making: A selective review for child psychiatrists
and psychologists. Journal of Child Psychiatry and Psychology, 46 (7), 675-689.
Gigerenzer, G., Gaissmaier, W., Kurz-Milke, E., Schwartz, L. M., & Woloshin, S. (2007). Helping
doctors and patients make sense of health statistics. Psychological Science in the Public Interest, 8,
No. 2.
Gurmankin, A. D., & Baron, J. (2005). How bad is a 10% chance of losing a toe? Judgments of
probabilistic conditions by doctors and laypeople. Memory and Cognition, 33, 1399-1406.
Read, D. & Loewenstein, G. (1999). Enduring pain for money: Decisions based on the perception and
memory of pain. Journal of Behavioral Decision Making, 12, 1-17.
Redelmeier, D. A., & Tversky, A. (1990). Discrepancy between medical decisions for individual patients
and for groups. New England Journal of Medicine, 322, 1162-1164.
Ritov, I., & Baron, J. (1990). Reluctance to vaccinate: omission bias and ambiguity. Journal of
Behavioral Decision Making, 3, 263-277.
Sieff, E.M., Dawes, R.M. & Loewenstein, G. (1999). Anticipated versus actual reaction to HIV test
results. American Journal of Psychology, 112 (2), 297-311.
Ubel, P. & Loewenstein, G. (1997). The role of decision analysis in informed consent: Choosing between
intuition and systematicity. Social Science and Medicine, 44, 647-656.