Download The Syllogism of Expert Testimony Edward J. Imwinkelried

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Forensic accountant wikipedia , lookup

Digital forensics wikipedia , lookup

Forensic chemistry wikipedia , lookup

Forensic psychology wikipedia , lookup

Forensic linguistics wikipedia , lookup

Forensic epidemiology wikipedia , lookup

Forensic entomology and the law wikipedia , lookup

Transcript
The Syllogism of Expert Testimony
Edward J. Imwinkelried,
Professor of Law
University of California,
Davis, School of Law
29
I.
THE SYLLOGISM OF EXPERT TESTIMONY
The Potential Uses of Expert Witnesses
A.. Witnesses to observed facts – FRE 602
B. Witnesses expressing lay opinions – FRE 701
C.. Expert witnesses presenting a general lecture to the trier of fact – FRE 702 (“an
opinion or otherwise”)
D. Expert witnesses testifying to opinions derived by applying a general theory or
technique to the specific facts of the case – FRE 702
1. The witness’s qualification as an expert – FRE 702
The syllogism
2. The general theory or technique that the expert relies on – FRE 702(c)
3. The case-specific facts or data that the expert is evaluating – FRE 703
4. The application of the general theory or technique to the facts or data – FRE
702(d)
5. The final conclusion derived by applying the general theory or technique to the
case-specific facts and data – FRE 704
II. THE EXPERT’S MAJOR PREMISE: THE VALIDITY OF THE GENERAL THEORY OR
TECHNIQUE
A. The traditional view
The general acceptance standard under Frye v. United States, 293 F. 1013 (D.C.
Cir. 1923)
1 PAUL C. GIANNELLI, EDWARD J. IMWINKELRIED, ANDREA ROTH &
JANE CAMPBELL MORIARTY, SCIENTIFIC EVIDENCE § 1.06 (5th ed.
2012)
-----General acceptance as a proxy for empirical validity
Bert Black, Francisco J. Ayala & Carol Saffran-Brinks, “Science and the Law in
the Wake of Daubert: A New Search for Scientific Knowledge,” 72 TEXAS
LAW REVIEW 715 (1994)
-----The exemption for “soft” science
People v. McDonald, 37 Cal.3d 351, 208 Cal.Rptr. 236, 690 P.2d 709 (1984)(“a
30
machine”)
B. The trend toward the empirical standard
Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579 (1993)
The progeny of Daubert
General Electric Co. v. Joiner, 522 U.S. 136 (1997)
Kumho Tire Co. v. Carmichael, 526 U.S. 137 (1999)
Weisgram v. Marley Co., 528 U.S. 440 (2000)
The evolution of Frye into a distinct minority view
SCIENTIFIC EVIDENCE, supra, at §§ 1.-14-.16
But California’s continued adherence to Frye
People v. Leahy, 8 Cal.4th 587, 882 P.3d 321, 323, 34 Cal.Rptr.2d 663 (1994)
C. The recent Sargon decision
Sargon Enterprises, Inc. v. University of Southern California, 55 Cal.4th 747,
288 P.3d 1237, 149 Cal.Rptr.3d 614 (2012)
David L. Faigman & Edward J. Imwinkelried, “Wading into the Daubert Tide:
Sargon Enterprises, Inc. v. University of Southern California,” 64 HASTINGS
LAW JOURNAL 1665 (2013)
1. The background
a.
In re Lockheed Litig. Cases, 23 Cal.Rptr.3d 762 (Ct.App. 2005),
petition for review granted, 110 P.3d 289 ( Cal. 2005), petition for
dismissed, 83 Cal.Rptr.3d 478 (2007)
California Evidence Code § 801(b)
“If a witness is testifying as an expert, his testimony in the form of
an opinion is limited to such an opinion as is:
(b) Based on matter . . . perceived by or personally known to the
witness or made known to him at or before the hearing, whether or
not admissible, that is of a type that reasonably may be relied upon
by an expert in forming an opinion on the subject to which his
testimony relates, unless an expert is precluded by law from using
such matter as a basis for his opinion.”
An amicus brief filed by two academics
(1)–California Evidence Code § 802 as controlling
“A witness testifying in the form of an opinion may state on direct
the reasons for his opinion and the matters . . . upon which it is
based, unless he is precluded by law from using such reasons or
31
matter as a basis for his opinion.”
(2)–The existence of judicial power under § 802 to enunciate
uncodified restrictions on the theory or technique the expert
proposes relying on
California Evidence Code § 160: “‘Law’ includes constitutional,
statutory, and decisional law.”
(3)–The exercise of the power to authorize California trial judges
to conduct a circumscribed inquiry whether, as a matter of logic,
the reasons cited by the expert support the expert’s conclusion that
the theory or technique is valid
The dismissal of the case
b.
The article restating the position of the amicus brief
Edward J. Imwinkelried & David L. Faigman, “Evidence Code
Section 802: The Neglected Key to Rationalizing the California
Law of Expert Testimony,” 42 LOYOLA L.A. LAW REVIEW
427 (2009)
2. The Sargon decision itself
Edward J. Imwinkelried, “California Case Upsets Equilibrium in on
Expert Testimony,” NATIONAL LAW JOURNAL, May 20, 2013, at 16.
The scope of the decision:
–approving citations to Daubert, Joiner, and Kumho
–adopting the essential teachings of Daubert, Joiner, and Kumho
Daubert: the judge’s gatekeeping role to ensure the reliability
of the testimony
Kumho: the across-the-board application of the reliability
requirement to all types of expert testimony
Joiner: the evaluation of the validity of an expert’s analogical
reasoning
–endorsing the view that § 802 authorizes the enunciation of uncodified
restrictions on the types of theories and techniques that the expert may
rely on
32
–endorsing the view that § 802 authorizes the type of circumscribed
inquiry urged by the law review article
–but stopping short of authorizing the trial judge to pass on the credibility
of the foundational testimony under Evidence Code § 405, as the federal
trial judge may do under Federal Rule 104(a)
–and indicating in footnote six that Frye is still good law
III. THE EXPERT’S MINOR PREMISE: THE TRUSTWORTHINESS OF THE EVIDENCE
OF THE CASE-SPECIFIC FACTS THAT THE EXPERT APPLIES THE THEORY OR
TECHNIQUE TO
A.
The accepted methods of supplying the expert with the case-specific facts to the
evaluated
1. The expert’s personal observation of the fact
2. The hypothetical question
a. The requirement for the prior presentation of admissible evidence of
every element of the hypothesis
b. Special instructions to disregard the opinion if the jury rejects any
element of the hypothesis
3. Reliance on out-of-court reports if it is the reasonable (customary) practice of
the expert’s specialty to consider reports from such sources
a. The conventional wisdom that such reports are received for a
nonhearsay purpose under Federal Rule 801(c). The theory is mental
input, the effect on the state of mind of the hearer or reader. The expert’s
receipt of the report makes the expert’s opinion better grounded and
therefore more reasonable.
b. A limiting instruction under Federal Rule 105
c. No necessity for the earlier presentation of admissible evidence of the
secondhand fact
Advisory Committee Note, Federal Rule 703
B.
The impact of Williams v. Illinois, 132 S.Ct. 2221, 183 L.Ed.2d 89 (2012) on the
treatment of opinions based on out-of-court reports of secondhand facts
Edward J. Imwinkelried, “The Gordian Knot of the Treatment of Secondhand
33
Facts Under Federal Rule of Evidence 703 Governing the Admissibility of Expert
Opinions: Another Conflict Between Logic and Law,” 3 UNIVERSITY OF
DENVER CRIMINAL LAW REVIEW 1 (2013)
1. The Williams opinions
a.
Justice Alito’s plurality opinion
(1) The plurality’s position that the DNA report was nontestimonial
(2) The plurality’s position that the report was used for a
nonhearsay purpose
(3) The plurality’s concession that the falsity of an asserted
secondhand fact that is an essential premise of the opinion renders
the opinion irrelevant
(4) The plurality’s approval of a jury instruction that “[i]f the
statements in these questions are not supported by the proof, then
the answers to the questions are entitled to no weight . . . .”
b. Justice Thomas’ concurrence
(1) The justice’s position that the DNA report was non-testimonial
But for a different reason than the plurality
(2) The justice’s position that the report was used to prove the truth
of the assertion
c. Justice Kagan’s dissent
(1) The dissent’s position that the report was testimonial
(2) The dissent’s position that the report used to prove the truth of
the assertion
d. The upshot
(1) A formal holding that the statement was non-testimonial
The five: The plurality plus Justice Thomas
(2) A forceful dictum by five justices that reports received under
34
Federal Rule of Evidence 403 are received for the truth of the
assertion
The five: Justice Thomas plus the dissenters
2. A critical evaluation of the merits of the dispute between the plurality and the
other justices over the status of reports used under Rule 703
a. The five justices’ error in thinking that reports used under Rule 703
are necessarily received for the truth of the assertion
The legitimacy of a non-hearsay theory
b. The plurality’s concession that if a secondhand report is an essential
premise of the expert’s opinion, as a matter of logic, the falsity of the
report renders the opinion invalid and therefore irrelevant,
A conditional opinion
3. The fallout from Williams
a. The need for a a jury instruction that if a secondhand report that is an
essential premise of the expert’s opinion is false, the jury must reject the
opinion itself
The dissent goes beyond that, and the plurality concedes that.the
falsity of an essential secondhand report renders the opinion
invalid and irrelevant.
b. The question whether, in the longterm, Williams will blur the lines
between hypothetical questions and questions based on 703 secondhand
facts: As in the case of hypothetical questions, will the lower courts begin
demanding independent, admissible evidence of the truth of the
secondhand report? DAVID H. KAYE, DAVID E. BERNSTEIN &
JENNIFER L. MNOOKIN, THE NEW WIGMORE: EXPERT
EVIDENCE § 4.12.7, at 70-71 (2014 Cum.Supp.)
The logic
Versus
The law
35
The Advisory Committee Note, F.R.E. 703
------The committee’s observation that outside court,
experts often make “life-and-death decisions” on the basis
of such reports. The report is entitled to more weight than
an attorney’s assertion during opening statement.
-----The committee’s view that the proponent should not be
required to go to the length of spending the time to present
admissible evidence of the content of the report (“with[out]
the expenditure of substantial time in producing and
examining various authenticating witnesses”)
Oliver Wendell Holmes, Jr., “The life of the law has not been logic . . . .”
THE COMMON LAW 1 (1881)
IV. THE EXPERT’S FINAL CONCLUSION: EXPOSING MEASUREMENTS AS MERE
UNCERTAIN ESTIMATES–AND VIRTUALLY MEANINGLESS UNLESS
ACCOMPANIED BY A MATHEMATICAL INDICATION OF THE UNCERTAINTY SUCH
AS A CONFIDENCE INTERVAL
Edward J. Imwinkelried, “Forensic Metrology: The New Honesty About the Uncertainty
of Measurements in Scientific Analysis,” in PROCEEDINGS OF THE FOURTH
INTERNATIONAL CONFERENCE ON EVIDENCE LAW AND FORENSIC
SCIENCE 23 (Beijing, China 2013), to be reprinted in ___JOHN MARSHALL LAW
REVIEW ___(forthcoming 2014).
36
THE IMPORTANCE OF FORENSIC METROLOGY IN PREVENTING
MISCARRIAGES OF JUSTICE: INTELLECTUAL HONESTY ABOUT THE
UNCERTAINTY OF MEASUREMENT IN SCIENTIFIC ANALYSIS
Edward Imwinkelried*
* Edward L Barrett, Jr. Professor of Law, University of California, Davis; coauthor,
GIANNELLI, IMWINKELRIED, ROTH & CAMPBELL MORIARTY, SCIENTIFIC
EVIDENCE (5th ed. 2012). This article is based in part on the author=s presentation at the Fourth
International Conference on Evidence Law and Forensic Science, held in Beijing, China on July
20-21, 2013. Reprinted with the kind permission of China University of Political Science and
Law.
37
A[I]t would be unreasonable to conclude that the subject of scientific testimony must be >known=
to a certainty; arguably, there are no certainties in science.@
BJustice Blackmun, Daubert v. Merrell Dow Pharmaceuticals, Inc.1
I.
THE RISK THAT MISUNDERSTOOD STATISTICAL TESTIMONY CAN
CAUSE A MISCARRIAGE OF JUSTICE
At the end of the 19th century, Justice Holmes predicted that Athe man of the future is the
man of statistics.@2 History has proven Holmes right. Statistical testimony is now in widespread
use in American courtrooms:3
Today, complex statistical evidence is being introduced in American courts in a wide
variety of both civil and criminal matters. [A] LEXIS search of district court opinions
using the words, Astatistic,@ Astatistics,@ or Astatistical,@ turned up 608 examples in the
years 1960 to 1969; 2,786 cases from 1970 to 1979; 4,364 cases from 1980 to 1989; and
3,015 from 1900 through July 31, 1995.4
Although the courts are now receptive to statistical testimony, initially the courts were
skeptical about the ability of lay triers of fact to critically evaluate such testimony.5 The courts
feared that Astatistics are subject to a wide range of manipulation@6Bmanipulation that could
trigger a miscarriage of justice. The skeptics often pointed to the California Supreme Court=s
famous 1968 decision in People v. Collins7 as proof of that danger.
In Collins, two defendants, an Afro-American husband and his Caucasian wife, were
1
590 U.S. 579, 590 (1993).
2
Oliver Wendell Holmes, The Path of the Law, 10 Harv.L.Rev. 457, 469 (1897).
3
Jonathan Koehler, The Probity/Policy Distinction in the Statistical Evidence
Debate, 66 Tul.L.Rev. 141 (1991).
4
United States v. Shonubi, 895 F.Supp. 460, 514, 516-19 (E.D.N.Y. 1995),
vacated, 103 F.3d 1085 (2d Cir. 1997).
5
C. McCormick, Evidence ' 210 (6th ed. 2006).
6
United States v. State of Oregon, 782 F.Supp.2d 502, 513 (D.Or. 1991), aff=d, 28
F.3d 110 (9th Cir. 1994).
7
68 Cal.2d 319, 438 P.2d 33, 66 Cal.Rptr. 497 (1968)
38
charged with robbery. At trial, the prosecution had difficulty establishing the defendant=s identity
as the perpetrators. The victim did not identify the defendants, and one eyewitness gave
somewhat equivocal testimony. To corroborate the identifications, the prosecutor called a
mathematics instructor at a state college as a witness. During the witness=s direct examination,
the prosecutor listed several characteristics that witnesses had attributed to the perpetrators and
that the defendants happened to match. The prosecutor then asked the witness to assume a
probability for each characteristic:
party yellow automobile
man with mustache
girl with ponytail
girl with blond hair
Negro man with beard
interracial couple in car
1/10
1/4
1/10
1/3
1/10
1/1,000.8
The prosecutor then invited the witness to apply the product rule to this data. According to the
product rule, if individual probabilities are independent, the probability that they will concur is
the product of the multiplication of the individual probabilities.9 The witness complied and
testified to the figure of 1/12,000,000. In the prosecutor=s mind, the witness=s testimony
established
the probability there was but one chance in 12 million that any couple possessed the
distinctive characteristics of the defendants. Accordingly, under this theory, it was to be
inferred that there could be but one chance in 12 million that defendants were innocent . .
. .10
On appeal, the California Supreme Court reversed. The court found numerous flaws in
the foundation for the witness=s testimony:
$
There was no basis for the individual probabilities that the prosecutor asked the witness
to assume. The prosecutor had pulled the numbers out of thin air.11
$
Although the product rule may be applied only when the individual probabilities are truly
independent, there was obviously an inadequate showing of independence in the court
below.12 Again, one factor referred to a man with a moustache while another mentioned a
man with a beard.
8
438 P.2d at 37 n. 10.
9
1 Paul C. Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane Campbell
Moriarty, Scientific Evidence ' 15.07[a], at 915-16 (5th ed. 2012).
10
438 P.2d at 37.
11
Id. at 38.
12
Id. at 39.
39
$
The prosecutor misstated the significance of the witness=s computation. Even if the
individual probabilities were accurate and the probabilities were independent, the
computation yielded only a random match probability.13 That computation answers the
question of how often you would have to dip into the relevant population to randomly
find another couple with the same set of characteristics?14 However, the prosecutor used
the figure as if it were a source probability.15 The latter probability addresses the
question of how probable it is that the defendants were the perpetrators.16 In order to
answer that question, the trier of fact would have to make an assumption about the size of
the relevant population.17 If only one person in a 1,000.000 in a population has a
characteristic but the relevant population size is 10,000,000, ten persons in the population
probably share the characteristic.
The court found that these errors were prejudicial, necessitating reversal. At the beginning of its
opinion, the court warned that mathematics is Aa veritable sorcerer in our computerized
society.@18 At the end of the opinion, the court asserted that A[u]ndoubtedly the jurors were
unduly impressed by the mystique of the mathematical demonstration but were unable to assess
its relevancy or value.@19
In a 2010 DNA case, McDaniel v. Brown,20 the United States Supreme Court struggled
with the question of whether a prosecutor=s misuse of a random match probability as a source
probability denied a defendant due process. McDaniel was a habeas proceeding rather than a
direct appeal. In McDaniel, the petitioner contended that in its arguments in the trial court the
government had committed A[t]he prosecutor=s fallacy . . . that the random match probability is
the same as the probability that the defendant was not the source of the DNA sample.@21 The
Court elaborated:
13
Id. at 40; 2 Paul C. Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane
Campbell Moriarty, Scientific Evidence ' 18.04[c][3], at 105-06 (5th ed. 2012).
14
Id. at 105.
15
Id. at ' 18.04[c][3], at 106-08 (5th ed. 2012).
16
Id.
17
Id.
18
438 P.2d at 33..
19
Id. at 41.
20
McDaniel v. Brown, 558 U.S. 120 (2010).
21
Id. at 128.
40
In other words, if a juror is told the probability a member of the general population would
share the same DNA is 1 in 10,000 (random match probability), and he takes that to mean
there is only a 1 in 10,000 chance that someone other than the defendant is the source of
the DNA found at the crime scene (source probability), then he has succumbed to the
prosecutor=s fallacy. It is further error to equate source probability with probability of
guilt, unless there is no explanation other than guilt for a person to be the source of the
crime-scene DNA. This faulty reasoning may result in an erroneous statement that, based
on a random match probability of 1 in 10,000, there is a .01% chance the defendant is
innocent or a 99.99% chance the defendant is guilty.22
The AState concede[d] that [the prosecution witness] Romero had overstated [the] probative
value@ of the DNA evidence.23 Ultimately, though, the Court rejected the petitioner=s claim.
The Court found that there was ample other evidence to establish the petitioner=s guilt. The
Court concluded that together with the inculpatory non-DNA evidence in the record, a proper
statistical evaluation of the DNA match furnished Apowerful evidence of guilt.@24
The common denominator in these two highly publicized cases was the courts= concern
that there might have been a miscarriage of justice because the lay trier of fact could have
misused a numerical figure testified to by an expert. That concern is not limited to celebrated
cases such as Collins and McDaniel. For that matter, the concern is not confined to DNA cases.
The issue is virtually ubiquitous in cases involving scientific testimony. As Part II explains, the
same issue arises whenever an expert witness relies on a measurementBa common occurrence
when experts take the stand.
II.
THE EVOLUTION OF THE COURTS= APPROACH TO RECOGNIZING THE
UNCERTAINTY OF STATISTICAL TESTIMONY ABOUT MEASUREMENTS
IN SCIENTIFIC ANALYSIS
22
Id.
23
Id. at 182.
24
Id.
41
In many cases, forensic scientists rely on measurements as a basis for their opinions. In
the words of one scientist, Ascience is measurement. If I cannot make measurements, I cannot
study a problem scientifically.@25 Toxicologists measure the concentration of the toxic drug found
in human bodies to determine whether the level of toxin was high enough to have caused the
person=s death.26 Accident reconstruction experts measure the length of skidmarks27 and yaw
marks28 to determine the velocity of a vehicle. In speeding cases, prosecutors rely on radar
measurements of the velocity of the vehicle driven by the defendant.29 Drug analysts measure the
weight of seized contraband drugs to determine whether the weight was large enough to violate
criminal statutes.30 Measurements have long been critical in drunk driving prosecutions. Many
jurisdictions have statutes that erect a presumption of intoxication when the measured blood
alcohol concentration exceeds a prescribed threshold.31 Moreover, many countries and states
have gone farther and adopted per se statutes proscribing the operation of a vehicle by a person
with a specified blood alcohol concentration.32 Under these statutes, the prosecution need not
show that the operator was impaired or intoxication; proof that the driver had a certain blood
alcohol concentration such as 0.08% is sufficient. In some jurisdictions, the driver is subject to
an enhanced sentence if the blood alcohol concentration exceeds an elevated level such as
0.15%.33 In all these cases, the testimony about the forensic scientist=s measurement is usually
the most important evidence in the case.
25
William H. George, Musical Acoustics Today, New Scientist, Nov. 1962, at 257.
26
2 Paul C. Giannelli , Edward J. Imwinkelried, Andrea Roth & Jane Campbell
Moriarity, Scientific Evidence ' 20.05[d], at 351-53 (5th ed. 2012).
27
Id. at ' 27.06[a].
28
Id. at ' 27.06[b].
29
John Jendzurski & Nicholas G. Paulter, Calibration of Speed Enforcement Downthe-Road Radars, 114 Journal of Research of the National Institute of Standards and Technology
137 (2009).
30
SWGDRUG, Measurement Uncertainty for Weight Determinations in Seized
Drug Analysis (2010).
31
2 Paul C. Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane Campbell
Moriarity, Scientific Evidence ' 22.01, at 473-76 (5th ed. 2012).
32
Id. at ' 22.01, at 361-62.
33
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part Two: Legal Analysis, in Understanding DUI Scientific Evidence 255, 323
(2011 ed.).
42
In the past, forensic scientists testifying about such measurements have often presented
the court with a single point value.34 When the expert does so, there is a distinct possibility that
the trier of fact will treat the testimony as an exact value.35 In the words of Washington State
Commissioner Paul Moon, AIt has been this court=s experience since 1983 that juries it has
presided over place heavy emphasis on the numerical value of blood alcohol tests. If an expert
testifies that a particular blood alcohol content measurement is value A, without stating a
confidence interval, it is this court=s opinion that the evidence is being represented as an exact
value to the trier of fact.@36
The problem is that metrology, the science of measurement, tells that there is an
unavoidable, inherent element of uncertainty in every measurement.37 If the expert does not give
the trier of fact some sense of that uncertainty, the statement of the single point valueBa
number38-- is incomplete39 and therefore potentially misleading.40 The trier could place undue
trust41 in the testimony about the numerical value. That mistake might affect the trier=s ultimate
34
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 220 (2011 ed.).
35
State of Washington v. Weimer, #7036A-09D (Snohomish Cty. Dist.Ct., Wash.
Mar. 23, 2010); Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part Two: Legal Analysis, in Understanding DUI Scientific Evidence 255, 311
(2011 ed.).
36
State of Washington v. Weimer, #7036A-09D (Snohomish Cty. Dist.Ct., Wash.
Mar. 23, 2010).
37
See Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of
Modern Approaches to Express Uncertainty in Measurement, 44 Metrologia 513 (2007); Jesper
Kristiansen & Henning Willads Peterson, An Uncertainty Budget for the Measurement of
Ethanol in Blood by Headspace Gas Chromatography, 28 Journal of Analytical Toxicology 456
(2004).
38
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 190, 225 (2011 ed.).
39
Id. at 191; Rod G. Gullberg, Measurement Uncertainty in Forensic Toxicology:
Its Estimation, Reporting and Interpretation, in Toxicity and Drug Testing 415, 415 (W. Acree
ed. 2012); Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result Interpretation.
Part Two: Legal Analysis, in Understanding DUI Scientific Evidence 255, 230 (2011 ed.).
40
Id.; State of Washington v. Weimer, #7036A-09D (Snohomish Ct. Dist. Ct.,
Wash. Mar. 23, 2010); State of Washington v. Fausto and Ballow, #C076949 and 9Y6231062
(King Cty. Dist.Ct., Wash. Sep. 21, 2010).
41
Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of Modern
43
decision,42 prompting the trier to either convict an innocent person or exonerate a guilty person.
Approaches to Express Uncertainty in Measurement, 44 Metrologia 513, 518 (2007).
42
State of Washington v. Fausto and Ballow, #C076949 and 9Y6231062 (King
Cty.Dist.Ct., Wash. Sep. 21, 2010).
44
However, metrology has not only demonstrated the inherent uncertainty of measurements.
It has also developed several methods of quantifying a measurement=s margin of error or
uncertainty.43 By using these methods, the testifying expert can put the trier in a much better
position to determine the appropriate evidentiary weight of the measurement.44 As we shall see
in Part IV, there are multiple approaches to quantifying the degree of error or uncertainty of a
measurement. The expert can draw on these approaches in order to prevent the jury from
ascribing inordinate weight to the testimony about the point value.
The purpose of this paper is to document the evolution of the law=s treatment of the
problem of uncertainty in forensic science and in particular uncertainty in measurements by
forensic scientists. Part I of this paper demonstrates that at first, many courts insisted that the
expert vouch for his or her opinion as a scientific or medical certainty. Part II points out that
later, as the courts began to realize the foolishness of such insistence, they permitted experts to
testify without opining that the point value was certainly exact. Part III shows that the evolution
of the law on this topic continues. Influenced by critiques from several prestigious scientific
organizations, a large number of courts now forbid the expert from couching his or her opinion as
a certainty. Finally, Part IV will describe an incipient judicial trend to require that in some way,
the expert quantity the degree of error or uncertainty in the expert=s measurement testimony.
This part discusses a number of recent cases as well as important developments in the field of
metrology. The paper concludes that this trend holds out the promise for more honest and open
cooperation between law and science.
A..
THE FIRST STAGE: THE ORIGINAL JUDICIAL INSISTENCE THAT
EXPERTS VOUCH FOR THEIR OPINIONS AS SCIENTIFIC OR MEDICAL
CERTAINTIES
In 1979, the International Association for Identification (I.A.I.) passed a resolution
making it professional misconduct for a a member latent print examiner to provide courtroom
testimony describing an identification as Apossible, probable, or likely@ rather than Acertain.@45 It
became a widespread practice at American trials for the expert=s proponent to invite the
43
Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of Modern
Approaches to Express Uncertainty in Measurement, 44 Metrologia 513 (2007).
44
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 218 (2011 ed.).
45
Nat=l Inst. Standards & Tech., Latent Print Examination and Human Factors:
Improving the Practice through a Systems Approach 350 (2012), citing I.A.I. Resolution VII.
Identification News 29 (Aug. 1979). See also Note, Daubert Rises: The (Re)applicability of the
Daubert Factors to the Scope of Forensics Testimony, 96 Minn.L.Rev. 1581, 1597 (2012)(AMost
forensic examiners are trained to either testify with absolute certainty or not at all. A firearmand-toolmark examiner testified that . . . >if two cartridge cases share the same magazine mark,
then one could say with one hundred percent certainty that the two cartridge cases had been
cycled through the same magazine.= [A] forensic shoe print analyst >offered a potential error rate
of zero for the method . . . .=@).
45
expert to testify that the expert=s opinion was a scientific or medical Acertainty.@ Eventually,
many courts elevated the practice to the status of a legal requirement.46 In some jurisdictions, it
was not merely that less certain testimony was legally insufficient to sustain a judgment in the
proponent=s favor; rather, if the expert refused to characterize his or her opinion in those terms,
the opinion was inadmissible as a matter of law.
That legal requirement reflected an idealized47 view of the physical world, which was
very widespread and shared by many, including judges and attorneys. According to this view,
the universe is orderly; and its phenomena are governed by invariable, determinist48 natural laws.
Those laws are not only mechanical and inexorable, but also discoverable by the scientific
method.49 A scientist could learn those natural laws by carefully applying the classic
Newtonian method of hypothesis formulation and empirical testing.
Newton had refined the scientific experimental technique in the late 17th and early 18th
centuries.50 The Age of Enlightenment began soon after.51 The supporters of the
Enlightenment were sanguine about the power of the human mind. One of the key events
triggering the Enlightenment was a translation of Newton=s Prinicipia into French with a preface
by Voltaire.52
46
State v. Holt, 246 N.E.2d 365 (Ohio 1969); Bert Black, The Supreme Court=s
View of Science: Has Daubert Exorcised the Certainty Demon?, 15 Cardozo Law Review 2129
(1994); Gregory Joseph, Less Than ACertain@ Medical Testimony, 1979 Medical Trial Technique
Quarterly 10 (1979); Gregory Joseph, Less Certain Medical Testimony, 14 Trial, Jan. 1978, at
51; Medical EvidenceBSufficiency of Expert=s Opinion, 17 Defense Law Journal 181 (1968).
47
Brief Amici Curiae of Physicians, Scientists, and Historians of Science in Support
of Petitioners at 11, Daubert v. Merrell Dow Pharmaceuticals, Inc. (U.S. No. 92-102).
48
Margaret G. Farrell, Daubert v. Merrell Dow Pharmaceuticals, Inc.: Epistemology
and Legal Process, 15 Cardozo Law Review 2183, 2194 (1994).
49
Anthony Z. Roisman, Conflict Resolution in the Courts: The Role of Science, 15
Cardozo Law Review 1945, 1950 (1994).
50
5 The Encyclopedia of Philosophy 489 (Paul Edwards ed. 1967).
51
Will Durant, The Story of Philosophy 199 (1961)
52
2 The Encyclopedia of Philosophy 519 (Paul Edwards ed. 1967).
46
AThe Age of Reason@ promoted the notion that armed with Newton=s systematic methodology,
the human intellect53 could identify and master the laws governing natural phenomena. Science
appeared to offer a Avision of total control.@54 Adherents of this view believed that proof of
scientific propositions was capable of attaining certainty and that scientific propositions could be
conclusively validated.55 In this simplistic universe, complete truth is attainable.56 In such a
universe, it is justifiable that the courts require that expert witnesses vouch for their opinions as
certainties.
B..
THE SECOND STAGE: THE COURTS= ABANDONMENT OF THE
REQUIREMENT THAT EXPERTS VOUCH FOR THEIR OPINIONS AS
CERTAINTIES
As Part I explained, the courts= original insistence that experts vouch for their opinions as
certainties reflected a particular, simplified view of the physical world. Eventually, new
scientific discoveries shattered that view.57 Those discoveries prompted the scientific
community to abandon the assumption
that the universe is entirely orderly. Quite to the contrary, they have confronted the
reality that the universe is at least partially chaotic and indeterminate. In 1927,
Heisenberg discovered Athat electrons could not be assigned a position in time and space
simultaneously, nor their future predicted by their present.@ His discovery led to the
53
Id.
54
Michael Crichton, Jurassic Park 313 (1990).
55
Brief Amici Curiae of Physicians, Scientists, and Historians of Science in Support
of Petitioners at 8, Daubert v. Merrell Dow Pharmaceuticals, Inc. (U.S. NO. 92-102).
56
Bert Black, The Supreme Court= View of Science: Has Daubert Exorcised the
Certainty Demon?, 15 Cardozo Law Review 2129, 2129-30 (1994).
57
Bert Black, A Unified Theory of Scientific Evidence, 56 Fordham L.Rev. 595,
616 (1988)(AAround the turn of the twentieth century, . . . advances in physiology and
psychology and the advent of the quantum and relativity theories in physics destroyed simple,
mechanistic certainty. Quantum theory tells us that certainty is a physical impossibility, relativity
that time is not absolute, and psychology that preconceptions color supposedly objective accounts
of the natural world@); Richard G. Halpern, Opening a New Door to Negotiation Strategy, 35
Trial, June 1999, at 22, 24 (Aphysicist Henri Poincare noticed that the orbits of certain planets did
not follow the paths predicted by the application of Newtonian physics. Poincare discovered that
for certain astronomical systems, typically those involving interactions among three or more
heavenly bodies, even the tiniest imprecisions in the measurement of a starting point would lead
to huge variations in the final resultBso huge that the measured prediction based on the principles
of Newtonian, or linear, physics would be no more accurate than a random position picked out of
a hat. He had discovered dynamical instability or chaos@).
47
formulation of the principle of uncertainty. That principle undermined the conception of
an Aorderly world in which [a] perfect understanding of cause and effect@ is possible.
[S]cience is now reconciled to the fact that some natural phenomena occur erratically.58
58
Edward J. Imwinkelried, Evidence Law Visits Jurassic Park: The Far-Reaching
Implication of the Daubert Court=s Recognition of the Uncertainty of the Scientific Enterprise, 81
Iowa Law Review 55, 60 (1995), citing Bert Black, The Supreme Court=s View of Science: Has
Daubert Exorcised the Certainty Demon?, 15 Cardozo Law Review 2129 (1994), Margaret G.
Farrell, Daubert v. Merrell Dow Pharmaceuticals, Inc.: Epistemology and Legal Process, 15
Cardozo Law Review 2183 (1994), and Brief Amici Curiae of Physicians, Scientists, and
Historians of Science in Support of Petitioners, Daubert v. Merrell Dow Pharmaceuticals, Inc.
(U.S. No. 92-102). See also Richard G. Halpern, Opening a New Door to Negotiation Strategy,
35 Trial, June 1999, at 22, 25 (AFirst stated by physicist Werner Heisenberg in 1927, the
Heisenberg uncertainty principle was . . . a major revelation . . . . The principle has three parts:
The act of observation changes any observed system irrevocably; precise measurement of any
system requires observation; because observation changes the system, accurate measurement is
impossible@).
48
Science not only reassessed its macrocosm view of the orderliness of the universe; it also
rethought the microcosm of the validation of individual scientific hypotheses. In some instances,
experts rely on deductive reasoning. If one assigns certain definitions to Atwo@ and Afour,@ one
can reason with inexorable deductive logic that two plus two equals four. However, in
investigational science exploring phenomena in the physical world, experts typically resort to
another mode of analysis, that is, inductive reasoning. Having observed specific instances of a
phenomenon, the scientist formulates a general hypothesis about the phenomenon and then
engages in controlled laboratory experimentation or systematic field observation to falsify or
validate the hypothesis.59 If the outcomes of numerous empirical tests of the hypothesis all
confirm the hypothesis, we can have increasing confidence in the validity of the hypothesis.60
However, we cannot regard as the hypothesis as Adefinitively confirmed because it is always
possible that an empirical test will some day demonstrate the theory to be incorrect.@61 Another
empirical test is always conceivable; and so long as that is true, a theoretical possibility of
falsification or disproof remains.62 In short, A[n]o amount of testing can establish that a
scientific theory is >true= in every conceivable circumstance.@63 When an expert relies inductive
reasoning to investigate the truth of an hypothesis, at most the hypothesis can be accepted
contingently64 or provisionally.65 Given the imperfect,66 incomplete67 state of our knowledge, the
59
Bert Black, Francisco J. Ayala & Carol Saffran-Brinks, Science and the Law in
the Wake of Daubert: A New Search for Scientific Knowledge, 72 Texas Law Review 715, 755
(1994).
60
Brief of the American Medical Association, American Medical
Association/Specialty Society Medical Liability Project as Amici Curiae in Support of
Respondent at 11, Daubert v. Merrell Dow Pharmaceuticals, Inc. (U.S. No. 92-102).
61
Id.
62
Bert Black, Francisco J. Ayala & Carol Saffran-Brinks, Science and the Law in
the Wake of Daubert: A New Search for Scientific Knowledge, 72 Texas Law Review 715, 764
(1994).
63
Clifton T. Hutchinson & Danny S. Ashby, Daubert v. Merrell Dow
Pharmaceuticals, Inc.: Redefining the Bases for Admissibility of Expert Scientific Testimony, 15
Cardozo Law Review 1875, 1885 (1994).
64
Bert Black, Francisco J. Ayala & Carol Saffran-Brinks, Science and the Law in
the Wake of Daubert: A New Search for Scientific Knowledge, 72 Texas Law Review 715,
Bruce S. Koukoutchos, Solomon Meets Galileo (And Isn=t Quite Sure What To Do With Him),
15 Cardozo Law Review 2237, 2253 (1994).
65
Brief Amici Curiae of Nicolas Bloembergen et al., at 9 n. 8, Daubert v. Merrell
Dow Pharmaceuticals, Inc. (U.S. No 92-102).
66
Brief of the Carnegie Commission on Science, Technology, and Government as
Amicus Curiae in Supprt of Neither Party at 22, Daubert v. Merrell Dow Phamaceuticals, Inc.
49
investigation can never validate the hypothesis as a certainty.68 As of July 2010, the
International Association for Identification changed its policy and expressly allowed members to
testify to qualified, uncertain opinions when such opinions are statistically defensible.69
(U.S. No. 92-102).
67
Bruce S. Koukoutchos, Solomon Meets Galileo (And Isn=t Quite Sure What To
Do With Him), 15 Cardozo Law Review 2237, 2253 (1994).
68
Vern R. Walker, The Siren Songs of Science: Toward a Taxonomy of Scientific
Uncertainty for Decisionmakers, 23 Connecticut Law Review 567 (1991).
69
Nat=l Inst. Standards & Tech., Latent Print Examination and Human Factors:
Improving the Practice through a Systems Approach (73 (2010), citing I.A.I. Resolution 2010-18,
dated July 16, 2010.
50
The changed views of the scientific community eventually had an impact on the legal
community. The courts gradually began relaxing the requirement that the expert affirm that he or
she could testify to their opinion as a scientific or medical certainty. In some jurisdictions, the
initial relaxation was selective. For example, although the court might insist that the expert
vouch for opinions about diagnosis70 or causation71 as a certainty, the courts accepted
prognosis72 opinions lacking that degree of certitude. Prognoses are futuristic predictions, and in
these jurisdictions it struck the courts as unrealistic to demand the same definiteness as in
opinions about past causation or present diagnosis.73 Other courts generally abandoned the
requirement.74 These courts reason that the indefiniteness of an opinion couched as a possibility
or probability reduces the weight of the testimony75 and renders the opinion more vulnerable to
discretionary exclusion,76 but they no longer enforce a rigid requirement automatically excluding
any opinion that is not phrased as a certainty.
70
Gregory Joseph, Less Than Certain Medical Testimony, 14 Trial, Jan. 1978, at 51-
71
Id. at 51.
72
Id. at 51-52.
73
Id. at 52.
52.
74
United States v. Oaxaca, 569 F.2d 518, 526 (9th Cir.), cert.denied, 439 U.S. 926
(1978); United States v. Spencer, 439 F.2d 1047, 1049 (2d Cir. 1971); Burke v. Town of
Walpole, 405 F.2d 66 (1st Cir. 2005); United States v. Glynn, 578 F.Supp.2d 567 (S.D.N.Y.
2008), cert.denied, 131 S.Ct. 2960, 180 L.Ed.2d 250 (2011); Huck v. State, 881 So.2d 1137,
1150 (Fla.Dist.Ct.App. 2004); State v. Boyer, 406 So.2d 143, 148 (La. 1981). .
75
In re Swine Flu Immunization Prods. Liability Litrigation, 533 F.Supp. 567, 578
(D.Colo. 1980).
76
Fed.R.Evid. 403, 28 U.S.C.A..
51
The Supreme Court itself recognized this insight. In Daubert, a large number of
scientists and scientific organizations submitted amicus curiae briefs.77 Many of those briefs
attempted to educate the members of the Court about the modern understanding of the limitations
of the scientific method. Drawing on several of those briefs,78 Justice Blackmun wrote: A[I]t
would be unreasonable to conclude that the subject of scientific testimony must be >known= to a
certainty; arguably there are no certainties in science.@79 That frank recognition by the Court has
prompted other courts to be more receptive to uncertain testimony such as statistical evidence. In
the final analysis, statistical testimony is overtly uncertain expert testimony.80 By way of
example, the expert concedes that she does not know the true value of a certain parameter of the
universe, but she has sampled the universe and computed a statistic that is an estimator of the
parameter. Several of the amicus briefs filed in the Daubert predicted that if the Court embraced
a more realistic view of the limits of the scientific method, the Court=s adoption of that view
would pave the way for the more liberal admissibility of a particular type of statistical testimony,
that is, testimony about the confidence intervals for point estimates.81 In the past, the courts
were often squeamish about explicitly acknowledging the uncertainty of testimony such as
scientific evidence.82 Some feared that doing so would undermine the public=s faith in the legal
system and their acceptance of the system=s outcomes as legitimate.83 However, once the
Supreme Court itself forthrightly conceded the inevitable uncertainty in scientific testimony, the
77
Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579, 581-82 *
(1993)(listing the amicus briefs).
78
Id. at 590.
79
Id.
80
Edward J. Imwinkelried, Evidence Law Visits Jurassic Park: The Far-Reaching
Implication of the Daubert Court=s Recognition of the Uncertainty of the Scientific Enterprise, 81
Iowa Law Review 55, 65 (1995).
81
Id. at 67 n. 131, citing Brief Amicus Curiae of Professor Alan R. Feinstein in
Support of Respondent at 3, 16, Daubert v. Merrell Dow Pharmaceuticals, Inc. (U.S. No. 92102), Brief Amici Curiae of Professors Kenneth Rothman et al. in Support of Petitioners at 7,
Daubert v. Merrell Dow Pharmaceuticals, Inc. (U.S. No. 92-102), and Brief of the United States
as Amicus Curiae Supporting Respondent at 6, Daubert v. Merrell Dow Pharmaceuticals, Inc.
(U.S. 92-102).
82
Daniel Shaviro, Statistical-Probability and the Appearance of Justice, 103 Harvard
Law Review 530, 534, 538, 547 (1989).
83
See generally Laurence H. Tribe, Trial by Mathematics: Precision and Ritual in
the Legal Process, 84 Harvard Law Review 1329 (1971); Charles Nesson, The Evidence or the
Event? On Judicial Proof and the Acceptability of Verdicts, 98 Harvard Law Review 1357
(1985).
52
lower courts, both federal and state, were less reluctant to make the same concession. Once they
had done so, it became clear that it was wrong-minded for the courts to continue to demand that
experts always vouch for their opinions as certainties. It was generally 84 permissible for the
expert to testify that the opinion he or she was asserting amounted to only a probability or
possibility.85
C.
THE THIRD STAGE: THE COURTS= PROHIBITION OF OPINIONS STATED
AS CERTAINTIES
As Part II explained, the concession of the inevitable uncertainty of investigational
science led to the courts= abandonment of the traditional rule that experts had to vouch for their
opinions as scientific or medical certainties. However, the logic of the concession leads beyond
the rejection of the traditional rule. If investigational science relying on inductive reasoning
cannot attain certainty, the courts should not only permit expressly uncertain expert testimony;
they also ought to forbid experts from testifying to purportedly certain opinionsBsuch opinions
are unwarranted as a matter of logic.
84
There are two caveats. The limited nature of the opinion might make the opinion
vulnerable to discretion exclusion under a statute such as Federal Rule of Evidence 403, and
without more an opinion worded as a mere possibility is legally insufficient to sustain the
proponent=s initial burden of production. Fed.R.Civ.P. 56, 28 U.S.C.A.. See Daubert v. Merrell
Dow Pharmaceuticals, Inc., 509 U.S. 579, 595-96 (1993).
85
Nat=l Inst. Standards & Tech., Latent Print Examination and Human Factors:
Improving the Practice through a Systems Approach 118 (2012)(Acourts do not normally demand
absolute certainty from scientists (or any other experts) . . . .@).
53
Within the past five years, two reports on forensic science, both coauthored by scientific
and legal authorities, drew this conclusion. Both reports attacked overstated expert opinions. In
2009, the National Research Council of the National Academies released Strengthening Forensic
Science in the United States: A Path Forward.86 The drafting committee included such members
as Judge Harry Edwards of the United States Court of Appeals for the District of Columbia
Circuit and Professor Constantine Gatsonsis, the Director of the Center for Statistical Sciences at
Brown University as well as two medical examiners and professors of biology, chemistry,
engineering, law, and physics. The report was highly critical of claims that Athe error rate for
fingerprint comparison [is] >essentially zero.=@87 The report stated flatly that the claim is Anot
scientifically plausible.@88 The report not only opposed the admission of expert opinions
couched as certainties; the report also went to the length of urging that A[a]ll results for every
forensic science method . . . indicate the uncertainty in the measurements that are made . . . .@89
The report formally recommended A[t]he development of quantifiable measures of uncertainty in
the conclusions of forensic analyses.@90 The report stated that A[a]s a general matter, laboratory
reports . . . should identify . . . the sources of uncertainty in the . . . conclusions along with
estimates of their scale . . . .@91 In particular, the report took the position that any measurement of
a subject=s blood alcohol concentration Aneed[s] to be reported, along with a confidence interval@
indicating the uncertainty of the measurement.92
86
Nat=l Research Council, Strengthening Forensic Science in the United States: A
Path Forward (2009).
87
Id. at 103. See Sandy Zabel, Fingerprint Evidence, 13 Journal of Law and Policy
143, 177 (2005)(Aexaminers have made assertions that fingerprints are >infallible,= that an
identification is >100 percent positive,= and that the >methodological error rate= for parts of the
process is zero@).
88
National Research Council, supra note 63, at 142. See also Sandy Zabel,
Fingerprint Evidence, 12 Journal of Law and Policy 177 (2005)(Ano scientific basis@).
89
National Research Council, supra note 63, at 184.
90
Id. at 190.
91
Id. at 186.
92
Id. at 117. See also David H. Kaye, David E. Bernstein & Jennifer L. Mnookin,
The New Wigmore: Expert Evidence 101 (2014 Cum.Supp.)(the English Law Commission
made a similar recommendation in a bill submitted to Parliament; the bill states that in deciding
the admissibility of expert testimony, the judge should consider “if the expert’s opinion relies on
the results of . . a . . . measurement . . . , whether the opinion takes proper account of matters,
such as the degree of precision or margin of uncertain, affecting the accuracy or reliability of
those results”).
54
The second report was published in 2012. In that year, the National Institute of Standards
and Technology=s Expert Working Group on Human Factors in Latent Print Analysis published
Latent Print Examination and Human Factors: Improving the Practice through a Systems
Approach.93 Like the National Research Council committee, the working group had a diverse
membership, including fingerprint examiners, computer scientists, cognitive researchers, law
professors, and statisticians. Like the National Research Council report, the N.I.S.T. report faults
experts= overstated claims that they could certainly attribute an impression to a single person Ato
the exclusion of all others in the world.@94 The report urged research into the development of a
probabilistic model for fingerprint testimony, similar to the model used in DNA cases.95 The
report observed that a Anumber of courts have curtailed the certainty with which a judgment of
individualization of toolmarks may be expressed.@96
93
Nat=l Inst. Standards & Tech., Latent Print Examination and Human Factors:
Improving the Practice through a Systems Approach (2012).
94
Id. at 72.
95
Id. at 85-86.
96
Id. at 120 n. 382, citing United States v. Willock, 696 F.Supp.2d 536 (D.Md.
2010), cert.denied sub nom. Smith v. United States, 132 S.Ct. 430, 181 L.Ed.2d 279 (2011);
United States v. Taylor, 663 F.Supp.2d 1170, 1180 (D.N.M. 2009), and Commonwealth v. Pytou
Heang, 458 Mass. 827, 942 N.E.2d 927 (2011).
55
In part inspired by critiques such as these two American reports,97 there is now a definite
judicial trend to forbid experts from testifying to purportedly certain opinions. If, as Justice
Blackmun declared in Daubert,98 it is impossible to attain certainty in investigational science, an
opinion couched as a certainty cannot qualify as reliable Ascientific . . . knowledge@ under Federal
Rule of Evidence 702; the available empirical data cannot support that knowledge claim. That
message has not been lost on the lower courts. As the N.I.S.T. report pointed out, several courts
have already forbidden purportedly certain opinions by toolmark experts.99 Similarly, in a 2010
opinion,100 a federal court prohibited a fingerprint examiner from testifying to such an opinion.
Furthermore, as we shall see in Part IV, in the same year a number of state courts have ruled that
a toxicologist may not describe a single point value for a subject=s blood alcohol concentration as
a certainty.101
D.
THE FOURTH STAGE: THE INCIPIENT JUDICIAL TREND TO INSIST
THAT THE EXPERT ACKNOWLEDGE AND PROVIDE A MEANINGFUL
QUANTITATIVE MEASURE OF THE UNCERTAINTY OF THE
MEASUREMENT
97
There is also a 2011 report from the English Law Revision Commission to the
same effect. David H. Kaye, David E. Bernstein & Jennifer L. Mnookin, The New Wigmore:
Expert Evidence ' 7.6.5, at 102-03 (2014 Cum.Supp.), citing Expert Evidence in Criminal
Proceedings in England and Wales para. 5.35(1) (2011). The commission submitted the report to
Parliament. The report urges Parliament to adopt legislation requiring the proponent of
testimony about a measurement to account for the Amargin of uncertainty.@
98
Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579, 590 (1993).
99
Note 70, supra.
100
United States v. Zajac, 748 F.Supp.2d 1327 (D. Utah 2010). See Paul Giannelli,
Fingerprints: Misidentifications, 20 Criminal Justice, Spr. 2005, at 50 (it is Aastounding@ that
experts still sometimes attempt to testify Athat there is a >zero error= rate in fingerprint
examination@); Tamara F. Lawson, Can Fingerprints Lie?: Re-weighing Fingerprint Evidence in
Criminal Jury Trials, 31 American Journal of Criminal Law 1, 7 (2003)(it is A[in]appropriate to
allow latent fingerprint expert witnesses to testify regarding their findings which are purported to
conclusively link an accused to a crime@); Note, Recognizing and Responding to a Problem with
the Admissibility of Fingerprint Evidence Under Daubert, 45 Jurimetrics Journal 41, 57
(2004)(the expert should not Atestify that a match conclusively ties a defendant to a print found at
the crime scene@; the state of the empirical research does not support Aan absolutist claim@).
101
State of Washington v. Fausto and Ballow, Case No. C076949 and 9Y6231062
(King Cty. Dist.Ct., Wash. Sep. 21, 2010); State of Washington v. Weimer, #7036A-09D
(Snohomish Cty. Dist.Ct., Wash. Mar. 23, 2010).
56
Measures of Error or Uncertainty That the Courts Have Already Approved
This paper began by reviewing the early judicial view that the expert must vouch for his
or her opinion as a scientific or medical certainty. Part III pointed out that today many courts not
only do not mandate such opinions; quite to the contrary, they forbid them. These courts
understandably preclude the expert from expressly misleading the trier of fact by claiming that
the expert=s conclusion is certainly correct. However, other courts have gone farther. Knowing
that a single measurement is inherently uncertain and that metrology recognizes several
quantitative measures of the margin of error or uncertainty, these courts demand102 that the expert
clarify the probative value of the measurement by proffering such a quantitative measure to
accompanying the testimony about the single point value. To date, these judicial demands have
taken two forms.
102
Rod G. Gullberg, Measurement Uncertainty in Forensic Toxicology: Its
Estimation, Reporting and Interpretation, in Toxicity and Drug Testing 415, 416 (W. Acree ed.
2012).
57
The Safety Margin or Band Gap Approach. The early cases opted for the safety margin
or band gap approach.103 In the 1940s the Swedish Supreme Court declared that in all drunk
driving prosecutions, the error or uncertainty of the blood alcohol concentration measurement
must be specified.104 Several American jurisdictions have followed suit. In evaluating the legal
sufficiency of the government=s evidence to sustain a conviction,105 courts in these states demand
that the inherent margin of error be deducted from the single point value. Suppose, for example,
that in a given jurisdiction, the law criminalizes operating a motor vehicle when the driver has a
blood alcohol concentration of 0.08% or more. Assume that in the instant case, the measurement
testimony is that the intoxilyzer reading indicated that the driver had a concentration of 0.081%.
However, suppose further that based on expert testimony, the court concludes that inherent
margin of error of the intoxilyzer is plus or minus 0.002%. On that set of assumptions, the
driver=s true concentration could be as low as 0.079%Bfalling below the 0.08% standard. After
considering all the evidence, the court would find the accused not guilty; the government=s
testimony is legally insufficient to support a finding that the accused=s blood alcohol
concentration was 0.08% or higher.
For its part, the intermediate appellate court in Hawaii has found that the inherent margin
of error for such blood alcohol readings is 0.0165%.106 The court stressed that as a matter of
interpretation, the trigger for the criminal prohibition is an actual blood alcohol concentration,
not a mere instrumental reading. The court reasoned that in order to present a legally sufficient
case, the government had to account for the margin of error in measurement.107 The court
elaborated:
In both of the cases at bar, the State has failed to establish a critical fact. The State
merely demonstrated that the reading of the breathalyzer machine was 0.10% for
Defendant Boehmer and 0.11% for Defendant Gogos. The inherent margin of error could
103
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 206-07 (2011 ed.). See also State v. Finch, 291 Kan. 665, 244 P.3d
673 (2011)(despite the margin of error, the trial judge should not have granted the accused=s
motion for acquittal; however, the testimony about the error margin was a permissible factor for
the jury to consider during deliberation); State v. McGinley, 229 N.J.Super. 191, 550 A.2d 1305
(1988)(since the breathalyzer has a substantial error margin, the blood alcohol measurement must
be adjusted), overruled on other grounds, State v. Downie, 117 N.J. 450, 569 A.2d 242 (1990).
104
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 214 (2011 ed.).
105
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part Two: Legal Analysis, in id. at 255, 266.
106
State v. Boehmer, 1 Haw.App. 44, 613 P.2d 916 (1980).
107
Id. at 918-19.
58
put both defendants= actual blood alcohol level below the level necessary for the
presumption to arise. The failure of the prosecution to establish beyond a reasonable
doubt that the actual weight of alcohol in defendants= blood was at least 0.10% required
the trial judge to ignore [any presumption based on the test result].108
108
Id. at 918.
59
The Nebraska Supreme Court has reached a similar result.109 The court has stopped short
of announcing that the blood alcohol concentration measurement must always be adjusted for a
fixed margin of error.110 The need for an adjustment Ais dependent upon the credible evidence in
each case.@111 However, the court has reiterated that when the credible evidence establishes Athe
margin of error,@ that figure must be Adeduct[ed@ from the reading to Areduce[] the test result . . .
.@112
Unlike the Nebraska and Hawaii courts, the Iowa Supreme Court refused to require that
the trial judge deduct the figure representing the inherent margin of error from the single point
value for the blood alcohol concentration in determining whether the government has met its
burden of going forward.113 While the Iowa court rejected the argument that the margin of error
should be subtracted, the Iowa legislature found the argument persuasive. That legislature then
amended its statute to read:
The results of a chemical test may not be used as the basis for a revocation of a person=s
driver license or nonresident operating privilege if the alcohol or drug concentration
indicated by the chemical test minus the established margin of error inherent in the device
or method used to conduct the chemical test is not equal to or in excess of the level
prohibited . . . .114
In reporting its results, the Virginia Department of Forensic Science has adopted the same
practice.115
109
State v. Bjornsen, 201 Neb. 709, 271 N.W.2d 839 (1978).
110
State v. Babcock, 227 Neb. 649, 419 N.W.2d 527, 530 (1988).
111
Id. at 530.
112
Id.
113
Nugent v. Dept. of Transportation, 390 N.W.2d 125 (Iowa 1986).
114
Iowa C.A. ' 321J.12(6). See also State v. Schuck, 22 Ohio St.3d 296, 490 N.E.2d
596, 598 (1986); State v. Prestier, 7 Ohio Misc.2d 36, 455 N.E.2d 24, 38-39 (Mun.Ct. 1982).
115
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 209 (2011 ed.).
60
The most recent decision in this vein is People v. Jabrocki,116 rendered in 2011 by a
Michigan trial court. In that case, District Judge Peter Wadel ruled that Ablood test results are not
reliable until the state police crime lab[oratory] calculates an uncertainty budget or error rate and
reports that calculation along with the blood test results.@ The court pointed out that the
standards of the American Society of Crime Laboratory Directors= Accreditation Board Arequires
the labs to develop an uncertainty budget for blood alcohol analysis.117 In addition, the court
quotes the 2009 National Academy of Sciences report on forensic science to the effect that A[a]ll
results of every forensic science method should indicate the uncertainty of the measurements that
are made . . . .@118 (In 2011, the English Law Commission made a similar recommendation.119)
Finally, the court cites the two Washington state decisions discussed in the ensuing paragraphs.
116
People of the State of Michigan v. Jeffrey James Jabrocki, 08-5461-FD (Dist.Ct.,
Mason Cty., May 6, 2011).
117
The court cites Jesper Kristiansen & Henning Willads Petersen, An Uncertainty
Budget for the Measurement of Ethanol in Blood by Headspace Gas Chromatography, 28 Journal
of Analytical Toxicology 456 (Sep. 2004). That article states that the uncertainty budget should
include:
----analytical standard uncertainty;
----biological variation (variation in both the density of blood and the matrix difference between
aqueous calibration samples and blood); and
-----traceability uncertainty related to the calibration standards.
Like the courts in some jurisdictions, the authors use the expression, a Asafety margin.@ Id. at
456, 462. Rod G. Gullberg, Measurement Uncertainty in Forensic Toxicology: Its Estimation,
Reporting and Interpretation, in Toxicity and Drug Testing 415, 428 (W. Acree ed.
2012)(discussing the concept of an uncertainty budget).
118
But see Christopher Boscia, Strengthening Forensic Alcohol Analysis in
California DUI Cases: A Prosecutor=s Perspective, 53 Santa Clara L.Rev. 733, 758-59
(2013)(Awhile little reported, NAS qualified its declaration about report requirements when it
said the reports >should describe, at a minimum, methods and materials, procedures, results, and
conclusions, and they should identify, as appropriate, the sources of uncertainty in the procedures
and conclusions along with estimates of their scale (to indicate the level of confidence in the
results). The NAS report did not list examples of when it would not be appropriate to report
sources of uncertainty in forensic science@).
119
David H. Kaye, David E. Bernstein & Jennifer L. Mnookin, The New Wigmore:
Expert Evidence ' 7.6.5, at 102, 107 (2014 Cum.Supp.)(the commission submitted a bill to
Parliament to require English courts to conduct a Daubert-style inquiry; one of the provisions of
the proposed bill reads that Aif the expert=s opinion relies on the results of the use of any method
(for instance, a test, measurement or survey),@ the judge should consider Awhether the opinion
takes proper account of matters, such as the degree of precision or margin of uncertainty,
affecting the accuracy or reliability of those results@; the report is Law Commission, Expert
Evidence in Criminal Proceedings in England and Wales para. 5.35(1)(2011), available at
http://www.justice.gov.uk/lawcommission/docs/lc325_Expert_Evidence_Report.pdf),
61
The confidence interval approach. While Hawaii, Iowa, and Nebraska have used the
safety margin approach to compensate for the uncertainty of measurements, Washington has
taken another approach.120 In Washington, the authorities administer two tests to each drunk
driving suspect and present testimony about the arithmetic mean or average of the two
measurements at trial.121 Two state courts have ruled that in order to convey a sense of the
uncertainty of the measurements to the trier, the prosecution expert must accompany that
testimony with an explanation of the confidence interval for the mean.
In March 2010 in State v. Weimer,122 Commissioner Paul Moon stated that testimony
about a single point value is misleadingly complete unless it is accompanied by evidence of the
confidence interval for the value. The commissioner cited the state version of Federal Rule of
Evidence 403 which empowers a trial judge to exclude an item of evidence if, in his or her
discretion, the judge believes that the probative dangers attending the evidence substantially
outweigh its probative value. As the commissioner explained, one of the recognized dangers
under Rule 403 is that the trier of fact will overvalue the item. Drawing on his previous
experience presiding over drunk driving trials, the commissioner concluded that if the trier does
not have the benefit of a quantified measure of the measurement=s uncertainty such as a
confidence interval, there is an intolerable risk that the jury will treat the point value as Aexact@
and attach undue weight to the value. Near the end of his opinion, Commissioner Moon tersely
asserted that A[t]o allow the test value into evidence without stating a confidence level violates@
Rules 403.
120
Rod G. Gullberg, Measurement Uncertain in Forensic Toxicology: Its Estimation,
Reporting and Interpretation, in Toxicity and Drug Testing 415, 417 (W. Acree ed. 2012)(there
are Amultiple ways@ to approach the problem of estimating uncertainty). To date, no firm
consensus has emerged. Id. at 451.
121
State of Washington v. Fausto and Ballow, Case No. C076949 and 9Y6231062
(King Cty.Dist.Ct. Sep. 21, 2010); Ted Vosk, Measurement Uncertainty: Forensic Metrology
and Result Interpretation. Part Two: Legal Analysis, in Understanding DUI Scientific Evidence
255, 312 (2011 ed.). Taking two measurements enables the analyst to determine whether the
subject was in the absorption or elimination phase at the time of the initial test. 2 Paul C.
Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane Campbell Moriarity, Scientific
Evidence ' 22.04, at 532-33 (5th ed. 2012).
122
State of Washington v. Weimer, #7036A-09D (Snohomish Cty. Dist.Ct., Wash.
Mar. 23, 2010).
62
In September of the same year, a three-judge panel handed down its decision in State v.
Fausto and Ballow.123 The panel asserted that standing alone, the testimony about the mean
measurement paints a Afalse picture.@ In the panel=s view, without an indication of the
uncertainty of the mean, the jury is likely to misinterpret the single point value. The panel
bluntly stated that A[g]iven the inherent variability of measurement, a statement of a measurement
result is incomplete (perhaps even meaningless) without an accompanying statement of the
estimated uncertainty of measurement.@ The panel acknowledged that there are numerous ways
of calculating uncertainty. Indeed, the court expressly cited the Hawaii and Nebraska decisions
adopting the safety net approach requiring testimony about the inherent margin of error.
However, in the instant case the panel decided to require the presentation of testimony about a
bias-corrected mean124 and a confidence interval for the mean. On the one hand, the panel did
not make a sweeping ruling that the prosecution must always present testimony about the
confidence interval. On the other hand, the panel mandated that the government laboratory
provide the defense with the confidence interval to enable the defense to present that testimony to
the jury:
[A]bsent a confidence interval, a Afinal@ breath-alcohol measurement is only a Abest
estimate@ of a person=s breath-alcohol level. Given . . . the discovery rules and E[vidence]
R[ule] 702, the State must provide Defendants with a confidence interval for each
Defendant=s breath-alcohol measurement. Absent this information, a defendant=s breathalcohol measurement will be suppressed.
In all the Washington decisions, the court selected a confidence interval as the most
appropriate measure of the point value=s uncertainty. Before turning to the future of forensic
metrology in the next section, it is important to clarify the notion of a confidence interval. A
confidence interval gives the statistician a sense of the dispersion of the values.125
$ The initial step in computing a confidence interval is identifying the sample statistic to
construct the interval around.126
123
Case No. C076949 and 9Y6231062 (King Cty. Dist.Ct. Sep. 21, 2010).
124
In this context, a bias-corrected mean is the mean adjusted for systematic error.
Systematic errors are those that consistently result in the under or overestimation of the true
value. Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result Interpretation, in
Understanding DUI Scientific Evidence 159, 168 (2011). Such errors remain constant or vary in
a predictable manner. Id. They can be caused by such factors as a calibration mistake or
environmental conditions. Id. A researcher can determine the bias by comparing the
measurements against a reference standard. Id. at 168-69. Having determined the bias, the
researchers corrects the point value to that extent. Id. at 169-70.
125
1 Paul C. Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane Campbell
Moriarity, Scientific Evidence ' 15.02 (5th ed. 2012); Rod G. Gullberg, Measurement
Uncertainty in Forensic Toxicology: Its Estimation, Reporting and Interpretation, in Toxicity and
Drug Testing 415, 415 (W. Acree ed. 2012).
126
See generally Douglas A. Skoog, Donald M. West & F. James Holler,
63
$ The next step is determining the standard error or deviation for the statistic. In effect, the
standard deviation measures the average deviation from the average127 or mean:
Fundamentals of Analytical Chemistry 26-30 (5th ed. 1988); Donald P. Land & Edward J.
Imwinkelried, Confidence Intervals: How Much Confidence Should the Courts Have in
Testimony About a Sample Statistic?, 44 Criminal Law Bulletin 257 (Mar.-Apr. 2008); Edward
J. Imwinkelried, Expert Witness: A Look at Intervals, National Law Journal, May 19, 2008, at
13.
127
David H. Kaye & David A. Friedman, Reference Guide on Statistics, in Reference
Manual on Scientific Evidence 211, 298 (3d ed. 2011); Douglas A. Skoog, Donald M. West & F.
James Holler, Fundamentals of Analytical Chemistry 26-30 (5th ed. 1988).
64
The statistician takes the difference between each value and the mean (the average) and
then squares the differences, adds these squares, and divides by N (or N-1). Squaring the
value is critical. If the statistician simply took the differences and added them, the sum
would be zero; there would be some values higher than the mean and others lower. The
zero would give no insight into the dispersion, squaring the difference yields a positive
number. The standard deviation incorporates all the elements of the set in its
computation, and its magnitude is directly related to the amount of variation in the
data.128
$
The third step is the choice of a confidence coefficient:
The coefficient equals one minus the significance level for the test. Thus, suppose the
expert wanted a significance level of .05, at which an outcome has a probability of
occurring by chance 5% of the time or less. He or she would choose a coefficient of .95.
Researchers commonly use one of three coefficientsB.68, .95, or .99. [W]hen the values
are normally distributed, 68.26% should fall within one standard error of the sample
statistic, 95.44% within two standard errors, and 99.73% within three standard errors.129
$
At this point, the statistician is ready to compute the confidence interval. There are
numerous formulae for computing the interval. As a generalization, though, the formula
incorporates the sample statistic, the sample size, the standard error, and the coefficient.130 There
are various ways of expressing the interval. For instance, the expert might note:
95% CONF (67; 62, 72).
In this notation, 67 is the sample statistic, perhaps a mean. The remaining entries indicate that
with 95% confidence, the interval runs from a lower boundary of 62 to an upper boundary of 72.
It is important to appreciate both what a confidence interval means and what it does not mean.
By way of example, it does not mean that the interval certainly includes the true value of the
population parameter or even that there is a certain probability that the interval includes the true
value.131 What then does the interval denote? The interval represents the proportion of similarly
constructed intervals that should capture the true value:
128
1 Paul C. Giannelli, Edward J. Imwinkelried, Andrea Roth & Jane Campbell
Moriarity, Scientific Evidence ' 15.02[b], at 831-33 (5th ed. 2012).
129
Donald P. Land & Edward J. Imwinkelried, Confidence Intervals: How Much
Confidence Should the Courts Have in Testimony About a Sample Statistic?, 44 Criminal Law
Bulletin 257, 263 (Mar.-Apr. 2008).
130
Edward J. Imwinkelried, Expert Witness: A Look at Intervals, National Law
Journal, May 29, 2008, at 13.
131
Donald P. Land & Edward J. Imwinkelried, Confidence Intervals: How Much
Confidence Should the Courts Have in Testimony About a Sample Statistic?, 44 Criminal Law
Bulletin 257, 270-71 (Mar.-Apr. 2008).
65
It is not so much that we have confidence in the particular interval computed from the
initial sample. Rather, we can have a degree of confidence in the formula, method, or
process. Suppose, for instance, that the expert employed a .95 confidence coefficient. In
the long run on average, 95% of the confidence intervals based on similarly sized,
similarly drawn samples from the same population will capture or cover the true value of
the population parameter.132
If the interval is narrow with a high confidence coefficient, it is justifiable to attach great weight
to the interval.133
In mid-2013, the Washington Court of Appeals handed down its decision in Fausto and
Ballow.134 In the appellate court=s words, the district court judge had ruled that Abreath tests are
categorically inadmissible unless the State introduces a corresponding confidence interval.@135
However, the court affirmed the King County Superior Court=s reversal of that ruling.
Significantly, the court pointed out that on appeal, the prosecution did not Achallenge[]@ the trial
judge=s finding that Aevery measurement is >uncertain= . . . . Even the best instruments yield only
an estimate of the true value. Uncertainty indicates a range in which the true value of a
measurement is likely to occur.@136 For that matter, the prosecution did not Adispute that
measurement uncertainty is recognized in all sciences and uncertainty measurements may be
helpful to the trier of fact in some circumstances.@137 However, the court concluded that the
uncertainty of the BrAC test result went to its weight rather than its admissibility.138 The court
emphasized that the defense has the opportunity to Apresent uncertainty evidence challenging
BrAC test results.@139 Yet, in closing, the court made it clear that it was rejecting only a Ablanket@
rule excluding test results unaccompanied by a confidence interval.140 In dictum at the end of its
132
Id. at 267-68.
133
Id. at 270. See ATA Airlines, Inc. v. Federal Exp. Corp., 665 F.3d 882, 889-96
97th Cir. 2011)(a wide confidence interval).
134
State of Washington v. King County District Court West Division; King County
District Court, East Division; King County District Court, South Division; Ballow; and Fausto,
17 Wm.App. 630, 307 P.3d 765 (2013). See also Christopher Boscia, Strengthening Forensic
Alcohol Analysis in California DUI Cases: A Prosecutor=s Perspective, 53 Santa Cl.L.Rev. 733
(2013).
135
307 P.3d at 766.
136
Id. at 769.
137
Id.
138
Id. at 770.
139
Id.
66
opinion, the court added that in a given caseBperhaps one with a test result close to the legal limit
and a wide confidence intervalBAthe trial court@ could Aexercis[e] its discretion . . . to exclude a[]
. . . BrAC test result on a case-by-case basis.@ The intermediate appellate court=s opinion may
not be the last word on the issue. The individual defendants are now petitioning the Washington
Supreme Court to review the issue.
The Future
140
Id.
67
As we have seen, several American jurisdictions have already mandated the use of
quantitative measures of error or uncertainty by forensic experts to ensure that testimony about a
point value does not mislead the trier of fact. To date, the courts have approved the use of
margins of error and confidence intervals. However, those figures do not exhaust the
possibilities. As the Washington court noted in Fausto,141 there are several quantitative
measures of error or uncertainty.
The safety margin and confidence interval approaches illustrate the classical approach.142
In evaluating the point value, the challenge is identifying and accounting for all the sources of
error or uncertainty. There are two types of problems: systematic errors (bias) and random
errors.143 Systematic errors tend to be unidirectional and consistently cause the point value to
under or overestimate the true value.144 If an instrument is improperly calibrated, it might
underestimate each measurement; or the temperature in a controlled environment could result in
consistent overestimations.145 In contrast, random errors such as operator mistakes occur in an
unpredictable fashion.146 The analyst approaches the two types of error in different manners:
141
State of Washington v. Fausto and Bellow, Case No. C076949 and 9Y6231062
(King Cty.Dist.Ct., Wash. Sep. 21, 2010).
142
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201 (2007).
143
Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of Modern
Approaches to Express Uncertainty in Measurement, 44 Metrologia 513, 515 (2007).
144
Douglas A. Skoog, Donald M. West & F. James Holler, Fundamentals of
Analytical Chemistry 13-14 (5th ed. 1988); Ted Vosk, Measurement Uncertainty: Forensic
Metrology and Result Interpretation. Part One: Measurement Results and Interpretation, in
Understanding DUI Scientific Evidence 159, 168 (2011 ed.).
145
Id.
146
Id. at 170.
68
$
The analyst attempts to identify and eliminate all systematic errors.147 It is virtually
impossible to know whether one has identified all the sources of systematic error.148 Often the
best approach is to compare the measurements in question against measurements with a reference
material.149 Based on that comparison, the analyst can compute a bias-corrected statistic.150
Otherwise, after making his or her best efforts to remove the sources of bias, the analyst has to
assume that all systematic error has been eliminated.151
$ The random error, though, is calculable.152 The analyst makes a large number of replicate
measurements to determine how they vary in an unpredictable manner.153 Given that set of
measurements, the analyst can compute a standard error or deviation154 (which, in turn, can be
employed to calculate a confidence interval).
However, there is no generally accepted method of combining systematic and random errors into
an overall error.155 Thus, under classical approach, the witness would testify to both a biascorrected statistic (to compensate for systematic error) and a separate confidence interval
(reflecting random error).
147
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation Quality
Assurance 201, 203-05 (2007).
148
Id. at 206.
149
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Results
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 168-69 (2011 ed.).
150
Id. at 169-70, 185.
151
Id. at 175, 179.
152
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 206.
153
Id. at 207; Ted Vosk, Measurement Uncertainty: Forensic Metrology and Results
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 170-71, 174, 176 (2011 ed.).
154
Id. at 171.
155
Id. at 178-79 (Athe two kinds of error are fundamentally different, requiring
different treatment@; Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy
and Description of Measurement (Preliminary Rationale for VIMS), 12 Accreditation and
Quality Assurance 201, 206 (2007).
69
Yet, modern metrology is moving away from the classical approach.156 Uncertainty
analysis is replacing error analysis.157 Today one of the foundations for uncertainty analysis is
the 1995 Guide to the Expression of Uncertainty in Measurement (GUM).158 The International
Committee for Weights and Measures (CIPM), the International Bureau of Weights and
Measures (BIPM), and several national metrology institutes collaborated to develop the
recommendations forming the core of the GUM approach.159 The essence of the approach is
combining systematic and random effects160 to generate a coverage interval that includes the true
value of the measurand to a certain probability.161 Just as the classical approach attempts to
identify systematic and random errors, the GUM approach considers two types of uncertainty.
The two types differ in the manner in which their numerical values are estimated:162
$ Type A uncertainty can be determined by the statistical analysis of a series of replicate
measurements or observation.163 The analyst can use several statistical techniques such as the
calculation of a standard deviation to quantify the random effects.164
$ Type B uncertainty is determined Aby means other than the statistical analysis of series of
observations.@165 The analyst exercises judgment based on such considerations as his or her
personal experience with the limits of the measurement technique employed.166
156
Id. at 206; Appendix, Understanding DUI Scientific Evidence 459, 514-18 (2011
ed.).
157
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Results
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 180 (2011 ed.).
158
Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of Modern
Approaches to Express Uncertainty in Measurement, 44 Metrologia 513, 513 (2007).
159
Id.
160
Id. at 520; Charles Erlich, Rene Dydkaer & Wolfgang Woger, Evolution of
Philosophy and Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation
and Quality Assurance 201, 209, 211 (2007).
161
Id. at 210-12.
162
Id. at 181.
163
Id. at 211; Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 211 (2011 ed.).
164
Id. at 194.
165
Id.
166
Id. at 194-95 (AIt relies on scientific knowledge, experience, and judgment in the
context of all available information to create a priori probability density functions based on the
70
degree of belief about effects that are not characterized, at least not completely, through
measurement itself. The information relied on may include: [q]uantity values from authoritative
publications and handbooks, [q]uantity values from reference material certifications, [c]alibration
certificates, [m]anufacturer specifications, [a]ccuracy classifications of a verified measuring
instrument, [l]imits deduced through personal experience, [and e]xperience with, or general
knowledge of the behavior and property of relevant materials, methods, and instruments@)
71
Type A evaluations are sometimes called Aobjective@ since they rest on direct analysis of
measurement results while Type B evaluations are termed Asubjective@ because they rely on
Aresearcher judgment and other information.@167 However, it cannot be assumed that Type A
evaluations are necessarily more reliable; while a Type A evaluation might be based on a small
number of measurements, a Type B evaluation could be informed by extensive experience.168
167
Id. at 196.
168
Id.
72
In any event, after conducting both evaluations, the analyst treats the Type A and Type B
standard uncertainties alike169 and uses Bayesian techniques170 to combine171 the two evaluations.
The purpose of the combination is to describe the true state of knowledge about the overall
uncertainty of the point value.172 What is the degree of belief about the accuracy of the point
value?173 The analyst can use an uncertainty budget, listing all the types of Type A and B
uncertainties and their magnitudes.174 The end result is a combined standard uncertainty and
169
Id. at 198.
170
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 212 (2007); Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 181 (2011). The theorem, devised by the Reverend Thomas Bayes,
permits the combination of probabilities, namely, the use of a new input to revise a prior
probability to yield a posterior probability. 1 Paul C. Giannelli, Edward J. Imwinkelried, Andrea
Roth & Jane Campbell Moriarty, Scientific Evidence ' 15.07[a], at 916-19 (5th ed. 2012), citing
Thomas Bayes, An Essay Toward Solving a Problem in the Doctrine of Chances, 53 Phil.
Transactions Royal Society 370 (1963). Importantly, the theorem enables the statistician to
combine an a prior probability estimated by lay jurors with genetic evidence to establish the
posterior probability. In this context, the statistician can put the systematic effect and random
effect Aon the same footing.@ Vosk, supra, at 182.
171
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 180, 182, 198 (2011 ed.)(ARegardless of whether Type A or Type B analysis
is engaged in, both are based on probability distributions, and the uncertainty so determined is
quantified as either a variance of standard deviation. When expressed as a standard deviation,
each quantified source constitutes a standard uncertainty. >The combined standard uncertainty of
a measurement result . . . is taken to represent the estimated standard deviation of the result=@);
Charles Erlich, Rene Dybkaer & Wolkgang Woger, Evolution of Philosophy and Description of
Measurement (Preliminary Rationale VIM3), 12 Accreditation and Quality Assurance 201, 209,
211 (2011 ed.).
172
Id. at 212; Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution of
Modern Approaches to Express Uncertainty in Measurement, 44 Metrologia 513, 517 (2007);
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result Interpretation. Part One:
Measurement Results and Interpretation, in Understanding DUI Scientific Evidence 159, 183,
186 (2011 ed.).
173
Id. at 181.
174
Id. at 197; Patrick T. Barone, DUI Defense: If Breath Alcohol Can=t Be Measured,
It Doesn=t Exist, 37 SADO Crim.Def. Newsletter, Nov. 2013, at 1, 10; Raghu Kacker, KlausDieter Sommer & Rudiger Kessel, Evolution of Modern Approaches to Express Uncertainty in
Measurement, 44 Metrologia 513, 521 (2007); Jesper Kristiansen & Henning Willads Petersen,
73
coverage or credible interval.175 With a specified level of confidence, 176 this Acoverage@177 or
Acredible@178 interval includes the true value. Thus, unlike a confidence interval,179 this coverage
interval includes the true value with a stated probability.180 Together, the best estimate and the
coverage interval completely characterize our knowledge of the quantity value sought.181
Although some commentators believe that it would be possible to use coverage intervals today,
others are of the view that the limited data with respect to breath alcohol analysis makes it
An Uncertainty Budget for the Measurement of Ethanol in Blood by Headspace Gas
Chromatography, 28 Journal of Analytical Toxicology 456 (Sep. 2004).
175
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 211 (2007); Raghu Kacker, Klaus-Dieter Sommer & Rudiger Kessel, Evolution
of Modern Approaches to Express Uncertainty in Measurement, 44 Metrologia 513, 519 (2007).
There is admittedly some confusion over the proper terminology. R. Willink, Coverage
Intervals and Statistical Coverage Intervals, 41 Metrologia L5, L6 (2004). For example, some
commentators appear to use the expression, Anominal coverage probability,@ to describe what
other statisticians refer to as the statistical confidence interval. Id.
176
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 203 (2011 ed.).
177
Id.; Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy
and Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 211-12 (2007); Ted Vosk, Measurement Uncertainty: Forensic Metrology and
Result Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 188, 203-04, 224 (2011 ed.).
178
Raghu Kacher, Klaus-Dieter Sommer & Rudriger Kessel, Evolution of Modern
Approaches to Express Uncerainty in Measurement, 44 Metrologia 513, 516 (2007); R. Willink,
Coverage Intervals and Statistical Coverage Intervals, 41 Metrologia L5, L6 (2004); E-mail
communication from Professor David H. Kaye (Nov. 12, 2012).
179
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 204 (2011 ed.).
180
Id.; Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy
and Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 211-12 (2007).
181
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result
Interpretation. Part One: Measurement Results and Interpretation, in Understanding DUI
Scientific Evidence 159, 186 (2011 ed.).
74
inappropriate to use the Bayesian approach to estimate the uncertainty in such analyses.182
182
While some believe that shifting to the Bayesian approach is desirable, others
question whether doing so would be an improvement over the classical approach, at least when
used in clinical or forensic laboratories. The classical methods of determining uncertainty are set
out in numerous textbooks on analytical chemistry and have been employed for well over 100
years.
One proponent of continued use of the classical approach is Dr. G. Simpson. Dr.
Simpson is a retired American research chemist who has specialized in concentration
measurements with a beam of light. He has published several articles in scientific journals,
including Clinical Chemistry, dealing with the margin of error in breath test results. In a series of
emails to the author in January and February 2013, Dr. Simpson stated that in order to determine
the reliability of a breath test result, the amount of error in the result must be known. This can be
done only by conducting experiments on a large number of subjects where rebreathing tests are
conducted to establish the difference between the test result and the rebreathing test result. Dr.
Simpson points out that although there are available simulator data, few data compare subjects=
measured BrAC with the actual BrAC for a particular subject. In his words, AIf the
measurements have not been done, there is no way to establish the margin of error.@ In Dr.
Simpson=s view, clinical and forensic laboratories should eschew the Bayesian approach and
continue to rely on the classical approach.
Likewise, in February and March 2013 emails, a leading European toxicologist, Dr. Frank
Martens of Belgium, pointed out that in his country, all the laboratories continue to employ the
classical approach rather than Bayesian statistics.
75
The GUM approach is not the only alternative to classical error analysis.183 For instance,
there is a hybrid approach, sometimes called the Conventional Value Hybrid Approach.184 This
is a two-step procedure. First, the analyst calibrates a measurement standard by using a Ahighlevel@ measuring system. Next, the analyst performs a second measurement on the calibrated
measurement standard by using a Alower-level@ procedure. AAn example of the CHVA is the use
of a standard weight to verify the performance of a balance. The weight is the (calibrated)
measurement standard, and the balance is the lower-level measuring instrument used to obtain
the measured quantity value . . . . The knowable measurement error is the difference between the
indication and the conventional quantity value of the weight that is placed on the balance.@185
The International Electrotechnical Commission has developed still another approach.186 The
pragmatic IEC approach concentrates on the compatibility of measurement results rather than any
true value.187 The focus is on the question of whether the measurements are sufficiently
compatible for operational purposes.188 For operational purposes, it is often unnecessary to
know the true value.189
III.
CONCLUSION
As Part II.A noted, at one time many courts subscribed to the silly view that an expert
opinion was admissible only if the expert was prepared to vouch that the opinion was certainly
true. In a series of steps, the courts have retreated from that view. Initially, as Part II.B
explained, the courts took the small step of ruling that it was permissible for an expert to testify
to an opinion couched as a mere probability or possibility. The courts began to realize that
especially in investigational science when experts rely on inductive reasoning, in principle an
expert can never validate any hypothesis as a certainty. Of course, as Part II.C demonstrated, the
183
Jack Wallace, Ten Methods for Calculating the Uncertainty of Measurement, 50
Science and Justice 182 (2010). The American Society of Crime Laboratory Directors Adoes not
prescribe a specific formula@ or Aa specific approach to estimating uncertainty of measurement.@
Ted Vosk, Measurement Uncertainty: Forensic Metrology and Result Interpretation. Part One:
Measurement Results and Interpretation, in Understanding DUI Scientific Evidence 159, 213
(2011 ed.).
184
Charles Erlich, Rene Dybkaer & Wolfgang Woger, Evolution of Philosophy and
Description of Measurement (Preliminary Rationale for VIM3), 12 Accreditation and Quality
Assurance 201, 215-17 (2007).
185
Id. at 216-17.
186
Id. at 213.
187
Id.
188
Id.
189
Id.
76
logic of that realization led farther; the logic dictated that the courts forbid overstated opinions
phrased as certainties. Part II.D documents the denouement of this new line of authority: judicial
mandates that whenever practical, the expert give the trier of fact a quantitative measure of the
error or uncertainty of the measurements underlying his or her opinion.
That denouement represents progress in two important respects. First, this incipient trend
promotes honesty in the courtroom. It is axiomatic that measurements are inherently uncertain.
As the Washington cases emphasize, it is misleading to present the trier of fact with only a single
point value. There is a grave risk that without the benefit of qualifying testimony, a miscarriage
of justice can occur because the trier will mistakenly treat the point value as exact and ascribe
undue weight to the evidence. The antidoteBthe necessary qualification--is a quantitative
measure of the margin of error or uncertainty.
Secondly, this trend should foster a more cooperative relationship between law and
science. Scientists often complain that the law forces them to use artificial standards and
pressures them to testify in terms that make them uncomfortable. Psychiatrists frequently assert
that the law compels them to use definitions of insanity that have no counterpart in their
professional discipline.190 Likewise, experts who know that their measurements are uncertain
are ill at ease opining that their opinion is a scientific or medical certainty.191 Like any other
witness, the expert takes an oath to tell Athe whole truth;@192 and an expert with any familiarity
with metrology knows that his or her testimony about the single point value is far from Athe
whole truth.@ The law should permit and encourage experts to Aemploy[] in the courtroom the
same level of intellectual rigor that characterizes the practice of an expert in the relevant field.@193
In Daubert, Justice Blackmun acknowledged that Athere are important differences between the
quest for truth in the courtroom and the quest for truth in the laboratory.@194 However, there is
one point on which judges, attorneys, and lawyer ought to agree: No expert should be permitted
to mislead the trier of fact.195 There is an unavoidable uncertainty in every measurement
190
T. L. Clanon, Lois Shawer & Douglas Kurdys, Less Insanity in the Courts, 68
American Bar Association Journal 824, 824 (July 1982)(in court, psychiatrists and psychologists
Aare asked to testify on issues that have no scientific definitions . . . .@; psychiatrists and
psychologists are Adissatis[fied@ because the law relies on Aan obsolete concept . . . of the
relationship between mental illness and crime@; the disconnect is so extreme that A[m]any
[scrupulous] psychiatrists and psychologists refuse to participate in criminal trials . . . .@).
191
Rod Gullberg, Professional and Ethical Considerations in Forensic Breath Alcohol
Testing Programs, 5 Journal of the Alcohol Testing Alliance 22 (2006).
192
State of Washington v. Fausto and Ballow, No. CO76949 and 9Y6231062
(King.Cty.Dist.Ct., Wash. Sep. 21, 2012); Ted Vosk, Measurement Uncertainty: Forensic
Metrology and Result Interpretation. Part Two: Legal Analysis, in Understanding DUI Scientific
Evidence 255, 312 (2011 ed.).
193
Kumho Tire Co., Ltd. v. Carmichael, 526 U.S. 137, 152 (1999).
194
Daubert v. Merrell Dow Pharmaceuticals, Inc., 509 U.S. 579, 596-97 (1993).
195
In the words of Federal Rule of Evidence 702(a), misleading testimony does not
77
underpinning an expert opinion, and it is intellectually dishonest for the expert to pretend
otherwise in court.
Ahelp the trier of fact to understand the evidence or to determine a fact in issue . . . .@
Furthermore, Federal Rule 403 recognizes that the risk of Amisleading the jury@ can justify the
exclusion of relevant, otherwise admissible evidence.
78