Download working draft - DAVID KAPLAN | Macquarie University

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Central pattern generator wikipedia , lookup

Clinical neurochemistry wikipedia , lookup

Sound localization wikipedia , lookup

Molecular neuroscience wikipedia , lookup

Cortical cooling wikipedia , lookup

Neural oscillation wikipedia , lookup

Neuroesthetics wikipedia , lookup

Neuroanatomy wikipedia , lookup

Convolutional neural network wikipedia , lookup

Neural coding wikipedia , lookup

Neuroeconomics wikipedia , lookup

Binding problem wikipedia , lookup

Neural correlates of consciousness wikipedia , lookup

Feature detection (nervous system) wikipedia , lookup

Artificial neural network wikipedia , lookup

Optogenetics wikipedia , lookup

Neural modeling fields wikipedia , lookup

Channelrhodopsin wikipedia , lookup

Mathematical model wikipedia , lookup

Cognitive neuroscience wikipedia , lookup

Neuroethology wikipedia , lookup

Synaptic gating wikipedia , lookup

Recurrent neural network wikipedia , lookup

Neuroinformatics wikipedia , lookup

Types of artificial neural networks wikipedia , lookup

Holonomic brain theory wikipedia , lookup

Neuropsychopharmacology wikipedia , lookup

Biological neuron model wikipedia , lookup

Neural engineering wikipedia , lookup

Development of the nervous system wikipedia , lookup

Metastability in the brain wikipedia , lookup

Neural binding wikipedia , lookup

Nervous system network models wikipedia , lookup

Neurophilosophy wikipedia , lookup

Transcript
Neural computation, multiple realizability, and the
prospects for mechanistic explanation
David Michael Kaplan
There is an ongoing philosophical and scientific debate concerning the nature of computational explanation in
the neurosciences. Recently, some have cited modeling work involving so-called canonical neural computations –
standard computational modules that apply the same fundamental operations across multiple brain areas – as
evidence that computational neuroscientists sometimes employ a distinctive explanatory scheme from that of
mechanistic explanation. Because these neural computations can rely on diverse circuits and mechanisms,
modeling the underlying mechanisms is supposed to be of limited explanatory value. I argue that these
conclusions about computational explanations in neuroscience are mistaken, and rest upon a number of
confusions about the proper scope of mechanistic explanation and the relevance of multiple realizability
considerations. Once these confusions are resolved, the mechanistic character of computational explanations can
once again be appreciated.
1. Introduction
Multiple realizability considerations have played a long-standing role in debates about reduction, explanation, and the
autonomy of higher-level sciences including psychology and functional biology. The concept of multiple realizability
(hereafter, MR) was initially introduced in the philosophy of mind to serve as an argument against mind-brain identity
theories, and to motivate a functionalist view about the mind. Since the 1970s, however, there has been a gradual shift
in focus beyond ontological issues raised by MR to epistemological (i.e., explanatory and methodological) ones
including whether psychological explanations and methods are autonomous from or reduce to those encountered in
lower-level sciences such as neuroscience. Fodor (1974) famously appeals to MR in an attempt to block the reduction of
higher-level theories and/or laws of the special sciences to those of lower-level sciences. Putnam (1975) invokes MR to
argue for the “autonomous” character of higher-level explanatory generalizations in psychology. Several decades later,
Fodor explicitly foregrounds the explanatory issues raised by MR, asserting that “the conventional wisdom in the
philosophy of mind is that psychological states are functional and the laws and theories that figure in psychological
explanations are autonomous [i.e., are not reducible to a law or theory of physics]” (1997, 149). These formulations of
the MR argument directly link up with debates in the philosophy of science concerning explanation and theory
reduction in ways that many early construals do not.1
Although philosophical excitement about MR and the prospects for the autonomy of psychology in its
traditional guise have faded to some degree in recent years, MR considerations are once again being invoked in a
different light in order to draw conclusions about the nature of explanation, and in particular, computational explanation
in neuroscience and psychology. More specifically, it has recently been argued by several authors (Carandini 2012;
1
For further discussion of the distinction between epistemological and ontological construals of reductionism, and the importance of the former over
the latter in scientific contexts, see Ayala (1968) Hoyningen-Huene (1989). Although these authors are explicitly concerned with reductionism in
biology, the same kinds of considerations are relevant in the present context. Finally, for examples of work on MR that prioritize explanatory over
ontological issues associated with reductionism, see many contributions from the philosophy of biology (e.g., Kitcher 1984; Rosenberg 2001; Sarkar
1992; Sober 1999; Waters 1990).
1
Carandini and Heeger 2012; Chirimuuta 2014) that certain kinds of computations performed in the brain — so-called
canonical neural computations — cannot be explained in mechanistic terms. Briefly, canonical neural computations
(hereafter CNCs) are defined as “standard computational modules that apply the same fundamental operations in a
variety of contexts” (Carandini and Heeger 2012, 51). According to these authors, the reason why CNCs cannot be
accommodated by the framework of mechanistic explanation – which is dominant across the life sciences – is that these
computations are associated with multiple neural circuits or mechanisms in the brain, which can vary from region to
region and from species to species. In other words, CNCs cannot be analyzed in mechanistic terms because they are
multiply realized. Instead, as Chirimuuta (2014) argues, modelling work involving CNCs must embody an entirely
different, non-mechanistic form of explanation. In this chapter, I argue against the necessity of this claim by showing
how MR fails to block the development of adequate mechanistic explanations of computational phenomena in
neuroscience. In this manner, I show how MR considerations are largely irrelevant to assessing the quality of
mechanistic explanations. Once these confusions are resolved, it becomes clear that CNC explanations can be properly
understood as mechanistic explanations.
The chapter is organized as follows. In Section 2, I discuss some traditional MR-based arguments for the
explanatory autonomy of psychology and briefly highlight their well-known limitations. In Section 3, I introduce the
concept of canonical neural computation and outline how MR considerations are supposed to raise problems for
mechanistic explanations of these phenomena. In Sections 4 and 5, I describe two different mechanistic explanations in
neuroscience that invoke the same neural computation to show how MR considerations, although operative in these
cases, are irrelevant to assessing the quality of the explanations provided. In Section 6, I show how persistent
confusions over the proper scope of mechanistic explanations serve to underwrite MR-based challenges to mechanistic
explanation of computational phenomena. Specifically, I elaborate how scope is not an appropriate norm on mechanistic
explanation and therefore even mechanistic models with highly restricted scope can in principle be explanatorily
adequate. This conclusion provides a foothold the view that mechanistic models of computational phenomena can
provide legitimate explanations, even in the presence of multiple realizability. In Section 7, I return to the specific
debate concerning the nature of CNC explanations and expose these same confusions about the relationship between
scope and mechanistic explanation. In Section 8, I summarize the mechanistic perspective on neural computation
defended in this chapter.
2 Traditional MR-based arguments for explanatory autonomy
The central idea behind traditional, MR-based arguments for autonomy is that the phenomena or kinds identified in the
theories of a higher-level science such as psychology can be multiply realized by heterogeneous sets of lower-level
realizers, and so will invariably be cross-classified by, and therefore cannot be explained in terms of, the phenomena or
kinds invoked in lower‐level theories (e.g., Fodor 1974, 1997; Putnam 1975). In virtue of this, higher-level phenomena
or kinds (and the theories and generalizations that invoke them) are said to enjoy a certain degree of explanatory
autonomy or independence from lower-level phenomena or kinds (and the theories or generalizations that invoke them).
Before proceeding, it is important to distinguish this kind of autonomy from methodological autonomy. A scientific
discipline X is methodologically autonomous from some lower-level discipline Y if X’s investigative methods, discovery
procedures, and most importantly, its taxonomic categories can vary independently of Y’s, or vice versa.2 Some
advocates of autonomous computational psychology implicitly embrace methodological autonomy without taking a
stance on explanatory autonomy. For example, the psychologist Philip Johnson-Laird embraces methodological
2
For a similar characterization of methodological autonomy, see Feest (2003).
2
autonomy several decades when boldly pronouncing that “[t]he mind can be studied independently from the brain.
Psychology (the study of the programs) can be pursued independently from neurophysiology (the study of the machine
and the machine code)” (Johnson-Laird 1983, 9).
It is important to keep these two notions of autonomy distinct because the evidence against the methodological
autonomy of psychology from neuroscience is relatively strong, whereas the case concerning the explanatory autonomy
of psychology remains far more uncertain. Keeley (2000), for example, argues compellingly that the discovery and
categorization of psychological kinds frequently depends on structural information about the brain delivered from the
neurosciences. Similarly, Bechtel and Mundale (1999) make a powerful case that the taxonomic enterprises of
psychology and neuroscience are not independent in the ways implied by the thesis of multiple realizability. In
particular, they show how the taxonomic identification of brain areas in neuroscience relies intimately on the
deployment of functional criteria and prior categorizations from psychology. They also demonstrate how efforts to
decompose psychological capacities or functions often depends on, and is sometimes refined by, structural information
about the brain and its organization. The functional and structural taxonomies of psychology and neuroscience thus
appear to reflect the complex interplay and interaction between the disciplines. The resulting picture is arguably one of
methodological interdependence, not autonomy. Consequently, even if the ontological thesis that psychological states
are (or can in principle be) multiply realized in different substrates, there is good reason to think this does not support
the methodological claim that neuroscience lacks usefulness in guiding psychological investigations, and vice versa. By
contrast, the status of the explanatory autonomy of psychology is far less settled.3 In what follows, the focus will
exclusively be on this form of autonomy.
As indicated above, many of the hallmark defenses of the explanatory autonomy of psychology and other
higher-level sciences invoke MR considerations in order to deny the possibility of reducing the theories or laws of the
higher-level science to those of the lower-level science (e.g., Fodor 1974; Putnam 1975).4 Proponents of these views
(and their critics) have generally assumed a traditional conception of theory reduction, according to which some higherlevel theory (or law) is said to be reducible to (and also explained by) some lower-level theory (or law) just in case the
former can be logically derived from the latter, together with appropriate bridge principles and boundary conditions
(e.g., Nagel 1961; Schaffner 1967). Since the terminology invoked by the higher- and lower-level theories often differ
in appreciable ways, so-called bridge principles connecting the terms or predicates of the two theories in a systematic
(one-to-one) manner are necessary to ensure the derivation can go through.5 The classical antireductionist strategy thus
involves showing how the bridge principle-building enterprise breaks down because higher-level phenomena are often
multiply realized by heterogeneous sets of lower-level realizers, implying an unsystematic (many-to-one) mapping
between the terms of the two theories. For example, Fodor, who famously frames the issue in terms of the challenges
MR poses for the logical positivist goal of the unity of science, states that:
The problem all along has been that there is an open empirical possibility that what corresponds to the natural kind
predicates of a reduced science may be a heterogeneous and unsystematic disjunction of predicates in the reducing
3
Before proceeding it is worth preempting a certain mistaken conclusion that this analysis might elicit. Although I am focusing here on challenges
raised about the standard conclusion of the MR argument — the autonomy of psychology — this does not imply that this is the exclusive or even
primary focus in the critical literature. Many authors have also challenged the MR argument by denying the MR premise itself. Although MR once
held the status as an unquestioned truth, as philosophers started to pay closer attention to empirical research in neuroscience, the evidence for multiple
realization began to appear less clear-cut than had been initially assumed (e.g., (Bechtel and Mundale 1999; Polger 2009; Shapiro 2000, 2008).
4
According to the traditional model, theory reduction implies that all the higher-level laws and observational consequences can be derived from
information contained in the lower-level theory. Strictly speaking, then, the higher-level (reduced) theory can make no non-redundant or
“autonomous” informational or explanatory contribution beyond that made by the lower-level (reducing) theory.
5
Because the lower-level theory will typically only apply over a restricted part of the domain of the higher-level theory (Nagel 1961) or at certain
limits (Glymour 1970; Batterman 2002), boundary conditions that set the appropriate range for the reduction are also typically required for the
derivation to succeed. For further discussion, see Kaplan (2015).
3
science, and we do not want the unity of science to be prejudiced by this possibility. Suppose, then, that we allow that
bridge statements may be of the form […] Sx ⇆P1x ∨ P2x ∨ … ∨ Pnx, where ‘P1x ∨ P2x ∨ … ∨ Pnx’ is not a
natural kind predicate in the reducing science. I take it that this is tantamount to allowing that at least some 'bridge laws'
may, in fact, not turn out to be laws, since I take it that a necessary condition on a universal generalization being lawlike
is that the predicates which constitute its antecedent and consequent should pick natural kinds.” (Fodor 1974, 107-108).
The major difficulty with this and many other traditional positions staked out on both sides of the debate over
explanatory autonomy is that they all commonly assume the appropriateness of what is now widely recognized as an
outdated and inapplicable law-based model of theory reduction and explanation (Rosenberg 2001).6 The problem, now
widely recognized among philosophers of the special sciences including biology and psychology, is the conspicuous
absence of lawful generalizations, either at the level of the reducing theory or the reduced theory (e.g., Machamer,
Darden, and Craver 2000; Woodward 2000).7 This is problematic for traditional reductionists because there is no scope
for a reduction without laws in either theory. Interestingly, the absence of laws also raises difficulties for traditional
defenders of autonomous psychological explanation because all recourse to laws, even at the level of the reduced
theory, becomes empty in this nomological vacuum. For this reason, philosophical interest in these debates has
diminished in recent years.
Despite diminished attention of late, MR considerations are once again being invoked in a different light in
order to draw conclusions about the nature of explanation, and in particular, computational explanation in neuroscience.
Critically, if these arguments about the distinctness and autonomy of computational explanation in neuroscience can
secure a foothold, then broader conclusions about the autonomy of computational explanations in psychology are also
within reach. In what follows, I will argue that even when rehabilitated from the outmoded framework of laws, MR
considerations are orthogonal to issues concerning explanation.
3 Canonical neural computation explanations and multiple realizability
Several recent authors (e.g., Carandini 2012; Carandini and Heeger 2011; Chirimuuta 2014) claim that certain kinds of
computations performed in the brain – canonical neural computations (CNCs) – cannot be explained in mechanistic
terms because they rely on a diversity of neural mechanisms and circuits. According to Carandini, a leading visual
neuroscientist, the brain “relies on a core set of standard (canonical) neural computations: combined and repeated across
brain regions and modalities to apply simpler operations to different problems.” (Carandini 2012, 508). The key idea is
that these neural computations are repeatedly deployed across different brain areas, different sensory modalities, and
even across different species.8 Prominent examples of CNCs include linear filtering and divisive normalization. Linear
filtering involves a computation of the weighted sum of synaptic inputs by linear receptive fields. Divisive
6
Explanation and theory reduction are closely linked in the traditional Nagelian model. Just as the D–N model ties successful explanation to the
presence of a deductive or derivability relationship between explanans and explanandum statements, so too successful reduction is tied to the presence
of a deductive relationship between two theories. Intertheoretic reduction is thus naturally treated as a special case of deductive-nomological
explanation. For further discussion, see Kaplan (2015).
7
To the extent that the notion of “law” continues to find uses in these scientific disciplines it is perhaps best characterized in terms of describing the
pattern, effect, or regularity to be explained, rather than as providing the explanation (e.g., Bechtel 2008; Cummins 2000; Wright and Bechtel 2006).
For example, Fitts’ law describes the tradeoff (negative correlation) between speed and accuracy in goal-directed, human motor behavior. The WeberFechner law describes the robust psychological finding that the just-noticeable difference between two sensory stimuli is proportional to their
magnitudes. These well-known generalizations mathematically describe but do not explain what gives rise to these widespread patterns or regularities
in human behavior and perception.
8
One assumption not addressed in this chapter is the foundational one concerning whether neural systems genuinely compute and process
information. While most computational neuroscientists assume that neural systems perform computations (canonical or otherwise), and so go beyond
what many other computational scientists assume about the phenomena they investigate when building computational models (e.g., computational
climate scientists do not think that climate systems compute in the relevant sense), this remains open to challenge. For further discussion of this
important issue, see Piccinini and Shagrir (2014).
4
normalization involves the activity of one neuron (or neural population) being normalized or adjusted by the activity of
other neurons in the same region. Below, I will focus exclusively on the example of divisive normalization.
The divisive normalization model was initially introduced to account for certain response properties of neurons
in primary visual cortex (V1) including contrast saturation and cross-orientation suppression (Heeger 1992; Carandini
and Heeger 1994; Carandini et al. 1997), which could not be handled by linear models of orientation selectivity (e.g.,
Hubel and Wiesel 1968). Cross-orientation suppression, for example, describes the phenomenon in which so-called
simple cells in V1 exhibit weaker responses to the superposition of two orientation stimuli than would be predicted by
the sum of the responses to each individual stimulus alone, even when one of the component stimuli evokes no response
at all (Figure 1).
Figure 1. Example of cross-orientation suppression in V1 neurons. (A) Grating stimuli are presented in
alignment with the neuron’s preferred orientation (top row; test stimuli), orthogonal to the neuron’s preferred
orientation (left column; mask stimuli), or by superimposing the two (remaining panels; plaid stimuli). (B)
Corresponding responses of a typical V1 neuron to the same stimulus set. Cross-orientation suppression is most
evident when non-preferred orientation grating stimuli are presented at high contrasts (bottom rows). For this
neuron, the superposition of two gratings (e.g., bottom right panel in A), one at the preferred-orientation (e.g.,
top right panel in A) and one in the null or orthogonal orientation (e.g., bottom left panel in A), evokes a smaller
response than does the preferred-orientation grating alone, even though the null orientation stimulus has no
effect on firing as compared to baseline levels. (Modified from Freeman et al. 2002)
Models of orientation selectivity such as that of Hubel and Wiesel, in which converging feedforward inputs from the
lateral geniculate nucleus (LGN) are combined by taking a weighted sum, fail to account for this and other non-linear
response properties. The divisive normalization model accommodates these non-linearities by including not only a term
for the feedforward or excitatory inputs (sometimes called the summative or simply receptive field), but also a term in
the divisor reflecting activity pooled from a large number of nearby neurons (sometimes called the suppressive field)
(equation 1). The overall neural response (output) thus reflects the ratio of activity between these two terms, and the
essential computational operation being performed is division. A generalized version of the normalization model is:
[1]
where Rj is the normalized response of neuron j; Dj in the numerator represents the neuron’s driving input; ∑j Dk
represents the sum of a large number of inputs, the normalization pool. The constants y, 𝜎, and n represent free
parameters fit to empirical data: y scales the overall level of responsiveness, 𝜎 prevents undefined values resulting from
division by zero, and n is an exponential term that modulates the values of individual inputs.
5
The model successfully accounts for response saturation in V1 neurons with increasing stimulus contrast
because terms in both the numerator and denominator positions scale proportionally with changes in stimulus contrast.
Increasingly larger driving/excitatory responses elicited by increasing contrast are kept neatly in check/effectively
suppressed (are “normalized” in the mathematical sense) because they are being divided by an increasingly larger
number in the denominator reflecting inputs from the entire suppressive field. The model also predicts cross-orientation
suppression effects: Responses of a V1 neuron to an optimally oriented stimulus (e.g., a grating stimulus aligned with
the neuron’s preferred orientation) are suppressed by superimposing a mask stimulus of different orientation. Even a
sub-optimal visual stimulus that is completely ineffective in eliciting a response such as a grating stimulus with an
orientation orthogonal to the neuron’s preferred orientation (the null orientation); Figure 1, left column) nonetheless
suppresses the response to a stimulus that optimally drives the neuron when the two stimuli are presented
simultaneously (Figure 1, top row). The normalization model accounts for this puzzling effect because suppression in
the denominator increases as a function of increasing contrast in both the preferred and non-preferred orientation
stimuli, whereas excitation in the numerator only increases with increasing contrast in the preferred orientation. While
the model's predictive successes in V1 are noteworthy, interest in divisive normalization has expanded considerably
because of its apparent ubiquity in nervous systems.
Critically, the normalization equation has since been applied successfully to a wide range of systems well
beyond mammalian V1. It has been used to characterize responses in invertebrate olfactory system (Olsen, Bhandawat,
and Wilson 2010); mammalian retina (Carandini and Heeger 2011); sub-cortical structures including the LGN (Bonin,
Mante, and Carandini 2005; Mante et al. 2005); higher visual cortical areas including V4 (Reynolds, Pasternak, and
Desimone 2000; Reynolds and Heeger 2009), area MT/V5 (Simoncelli and Heeger 1998; Rust et al. 2006), and IT
(Zoccolan, Cox, and DiCarlo 2005); and in non-visual cortical areas including auditory cortex (Rabinowitz et al. 2011).
The model has even been applied to describe certain operating characteristics of the invertebrate olfactory system.
Simply stated, the same computation is performed in multiple neural circuits across a diversity of brain regions and
species. This, we are told, seriously limits the prospects of providing a suitable mechanistic explanation of this
phenomenon. Carandini has the following to say about canonical neural computation and the relevance of underlying
mechanistic details:
Crucially, research in neural computation does not need to rest on an understanding of the underlying biophysics. Some
computations, such as thresholding, are closely related to underlying biophysical mechanisms. Others, however, such as
divisive normalization, are less likely to map one-to-one onto a biophysical circuit. These computations depend on
multiple circuits and mechanisms acting in combination, which may vary from region to region and species to species.
In this respect, they resemble a set of instructions in a computer language, which does not map uniquely onto a specific
set of transistors or serve uniquely the needs of a specific software application. (Carandini 2012, 508)
Elsewhere, Carandini and Heeger echo the same point:
[T]here seem to be many circuits and mechanisms underlying normalization and they are not necessarily the same
across species and systems. Consequently, we propose that the answer has to do with computation, not mechanism.
(Carandini and Heeger 2012, 60)
Given the apparent boldness of these claims, it is unsurprising that philosophers have started to take notice. For
example, Mazviita Chirimuuta (2014), seeking to build on considerations of this kind, develops a full-blown argument
against mechanistic analyses for these kinds of neural computations. Canonical neural computation, she says, “creates
difficulties for the mechanist project [because these] computations can have numerous biophysical realisations, and so
straightforward examination of the mechanisms underlying these computations carries little explanatory weight.”
6
(Chirimuuta 2014, 127). CNC modeling is instead supposed to embody a “distinct explanatory style”, and “[s]uch
explanations cannot be assimilated into the mechanistic framework” (Chirimuuta 2014, 127). The claim here is clear:
CNCs cannot be accommodated by the framework of mechanistic explanation, and must therefore embody a distinctive
non-mechanistic explanatory form, because these computations are associated with multiple neural circuits or
mechanisms in the brain. In other words, CNCs are multiply realized, and therefore cannot be analyzed in mechanistic
terms.
Chirimuuta
explicitly
formulates
the
key
premise
as
follows:
[The] most important characteristic of normalization as a CNC is that there is now good evidence that normalization is
implemented by numerous different biophysical mechanisms, depending on the system in question…In other words,
normalization is multiply realized. (Chirimuuta 2014, 138-139)
According to Chirimuuta, the conclusion we are supposed to draw from this premise — that the same computations
performed across multiple neural circuits, areas, and species — is that CNCs cannot be characterized properly in
mechanistic terms and instead require a distinctive, non-mechanistic explanatory framework. Chirimuuta therefore
defends what she calls a “distinctness of computation” thesis about CNCs, according to which “ there can be principled
reasons for analyzing neural systems computationally rather than mechanistically.” (2014, 139). One clear-cut way of
defending the “distinctness” thesis would be to show that CNC explanations fail to satisfy various norms on adequate
mechanistic explanations. Indeed, this is precisely what Chirimuuta sets out to do. She argues that CNC explanations
“should not be subject to mechanistic norms of explanation” (2014, 131). In particular, she identifies Kaplan and
Craver’s (2011) Model–Mechanism–Mapping (3M) principle as the central mechanistic norm from which CNC
explanations evidently depart. As a reminder, the 3M principle states:
3M: A mechanistic model of a target phenomenon explains that phenomenon to the extent that (a) the elements in the
model correspond to identifiable components, activities, and organizational features of the target mechanism that
produces, maintains, or underlies the phenomenon, and (b) the pattern of dependencies posited among these elements in
the model correspond to causal relations among the components of the target mechanism. (Kaplan 2011; Kaplan and
Craver 2011)
The objective of 3M is to consolidate some of the major requirements on explanatory mechanistic models and highlight
the primary dimensions along with mechanistic models may be assessed. Explanatory mechanistic models reveal
aspects of the causal structure of the mechanism – the component parts, their activities, and organization – responsible
for the phenomenon of interest. Mechanistic models are explanatorily adequate to the extent that they accurately and
completely describe or represent this structure.9 Chirimuuta claims that CNC models fail to satisfy 3M, and therefore
cannot be assimilated into the mechanistic framework. She characterizes the conclusion of her critical analysis as
follows10:
Along the way we have seen that the mechanists’ 3M requirement […] sit[s] oddly (to put it mildly) with the new
interpretation and status of the normalization model. For Carandini and Heeger (2012) no longer think that we should
9
Accuracy and completeness requirements on mechanistic models are implicit in 3M. It is beyond the scope of this chapter to address how the norms
of accuracy and completeness are central to assessing mechanistic explanations. For further discussion, see Craver and Darden (2013).
10
It is important to acknowledge that Chirimuuta (2014) not only develops a negative argument against embracing a mechanistic perspective on CNC
explanation, she also develops a positive characterization of what distinct kind of non-mechanistic explanatory form CNC models supposedly
encompass — what she terms an informational minimal model (or I-minimal model). Her positive view is interesting and merits consideration on its
own terms, but is beyond the scope of this chapter.
7
interpret the normalization model mechanistically, as describing components and dynamics of a neural mechanism. […]
[A]ny digging down to more mechanistic detail would simply lead us to miss the presence of CNC’s entirely, because
of their many different realizations. (140)
Here is one plausible way of summarizing the entire argument:
P1. If a given model is a mechanistic explanation, then it satisfies 3M.
P2. If a given phenomenon is multiply realizable, then a model of that phenomenon will not satisfy 3M.
P3. Canonical neural computations such as divisive normalization are multiply realizable.
P4. Models of canonical neural computations do not satisfy 3M. (Inference from P2 and P3)
C. Models of canonical neural computations are not mechanistic explanations. (Inference from P1 and
P4)
As should be obvious now, P2 is the crucial premise in the argument. P2 identifies MR as a sufficient condition for a
given model failing to satisfy 3M. There are at least two possible interpretations of P2 that are consistent with the letter
of 3M. One way of interpreting P2 is that it outlines a sufficient condition for a given model to count as a bad (or
inadequate) mechanistic explanation. Another interpretation is that is provides a sufficient condition for not counting as
a mechanistic explanation at all (good or bad). The latter interpretation seems to be the one that Chirimuuta prefers. In
what follows, I will make a case for rejecting (both interpretations of) P2 by showing how MR considerations are
irrelevant to assessing the quality of a mechanistic 0explanation.11 Once this barrier to understanding is removed, it will
be clear how CNC explanations are properly characterized as mechanistic explanations. Although Chirimuuta develops
an interesting positive proposal centered on what she terms an informational minimal model (or I-minimal model), the
focus of the present chapter is on her negative thesis.
4. Sound localization: Birds
Specific cases of CNCs are too controversial to be useful for the purposes of illustrating how MR considerations are
irrelevant to assessing the adequacy of mechanistic explanations. Instead, I want to focus on one of the most successful
explanatory models in all of systems neuroscience, the neural circuit model of auditory sound localization in the barn
owl. For reasons that will become evident shortly, this is a paradigmatic example of a mechanistic explanation.
Nevertheless, the capacity to localize airborne sounds is implemented in a diversity of neural mechanisms across
different species throughout animal kingdom. Yet, I will argue, this does not in any way compromise the explanatory
status of the model. Nor does it affect the adequacy of any other mechanistic explanations for sound localization in
different systems or species. This example helps to show how MR considerations are irrelevant to assessing mechanistic
explanations.
Many species of birds are capable of accurately localizing sounds on the basis of auditory cues alone such as
during flight in complete darkness. These animals exploit the different arrival times of a sound at the two ears (Figure
2). Although these interaural time differences (ITDs) may only be microseconds apart, birds have evolved an exquisite
11
There is another route to undermining P2 on entirely different grounds from those developed in the present chapter. This strategy involves
challenging the basic premise of multiple realization itself. In particular, following a line of argument first developed by Mundale and Bechtel (1999),
one could argue that the initial plausibility of MR claims (at the heart of the challenge from canonical neural computation) rests on a spurious
mismatch between a coarse-grained functional individuation criterion (for establishing the sameness of the computation being performed) with an
excessively fine-grained structural individuation criterion (for establishing the difference of underlying brain mechanisms). Although it will not be
pursued here, at least for the case of divisive normalization, the granularity objection appears to have teeth.
8
strategy to detect and use ITDs to localize sounds. Unraveling precisely how ITDs are computed in the brain is one of
the great success stories of modern neuroscience.
Figure 2. Sound localization in birds (A) and mammals (B). Both rely on computing the difference between the
arrival times at the two ears (ITD) to localize airborne sounds in the horizontal plane.
More than 50 years ago, the psychologist Lloyd Jeffress proposed an elegant computational model for sound
localization in the horizontal plane (Jeffress 1948). The Jeffress model involves three basic elements: (e1) converging
time- or phase-locked excitatory inputs from both ears, (e2) coincidence detectors that respond maximally when signals
arrive from each ear simultaneously, and (e3) an arrangement of delay lines with systematically varying lengths from
each ear so that different coincidence detectors encode different ITDs (Figure 3a).12 Since neural transmission delay
time is directly proportional to an axon’s length, tuning for different ITDs can be achieved by having axonal “delay
lines” of systematically varying lengths project from each ear onto different individual coincidence detectors. A final,
related detail of the model is that the set of coincidence detectors are topographically organized such that adjacent
coincidence detectors represent adjacent locations in the horizontal plane (Figure 3a, schematically represented by
grayscale coding along the sagittal plane). Strikingly, Jeffress developed the model to account for a body of human
psychophysical data on sound localization, and did so in the absence of information about the nature of the underlying
brain mechanisms.13
Remarkably, all of the major elements of the Jeffress delay line model (e1-e3) have now been confirmed in the
barn owl (Carr and Konishi 1990; Konishi 2003; Pena et al. 2001).14 Careful behavioral, anatomical, and physiological
investigations have revealed a neural circuit for computing ITDs involving delay line-based coincidence detection of
signals from the two ears. More specifically, so-called bushy cells in the left and right nucleus magnocellularis (NM)
send time-locked excitatory inputs from each ear, implementing e1 of the Jeffress model. Neurons in the nucleus
laminaris (NL), the first station of binaural processing in the avian auditory brainstem, are maximally responsive when
ipsilateral and contralateral input signals arrive simultaneously. In other words, these neurons perform coincidence
detection, implementing e2 of the model. Individual NL neurons tuned to the same characteristic frequency show
12
Temporally coincident inputs can be precisely defined in the model as occurring when the sum of acoustic and neural transmission delays
originating from one ear equals that from the other ear: Ai + Ni = Ac + Nc, where A indicates the auditory input signal, N indicates the neural
transmission delay, and subscripts i and c indicate ipsilateral and contralateral, respectively. For further discussion, see Konishi (2003).
13
As the original Jeffress model involved a set of minimally constrained mechanistic conjecture about how sound localization might be performed, it
constitutes a clear example of a how-possibly mechanistic model (Craver 2007; Kaplan 2011).
14
Although the model was originally verified in the owl, extremely similar observations have been confirmed in most other bird species (Grothe and
Pecka 2014).
9
different preferred ITDs in virtue of differences in the axonal conduction delays from each ear.15 More specifically,
since the time delay of neural conduction through an axon is directly proportional to its length, neural tuning for
different ITDs can be achieved by having axons (“delay lines”) of systematically varying lengths project from each ear
onto individual NL neurons. A matrix of NL neurons receive inputs from axons of systematically varying lengths (and
with systematically varying interaural delays) project along the length of the nucleus, implementing e3 of the Jeffress
delay line model (Figure 3a).16 Because different NL neurons have different preferred ITDs, a population code is
required to represent the entire physiologically relevant range of ITDs (Figure 3b)17, and by extension, the entirety of
horizontal auditory space.18
Figure 3. Neural computation of interaural time differences (ITDs) in birds and mammals. (A) Jeffress-type
computational mechanism observed in the nucleus laminaris (NL) of birds involving coincidence detection of
excitatory inputs from the two ears. (B) Distribution of different preferred ITDs across a population of
narrowly-tuned individual neurons in one hemispheric NL. Shaded area indicates physiologically relevant range
of ITDs. (C) Computational mechanism in mammals involving precisely timed hyperpolarizing inhibition that
adjust the timing of excitatory inputs to coincidence detector neurons in the medial superior olive (MSO). (D)
For a given frequency band, the ITD tuning of the population of MSO neurons in the left MSO is the inversion of
that of the corresponding population of neurons in the right MSO. (Source: Grothe 2003)
Critically, the neural circuit model of sound localization developed by Konishi and colleagues is a
paradigmatic example of a mechanistic explanation in computational neuroscience. Its explanatory status is beyond
reproach within the neuroscience community, and more importantly, it exhibits all the hallmarks of an adequate
15
Frequency tuning is observed in NL and in many other auditory neurons mainly because hearing begins when the cochlea mechanically filters
incoming sounds into separate frequency components. Consequently, all output signals from the cochlea are already broken down or filtered
according to their various frequencies.
16
To make the critical role played by delay lines and coincidence detection clear, it is helpful to consider how different arrangements of delays from
the two ears give rise to coincident inputs from different locations in auditory space. For example, axons of equal length projecting from each ear onto
a coincidence detector neuron in NL will have equal internal neural conduction delays (Figure 3a, central column positions); and consequently will
give rise to coincident inputs only when sounds are emitted from straight ahead and reach both cochlea at the same time (i.e., when ITD = 0). Because
these neurons fire maximally for ITDs of zero, they are said to be tuned or have a best or preferred ITD of zero. By contrast, an NL neuron receiving a
short axonal projection from the left (ipsilateral) ear and a long axonal projection from the right (contralateral) ear, will receive coincident inputs and
exhibit tuning only for sounds coming from the right auditory hemifield. This is because signals from the right ear must travel a longer path length
compared to those from the left and therefore have internal transmission delays that precisely offset the delay in the arrival time at the two ears.
Conversely, a NL neuron with a short axonal projection from the right ear and a long axon from the left ear, will receive coincident inputs and exhibit
tuning only for sounds coming from the left auditory hemifield.
17
What comprises the physiologically relevant range of ITDs is determined by factors such overall head size, and more specifically, the distance
between the ears.
18
It turns out that these neurons are also topographically organized in NL, such that neurons in adjacent positions in NL code for spatially adjacent
locations in contralateral auditory space, thereby implementing another detail of the Jeffress model (Figure 3a).
10
mechanistic explanation. In particular, it readily satisfies 3M. There is a clear mapping from elements of the model onto
all key aspects of the target mechanism in the avian auditory system. The parts, activities, and organization identified in
the model are implemented by corresponding structures, activities, and organizational features in the avian brainstem.
NL neurons function as coincidence detectors and axons from NM neurons serve as delay lines. Different individual
coincidence detector neurons in NL exhibit different response or tuning properties such that across the population the
full range of physiologically relevant ITDs are encoded. Finally, the temporal and spatial organization depicted in the
model is precisely reproduced in the avian brainstem circuitry. As described in detail above, the timing of excitatory
inputs received by individual NL coincidence detector neurons reflects the exquisite spatial organization of axons (i.e.,
component parts) that are systematically arranged such that their lengths generate neural transmission delays that
precisely offset specific ITDs. This organization is essential to their ITD tuning properties.
One way of elucidating why mechanistic models have explanatory force with respect to the phenomenon they
are used to explain appeals to the same sorts of considerations that one might plausibly appeal to for other kinds of
causal explanations. Along these lines, adequate mechanistic explanations allow us to answer a range of what-if-thingshad-been-different questions (or w-questions) just as causal explanations do (Kaplan 2011; Kaplan and Craver 2011;
Woodward 2003). Purely descriptive models may either fail to provide answers to w-questions and offer no explanation
at all, or answer only a very restricted range of w-questions and offer superficial explanations.. By contrast, deeper
mechanistic explanations afford answers to a broader range of w-questions concerning interventions on the target
mechanism than do more superficial ones. The model for sound localization in the barn owl is an example of a deeper
mechanistic explanation. The model allows us to answer a multitude of w-questions. For example, how would the
response profile of a given coincidence detector neuron in NL change if we intervened to vary the length of either the
contralateral or ipsilateral axonal delay line through which it receives its inputs? How would the neural circuit perform
if excitatory inputs from one of the two ears were completely or partially eliminated? Which individual coincidence
detector neuron in NL would respond if we artificially set the ITD value of a sound to x microseconds? And so on.
Mechanistic models explain because they deliver answers to these and other w-questions.
5. Sound localization: Mammals
For many years, it was thought that all birds and mammals perform sound localization by relying on the same Jeffresstype mechanism involving coincidence detection of excitatory inputs coming from the two ears. However, mounting
evidence now suggests that mammals compute ITDs and thereby localize sounds using a different underlying
mechanism. It is becoming increasingly clear that mammals (unlike birds) do not rely on a population of neurons with a
precise arrangement of axonal delay lines from the two ears in order to compute ITDs and perform auditory localization
(Ashida and Carr 2011; Brand et al. 2002; Grothe 2003; McAlpine and Grothe 2003; McAlpine, Jiang, and Palmer
2001; Myoga et al. 2014)19 Instead, the emerging picture attributes a major role for synaptic inhibition in the processing
of ITDs (Figure 3c). More specifically, precisely timed hyperpolarizing inhibition controls the timing of excitatory
inputs reaching binaural coincidence detector neurons in the mammalian auditory brainstem structure known medial
superior olive (MSO).20 Inhibition slows the transmission of excitatory inputs to the MSO in such a way as to precisely
offset the difference in arrival time at the two ears arising from the specific location of a sound source. This altered
temporal sensitivity of binaural neurons in the MSO provides the basis for ITD computation. The fact that excitatory
19
The synaptic inhibition model has primary been described in gerbils and guinea pigs. However, there is evidence that similar neural mechanisms for
ITD computation are at work in other mammals including cats and possibly even humans. For additional discussion, see Grothe (2003) and
Thompson et al. (2006).
20
MSO receives bilateral excitatory inputs from so-called spherical bushy cells (SBCs) in both ventral cochlear nuclei (VCNs). For additional
discussion of the neural circuit underlying sound localization in mammals, see Grothe (2003).
11
inputs from both ears would reach the MSO without any significant interaural conduction delays (and thus would
always coincide at ITDs of zero) in the absence of inhibition, clarifies its role in ITD computation. Mammalian sound
localization therefore reflects the convergence of bilateral excitatory and exquisitely timed inhibitory inputs onto
coincidence detector neurons in the MSO (Figure 3c).
The mechanism underlying ITD computation and sound localization in mammals differs from the mechanism
observed in birds in several major ways. First, even though the inhibitory mechanism involves similarly functioning
parts (i.e., neurons serving as coincidence detectors), the functional activities – tuning properties – of coincidence
detector neurons in MSO are fundamentally different from those observed in the avian auditory system. Specifically,
individual MSO neurons do not have different preferred ITDs distributed across the entire relevant range of ITDs, as is
found in birds (Figure 3b). In mammals, all MSO neurons tuned to the same frequency band within each hemispheric
MSO exhibit the same ITD tuning, and horizontal sound location is read out from the population-averaged firing rate
across two broadly tuned spatial channels – one for each hemispheric MSO (Figure 3d). What this means is that a
change in the horizontal position of a sound source will induce a specific pattern of change in population activity in one
hemisphere and an corresponding change of opposite sign in the population activity in the other hemisphere. For
example, a sound moving away from the midline (where ITD = 0), might serve to increase activity in the contralateral
MSO and correspondingly decrease activity in the ipsilateral MSO, thereby indicating that the sound source has shifted
to a more lateral position. It is therefore the relative difference between the two hemispheric channels (the relative
activity across the entire population of MSO neurons) that encodes ITD information and indicates the horizontal
location of a sound source. Because individual MSO neurons within each hemisphere carry similar information, this
argues against the local coding strategy observed in birds in which each individual neuron encodes information about
different ITDs, and instead strongly implies that a population code is used to represent ITDs (Lesica, Lingner, and
Grothe 2010; McAlpine, Jiang, and Palmer 2001).
Second, the mammalian mechanism for ITD computation involves different parts doing different things. As
indicated above, in addition to the excitatory projections originating from the left and right cochlea, MSO neurons also
receive bilateral inhibitory projections from other structures in the auditory system that are highly specialized to
preserve the fine temporal structure of auditory stimuli with high precision.21 These structures generate temporally
accurate patterns of inhibition that precisely control the timing of excitatory inputs reaching the MSO.
Third, and perhaps most obviously, certain key parts and properties of the mechanism for computing ITDs in
birds are simply missing from the inhibitory mechanism found in mammals. Specifically, there are no axons serving as
delay lines. Axon lengths are roughly equivalent such that excitatory inputs from both ears would reach the MSO
without any appreciable interaural conduction delay in the absence of inhibition, and thus always coincide at ITDs of
zero. It is only through precise inhibitory control over excitatory timing that appropriate ITD tuning is achieved.
Despite these profound differences in underlying mechanisms, the inhibitory control model similarly provides
an adequate mechanistic explanation of sound localization in mammals. Like the delay-line model, it straightforwardly
satisfies 3M. The inhibitory control model describes the parts, activities, and organizational features of the target
mechanism in the mammalian nervous system underlying ITD computation and sound localization.
6 Multiple realizability, scope, and mechanistic explanation
21
Glycinergic neurons in the medial nucleus of the trapezoid body (MNTB) provide the main source of hyperpolarizing inhibition to the mammalian
MSO. For further discussion, see Grothe (2003).
12
The upshot of the preceding discussion should now be clear. Together, the cases described above provide a strong
counterexample to P2 in Chirimuuta’s argument – the claim that if a given phenomenon is multiply realizable, then
models of that phenomenon will fail to satisfy 3M. Here we have just seen that the phenomenon of ITD computation (or
sound localization) is multiply realized, and yet both the delay line model for birds and the inhibitory control model for
mammals each individually satisfy the requirements of 3M. Both provide adequate mechanistic explanations. Indeed,
both are widely regarded as paradigmatic instances of mechanistic explanations. Hence, at least for these cases, MR
appears irrelevant to assessing their explanatory status. This result should be prima facie puzzling given the strong
parallels between these models and models of CNCs. Before returning to consider CNCs, it will help to explore a bit
further how these argumentative moves play out in the context of models of ITD computation canvassed above.
Suppose that instead of being satisfied that both the delay line model and the inhibitory control model
independently provide adequate mechanistic explanations of sound localization (because each independently satisfies
3M), an additional requirement is imposed to the effect that there must be a single, unifying explanation to cover both
birds and mammals since both perform the same computation (i.e., ITD computation). To be clear, what is being
demanded here is a wide-scope explanation that unifies or subsumes all known systems in which a given computation is
performed. Although a detailed comparison goes well beyond the purview of the current chapter, this requirement
reflects deep similarities with unificationist accounts of explanation (e.g., Kitcher 1981). The general idea behind
unificationist views is that scientific explanation is a matter of providing a unified account of a range of different
phenomena. For ease of reference, we might therefore call this additional requirement or principle the unification
principle (UP). Because no sufficiently wide-scope or unifying mechanistic explanation will be forthcoming in the case
of ITD computation – the underlying mechanisms in birds and mammals are known to differ, after all – taking UP on
board entails the rejection of the mechanistic approach in favor of some other non-mechanistic form of explanation. But
is this inference justified? The short answer is “no”.
There are two things wrong with this line of argument. First, it trades on a confusion over the proper scope of
mechanistic explanations, and subsequently, neglects the important possibility of constructing “local” or narrow-scope
mechanistic explanations. Second, it problematically assumes that more unifying models – those with wider scope –
will always be explanatorily superior to those that are less unifying. I will address each in turn.
6.1 Scope and mechanistic explanation
Attempts to impose UP reflect a deep confusion over the proper scope of mechanistic explanations. Scope concerns
how many systems or how many different kinds of systems there actually are to which a given model or generalization
applies. In the present context, UP implies that models of ITD computation cannot successfully explain unless they
have relatively wide scope – i.e., unless there is a model (mechanistic or otherwise) that uniformly captures the common
computation performed across multiple neural circuits and taxa. Critically, although it is true that wide-scope
mechanistic explanations will be unavailable if MR obtains, it does not follow that the mechanistic approach must be
abandoned (and some other explanatory scheme embraced), as is implied by P2 in Chirimuuta’s argument. This is
because wide-scope explanations are not the only type of mechanistic explanation that can be provided for neural
computations. In particular, as the cases of ITD computation canvassed above clearly indicate, perfectly adequate
mechanistic explanations can be constructed that capture some but not necessarily all instances in which the same ITD
computation is being performed. The delay line model provides an adequate mechanistic explanation even though its
scope only includes the biological taxa of birds but not mammals. Similarly, the inhibitory control model provides an
adequate mechanistic explanation even though its scope only includes mammals but not birds. These models explain
how a given computation is implemented in a specific system or type of system without covering all systems in which
13
the computation is implemented. Crucially, they show how the scope of mechanistic models can be relatively narrow or
“local” without this restricting or compromising their explanatory power. They are perfectly adequate for the range of
systems they do in fact cover. Why?
Although scope is a dimension along which mechanistic models can and do vary (Craver and Darden 2013), it
is not an appropriate norm for evaluating mechanistic explanations. This is implicit in the 3M principle. 3M clarifies
how mechanistic models are to be judged as explanatorily adequate based on the extent to which they accurately and
completely describe the causal structure of the mechanism — the component parts, their activities, and organization —
responsible for the phenomenon of interest (Kaplan 2011; Kaplan and Craver 2011). The quality of the explanation
provided does not in any way depend on how widely the model applies. Consequently, our assessments of mechanistic
explanations should be insensitive to scope.
Despite the fact that adequate mechanistic explanations can and frequently do describe highly conserved
biological or neural mechanisms (e.g., DNA or the selectivity filter mechanism in potassium channels; Doyle et al.
1998; MacKinnon et al. 1998), resulting in wide-scope mechanistic explanations, this is not required. Instead, just like
the mechanistic explanations of ITD computation described above, the scope of explanatory mechanistic models is
often highly restricted in biological and neurobiological contexts. Along these lines, Bechtel (2009) maintains that
mechanistic accounts are often “highly particularized” in that they “characterize the mechanism responsible for a
particular phenomenon in a specific biological system” (Bechtel 2009, 762). At the limit, a mechanistic model might
aim to capture only a single case (e.g., the unique mechanism responsible for producing the specific pattern of cognitive
deficits exhibited by patient H.M.; Annese et al. 2014) or a recurring mechanism found in only one species (e.g., the
exotic mechanism underlying color vision in the mantis shrimp; Thoen et al. 2014). These “highly particularized” or
scope-restricted models can nonetheless provide perfectly adequate mechanistic explanations – by satisfying 3M – even
though their application may be restricted to extremely narrow domains.
Of course, things could have turned out differently in the cases of ITD computation surveyed above due to
differences in patterns of evolution, drift, or both. Under some counterfactual evolutionary scenarios, a common circuit
design across these taxa might have resulted instead of the actual divergent mechanisms observed today. Although these
differences would invariably change the scope of our computational models, such alterations in scope are (or at any rate
should be) inconsequential to how we evaluate these models as mechanistic explanations.
6.2 Unification and explanation
This deflationary view about the relationship between model scope and explanatory power is by no means unique or
idiosyncratic to the mechanistic framework. Although the view that wide scope is not an important property of
explanatory models places the mechanistic approach at odds with traditional covering law and unificationist accounts of
explanation, interventionist approaches to causal explanation similarly reject scope as a reliable indicator of explanatory
depth or power (Hitchcock and Woodward 2003; Woodward 2003). According to the interventionist perspective,
explanatory depth reflects the extent to which a given generalization or model provides resources for answering a
greater range of what-if-things-had-been-different questions about the phenomenon of interest. This in turn is argued to
depend on how wide the range of interventions is under which a given generalization is invariant. One generalization or
model thus provides a better or “deeper” explanation than another because it is invariant under a wider range of
interventions than the other, not because it has wider scope. Hence, both mechanistic and interventionist perspectives
provide arguments for thinking that scope and explanatory power can and often do vary independently of one another.
Along similar lines, Sober (1999) rejects the assumption underlying UP, namely, that relatively abstract,
unifying explanations are always better and are to be preferred over more detailed, disunified explanations. In
14
particular, he contends that there is “no objective reason ” (p.551) to prefer unified over disunified explanations.
According to unificationists like Kitcher, explanations involving an abundance of “micro-details” (or mechanistic
details) are “less unified” than explanations that abstract from such detail because the latter and not the former can more
easily apply across a wide range of systems with different underlying realizations. Sober argues that it is the nature of
our explanatory interests that determines which type we will prefer in any given situation. In some explanatory contexts,
highly detailed, narrow-scope explanations may be better or more appropriate. In other contexts, more abstract, widescope explanations may be preferable. Crucially, there is no automatic preference or requirement for explanations
involving abstraction away from micro-detail, as is implied by UP.
7. Canonical neural computation explanations reconsidered
We now have the appropriate resources to reconsider the claim that CNCs cannot be analyzed in mechanistic terms and
therefore require a distinct explanatory framework. In particular, we can now see how arguments against the
mechanistic analyses of CNCs trade on precisely the same confusion as the one identified above concerning the proper
scope of mechanistic explanations. Specifically, these arguments start by pointing out that wide-scope mechanistic
explanations of CNCs will invariably fail in virtue of the diversity of mechanisms across different neural circuits and
species that implement the computation. From this, together with an assumed premise embodying something highly
similar to UP, a conclusion is drawn to the effect that CNCs cannot be assimilated into the mechanistic framework
(Chirimuuta 2014). However, for reasons canvassed above, this argumentative strategy assumes that only wide-scope
explanations are appropriate, and it consequently neglects the possibility of developing narrow-scope mechanistic
explanations of computational phenomena. Yet, as the discussion of ITD computation shows, narrow-scope mechanistic
explanations can in principle be provided, even in contexts where MR is operative.
After reviewing evidence that divisive normalization computations are implemented differently across
different systems, Carandini and Heeger (2012) claim that: “it is unlikely that a single mechanistic explanation will hold
across all systems and species: what seems to be common is not necessarily the biophysical mechanism but rather the
computation” (2012, 58). On a strong reading, Carandini and Heeger (2012) appear to embrace UP. Chirimuuta
reinforces this interpretation when she claims that they “no longer think that we should interpret the normalization
model mechanistically, as describing components and dynamics of a neural mechanism” (Chirimuuta 2014, 140). She
interprets them as embracing the view that unless a wide-scope explanation of divisive normalization is forthcoming, all
manner of mechanistic explanation will be out of reach. Independently of whether or not this is the correct interpretation
of Carandini and Heeger’s view22, Chirimuuta explicitly endorses this position as her own. She takes the fact that no
single mechanistic explanation can be provided that subsumes all divisive normalization phenomena – i.e., that
describes the full range of circuits and systems that perform the same (divisive normalization) computation – as
evidence that “[s]uch explanations cannot be assimilated into the mechanistic framework” (2014, 127). But this
assumption of UP is just as problematic in the context of models of CNCs as it was in the context of models of sound
localization.
As was true in the case of ITD computation, scope-restricted explanations of divisive normalization are either
22
Carandini and Heeger’s (2012) claim is compatible with both a weaker and stronger reading, only the latter of which implies a commitment to UP.
According to the weaker reading, although a single mechanistic explanation is unavailable, multiple individual mechanistic explanations may still be
available or forthcoming. On the stronger reading, the fact that no single mechanistic explanation can be given implies that no mechanistic
explanation of any kind is available or forthcoming. Although determining which reading is more closely aligned with their actual view goes beyond
the scope of this chapter, based on the broader context, they appear to endorse the weaker reading. Specifically, they repeatedly emphasize how
research into canonical neural computations such as divisive normalization involves the discovery and articulation of underlying mechanisms, stating
that “[a] key set of questions concerns the circuits and mechanisms that result in normalization” and how “[u]nderstanding these circuits and
mechanisms is fundamental” (2012, 60).
15
currently available or forthcoming. 23 For example, neuroscientists have already developed a robust narrow-scope
mechanistic explanation for how divisive normalisation computations are performed in the fruitfly (Drosophila)
olfactory system There is strong evidence that, in this specific system, normalization is implemented by GABAmediated inhibition of presynaptic connections between neurons in the fly antennal lobe (Olsen and Wilson 2008;
Olsen, Bhandawat, and Wilson 2010). Similarly robust narrow-scope explanations have been developed for divisive
normalization in the fly visual system and the mammalian retina, and the mechanisms thought to be involved differ in
important respects. Although uncertainty remains about the specific mechanism underlying divisive normalization in
mammalian primary visual cortex (V1) – the original phenomenon for which the normalization model was developed –
candidate narrow-scope mechanistic explanations of this phenomenon continue to be the subject of intense experimental
investigation in contemporary neuroscience. One plausible narrow-scope explanation posits a form of shunting
inhibition – increases in membrane conductance without the introduction of depolarising or hyperpolarizing synaptic
currents – as the mechanism underlying divisive normalization in V1 (Chance, Abbott, and Reyes 2002). Other narrowscope explanations posit mechanisms of synaptic depression (Abbott et al. 1997) or intrinsic neural noise (Carandini
2007; Finn, Priebe, and Ferster 2007). The fact that we do not at present have a complete mechanistic explanation for
this phenomenon, or that the evidence for one plausible mechanism over another remains mixed, is inconsequential.
What matters is that progress is being made, or can in principle be made, towards suitable narrow-scope mechanistic
explanations of divisive normalization phenomena and other kinds of CNCs more generally. And we should have every
reason to think that the question of how each of these neural systems performs computations will ultimately succumb to
this kind of mechanistic analysis.
Assimilating CNCs into the mechanistic framework does not require, as Chirimuuta and others have argued,
the provision of a single mechanistic explanation to cover all instances in which one and the same computation is
performed – a wide-scope mechanistic explanation – but rather only requires the provision of multiple suitable narrowscope explanations. Once recognized, the irrelevance of MR considerations becomes clear. For narrow-scope
mechanistic explanations of computational phenomena, the objective is to reveal the particular implementing
mechanism underlying a given neural computation in a specific system (or type of system), independently of whether
multiple other possible mechanisms also might have performed the computation, or whether other actual mechanisms
might be performing the same computation in other neural systems or in other species.
8. Conclusion
As the mechanistic perspective comes to occupy a dominant position in philosophical thinking about explanation across
the biological sciences including all areas of neuroscience, questions will naturally arise about its scope and limits.
Along these lines, some authors have recently appealed to modeling work involving canonical neural computations to
argue that some explanations in neuroscience fall outside the bounds of the mechanistic approach. Because these neural
computations can rely on diverse circuits and mechanisms, modeling the underlying mechanisms is supposed to be of
limited explanatory value. At its core, this is a challenge from multiple realizability. In this chapter, I argue that these
anti-mechanistic conclusions about canonical neural computation explanations are mistaken, and rest upon confusions
about the proper scope of mechanistic explanation and the relevance of multiple realizability considerations. More
specifically, I maintain that this confusion stems from a failure to appreciate how scope is not an appropriate norm on
23
For reasons that go beyond the scope of this chapter to address, the underlying mechanisms cited in these explanations of normalization are highly
similar in many respects and bear a family resemblance to one another. It seems entirely plausible to think that even though there is no single common
normalization mechanism, the fact that all these systems perform a common computational operation nevertheless places some important constraints
on the nature of the underlying mechanisms. Consequently, we might reasonably expect a family of highly similar mechanisms (and in turn,
mechanistic explanations) whenever a common computational operation is being performed across systems.
16
mechanistic explanation and therefore even mechanistic models with highly restricted scope can in principle be
explanatorily adequate. Once appreciated, it becomes clear how mechanistic models of computational phenomena can
provide legitimate explanations, even in the presence of multiple realizability.
Although there is a long history of wielding multiple realizability considerations to successfully challenge
dominant philosophical positions including mind-brain type identity theory and reductionism in biology, throughout this
chapter I have argued that such considerations does not create any insuperable obstacles for adopting a mechanistic
perspective about computational explanation in neuroscience.
References
Abbott, L. F., J. A. Varela, K. Sen, and S. B. Nelson. 1997. “Synaptic Depression and Cortical Gain Control.” Science
(New York, N.Y.) 275 (5297): 220–24.
Annese, J., Schenker-Ahmed, N. M., Bartsch, H., Maechler, P., Sheh, C., Thomas, N., … Corkin, S. (2014).
examination of patient H.M.’s brain based on histological sectioning and digital 3D reconstruction. Nature
Communications, 5, 3122.
Ashida, Go, and Catherine E. Carr. 2011. “Sound Localization: Jeffress and beyond.” Current Opinion in Neurobiology
21 (5): 745–51.
Ayala, Francisco J. 1968. “Biology as an Autonomous Science.” American Scientist, 207–21.
Batterman, Robert W. 2002. The Devil in the Details: Asymptotic Reasoning in Explanation, Reduction, and
Emergence. Oxford University Press Oxford.
Bechtel, W. 2008. Mental Mechanisms: Philosophical Perspectives on Cognitive Neuroscience. Lawrence Erlbaum.
Bechtel, William. 2009. “Generalization and Discovery by Assuming Conserved Mechanisms: Cross Species Research
on Circadian Oscillators.” Philosophy of Science, 762–73.
Bechtel, William, and Jennifer Mundale. 1999. “Multiple Realizability Revisited: Linking Cognitive and Neural
States.” Philosophy of Science, 175–207.
Bonin, Vincent, Valerio Mante, and Matteo Carandini. 2005. “The Suppressive Field of Neurons in Lateral Geniculate
Nucleus.” The Journal of Neuroscience 25 (47): 10844–56.
Brand, Antje, Oliver Behrend, Torsten Marquardt, David McAlpine, and Benedikt Grothe. 2002. “Precise Inhibition Is
Essential for Microsecond Interaural Time Difference Coding.” Nature 417 (6888): 543–47.
Carandini, Matteo. 2007. “Melting the Iceberg: Contrast Invariance in Visual Cortex.” Neuron 54 (1): 11–13.
doi:10.1016/j.neuron.2007.03.019.
———. 2012. “From Circuits to Behavior: A Bridge Too Far?” Nature Neuroscience 15 (4): 507–9.
doi:10.1038/nn.3043.
Carandini, Matteo, and David J. Heeger. 1994. “Summation and Division by Neurons in Primate Visual Cortex.”
Science 264 (5163): 1333–36.
———. 2012. “Normalization as a Canonical Neural Computation.” Nature Reviews Neuroscience 13 (1): 51–62.
Carandini, Matteo, David J. Heeger, and J. Anthony Movshon. 1997. “Linearity and Normalization in Simple Cells of
the Macaque Primary Visual Cortex.” The Journal of Neuroscience 17 (21): 8621–44.
Carr, C. E., and M. Konishi. 1990. “A Circuit for Detection of Interaural Time Differences in the Brain Stem of the
Barn Owl.” The Journal of Neuroscience 10 (10): 3227–46.
Chance, Frances S., L. F. Abbott, and Alex D. Reyes. 2002. “Gain Modulation from Background Synaptic Input.”
Neuron 35 (4): 773–82.
Chirimuuta, M. 2014. “Minimal Models and Canonical Neural Computations: The Distinctness of Computational
Explanation in Neuroscience.” Synthese 191 (2): 127–53.
Craver, Carl F., and Lindley Darden. 2013. In Search of Mechanisms: Discoveries across the Life Sciences. Chicago  ;
London: University Of Chicago Press.
Craver, C. F. Explaining the Brain. Mechanisms and the Mosaic Unity of Neuroscience 2007. New York: Oxford
University Press.
Cummins, Robert. 2000. “How Does It Work?‘ Versus’ What Are the Laws?": Two Conceptions of Psychological
Explanation.” Explanation and Cognition, 117–44.
Feest, Uljana. 2003. “Functional Analysis and the Autonomy of Psychology.” Philosophy of Science 70 (5): 937–48.
Finn, Ian M., Nicholas J. Priebe, and David Ferster. 2007. “The Emergence of Contrast-Invariant Orientation Tuning in
Simple Cells of Cat Visual Cortex.” Neuron 54 (1): 137–52. doi:10.1016/j.neuron.2007.02.029.
Fodor, Jerry. 1997. “Special Sciences: Still Autonomous after All These Years.” Noûs 31 (s11): 149–63.
Fodor, Jerry A. 1974. “Special Sciences (or: The Disunity of Science as a Working Hypothesis).” Synthese 28 (2): 97–
115.
Freeman, Tobe CB, Séverine Durand, Daniel C. Kiper, and Matteo Carandini. 2002. “Suppression without Inhibition in
Visual Cortex.” Neuron 35 (4): 759–71.
Glymour, Clark. 1970. “On Some Patterns of Reduction.” Philosophy of Science 37 (3): 340–53.
17
Grothe, Benedikt. 2003. “New Roles for Synaptic Inhibition in Sound Localization.” Nature Reviews Neuroscience 4
(7): 540–50.
Grothe, Benedikt, and Michael Pecka. 2014. “The Natural History of Sound Localization in Mammals–a Story of
Neuronal Inhibition.” Frontiers in Neural Circuits 8.
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC4181121/.
Heeger, David J. 1992. “Normalization of Cell Responses in Cat Striate Cortex.” Visual Neuroscience 9 (02): 181–97.
Hitchcock, Christopher, and James Woodward. 2003. “Explanatory Generalizations, Part II: Plumbing Explanatory
Depth.” Noûs 37 (2): 181–99.
Hoyningen-Huene, Paul. 1989. “Epistemological Reductionism in Biology: Intuitions, Explications, and Objections.” In
Reductionism and Systems Theory in the Life Sciences, 29–44. Springer.
http://link.springer.com/chapter/10.1007/978-94-009-1003-4_3.
Hubel, David H., and Torsten N. Wiesel. 1968. “Receptive Fields and Functional Architecture of Monkey Striate
Cortex.” The Journal of Physiology 195 (1): 215–43.
Jeffress, Lloyd A. 1948. “A Place Theory of Sound Localization.” Journal of Comparative and Physiological
Psychology 41 (1): 35.
Johnson-Laird, Philip N. 1983. Mental Models: Towards a Cognitive Science of Language, Inference, and
Consciousness. 6. Harvard University Press.
Kaplan, David Michael. 2015. “Explanation and Levels in Cognitive Neuroscience.” In Handbook of Neuroethics,
edited by Jens Clausen and Neil Levy, 9–29. Springer Netherlands.
http://link.springer.com/referenceworkentry/10.1007/978-94-007-4707-4_4.
Kaplan, David Michael, and Carl F. Craver. 2011. “The Explanatory Force of Dynamical and Mathematical Models in
Neuroscience: A Mechanistic Perspective*.” Philosophy of Science 78 (4): 601–27.
Kaplan, D.M. 2011. “Explanation and Description in Computational Neuroscience.” Synthese, 1–35.
Keeley, Brian L. 2000. “Shocking Lessons from Electric Fish: The Theory and Practice of Multiple Realization.”
Philosophy of Science, 444–65.
Kitcher, P. (1981). Explanatory unification. Philosophy of Science, 507–531.
Kitcher, Philip. 1984. “1953 and All That. A Tale of Two Sciences.” The Philosophical Review, 335–73.
Konishi, Masakazu. 2003. “Coding of Auditory Space.” Annual Review of Neuroscience 26 (1): 31–55.
Lesica, Nicholas A., Andrea Lingner, and Benedikt Grothe. 2010. “Population Coding of Interaural Time Differences in
Gerbils and Barn Owls.” The Journal of Neuroscience: The Official Journal of the Society for Neuroscience 30
(35): 11696–702. doi:10.1523/JNEUROSCI.0846-10.2010.
Machamer, Peter, Lindley Darden, and Carl F. Craver. 2000. “Thinking about Mechanisms.” Philosophy of Science, 1–
25.
Mante, Valerio, Robert A. Frazor, Vincent Bonin, Wilson S. Geisler, and Matteo Carandini. 2005. “Independence of
Luminance and Contrast in Natural Scenes and in the Early Visual System.” Nature Neuroscience 8 (12):
1690–97.
McAlpine, David, and Benedikt Grothe. 2003. “Sound Localization and Delay Lines–do Mammals Fit the Model?”
Trends in Neurosciences 26 (7): 347–50.
McAlpine, David, Dan Jiang, and Alan R. Palmer. 2001. “A Neural Code for Low-Frequency Sound Localization in
Mammals.” Nature Neuroscience 4 (4): 396–401.
Myoga, Michael H., Simon Lehnert, Christian Leibold, Felix Felmy, and Benedikt Grothe. 2014. “Glycinergic
Inhibition Tunes Coincidence Detection in the Auditory Brainstem.” Nature Communications 5: 3790.
doi:10.1038/ncomms4790.
Nagel, Ernest. 1961. The Structure of Science: Problems in the Logic of Scientific Explanation. Vol. 1. Harcourt, Brace
& World New York.
Olsen, Shawn R., Vikas Bhandawat, and Rachel I. Wilson. 2010. “Divisive Normalization in Olfactory Population
Codes.” Neuron 66 (2): 287–99. doi:10.1016/j.neuron.2010.04.009.
Pena, Jose Luis, Svenja Viete, Kazuo Funabiki, Kourosh Saberi, and Masakazu Konishi. 2001. “Cochlear and Neural
Delays for Coincidence Detection in Owls.” The Journal of Neuroscience 21 (23): 9455–59.
Piccinini, Gualtiero, and Oron Shagrir. 2014. “Foundations of Computational Neuroscience.” Current Opinion in
Neurobiology 25: 25–30.
Polger, Thomas W. 2009. “Evaluating the Evidence for Multiple Realization.” Synthese 167 (3): 457–72.
Putnam, Hilary. 1975. “Mind, Language and Reality (Vol. 2).” Philosophical Papers.
Rabinowitz, Neil C., Ben DB Willmore, Jan WH Schnupp, and Andrew J. King. 2011. “Contrast Gain Control in
Auditory Cortex.” Neuron 70 (6): 1178–91.
Reynolds, John H., and David J. Heeger. 2009. “The Normalization Model of Attention.” Neuron 61 (2): 168–85.
Reynolds, John H., Tatiana Pasternak, and Robert Desimone. 2000. “Attention Increases Sensitivity of V4 Neurons.”
Neuron 26 (3): 703–14.
Rosenberg, Alex. 2001. “Reductionism in a Historical Science.” Philosophy of Science, 135–63.
Rust, Nicole C., Valerio Mante, Eero P. Simoncelli, and J. Anthony Movshon. 2006. “How MT Cells Analyze the
Motion of Visual Patterns.” Nature Neuroscience 9 (11): 1421–31.
Sarkar, Sahotra. 1992. “Models of Reduction and Categories of Reductionism.” Synthese 91 (3): 167–94.
Schaffner, Kenneth F. 1967. “Approaches to Reduction.” Philosophy of Science, 137–47.
18
Shapiro, Lawrence A. 2000. “Multiple Realizations.” The Journal of Philosophy 97 (12): 635–54.
———. 2008. “How to Test for Multiple Realization.” Philosophy of Science 75 (5): 514–25.
Simoncelli, Eero P., and David J. Heeger. 1998. “A Model of Neuronal Responses in Visual Area MT.” Vision
Research 38 (5): 743–61.
Sober, Elliott. 1999. “The Multiple Realizability Argument against Reductionism.” Philosophy of Science, 542–64.
Thoen, H. H., How, M. J., Chiou, T.-H., & Marshall, J. (2014). A different form of color vision in mantis shrimp.
Science (New York, N.Y.), 343(6169), 411–413.
Thompson, Sarah K., Katharina von Kriegstein, Adenike Deane-Pratt, Torsten Marquardt, Ralf Deichmann, Timothy D.
Griffiths, and David McAlpine. 2006. “Representation of Interaural Time Delay in the Human Auditory
Midbrain.” Nature Neuroscience 9 (9): 1096–98. doi:10.1038/nn1755.
Waters, C. Kenneth. 1990. “Why the Anti-Reductionist Consensus Won’t Survive: The Case of Classical Mendelian
Genetics.” In PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association, 125–39.
JSTOR. http://www.jstor.org/stable/192698.
Woodward, J. 2000. “Explanation and Invariance in the Special Sciences.” The British Journal for the Philosophy of
Science 51 (2): 197–254.
———. 2005. Making Things Happen: A Theory of Causal Explanation. Oxford University Press, USA.
Wright, Cory D., and William Bechtel. 2006. “Mechanisms and Psychological Explanation.” Handbook of Philosophy
of Psychology and Cognitive Science, 31–79.
Zoccolan, Davide, David D. Cox, and James J. DiCarlo. 2005. “Multiple Object Response Normalization in Monkey
Inferotemporal Cortex.” The Journal of Neuroscience 25 (36): 8150–64.
19