Hostname: page-component-78c5997874-t5tsf Total loading time: 0 Render date: 2024-11-10T11:17:04.123Z Has data issue: false hasContentIssue false

Phenomenological Laws and Mechanistic Explanations

Published online by Cambridge University Press:  11 April 2023

Gabriel Siegel*
Affiliation:
Department of Philosophy, Washington University in St. Louis
Carl F. Craver
Affiliation:
Department of Philosophy, Washington University in St. Louis
*
Corresponding author: Gabriel Siegel; Email: gabriel.siegel@wustl.edu
Rights & Permissions [Opens in a new window]

Abstract

In light of recent criticisms by Woodward (2017) and Rescorla (2018), we examine the relationship between mechanistic explanation and phenomenological laws. We disambiguate several uses of the phrase “phenomenological law” and show how a mechanistic theory of explanation sorts them into those that are and are not explanatory. We also distinguish the problem of phenomenological laws from arguments about the explanatory power of purely phenomenal models, showing that Woodward and Rescorla conflate these problems. Finally, we argue that the temptation to pit mechanistic and interventionist theories of explanation against one another occludes important and scientifically relevant research questions.

Type
Article
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of the Philosophy of Science Association

Introduction

Mechanistic explanation has been a major preoccupation of recent work in philosophy of science (e.g., Machamer, Darden and Craver Reference Machamer, Darden and Craver2000; Craver Reference Craver2006, Reference Craver2007; Bechtel and Abrahamsen Reference Bechtel and Abrahamsen2005; Glennan Reference Glennan2002, Reference Glennan2017). Central to this development has been an emphasis on explanations in which the explanandum phenomenon is the behavior or property of a mechanism as a whole and the explanans involves the organized activities of the mechanism’s components. A primary driver of this work has been to build a model of explanation that adequately describes explanatory practices in the special sciences. Such theorists built on the well-known criticisms of the covering law (Hempel Reference Hempel1965) and unificationist (Kitcher Reference Kitcher, Kitcher and Salmon1989) models of explanation (by, e.g., Scriven Reference Scriven1975 and Salmon Reference Salmon1984), and built upon positive steps away from classical models of reduction by, e.g., Wimsatt (Reference Wimsatt1972, Reference Wimsatt1974), Lycan (Reference Lycan and Lycan1990), and Simon (Reference Simon1969). They sought to characterize mechanistic explanation in terms that could be given clear philosophical expositions (e.g., causation, organization, component), escaping once-prevalent suspicion that the term “mechanism” could be given no general and useful formulation (e.g., Schaffner Reference Schaffner1993).

One common criticism of mechanistic approaches is that they deny that so-called phenomenological laws, such as the ideal gas law, have explanatory power (Woodward Reference Woodward and Kaplan2017; Rescorla Reference Rescorla2018).Footnote 1 The argument is simple:

  1. 1. Mechanists hold that for a model to count as explanatory of a phenomenon P, that model must describe the mechanism for P.

  2. 2. Mechanists hold that phenomenological laws describe regularities and do not describe the mechanisms by virtue of which those regularities hold.

  3. C. Mechanists hold that phenomenological laws are not explanatory; for example, one cannot explain a helium balloon’s expansion by citing the ideal gas law and the fact that one raised the temperature of the gas inside the balloon.

According to this argument, the mechanistic view declares illegitimate a mundane and unproblematic bit of everyday scientific explanatory practice. Given that phenomenological explanations are ubiquitous in the special sciences, this counts heavily against the prized descriptive adequacy of mechanistic views.

We will argue that this argument rests on a few simple, but subtle, confusions and equivocations. Some of these confusions are exegetical; as we’ll argue, Premise 1 is false. But the tendency to believe Premise 1 traces to some deeper confusions about explanation. In particular, these critics conflate two philosophical problems: the problem of phenomenological laws and the problem of phenomenal models. This conflation, we believe, traces ultimately to a failure to digest an insight of mechanical philosophers since Salmon: that there are both etiological and constitutive aspects of explanation, each with different norms and associated practices.

The failure to mark this distinction is often a blind-spot for interventionist theories of explanation. Perhaps most fundamentally, the errors packaged in this argument derive from a mistakenly adversarial casting of the intellectual terrain: many advocates of mechanistic explanation (e.g., Craver Reference Craver2007; Kaplan Reference Kaplan2011), understood as we defend it here, did not offer their views as competitors to interventionist theories of causal explanation but rather as extensions and companions to that view.Footnote 2 Most centrally, these mechanists took themselves to be filling in the interventionist blind-spot for constitutive explanations. Once this blind-spot is addressed, a progressive research field opens in a space once crowded out by apparent conflict.

In Section 1, we consider some different ways of thinking about phenomenological laws, each with different implications for their status as explanatory. We use this to lay out a plausible formulation of the critics’ target argument, and then to respond to it. In Section 2, we consider the altogether distinct problem of phenomenal models, which arises only in the context of understanding constitutive explanations. We explain why mechanists object to phenomenal models as explanations and argue that this is a reasonable view. In Section 3, we return to Woodward (Reference Woodward and Kaplan2017) and Rescorla’s (Reference Rescorla2018) shared critique and use the conceptual solvents introduced in Sections 1 and 2 to counter their arguments. This is important for allowing key and unresolved issues on scientific explanation to take form. In Section 4, we consider two among many: Developing a theory of constitutive explanatory relevance and providing an analysis of what it means to find the appropriate level of explanation. We emphasize that a collaborative view of the relationship between interventionist and mechanistic views opens such progressive research questions. Otherwise, these questions are foreclosed by the adversarial approach our arguments are designed to dampen.

1. Phenomenological Laws

Before discussing the relationship between mechanistic explanation and phenomenological laws, we consider some of the various ways the term ‘phenomenological laws’ has been used. The different uses of the term are often connected with different traditions in the philosophy of science. Thus, the “problem” of phenomenological laws is not one problem but many. We start by describing various ways that the idea of phenomenological laws has been understood, and we discuss how, on these different understandings, the fact that the law is “phenomenological” is supposed to impact its status as an explanation.

Nancy Cartwright (Reference Cartwright1983), for example, holds that ‘phenomenological’ laws contrast with ‘theoretical’ laws. In this sense, which traces to the era of logical positivism, phenomenological laws are law statements articulated exclusively in terms describing observable features of the world. Theoretical laws, in contrast, posit theoretical entities that explain the regularities among observables. Nothing in this notion so far indicates one of the problems of phenomenological laws, as there is no general rule that bona fide explanations must posit theoretical entities. For example, the car stopped running (observable) because there was no gas in the tank (observable).

Next, consider the caricature view that phenomenological laws express relations among observations, our perceptions of worldly phenomena, rather than among observables, the objects of those perceptions. This caricature, perhaps endorsed by Mach (Reference Mach1914), really puts the “phenomenology” in “phenomenological.” But, if we restrict our attention to explanations outside psychology and other sciences that do, in fact, concern themselves with relations among perceptions, it is clear why phenomenological laws in this sense would seem explanatorily suspect.Footnote 3 Namely, the psychological registration of worldly phenomena is, in most cases, epiphenomenal to their occurrence. The car would run out of gas even if nobody looked in the tank. If one holds, reasonably, that explanations (outside of psychology, economics, etc.) cite features of the world rather than our psychological representations of those features, this positivist caricature is at least prima facie suspect.

This caricature does not, however, capture the standard examples motivating the contemporary discussion. Wesley Salmon, a key defender of causal-mechanical views of explanation, considers a child who asks why the ocean waves are creeping slowly toward her sandcastle. He opines, “A very primitive explanation might consist in informing it [the child] of the regular way in which the tides advance and recede” (Reference Salmon1998, 60). This “primitive explanation” is phenomenological in that it describes an observable regularity without revealing the laws or causes in virtue of which it holds: tides in, tides out. Unlike the positivist reading, Salmon’s phenomenological laws describe relations among worldly magnitudes, rather than among perceptions; but they are epiphenomenal with respect to the explanandum. The low tide does not cause the high tide; rather, both the low tide and the high tide are distal indicators of the causes at work (i.e., the relative position of the moon and the earth). Phenomenological laws in this sense describe mere correlations. To use another of Salmon’s examples, there is a phenomenological regularity between falling barometric readings and storms. Again, this phenomenological law fails to cite the common cause (air pressure) of both. The falling mercury is epiphenomenal with respect to storms; the storm would happen even if all barometers were pulverized. Phenomenological laws in this sense express correlations ultimately explained by factors unmentioned in the law statement.

If the phenomenological nature of phenomenological laws is understood positivistically or in terms of merely correlational regularities, then the factors cited in the phenomenological law are epiphenomenal, and so explanatorily irrelevant, to the explanandum phenomenon, as both mechanists and interventionists should insist. If you intervene ideally to prevent low tide, that is, leaving everything else as it is, without disrupting the system’s causal structure, it will make no difference to the occurrence of the high tide. Consistent with Premise 2, certain phenomenological law statements are phenomenological because they fail to describe the operative mechanisms that explain why things appear as the statement describes them. But importantly, they also fail to describe causes or constitutively relevant factors producing the explanandum phenomenon. Instead, they describe epiphenomena, correlates, relations among indicators, all causally irrelevant to (though regularly associated with) that explanandum.

For this reason, we should revise the first premise of the target argument: the mechanist need not reject the explanatory value of phenomenological laws on the grounds that they fail to describe mechanisms; rather, they might reject the explanatory power of phenomenological laws on the grounds that they fail to describe explanatorily relevant factors. And what are those?

Many mechanists, such as Craver, Kaplan (e.g., see Craver Reference Craver2007; Craver and Kaplan Reference Craver and Kaplan2020) and Glennan (Reference Glennan2017), follow Salmon Reference Salmon1984) articulated his view, explanatorily relevant features might be either parts of the causal nexus in the past light-cone of the explanandum phenomenon (the “etiological aspect” of scientific explanation) or parts of the underlying mechanism for the phenomenon (the “constitutive aspect”). A more accurate formulation of Premise 1 would therefore read:

Premise 1*. MechanistsFootnote 4 hold that for a model to count as explanatory for a phenomenon P, that model must describe either a causally relevant antecedent to the occurrence of P or a constitutively relevant feature of the mechanism underlying P.

Premise 1 exhibits a blindness toward the distinction between etiological and constitutive explanation and to the fact that several mechanists since Salmon have distinguished and acknowledged both causal (etiological) and constitutive explanations.

Consider an example. Let P be the phenomenon that the kidney regulates plasma osmolality. Etiological explanations might explain P by describing kidney development or how evolution by natural selection shaped the osmoregulatory mechanisms in the kidney. Constitutive explanations, on the other hand, explain P by describing how the activities of parts of the kidney (the glomeruli, the loop of Henle, etc.) are organized, in the here and now, such that the kidney regulates ion concentrations. These are two different ways of situating P in the causal nexus: from before and from within. Mechanists can and do embrace both.

To return to the progress made in Premise 1* (over Premise 1), consider a third sense in which some laws are said to be phenomenological, namely, that they are derivative as opposed to fundamental. Mill (Reference Mill1906) used “phenomenological” in this way. All the phenomenological laws we have considered up to this point are derivative in this sense; they are regularities that depend for their existence on explanatory facts unmentioned in the law statement. But now focus on a specific subset of derivative phenomenological laws, namely, those that are derivative but that nonetheless contain variables that refer to explanatorily relevant factors. Such laws are “phenomenological” insofar as they are derivative, but not insofar as they lack causally relevant information. They are thus not representative of one of the problems of phenomenological laws.

This is the kind of phenomenological law epitomized by the ideal gas law: pV = nRT. This equation describes a generalizable relationship between pressure (p), volume (V), temperature (T), the amount of substance (n) and the ideal gas constant (R). The ideal gas law equation can be used to represent relations among observable properties of actual gases. For example, it describes how pressure and volume change as a function of temperature. This law is famously derivative, a derivative consequence of more fundamental laws of statistical mechanics and the molecular theory of gases. It describes relations among magnitudes without describing the mechanism in virtue of which those relations hold. According to Premise 1 (versus Premise 1*), the mechanists should count that law as non-explanatory. But we need to slow down: non-explanatory of what?

Suppose we want to explain why Anna’s balloon popped. And suppose we learn that she had been heating it with a hair dryer. The ideal gas law can now do some serious explanatory work, for in this employ, it describes a causal regularity among magnitudes (rather than epiphenomena). One can defend the explanatory value of this law from the perspective of many different theories of scientific explanation (for a review, see Woody Reference Woody2013). Salmon, in contrast, rejected its explanatory value (see Salmon 1984, 136, Reference Salmon1998, 55). This is because he held that etiological explanations describe causes and, further, that causation involves mark transmission or exchanges of conserved quantities linking cause to effect in an unbroken chain of intersections in spacetime. In the case of gases, these marks and quantities are borne by the molecules composing the gas. The gas law abstracts away from these causal mechanisms, and so fails on his account to describe why these variables stand in this relation to one another.

But Salmon arguably need not have said this, even on the assumption of his contact-action theory of causation. He might rather have tried to develop an account of how higher-level causal relevance is possible in a world that is, fundamentally, the propagation and transmission of conserved quantities and contact action. This would require making temperature (and not just the motion of molecules) part of the past light-cone of the balloon’s popping. Lewis (Reference Lewis1986) does precisely this: the heating is explanatory in virtue of being a counterfactually relevant part of the causal history of the popping.

Contemporary mechanists such as Craver and Kaplan adopt neither Salmon’s nor Lewis’ approach to this problem exactly, though they are closer to Lewis’ than to Salmon’s. They embrace, instead, Woodward’s (Reference Woodward2003) theory of causal (and so explanatory) relevance (see Craver Reference Craver2007, Chapters 3 and 6; Craver and Kaplan Reference Craver and Kaplan2020; Kaplan and Craver Reference Kaplan and Craver2011; Kaplan and Bechtel Reference Kaplan and Bechtel2011). According to the interventionist view, the gas law is explanatory because you can manipulate the volume of the balloon by intervening ideally (in a sense that need not detain us) on the temperature of the gas. That’s what Anna does with her ideal hairdryer. The ideal gas law is explanatory because it provides counterfactual dependency information, answering what-if-things-had-been-different questions, i.e., ‘w-questions’ (Woodward 2003). The ideal gas law indicates roughly how volume would have been different given ideal interventions on temperature, namely, in accordance with pV = nRT.

On Salmon’s way of fleshing out the causal interpretation of this law, the gas law statement describes superficial features and leaves out the crucial causal work done by gas molecules (see Salmon Reference Salmon1998, 56). But on Woodward’s and Lewis’ way of cashing it out, there is no problem: causation need not involve contact action; omissions and double preventions can be causally relevant; causes are difference-makers. Craver, Kaplan, and Glennan all adopt this interpretation themselves, so there is no disagreement about how to handle etiological explanations that appeal to phenomenological laws.

As we’ll see in the next section, Premise 1 would appear to be true of mechanistic theories of constitutive explanation even if, as we’ve seen, it is not true of their theories of etiological explanations. Perhaps this explains Woodward’s and Rescorla’s subtle confusion. Concerning both kinds of explanation, Premise 1* is a much more accurate summary of the mechanist approach we endorse (e.g., mechanists such as Craver, Kaplan and Glennan).

2. Phenomenal Models and Mechanistic Explanations

Because the ideal gas law is a derivative law, it expresses a relationship among magnitudes that has a deeper explanation. We are thus faced with the following question: Why is the ideal gas law (approximately) true (when it is)? When we ask this question, the law describes the explanandum phenomenon, not the explanans. It is in this kind of explanatory context, where a constitutive explanation is called for, that the problem of phenomenal models arises. This problem is altogether different from the problems of phenomenological laws. Indeed, the “problem” of phenomenal models is not so much a problem as it is an apparent truism about explanation: explanandum phenomena don’t explain themselves, or more accurately, you can’t explain something merely by describing it differently. Only some redescriptions have explanatory content.

What, exactly, is a phenomenal model? Craver describes phenomenal models as, “complete black boxes; they reveal nothing about the underlying mechanisms and so merely ‘save the phenomenon’ to be explained” (Reference Craver2006, 360). He says that, “the signature of a phenomenal model is that it describes the behavior of the target system without describing the ontic structures that give rise to it (Reference Craver, Kaiser, Scholz, Plenge and Hüttemann2014, 38-39).”Footnote 5 More recently, Kaplan and Craver define phenomenal models as follows: “a model, M, is a phenomenal model of a mechanism if and only if it describes the inputs to, modulators of, and outputs from a mechanism without describing its relevant internal causal structure” (Reference Craver and Kaplan2020, 296).Footnote 6 A phenomenal model (or law, if that be preferred) is phenomenal in virtue of what it leaves out: it describes the behavior of the mechanism without describing anything about how the mechanism works. The root, “phenom,” can lead one to mistake these two, altogether distinct concepts.

To mark the divide between purely phenomenal models and constitutive explanatory models, Kaplan and Craver (Reference Kaplan and Craver2011) impose the “model-to-mechanism mapping” (3M) requirement on constitutive explanations: for the model to go beyond a purely phenomenal redescription, it must reveal some details about the underlying mechanism. This requirement does not assert (despite what some critics allege) that an explanatory model must describe all those details. Nor does it assert that a model is more explanatory to the extent that it describes more of those details (see Craver and Kaplan Reference Craver and Kaplan2020). Rather, it simply marks the divide between models that function only to fix the explanandum phenomenon and models that provide information about the constitutive explanation.

The problem of phenomenal models arises only in the context of constitutive explanations and does not arise for etiological explanations. The problem with phenomenal models in this context is that they, by definition, provide no information about how the activities of component entities are organized such that the relation specified in the explanandum holds. Why does the sleeping pill help you fall asleep? Because it has a “dormitive virtue”. How does the hippocampus encode memories? Because it has the function by which memories are encoded. These putative constitutive explanations are hollow; they are purely phenomenal in the sense that they merely redescribe the explanandum phenomenon without telling us anything about the mechanism, the nexus of causes in virtue of which the explanandum holds true.

We must be mindful here of the distinction between etiological and constitutive explanations. If one wants to know why Anna’s balloon popped, the gas law can be used as part of an etiological explanation, along with her hairdryer. But if someone asks why gases expand when heated, the ideal gas law is explanatorily impotent. It exchanges a verbal description of the thing that we want to explain for a more precise, mathematical characterization of the same phenomenon. That is a kind of intellectual progress, but it doesn’t explain why a change in temperature should yield a change in volume. In the context of constitutive explanations, phenomenal models provide no explanatory nourishment beyond that contained in the request for explanation. To take them as explanatory is to run in an explanatory circle, to explain a phenomenon by appealing to that very phenomenon.Footnote 7

To be sure, phenomenal models are not useless simpliciter. We saw above that they can be deployed in etiological explanations (e.g., the ideal gas law). But even in constitutive explanations, one can make explanatory progress by characterizing the explanandum phenomenon. For David Marr (Reference Marr1982) the primary value of computational-level descriptions in cognitive science is that they characterize precisely the cognitive operations one wants to explain; they fix the target of the explanation. One way to start the search for a mechanism (Marr thought it was the best way) is to study the phenomenon, to master the constraints on any acceptable explanation for it, and to use that model as a spring-board for conjecturing and eliminating possible and impossible explanations for it (see Bechtel and Richardson Reference Bechtel and Richardson2010; Craver and Darden Reference Craver and Darden2013; see also Shagrir Reference Shagrir2010). Phenomenal models can also help scientists and students conceptualize a phenomenon and guide their reasoning about it. As Woody argues, “by providing selective attention to certain gas properties and their relations and ignoring other aspects of actual gas phenomena, the ideal gas law effectively instructs chemists in how to think about gases as gases” (Reference Woody2013, 1574).

Return now to our target argument, this time with constitutive rather than etiological explanations in mind. Again, Premise 1 is not so much incorrect as imprecise. Premise 1* is better, though here we focus on its second disjunct, i.e., on constitutive explanations. Premise 2 is also correct; but we should be clear that we are talking about merely phenomenal models as Craver and Kaplan define them: they describe the explanandum phenomenon but, by definition, fail to include details about the mechanism by virtue of which the phenomenon holds. So, it is clear that we should reject the conclusion C. If all parties agree with the apparent truism that merely re-describing an explanandum phenomenon does not suffice to explain that phenomenon (and we have seen little opposition to that thought),Footnote 8 then the problem of phenomenal models, described by mechanists, cannot be used to support C. While mechanists hold that phenomenological laws, such as the ideal gas law, are explanatorily impotent in the context of constitutive explanations, this does not entail that they are non-explanatory simipliciter (e.g., as etiological explanations). In this way, C is transparently false.

3. Phenomenal Models in the Cognitive Neurosciences: The Mechanist’s Response

Woodward (Reference Woodward and Kaplan2017) and Rescorla (Reference Rescorla2018) especially criticize the mechanistic framework as applied to the cognitive neurosciences, a domain Craver and Kaplan emphasize in their development of the view. If the framework fails there, this is a serious problem. However, as we now explain, Woodward and Rescorla’s criticisms do not reflect an appreciation of the ideas in the previous two sections. Neither acknowledges the distinction between etiological and constitutive explanation and, as a result, each embraces the confusing ambiguity of Premise 1. Coupled with an inaccurately oppositional framing of the dialectical situation, these missteps lead to fundamental confusions about scientific explanation in these sciences and beyond.

3.1 Rescorla offers his version of the argument to support his interventionist analysis of explanation in psychology: “scientific psychology causally explains mental and behavioral outcomes by specifying how those outcomes would have been different had an intervention altered various factors, including relevant psychological states” (Reference Rescorla2018: 1920).

Rescorla uses Bayesian models of perception to exemplify the power of his approach. In general terms, the perceptual system represents distal features like shape, color, size, spatial location, etc., on the basis of proximal sensory stimulation (e.g., light on the retina). These models are Bayesian in the sense that they rely in their estimations on the prior probabilities of these features and on the likelihood of the retinal stimulation given the presence or absence of these features. We suppose with Rescorla that these models accurately describe the relationship between these proximal stimuli (inputs) and the downstream perceptual representations (outputs).

Suppose we are giving an etiological explanation of why John perceives a green circle. The Bayesian model tells us how the perceptual estimate would have been different given different proximal stimulation, different background priors, and different likelihood values (e.g., if the subject had been presented with different patterns of lighting in the past) (Reference Rescorla2018, 1926). Hence, under the interventionist criteria, Bayesian models play an explanatory role because they allow us to answer a set of w-questions: how would the estimate differ if the pattern of light on the retina were different? How would it differ if one’s life experience had resulted in different priors? The ability to answer such w-questions, as we discussed in Section 2, is the mark of the Bayesian model’s explanatory power.

So far, mechanists and interventionists can agree, given that mechanists adopt the same model of causal relevance as Rescorla (as do Craver, Kaplan and Glennan). But Rescorla goes further and argues that Bayesian models are “highly non-mechanistic” (Reference Rescorla2018, 1935) and thus are ruled illegitimate by mechanistic theories of explanation. Exemplifying our target argument, he begins with the ambiguous Premise 1, asserting that mechanists require all explanations (rather than just constitutive explanations) to describe underlying mechanisms. His Premise 2 is that Bayesian models are phenomenal models; they describe regularities and do not describe the neural mechanisms in virtue of which those regularities hold. According to Rescorla, mechanists should falsely view Bayesian models as non-explanatory; he concludes C.

Rescorla drives home his argument about Bayesian models by comparing them to the ideal gas law (IGL):

A recurring worry facing the mechanistic conception of scientific explanation is that many successful scientific explanations seem non-mechanistic. Consider the ideal gas law: pV = nRT, [IGL]…From a mechanistic perspective, [IGL] is not explanatory. A genuinely mechanistic explanation must instead deploy statistical mechanics, which describes the gas as a collection of tiny interacting particles. I acknowledge that statistical mechanics can augment the explanatory power of …[IGL]. It does not follow (and is not true) that [IGL] itself is unexplanatory. (Rescorla Reference Rescorla2018, 1914).

In short, Rescorla’s argument is a near perfect instantiation of our target argument, embodying all the elementary yet subtle errors we have now exposed.

First, ask the crucial starting question: What is being explained? Is IGL supposed to explain the popping of Anna’s balloon? Or is to supposed to explain why gases expand when heated? If the explanandum is the former, that is an etiological explanation, and IGL has considerable explanatory power. It is false that Craver, Kaplan and Glennan, for example, believe otherwise (see Premise 1*). If the explanandum is the latter, then we are dealing with a constitutive explanation, IGL is explanatorily impotent, and mechanists are correct in so judging.

These simple mechanist responses apply, mutatis mutandis, to the Bayesian models. What is being explained? Are we trying to explain why an organism makes a particular perceptual judgement in response to a stimulus? In that case, the model tells us that we should look to the proximal stimulus, to their cumulative experience as stored in their priors, to their assessments of the likelihoods, etc. That is an etiological explanation. The Bayesian model (we presume) expresses a true causal generalization about the explanandum, and mechanists see this as a legitimate causal explanation. According to mechanists like Craver and Kaplan, causal phenomenological laws can describe appropriately constrained counterfactual dependency relations between magnitudes or properties. Are we instead trying to explain why the visual system behaves approximately in accordance with Bayes’ rule? If so, the Bayesian model leaves us explanatorily malnourished. In the context of constitutive explanations, phenomenological laws are by definition mere redescriptions; they do not explain why the counterfactual dependency relations they describe hold. If we are looking for a constitutive explanation, this will require details about the underlying mechanisms. Nothing in Rescorla’s argument gives any reason to deny this.

The distinction between etiological and constitutive explanation matters, not just in philosophy, but in the science itself. Consider again Rescorla’s discussion of the success of Bayesian models of perception. As he notes, such models accurately describe counterfactual relationships between prior probabilities and perceptual estimates. If so, these models can function as part of an etiological explanation. However, a Bayesian model does not explain why such counterfactual relationships hold. That explanation might itself be supplied etiologically (in terms of development or evolution) or constitutively, in terms of how parts of the brain and other physiological systems involved are organized to instantiate that set of counterfactual relations. Indeed, scientists are beginning to understand the mechanisms that produce perceptual estimations, as described by Bayesian models.Footnote 9 For example, the capacity for calculating prior probabilities is thought to involve neural tuning to common physical features of the environment. In a recent review, de Lange et al. (Reference De Lange, Heilbron and Kok2018) note that “prior expectations (or priors) are likely to be learnt over relatively long timescales, leading them to become encoded in the tuning properties of our sensory cortices” (Reference De Lange, Heilbron and Kok2018, 766). Furthermore, they note that future research should investigate “how all the prediction signals from different sources are ultimately combined during perceptual inference…One possibility is that all regions send their predictions to a shared ‘blackboard’ that resides in the primary visual cortex, facilitating the combining of different priors” (Reference De Lange, Heilbron and Kok2018, 775).

Here, De Lange et al. are calling for mechanistic details about how components and their activities are organized to bring about the phenomena described by Bayesian models of perception. By including mechanisms of neuronal tuning in a model for the capacity of perceptual estimation, we begin to grasp how prior interactions with the physical environment influence the expectation of distal features. (We leave for now whether “tuning” is more than a promissory note, a neural dormitive virtue). In other words, this is at least a step toward a constitutive explanation of the phenomena described in Bayesian models of perception. And the subsequent development of these models shows scientists responding to an explanatory need Bayesian models themselves cannot satisfy.

Rescorla conflates two distinct forms of causal-mechanical explanation. Interventionism has often concerned itself exclusively with etiological explanations, and mechanists have tended to focus on constitutive explanations. To compare these accounts with respect to etiological explanations alone is inappropriate. This is why Premise 1 must be revised into Premise 1*. Since it is implausible that Rescorla thinks phenomenological laws explain themselves, it’s more likely that he simply has not made use of the etiological/constitutive distinction as it should be applied in the current context.

3.2 Woodward (Reference Woodward and Kaplan2017) provides a similar argument to Rescorla’s. Woodward writes:

Consider the very common use of models involving recurrent networks with auto-associative features to explain phenomena like retrieval of memories from partial cues…Such models obviously abstract away from many neural details, and in this respect are relatively non-mechanistic in Craver’s sense. On my view, however, we should not conclude that they are unexplanatory for this reason alone. Instead, their explanatory status depends on whether they accurately capture the dependency relations in real neural structures (Reference Woodward and Kaplan2017, 88-89).

Woodward’s exegetical error is analogous to Rescorla’s and traces to a basic misreading of Craver and Kaplan’s views. Woodward plausibly construes such models as providing etiological, as opposed to constitutive, explanations. It is also plausible that he thinks the models are true (enough) in the relevant respects of the actual systems (e.g., do neural mechanisms deploy auto-association?).Footnote 10 If the explanation is etiological, and if the descriptive and counterfactual claims in these models are true (enough), then mechanistic views, as we’ve discussed, should entail that these models express explanatorily relevant information.

On the other hand, if the explanation he had in mind was constitutive, and neural network models are still true (enough) descriptions of the components and processes involved in “retrieval of memories from partial cues,” then again, the mechanist view entails that the model contains explanatory information. If, however, the models are merely accurate summaries of the inputs and outputs of retrieval processes, and the components in the model bear no (or are intended to bear no) relation to actual neural mechanisms (i.e., to their parts, activities or organizational features), then it would seem everyone should agree that, within the context of constitutive explanation, they are phenomenal models with no explanatory import. In other words, if the model doesn’t provide constitutive information about the causal mechanism underpinning memory retrieval from cues, then it is merely phenomenal and so explanatorily vacuous, like a dormitive virtue.

However, if the model at least represents organizational features of the relevant memory retrieval mechanism (as a plausible reading of Woodward suggests), even if the model abstracts away from details regarding specific neural parts, then this is not a merely phenomenal model. Rather, it is a form of genuine constitutive explanation, however incomplete and unsatisfying it may be. In such a case, mechanists such as Craver and Kaplan would not dismiss the model as non-explanatory, but rather would say something about the quantity and quality of causal information the model supplies.

Woodward provides another argument that doesn’t appeal to any particular example of a phenomenological law. This involves a way to demarcate “purely descriptive or phenomenological models, and, on the other hand, explanatory models” (Reference Woodward and Kaplan2017, 90). In this case, Woodward discusses how to demarcate genuinely explanatory from non-explanatory, i.e., merely correlational or observational, models. This is indicative of some of the problems of phenomenological laws discussed in Section 1. Woodward falsely claims that mechanists use mechanistic details as the distinguishing factor (Reference Woodward and Kaplan2017, 90), i.e., that they embrace Premise 1. Woodward then provides a variety of ways in which a model might fail to be explanatory (Reference Woodward and Kaplan2017, 89-94). These are roughly ways in which the interventionist criteria for explanation are not met. Woodward summarizes:

We can explain what is explanatorily defective about such models in terms of violations of basic interventionist/dependency requirements on explanation without invoking the idea that all explanations must be mechanistic. To the extent that a model avoids the problems described under 1–5 above, and satisfies the interventionist constraints on explanation, it will count as explanatory even if it fails to be mechanistic (Reference Woodward and Kaplan2017, 94).

But, to reiterate, mechanists such as Craver and Kaplan adopt the interventionist framework for explanatory relevance. Several mechanists can and do agree that interventionist criteria supply a key notion of causal (and so explanatory) relevance and so are useful for sorting good from bad explanations (see, e.g., Craver Reference Craver2007, Chapter 3).

Any model that meets the interventionist criteria will contain counterfactual dependency information. It will thus outline part of the causal structure of the world and hence be potentially explanatory of something. But, again, of what? If the counterfactual dependency information correctly describes the relation between an effect and its cause, then such models can be used in etiological explanations of the effect. But does the model provide a constitutive explanation of why that counterfactual relationship holds? That depends on whether the model contains, in addition, details about the mechanism by which that relationship holds. If we are trying to explain a phenomenon P, and our model describes P alone, without giving any of the mechanistic details in virtue of which P occurs or any of the causes of P, then the mechanist is committed to that explanation’s explanatory impotence.

3.3 How did this confusion arise? Above we hinted at a possible historical source in Salmon Reference Salmon1984) was, as far as we know, the first to distinguish etiological and constitutive aspects of scientific explanation (unlike his contemporary, David Lewis), he dedicated the vast majority of his attention to etiological explanation and, in fact, never developed a working model of constitutive explanation.

There is good reason for that: his account of causal (and so explanatory) relevance does not apply to (indeed, cannot apply to) the constitutive explanatory relation between parts and wholes. There are general reasons to be suspicious of the idea of part-whole causal interactions, as many philosophers have emphasized (see, for example, Lewis Reference Lewis1986; Craver and Bechtel Reference Craver and Bechtel2007; Kim Reference Kim and Beckermann1992). But whatever general suspicions there are, Salmon had a particularly acute difficulty (as Craver Reference Craver2007 notes): his understanding of causation involved causal processes intersecting in space-time, exchanging either marks (Salmon Reference Salmon1984) or conserved quantities (Salmon Reference Salmon1997 ) with one another, and carrying those marks or quantities beyond the intersection point. It is at least very hard to make sense of a whole receiving conserved quantities it did not already have from its parts or to understand how a whole might come to carry a mark it did not have, though one of its parts had from the start. And given that wholes and parts are always everywhere intersecting one another in space and time, it is hard to see how they would come to intersect or continue to carry marks and conserved quantities once those intersections cease. In short: His view of causation prohibits interlevel interactions; interactions are the basis of his theory of relevance; constitutive explanation must be utterly mysterious.

Phenomenal models describe regularities without describing the billiard-ball style causal interactions that, on his account, fully constitute the explanation. So, Salmon was thus primed to assert that phenomenal models are universally explanatorily vacuous (even in etiological contexts). If one presumes incorrectly that contemporary mechanists (specifically Craver and Kaplan) follow Salmon in this regard (and indeed, some have, see Machamer Reference Machamer2004; Bogen Reference Bogen2005), one might conclude that they similarly must insist on the explanatory inadequacy of phenomenal models.

Perhaps, that is, Woodward has simply not turned the page from Salmon to contemporary mechanists, and so finds himself responding to echoes of a previous generation. Consider how Woodward’s (Reference Woodward1989) review of Salmon Reference Salmon1984) parallels his critiques of Craver thirty years later:

It is a central claim of much theory in cognitive psychology that such accounts can provide genuine explanations even though they do not describe in detail the operation of neurophysiological or biochemical mechanisms and even though similar information-processing strategies may have interestingly different neurophysiological realizations in different subjects. None of these explanations seems to explain by tracing in detail continuous causal processes or underlying physical mechanisms…I think it is at least not obvious…how Salmon can avoid the conclusion that many of the above theories are pretty dubious as explanations, in virtue of their apparent failure to specify continuous causal processes (Reference Woodward1989, 366).

Woodward argues that Salmon cannot accommodate the use of phenomenal models in cognitive psychology. Here we are using “phenomenal models” in the sense described in Section 2. Unlike Craver and Kaplan, Salmon does seem to hold that explanatory models must specify the underlying causal processes underpinning some phenomenon. However, Craver and Kaplan break with this idea and embrace an interventionist account of causation as a useful component of their view of mechanistic explanation. Woodward’s criticisms of Salmon cannot simply be transferred, mutatis mutandis, to the work of contemporary mechanists.

In summary, Woodward (Reference Woodward and Kaplan2017) and Rescorla (Reference Rescorla2018) both rely on the target argument in their criticism of mechanistic views of explanation in cognitive neuroscience. In doing so, they fail to mark the crucial distinction between etiological and constitutive explanations, perpetuating a blind-spot in interventionist theories generally concerning constitutive explanation. Given the ubiquity and import of constitutive explanations, not just in cognitive neuroscience but in science generally, this blind-spot renders interventionism incomplete as a theory of explanation in the special sciences. The tendency to cast interventionist and mechanistic theories as opponents prevents the interventionist from hearing what mechanists since Salmon have had to say about this central and neglected aspect of scientific explanation.

4. Beyond Conflict

In this section, we demonstrate the importance of recognizing that interventionism and mechanistic accounts should not be seen as competitors. To illustrate this importance, we describe two issues related to scientific explanation: the problem of constitutive explanatory relevance and the problem of finding the appropriate level of explanation. In our view, neither interventionism nor mechanism can solve these problems on their own; they require collaboration.

The adversarial framing, of mechanists pitted against interventionists, is not unique to interventionists such as Woodward and Rescorla; it has been invited by writings of certain mechanists, for example, by Machamer (Reference Machamer2004), Bogen (Reference Bogen2005), and (at times) Darden (Reference Darden2008). These mechanists describe interventionist counterfactuals as “behaviorist” input-output relations, impoverished in articulating how the input is transformed into output. Sometimes this contrast is expressed in an alliance between mechanistic theories and Anscombe’s (Reference Anscombe1971) view that “causation” is a philosophical abstraction that gains content only in diverse activities such as burning, scraping, and pushing. Whatever the merits of such substantive and singular theories of causation for thinking about discovery and about the metaphysics of causation, these notions have proved to be of limited use in articulating the idea of explanatory relevance (see Craver Reference Craver2007, Chapter 2).

In contrast with mechanists such as Machamer and Bogen, explicit appeal to interventionist counterfactuals lies at the causal heart of the theory of mechanism supplied by Craver (Reference Craver2007), Kaplan (Reference Kaplan2011), and Glennan (Reference Glennan2017) and is the basis for their discussion of higher-level explanations (see Craver Reference Craver2007, Chapter 6) and higher-level laws, such as dynamical equations (see Kaplan and Bechtel Reference Kaplan and Bechtel2011). The novelty in the work of these interventionist-mechanists lies not in the contrast with Woodward’s view but rather their efforts to supplement it with an account of constitutive explanation. Woodward’s interventionism, like its counterfactual antecedents (Lewis Reference Lewis1986), has appeared blind to constitutive explanation generally. Seen in this light, mechanists have something to offer interventionists; they are not adversaries.Footnote 11 A theory of explanation that has nothing to say about constitutive explanation simply cannot be adequate to the cognitive neurosciences.

Consider, then, some research topics that emerge once competition is abandoned for symbiosis. First, central to a theory of constitutive explanation is an account of constitutive explanatory relevance. Causal accounts of etiological explanation (including interventionists and mechanists alike) tend to identify etiological explanatory relevance with causal relevance: to be etiologically explanatorily relevant to some event is to be part of the causal history of that event (Lewis Reference Lewis1986), to be a causal process in the past-light cone of the event (Salmon Reference Salmon1984), or to be related via interventionist counterfactuals to the event (Woodward 2003). Yet when we turn our attention to constitutive explanations, in which the behavior of a whole (e.g., a gas) is explained in terms of the activities and arrangements of its parts (e.g., the molecules), one cannot straightforwardly appeal to causal relations between the explanans and the explanandum. This is because each of these authors either explicitly rejects the appropriateness of causal language for describing the relationship between parts and wholes or ignores the topic entirely.

For example, Lewis (Reference Lewis1986) explicitly rejects the possibility of causal relations between parts and wholes, claiming it “will not do” to assert that speaking the sentence causes the speaking of its first half. This is so despite the fact that Lewis includes the parts of an event in his definition of the causal history of the event and asserts that to explain is to describe a portion of its causal history. As discussed briefly above, neither of Salmon’s views of causation, the mark transmission theory nor the conserved quantity theory, can countenance part-whole causal relations.

Woodward’s interventionist theory, as currently articulated, applies to cases in which the value of one variable is explained in terms of the value of another variable. It is not equipped to discuss the relationship between a variable describing a part and an input and output regularity characteristic of the whole. To hammer this topic into the standard form of an interventionist explanation, one might suppose that one can describe a regularity (such as completing memories from partial prompts) as a variable, or perhaps as a capacity (such as memory completion from prompts, to use Woodward’s example) that can be on or off.

Now consider some possible interventionist counterfactuals relating that capacity to other variables: (i) when we ideally intervene to engage the capacity, a brain region is differentially activated in an MRI scanner; (ii) when we ideally remove brain region X, subjects can no longer engage the capacity; (iii) when we ideally stimulate brain region X, subjects report apparently complete memories. Anyone who has spent time thinking critically about findings in cognitive neuroscience knows that these counterfactuals can each be true without saying anything explanatorily relevant about the mechanisms of memory completion. (i) can be true even if the brain region is irrelevant, for example, its activation is downstream from a relevant part. (ii) can be true, but due to swelling in an area adjacent to the brain lesion. (iii) can be true due to spreading activation from brain region X to neighboring regions. Our point is not that the interventionist cannot find solutions to these problems (mechanist interventionists such as Craver have invested significant energy in that topic). Rather it is this: one cannot simply assume that the interventionist theory developed to handle etiological explanations applies without further elaboration to its constitutive cousin. There are too many true interventionist counterfactuals that are explanatorily empty.Footnote 12

One of the major advances of the mechanistic theory of explanation is the centering of this topic in discussions of explanation in the special sciences. There are now many options for understanding constitutive relevance, some of which are interventionist in character (Craver Reference Craver2007; Craver, Glennan and Povich Reference Craver, Glennan and Povich2021; Harinen Reference Harinen2018; Krickel Reference Krickel2018; Prychitko Reference Prychitko2021; Ylikoski Reference Ylikoski2013), others of which are more conceptual and make use of Woodward’s “requirement of independent fixability” (Baumgartner and Casini Reference Baumgartner and Casini2017), and others of which hope to do all the work required of a theory of constitutive relevance with suitably restricted non-interventionist counterfactuals (Couch Reference Couch2011; Harinnen Reference Harinen2018). Which of these views is ultimately successful remains a matter of active discussion.

Suppose now that an interventionist solution to this problem is viable. Another important question, of some relevance to scientists, is this: is there a correct level at which a given constitutive explanation should stop? Do all constitutive explanations bottom out in fundamental physics? Or are some explanatory problems solved at intermediate levels? Every textbook in cognitive neuroscience, even those authored and edited by die-hard cellular-molecular reductionists, recognizes the importance of neuroscientific work at multiple “levels of organization.” Is there a general answer to the question of how we decide on the right level for solving a given constitutive explanatory problem?

Interventionists have addressed an analogous problem for etiological explanation, such as the “problem of mental causation” (e.g., Heil and Mele Reference Heil and Mele1993; Kim Reference Kim1998), or biological causes, or the problem of social-level causes (Hedstrom and Ylikoski Reference Hedström and Ylikoski2010). They do so by appealing to the proportionality of a cause to its effect (Yablo Reference Yablo1992) and to the specificity and sensitivity of the cause to the effect (Woodward Reference Woodward2010). To what extent can solutions developed for thinking about the correct way to formulate an etiological explanation be retooled in the interlevel domain of constitutive explanations? This remains an open question. Yet it is a question that remains invisible so long as we refuse to acknowledge the crucial distinction between etiological and constitutive forms of explanation.

5. Conclusion

The tendency to confuse the problem of phenomenological laws with the problem of phenomenal models, as well as the failure to distinguish etiological from constitutive causal-mechanical explanations, has led to copious confusion in debates about the nature of scientific explanation. Our aim has been to expose and undo these confusions and chart a path forward. A key upshot is that the mechanistic view can be and should be (as opposed to Salmon Reference Salmon1984; Machamer Reference Machamer2004; Bogen Reference Bogen2005) viewed as a helpful companion, rather than a competitor to, the interventionist framework as an aspect of a causal-mechanical explanatory worldview. By distinguishing these problems and developing the mechanistic response to each, we have shown that the criticisms of Woodward and Rescorla are premised on simple misunderstandings of the (interventionist) causal-mechanical view and its significance. Removing this misunderstanding is the first step toward a productive synthesis that promises fecund exploration in coming years.

Acknowledgments

We thank Stuart Glennan, David Kaplan, Arnon Levy, Mark Povich, James Woodward and two anonymous referees from Philosophy of Science.

Footnotes

1 For related criticisms, also see Weiskopf (Reference Weiskopf2011), Woodward (Reference Woodward2013), Shapiro (Reference Shapiro2020), and Ross (Reference Rossforthcoming).

2 We focus on mechanists such as Craver and Kaplan because the critics explicitly name them as targets. Perhaps this argument, however, finds a better target in other expressions of the mechanistic view of explanation, such as Salmon (1984; Reference Salmon1998), Glennan (Reference Glennan2005), Machamer (Reference Machamer2004), Bogen (Reference Bogen2005), though even these authors could distinguish the metaphysics of causation and theories of explanatory causal relevance. Importantly they might (and Glennan does) conceive the latter of these as interventionist in nature.

3 We still acknowledge the importance of the logical positivism that followed on Mach’s influential perspective.

4 As one reviewer points out, mechanists such as Levy and Bechtel might reject both Premise 1 and Premise 1*. In fact, Levy (personal communication) embraces 1*, but acknowledges that there might be, in principle, ways of supplying “understanding” via explanatory representations that are neither causal or constitutive (see Levy Reference Levy, Levy and Godfrey-Smith2020). In the strictest sense, we use the term “mechanist” here to refer to theorists such as Craver, Kaplan and Glennan.

5 Also see Glennan (Reference Glennan2002, Reference Glennan2005) for a related discussion of the distinction between phenomenal and mechanistic descriptions.

6 Precisely this view is articulated in Kaplan and Hewitson’s (Reference Kaplan, Hewitson, Calzavarini and Viola2021) discussion of Bayesian models, a paper we discovered late in drafting this article. It is entirely complementary to the more abstract views expressed here.

7 Stuart Glennan (personal communication) points out that this point can be put differently by distinguishing the explanandum phenomenon rather than types of explanation: are we explaining the balloon’s popping (an event) or are we explaining why gasses expand when heated (a regularity). We express it in terms of kinds of explanations because regularities can also both be given etiological and constitutive explanations.

8 One reviewer reads Silverstein, Chemero and Bechtel as providing opposition to this claim. Anyone who supports that view must distinguish the kinds of redescription involved in the case of dormitive virtues from the explanatory kinds of redescription.

9 For a detailed example see Kaplan and Hewitson (Reference Kaplan, Hewitson, Calzavarini and Viola2021). Also, Woodward (personal communication) suggests that facts about the constitutive explanation are the initial/boundary conditions that must hold for the explanatory generalization to be true. This is correct, though there are many more initial/boundary conditions than there are mediators of the input-output relation, i.e., mechanisms. Regardless, this thought perhaps provides an important indicator of work to be done to properly integrate mechanistic approaches with the interventionist framework and vice versa.

10 Woodward confirmed these two assumptions in personal correspondence.

11 In fact, even mechanistic critics of Woodward (e.g., Machamer and Bogen) could, like Glennan (Reference Glennan2017) accept interventionism as an account of explanatory relevance.

12 Also see Craver (Reference Craver, Kaiser, Scholz, Plenge and Hüttemann2014) and Povich and Craver (Reference Povich, Craver, Glennan and Illari2017) who offer a separate reason for rejecting interlevel causation.

References

Anscombe, G. E. M. 1971. Causality and Determination: An Inaugural Lecture. CUP Archive.Google Scholar
Baumgartner, Michael, and Casini, Lorenzo. 2017. “An Abductive Theory of Constitution.” Philosophy of Science 84(2): 214233. https://doi.org/10.1086/690716 CrossRefGoogle Scholar
Bechtel, William and Abrahamsen, Adele. 2005. “Explanation: A Mechanist Alternative.” Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences 36(2): 421441. https://doi.org/10.1016/j.shpsc.2005.03.010 CrossRefGoogle ScholarPubMed
Bechtel, William and Richardson, Robert C. 2010. Discovering Complexity: Decomposition and Localization as Strategies in Scientific Research. Cambridge: MIT Press. https://doi.org/10.7551/mitpress/8328.001.0001 CrossRefGoogle Scholar
Bogen, Jim. 2005. “Regularities and Causality; Generalizations and Causal Explanations.” Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences 36(2): 397420. https://doi.org/10.1016/j.shpsc.2005.03.009 CrossRefGoogle ScholarPubMed
Cartwright, Nancy. 1983. How the Laws of Physics Lie. Oxford: OUP. https://doi.org/10.1093/0198247044.001.0001 CrossRefGoogle Scholar
Couch, Mark B. 2011. “Mechanisms and Constitutive Relevance.” Synthese 183: 375388. https://doi.org/10.1007/sl1229-011-9882-z CrossRefGoogle Scholar
Craver, Carl F. 2006. “When Mechanistic Models Explain.” Synthese 153(3): 355376. https://doi.org/10.1007/s11229-006-9097-x CrossRefGoogle Scholar
Craver, Carl F. 2007. Explaining the Brain: Mechanisms and the Mosaic Unity of Neuroscience. Oxford: OUP.10.1093/acprof:oso/9780199299317.001.0001CrossRefGoogle Scholar
Craver, Carl F. 2014. “The Ontic Account of Scientific Explanation.” In Explanation in the Special Sciences, eds. Kaiser, Marie, Scholz, Oliver, Plenge, Daniel, and Hüttemann, Andreas, 2752. Dordrecht: Springer. https://doi.org/10.1007/978-94-007-7563-3_2 CrossRefGoogle Scholar
Craver, Carl F. and Bechtel, William. 2007. “Top-Down Causation Without Top-Down Causes.” Biology & Philosophy 22(4): 547563. https://doi.org/10.1007/s10539-006-9028-8 CrossRefGoogle Scholar
Craver, Carl F. and Darden, Lindley. 2013. In Search of Mechanisms: Discoveries across the Life Sciences. Chicago: University of Chicago Press.10.7208/chicago/9780226039824.001.0001CrossRefGoogle Scholar
Craver, Carl F., Glennan, Stuart, and Povich, Mark. 2021. “Constitutive Relevance & Mutual Manipulability Revisited.” Synthese 199(3): 88078828. https://doi.org/10.1007/s11229-021-03183-8 CrossRefGoogle Scholar
Craver, Carl F. and Kaplan, David M.. 2020. “Are More Details Better? On the Norms of Completeness for Mechanistic Explanations.” The British Journal for the Philosophy of Science 71(1): 287319. https://doi.org/10.1093/bjps/axy015 CrossRefGoogle Scholar
Darden, Lindley. 2008. “Thinking Again about Biological Mechanisms.” Philosophy of Science 75(5): 958969. https://doi.org/10.1086/594538 CrossRefGoogle Scholar
De Lange, Floris P., Heilbron, Micha, & Kok, Peter. 2018. “How do Expectations Shape Perception?Trends in Cognitive Sciences 22(9): 764779. https://doi.org/10.1016/j.tics.2018.06.002 CrossRefGoogle ScholarPubMed
Glennan, Stuart. 2002. “Rethinking Mechanistic Explanation.” Philosophy of Science, 69(S3): S342S353. https://doi.org/10.1086/341857 CrossRefGoogle Scholar
Glennan, Stuart. 2005. “Modeling Mechanisms.” Studies in History and Philosophy of Science Part C: Studies in History and Philosophy of Biological and Biomedical Sciences 36(2): 443464. https://doi.org/10.1016/j.shpsc.2005.03.011 CrossRefGoogle ScholarPubMed
Glennan, Stuart. 2017. The New Mechanical Philosophy. Oxford: OUP.10.1093/oso/9780198779711.001.0001CrossRefGoogle Scholar
Harinen, Totte. 2018. “Mutual Manipulability and Causal Inbetweenness.” Synthese 195(1): 3554. https://doi.org/10.1007/s11229-014-0564-5 CrossRefGoogle Scholar
Hedström, Peter, and Ylikoski, Petri. 2010. “Causal Mechanisms in the Social Sciences.” Annual Review of Sociology 36: 4967. https://doi.org/10.1146/annurev.soc.012809.102632 CrossRefGoogle Scholar
Heil, John, & Mele, Alfred. 1993. Mental Causation. Oxford: Clarendon Press.10.1093/oso/9780198239291.001.0001CrossRefGoogle Scholar
Hempel, Carl G. 1965. Aspects of Scientific Explanation. New York: Free Press.Google Scholar
Kaplan, David M. 2011. “Explanation and Description in Computational Neuroscience.” Synthese 183(3): 339373. https://doi.org/10.1007/s11229-011-9970-0 CrossRefGoogle Scholar
Kaplan, David M. and Bechtel, William. 2011. “Dynamical Models: An Alternative or Complement to Mechanistic Explanations?Topics in Cognitive Science 3(2): 438444. https://doi.org/10.1111/j.1756-8765.2011.01147.x CrossRefGoogle ScholarPubMed
Kaplan, David M. and Craver, Carl F.. 2011. “The Explanatory Force of Dynamical and Mathematical Models in Neuroscience: A Mechanistic Perspective.” Philosophy of Science 78(4): 601627. https://doi.org/10.1086/661755 CrossRefGoogle Scholar
Kaplan, David M. and Hewitson, Christopher L.. 2021. “Modelling Bayesian Computation in the Brain: Unification, Explanation, and Constraints.” In Neural Mechanisms, eds. Calzavarini, Fabrizio and Viola, Marco, 1133. Cham: Springer. https://doi.org/10.1007/978-3-030-54092-0_2 CrossRefGoogle Scholar
Kim, Jaegwon. 1992. “‘Downward Causation’ in Emergentism and Nonreductive Physicalism.” In Emergence or Reduction?, ed. Beckermann, Ansgar, 119138. De Gruyter.Google ScholarPubMed
Kim, Jaegwon. 1998. Mind in a Physical World: An Essay on the Mind-Body Problem and Mental Causation. Cambridge: MIT Press. https://doi.org/10.7551/mitpress/4629.001.0001 CrossRefGoogle Scholar
Kitcher, Philip. 1989. “Explanatory Unification and the Causal Structure of the World.” In Scientific Explanation, eds. Kitcher, Philip, Salmon, Wesley, 410505. Minneapolis: University of Minnesota Press.Google Scholar
Krickel, Beate. 2018. “Saving the Mutual Manipulability Account of Constitutive Relevance.” Studies in History and Philosophy of Science Part A 68: 5867. https://doi.org/10.1016/j.shpsa.2018.01.003 CrossRefGoogle ScholarPubMed
Levy, Arnon. 2020. “Metaphor and Scientific Explanation.” In The Scientific Imagination, eds. Levy, Arnon, Godfrey-Smith, Peter, 281303. Oxford: OUP.10.1093/oso/9780190212308.001.0001CrossRefGoogle Scholar
Lewis, David K. 1986. “Causal Explanation.” In Philosophical Papers: Volume II. Oxford: OUP.Google Scholar
Lycan, William G. 1990. “The Continuity of Levels of Nature.” In Mind and Cognition: A Reader, ed. Lycan, William, 7796. Cambridge: Blackwell.Google Scholar
Mach, Ernst. 1914. The Analysis of Sensations, and the Relation of the Physical to the Psychical. Open Court Publishing Company.Google Scholar
Machamer, Peter. 2004. “Activities and Causation: The Metaphysics and Epistemology of Mechanisms.” International Studies in the Philosophy of Science 18(1): 2739. https://doi.org/10.1080/02698590412331289242 CrossRefGoogle Scholar
Machamer, Peter, Darden, Lindley, and Craver, Carl F.. 2000. “Thinking about Mechanisms.” Philosophy of Science 67(1): 125.10.1086/392759CrossRefGoogle Scholar
Marr, David. 1982. Vision. Cambridge: MIT press.Google Scholar
Mill, John S. 1906. A System of Logic, Ratiocinative and Inductive. Longmans Green.Google Scholar
Povich, Mark and Craver, Carl F.. 2017. “Mechanistic Levels, Reduction, and Emergence.” In The Routledge Handbook of Mechanisms and Mechanical Philosophy, eds. Glennan, Stuart and Illari, Phyllis, 185197. Routledge.10.4324/9781315731544-14CrossRefGoogle Scholar
Prychitko, Emily. 2021. “The Causal Situationist Account of Constitutive Relevance.” Synthese 198(2): 18291843. https://doi.org/10.1007/s11229-019-02170-4 CrossRefGoogle Scholar
Rescorla, Michael. 2018. “An Interventionist Approach to Psychological Explanation.” Synthese 195(5): 19091940. https://doi.org/10.1007/s11229-017-1553-2 CrossRefGoogle Scholar
Ross, Lauren N. Forthcoming. “Cascade versus Mechanism: The Diversity of Causal Structure in Science.” British Journal for the Philosophy of Science. https://doi.org/10.1086/723623 Google Scholar
Salmon, Wesley C. 1984. Scientific Explanation and the Causal Structure of the World. Princeton: Princeton University Press.Google Scholar
Salmon, Wesley C. 1997. “Causality and Explanation: A Reply to Two Critiques.” Philosophy of Science 64(3): 461–477.10.1086/392561CrossRefGoogle Scholar
Salmon, Wesley C. 1998. Causality and Explanation. Oxford: OUP.10.1093/0195108647.001.0001CrossRefGoogle Scholar
Schaffner, K. F. 1993. Discovery and Explanation in Biology and Medicine. Chicago: University of Chicago Press.Google Scholar
Scriven, Michael. 1975. “Causation as Explanation.” Nous 3–16. https://doi.org/10.2307/2214338 CrossRefGoogle Scholar
Shagrir, Oron. 2010. “Marr on Computational-Level Theories.” Philosophy of Science 77(4): 477500. https://doi.org/10.1086/656005 CrossRefGoogle Scholar
Shapiro, Lawrence A. 2020. “Mechanism or Bust? Explanation in Psychology.” The British Journal for the Philosophy of Science 68(4): 10371059. https://doi.org/10.1093/bjps/axv062 CrossRefGoogle Scholar
Simon, Herbert A. 1969. The Sciences of the Artificial. Cambridge: MIT Press.Google Scholar
Weiskopf, Daniel A. 2011. “Models and Mechanisms in Psychological Explanation.” Synthese 183(3): 313338. https://doi.org/10.1007/s11229-011-9958-9 CrossRefGoogle Scholar
Wimsatt, William C. 1972. “Complexity and Organization.” PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association, 67–86. https://doi.org/10.1086/psaprocbienmeetp.1972.3698961 CrossRefGoogle Scholar
Wimsatt, William C. 1974. “Reductive Explanation: A Functional Account.” PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association, 671–710. https://doi.org/10.1086/psaprocaswbienmeetp.1974.495833 CrossRefGoogle Scholar
Woodward, James. 1989. “The Causal Mechanical Model of Explanation.” Minnesota Studies in the Philosophy of Science 13: 357383.Google Scholar
Woodward, James. 2003. Making Things Happen: A Theory of Causal Explanation. Oxford: OUP.Google Scholar
Woodward, James. 2010. “Causation in Biology: Stability, Specificity, and the Choice of Levels of Explanation.” Biology & Philosophy 25: 287318.10.1007/s10539-010-9200-zCrossRefGoogle Scholar
Woodward, James. 2013. “Mechanistic Explanation: Its Scope and Limits.” In Aristotelian Society Supplementary Volume 87(1): 3965. Oxford: Blackwell Publishing.10.1111/j.1467-8349.2013.00219.xCrossRefGoogle Scholar
Woodward, James. 2017. “Explanation in Neurobiology.” In Explanation and Integration in Mind and Brain Science, ed. Kaplan, David M., 70100. Oxford: OUP. https://doi.org/10.1093/oso/9780199685509.003.0004 Google Scholar
Woody, Andrea I. 2013. “How is the Ideal Gas Law Explanatory?Science & Education 22(7): 15631580. https://doi.org/10.1007/s11191-011-9424-6 CrossRefGoogle Scholar
Yablo, Stephen. 1992. “Mental Causation.” The Philosophical Review 101(2): 245280. https://doi.org/10.2307/2185535 CrossRefGoogle Scholar
Ylikoski, Petri. 2013. “Causal and Constitutive Explanation Compared.” Erkenntnis 78(2): 277297. https://doi.org/10.1007/s10670-013-9513-9 CrossRefGoogle Scholar