This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
the evidence supports p just in case (a) p ε A, (b) p is consistent with E, and (c) for any q ε A different from p, q is inconsistent with E. (Remark: for cases that involve probabilistic answers, the notion of consistency has to be replaced by a statistical analog.) Now there are some pure cases in the practice of science that look just like this;
92
Philosophy of Science Matters
that’s what has fueled the abiding interest among scientists in the notion of a crucial experiment, and it’s a large part of what has made Karl Popper’s views so attractive to researchers (Popper 1959). In general, however, matters are more complicated. The simple version of eliminativism yields a different conception of evidential support (one that is allied to the idea of increase in probability, against which Achinstein argues). I suggest that this should be seen in explicitly comparative terms: in a context of judgment, the evidence favors p over q just in case (a) p, q ε A, and (b) p is consistent with E and q is inconsistent with E. Obviously, when the evidence supports p, it favors p over each other member of A, but in cases of incomplete elimination— when more than one member of A is consistent with E—no statement is supported, although some statements may be favored over others. In a context of judgment, a selection procedure is deficient if there’s some p ε A consistent with the total available evidence such that it’s guaranteed that p will continue to be consistent with any statements added through the selection procedure. A selection procedure is ideal if it is guaranteed to yield statements that will be inconsistent with all except (possibly) one of the members of A that are currently consistent with the total available evidence. Unfortunately, simple eliminativism is too simple. The trouble is that the majority of scientific contexts are ones in which the available evidence is inconsistent with all the answers. Sometimes the solution is to expand the set of answers, to think of a possibility nobody has considered before. Often, however, what needs to be done is to revise statements one was previously inclined to accept, showing how there’s a modification E* of E that (a) doesn’t flout the rules for accepting statements on the basis of observation and (b) preserves the explanatory and predictive successes of E. Call such a modification of E an acceptable modification. Now if we allow for acceptable modifications to support hypotheses in cases where we start from inconsistency with all the answers, we ought also to permit “rescues” by acceptable modification when one or several answers are consistent with the available evidence. The accounts of support and favoring have to be revised to accommodate this. Improved eliminativism must thus be developed along the following lines. In a context of judgmentthe evidence supports p just in case (a) p ε A, (b) p is consistent with E or with some acceptable modification E* of E, and (c) for all q in A distinct from p, and for every acceptable modification E* of E, q is inconsistent both with E and E*. Similarly the evidence favors p over q just in case (a) p, q ε A, (b) p is consistent with E or some acceptable modification of E, and (c) q is inconsistent both with E and any acceptable modification of E.
On the Very Idea of a Theory of Evidence
93
But this is still problematic. Scientists are not always in a position to specify an acceptable modification of E that will satisfy the (b) clauses, and they’re almost never in a position to survey all the acceptable modifications of E and show that the (c) clauses obtain. Often their predicament is that they have specified an acceptable modification of E for which some of the previous inconsistencies have been resolved, and for which there’s a possibility that the remaining inconsistencies might be tackled by further modification; on the other hand, all the acceptable modifications of which they (and their rivals) have been able to think don’t manage to resolve the inconsistencies for alternative answers. Thus a standard strategy of scientific argumentation is to identify ways of solving what have appeared to be puzzles for your favored view and to multiply the difficulties for your rival. Any number of claims about evidential support in the sciences exhibit this strategy. Back now to my four examples. I’ve tried to show in previous work (Kitcher 1993, 263–272) how Darwin’s research on biogeography proceeds by a complex eliminative argument; effectively Darwin eliminates doubts about his hypothesis in terms of descent with modification by showing how it’s acceptable to amend commonsensical views about the transportation of organisms, and thereby remove some of the sources of inconsistency. Darwin got rid of a lot of the trouble, but, even after his detailed work, there were still puzzles to be mopped up later (some of them only resolved after the acceptance of continental drift, others that still remain). The track record of puzzle removal gives reason to believe that the appropriate (b) clause is satisfied; the track record of Creationist biogeography suggests that the main rival is still as badly off as it was in 1859. At first sight, one might think that the case of Hertz shows simple elimination at work. But this would be a mistake. Hertz does consider the possibility that cathode rays consist of charged particles and that the result of his experiment might still be null. The problem is that he doesn’t consider the right modification of the total available evidence—the one Thomson offered in 1897. It’s not that Hertz is lazy, or unimaginative, or irresponsible in any way; he’s unlucky in that a rather subtle possibility doesn’t occur to him. The molecular biologist of my third example knows that something is wrong with the way she’s setting up the problem. She wants to know a good combination of enzymes for tackling her favorite organism. So she considers various generalizations of the form “such-and-such combination is good for so-and-so kinds of organisms.” The conflicting reports she gets teach her that there’s a finer grain of classification that needs to be introduced, but she doesn’t yet know what that is. If she were interested in the abstract problem, then she could explore various
94
Philosophy of Science Matters
possibilities. But that’s not really what she’s after. Since her goal is to obtain a good set of DNA fragments, she goes to work on the basis of whatever seems most promising, prepared to tinker, and even to try alternatives. The student of example 4 is in an equally problematic context. She knows that not all of the reports about Wolfram she’s received can be authoritative, but she doesn’t have a way of deciding which ones to trust and which to throw out. All she can do is take the plunge. (I suspect that situations like this, in which one just has to forge ahead, are quite common in the practice of the sciences. This may by no means be a bad thing; for if investigators are led to make different choices—whether because of differences in temperament, the appeal of social rewards, or whatever—the community inquiry may benefit from their diverse explorations; I’ve explored this conception of “social learning” in Kitcher 1993, 303–389).
4. So how would the deans respond to this? I think (hope) they’d feel that the account I’ve sketched stays closer to the phenomenology of scientific judgment than philosophers typically manage. But I also think they’d be skeptical about how illuminating it is. Once again, the deans speak: “Okay, this is a reasonable description of what we do; but everything clear and precise you’ve told us we already knew. The idea of inconsistency is pretty obvious (at least when probabilistic hypotheses aren’t involved, and you’ve ducked that issue). What’s hard, and where we might look for advice in a theory of evidence, is in understanding how contexts of judgment are properly set up, how we decide what an “acceptable modification” is, and, above all, how we properly make the judgments that there’s a way of eliminating the inconsistencies that beset one answer and no way of doing the same for its rivals, given that you admit, as you should, that we can never solve all the problems. That’s where we want a theory. Come back when you have one.” I agree. Is any such theory possible? Let me suggest an analogy. The elements of tactics in chess—pins, forks, skewers, and so forth—are very well understood. We’re unlikely, I think, to improve our account of them, to arrive at a deeper and more precise theory. By the same token, I think we have a clear view of the tactics of scientific reasoning; if eliminativism is right, it’s fundamentally a matter of spotting inconsistencies, and we’re not going to get much more by way of a theory of that. The trouble in both cases lies in understanding strategy—the sort of thing that grandmasters understand about space, pawn structure, weaknesses of various kinds
On the Very Idea of a Theory of Evidence
95
and so forth, and the sort of thing scientists understand about when inconsistencies really matter and when they can be trusted to get sorted out in due time. Studies of chess strategy are by no means as precise as the explanations of tactics. I’m not sure that we philosophers can produce much by way of precise theory about the strategy of scientific argument. What we can manage, on occasion, is analogous to the analyses of games that expert chess players are able to provide: that is, a reconstruction of the considerations that have led a particular scientific community to a judgment. Doing that is both non-trivial and potentially valuable. Think, for example, of controversies that occupy the public: the IQ debate, the disputes over evolutionary theory, the claims made on behalf of human sociobiology or evolutionary psychology. It would be wonderful if we had an illuminating reconstruction of the scientific consensus on anthropogenic climate change. I contend that providing insights of the kind just mentioned requires no formal account of evidence. Yet, for reasons that might move both my imagined deans, a formal theory, if we could get it, would be a good thing to have. The Book of Evidence develops a theory that provides intricate and interesting solutions to problems that have worried philosophers for several decades. For that it deserves to be celebrated. But I don’t think it answers the dean’s challenge. To do so would require a very different kind of theory. I wish I knew how to provide it. REFERENCES Achinstein, P. 2001. The Book of Evidence. New York: Oxford University Press. Buchwald, J. Z. 1994. The creation of scientific effects: Heinrich Hertz and electric waves. Chicago: University of Chicago Press. Darwin, C. 1964. The Origin of Species. Cambridge, Mass.: Harvard University Press. Kitcher, P. 1993. The Advancement of Science. New York: Oxford University Press. Kuhn, T. 1962. The Structure of Scientific Revolutions. Chicago: University of Chicago Press. Popper, K. 1959. The Logic of Scientific Discovery. New York: Basic Books. Wolfram, S. 2002. A New Kind of Science. Champaign, Ill.: Wolfram Media.
8 Mill on the Hypothetical Method: A Discussion of Achinstein’s Defense of Mill and Newton on Induction Frederick M. Kronz
1. INTRODUCTION Following Newton and Mill, Peter Achinstein maintains in Evidence, Explanation, and Realism that there are universal rules of induction (Achinstein 2010). He also maintains that such rules may be formalized, but that their validity cannot be determined formally, as in deduction. More precisely, he maintains that inductive inferences are warranted by material facts or empirical assumptions, in contrast with deductive inferences in logic and mathematics, which can be evaluated by formal means alone. These elements of Achinstein’s view are provisionally accepted. Two additional components of his view, that induction and hypotheticodeduction are mutually exclusive and that induction is the core of the scientific method, are not. The corresponding themes defended here are that induction and hypothetico-deduction are mutually complementary, that one is not subsidiary to the other, and that the scientific context determines which of the two modes is most appropriate. In doing so, I argue that neither Mill nor Newton (following Mill, I focus on Newton’s actual practice of science) is really averse to the use of hypotheses in science, contrary to what Achinstein suggests. I also provide a new formal characterization of hypothetico-deduction that incorporates key insights of Mill, Whewell, and others. This development is in support of Achinstein’s general approach. The upshot is that Achinstein should regard both hypothetico-deduction and induction (in the narrow sense, as characterized by Newton and Mill) as distinct universal and abstract rules (or sets of rules) of induction (in the broad sense, which includes any empirically defensible mode of non-deductive reasoning).
96
Mill on the Hypothetical Method
97
2. ACHINSTEIN ON INDUCTION Achinstein (2010) defends the inductivist views of Newton and Mill against those of Whewell. He focuses on the rules that each uses to characterize induction: Newton’s Rules for the Study of Natural Philosophy and Mill’s Deductive Method, which essentially involves his Four Methods of Experimental Inquiry. Achinstein’s take on the inductivists is presented later in this section after I briefly discuss Mill’s Methods. Since Mill acknowledges an essential (though subsidiary) role for the Hypothetical Method and sees clear parallels between it and his Deductive Method, they are presented together in the next section. Mill’s Four Methods of Experimental Inquiry are his Method of Agreement, Method of Difference, Method of Residues, and Method of Concomitant Variation (Mill 1879, 278–92).1 The first two are the core methods. In presenting the Four Methods, Mill puts forth Five Canons of Reason; the third canon combines the Method of Agreement and the Method of Difference to form a hybrid method that is often characterized in the literature as a fifth method (referred to by Mill as the Indirect Method of Difference, and as the Joint Method of Agreement and Difference). The Method of Residues and the Method of Concomitant Variation are also derivative with respect to the first two. In presenting these methods, he uses capital letters ABC (short for A, B, C, …) to denote antecedent conditions and small letters abc (short for a, b, c, …) to denote consequent conditions. He also distinguishes causal laws (laws of cause and effect) from other laws, such as phenomenological laws and accidental generalizations. The ultimate goal in using the methods (according to Mill) is to obtain causal laws; he regards the Method of Difference to be of paramount importance for attaining that goal. It is unnecessary to elaborate further on Mill’s Methods for what follows.2 In developing his interpretation of Newton’s and Mill’s rules, Achinstein claims that they are contrary to other methodologies, particularlyWhewell’s method, which involves the use of hypotheses. However, there is very little in Newton’s Rules of Reasoning per se and nothing in Mill’s characterization of the Deductive Method that precludes the use of hypotheses in science. It is necessary to look elsewhere in their writings in order to make a determination as to their respective views. As it turns out, their expressed views diverge drastically: Newton expresses adamant opposition to any use of hypotheses in science, whereas Mill regards the Hypothetical Method as essential to science. But expressed views can be misleading. As Mill is careful to point out, one must distinguish between what Newton says about hypotheses and what he does with them in
98
Philosophy of Science Matters
practicing science. Indeed, Mill makes a good case (presented below) that Newton does not deprive himself of the use of hypotheses in the Principia and elsewhere. In the Third Book of his famous treatise, Newton puts forth four methodological rules. The fourth is the only one that has direct bearing on the role of hypotheses in science. Rule 4: In experimental philosophy we are to look upon propositions inferred by general induction from phenomena as accurately or very nearly true, notwithstanding any contrary hypotheses that may be imagined, till such time as other phenomena occur, by which they may either be made more accurate, or liable to exceptions. (Newton 1999, 943)
It gives substantial epistemic status to propositions that are inferred by general induction from phenomena and dictates against any attempt to use hypotheses to undermine that status. However, it says nothing about the use of hypotheses in other situations. For example, it does not preclude their use to account for phenomena not amenable to general induction nor does it preclude their use to enhance the epistemic status of propositions obtained via an induction from phenomena. Thus, Achinstein overstates the case concerning Newton’s Rules; they really do not exclude the Hypothetical Method. Although Newton’s Rules do not dictate broadly against the use of hypotheses in science, he does issue much more general declarations elsewhere that do. His clearest and best known statement to that effect is the following quotation from the General Scholium, which immediately follows his characterization of his Four Rules; it is known as Newton’s Hypothesis non fingo (a Latin expression meaning “I do not feign hypotheses”). I have not as yet been able to discover the reason for these properties of gravity from phenomena, and I do not feign hypotheses. For whatever is not deduced from the phenomena must be called a hypothesis; and hypotheses, whether metaphysical or physical, or based on occult qualities, or mechanical, have no place in experimental philosophy. In this philosophy particular propositions are inferred from the phenomena, and afterwards rendered general by induction. (Newton 1999, 943)
The declaration above is not a corollary to Newton’s Rules; it is a separate claim. One could elevate it to the status of a Rule and thereby reject the Hypothetical Method, the approach ascribed to Achinstein. Or, one could follow Mill and adopt a mitigated stance towards Newton’s declaration and accept the Hypothetical Method, which is the approach followed here. Interpreting Mill as an advocate of the Hypothetical Method has ramifications for interpreting the Mill–Whewell debate. Achinstein characterizes it as a debate about which of two mutually incompatible methods
Mill on the Hypothetical Method
99
(induction or hypothetico-deductivism) should be used by scientist. A more subtle interpretation is proposed here, that they regard the two methods as mutually complementary while giving prominence to one over the other. Mill regards the Hypothetical Method as subsidiary to Induction, while Whewell does the reverse.3 The upshot for Achinstein is that he should interpret Mill as promoting both methods. Moreover, he should regard each of them as corresponding to a universal, abstract rule of induction (in the broad sense). It is also recommended that neither method be regarded as subordinate to the other (contrary to the contrasting views of Mill and Whewell). Both methods are needed to do science, and it is the scientific context that determines which is appropriate. This is an attractive view, and meshes well with the fundamental structure of deductive logic, a system of independent and mutually complementary rules. It is reasonable to suppose, following Achinstein, that the analogy with deductive logic breaks down over validity.
3. MILL ON THE DEDUCTIVE METHOD AND THE HYPOTHETICAL METHOD Mill characterizes the Deductive Method in Section 3.11.1: The mode of investigation which . . . remains to us as the main source of the knowledge we possess or can acquire respecting the conditions and laws of recurrence of the more complex phenomena, is called, in its most general expression, the Deductive Method, and consists of three operations—the first, one of direct induction; the second, of ratiocination; the third, of verification. (Mill 1879, 325)
He focuses on causal laws in elaborating on the three components of the Deductive Method. A direct induction is an inference to a law governing a separate cause that shares in bringing about an effect; his Four Methods are used in making such inferences. The second stage essentially involves deductive reasoning; it consists in calculating how the separate causes (from the first stage) operate in combinations. The third is a comparison of the calculated effects of combined causes (from the second stage) with known phenomena or the empirical laws known to characterize those phenomena. Mill notes that the Hypothetical Method has fairly close counterparts to the second and third stages of his Deductive Method; once a hypothesis is formulated, one deduces observational consequences from it and then compares them with the phenomena. They differ drastically on the first stage; in the Hypothetical Method one conjectures a hypothesis
100
Philosophy of Science Matters
whereas in the Deductive Method one infers a general law from observations. The Deductive Method does not preclude the use of hypotheses. Mill says that it is the main source of scientific knowledge; he does not say nor does he imply that it is the only legitimate method for science. There is nothing in what Mill says in characterizing the method that precludes the use of hypotheses in science. Indeed, not only is it consonant with the passage cited above for hypotheses to play an essential though subsidiary role with respect to the Deductive Method, but that is precisely the view that he explicitly puts forth several chapters later (in Section 3.14.5). This function, however, of hypotheses, is one which must be reckoned absolutely indispensable in science. When Newton said, "Hypotheses non fingo," he did not mean that he deprived himself of the facilities of investigation afforded by assuming in the first instance what he hoped ultimately to be able to prove. Without such assumptions, science could never have attained its present state: they are necessary steps in the progress to something more certain; and nearly everything which is now theory was once hypothesis. (Mill 1879, 353)
Mill characterizes this indispensible function in the paragraph immediately preceding this passage: A hypothesis can serve to suggest observations and experiments that later serve to provide independent evidence for that hypothesis, if such evidence is obtainable. The passage itself is striking for several reasons. It provides a drastically mitigated interpretation of Newton’s claims about the role of hypotheses in science, and in doing so suggests that Newton really did make use of hypotheses in doing science (despite what he says in his famous declaration). It also provides a sufficient reason for regarding the Hypothetical Method as complementary to (rather than incompatible with) the Deductive Method, given that science could not have obtained its present state without it. The passage under discussion might be taken to suggest that Mill actually regards the Hypothetical Method as being on a par with, rather than subsidiary to, the Deductive Method, but that would clash with the passage cited earlier in which he says that the Deductive Method is the “main source” of scientific knowledge. This apparent tension is substantially mitigated by his assertion that the proper use of the Hypothetical Method must include a key element of the Deductive Method, the Method of Difference (broadly construed—see Section 2 of this essay above), for it to work properly. An elaboration of that point follows. In Section 3.14.4 (the section preceding the passage cited just above), Mill explains that Newton uses the Hypothetical Method when he conjectures that there is a force that instantaneously deflects a planet from its
Mill on the Hypothetical Method
101
rectilinear path at each instant and tends directly toward the sun so that the planet sweeps equal areas in equal times in its curved path about the sun. He suggests that Newton proceeds similarly when he conjectures further that the force varies inversely as the square of the distance in order to deduce Kepler’s other two laws (that planets move in elliptical orbits and the harmonic law). He also makes a crucially important point that contrasts with Newton’s expressed views and serves to distinguish his methodological views from Whewell’s with regard to the Hypothetical Method: namely, that Newton uses the Method of Difference to legitimize his use of the Hypothetical Method and thereby to “fulfill the conditions of a complete induction.” Now, the Hypothetical Method suppresses the first of the three steps [of the Deductive Method], the induction to ascertain the law, and contents itself with the other two operations, ratiocination and verification, the law which is reasoned from being assumed instead of proved. This process may evidently be legitimate on one supposition, namely, if the nature of the case be such that the final step, the verification, shall amount to and fulfill the conditions of a complete induction . . . It is thus perfectly possible, and indeed is a very common occurrence, that what was an hypothesis at the beginning of the inquiry, becomes a proved law of nature before its close. But in order that this should happen, we must be able, either by deduction or experiment, to obtain both the instances which the Method of Difference requires. (Mill 1879, 350–1)
The Method of Difference requires positive instances (such as ABC and abc) and negative instances (such as Ā BC and ābc). Mill explains that Newton uses it to show the centripetal (center-directed) character of gravitational forces; Newton provides positive instances by showing that centripetal forces yield Kepler’s first law (equal areas swept out in equal times), and negative instances by showing that non-centripetal forces violate the law.4 Mill explains that Newton also does so in showing the inverse square character of gravitational forces; he shows that Kepler’s second and third laws hold for inverse square forces (positive instances) but not for non-inverse square forces (inferred negative instances). In the last paragraph of Section 3.14.4 Mill suggests that he has effectively provided (via his Method of Difference) an explication of Newton’s notion of a vera causa, a Latin term that means true cause. Newton uses this term in his first rule of reasoning, but he never indicates what serves to promote a causal hypothesis to a vera causa despite his having used multiple evidentiary modes in supporting causal hypotheses. In any case, the upshot of the paragraph at issue seems to be that when there is a demonstrated difference for a causal hypothesis, that demonstration
102
Philosophy of Science Matters
constitutes a proof that the hypothetical cause is a true cause (according to Mill, that is). In particular, Newton’s demonstrated differences for his law of gravitation show gravity to be a true cause. It is worth mentioning that knowing that an antecedent condition is a true cause and knowing how to describe the causal relationship mathematically are not sufficient for understanding how the cause operates. Newton makes it very clear that he does not understand how it is that one body instantaneously influences another at a distance, and asserts that he will “feign no hypothesis” in that regard. The section of Mill’s logic under discussion (Section 3.14.4) is a crucial one, though it is rather obscure. For example, he introduces the distinction between a real cause and a fictitious one, and then puts forth the claim that there are real causes for which the associated laws are merely supposed, and fictitious causes for which the associated phenomena are characterized by known laws. A bit later he mentions another distinction between “already ascertained” causes and unknown causes, but he does not indicate what it means for a cause to be “already ascertained.” It is not clear whether that means known or real or something else, and if known whether that means observed or inferred. A careful exegesis of the associated text cannot be provided here. For the purposes of this section of the paper the general thrust will suffice, which seems to be that one must sufficiently articulate the precise mode of dependence of factors in a hypothetical law in order to be in a position to use the Method of Difference either to prove or to disprove the law and, in the case of hypothetical causal laws, to show whether the associated hypothetical cause is a true cause. It appears, then, to be a condition of the most genuinely scientific hypothesis, that it be not destined always to remain an hypothesis, but be of such a nature as to be either proved or disproved by comparison with observed facts. This condition is fulfilled when the effect is already known to depend on the very cause supposed and the hypothesis relates only to the precise mode of dependence; the law of variation of the effect according to the variations in the quantity or in the relations of the cause. (Mill 1879, 352)
In contrast with Mill, Whewell explicates Newton’s notion of a vera causa using his notion of a Consilience of Inductions. In explicating Newton’s first rule he says the following. When the explanation of two kinds of phenomena, distinct, and not apparently connected, leads to the same cause, such a coincidence does give reality to the cause, which it has not while it merely accounts for those appearances which suggested the supposition. This coincidence of propositions inferred from separate classes of facts, is exactly what we noticed . . . as one of the most
Mill on the Hypothetical Method
103
decisive characteristics of a true theory, under the name of Consilience of Inductions. (Whewell 1860, 190–1)
In the text immediately following this passage, Whewell points to Newton’s use of the inverse square force to explain both planetary orbits and the moon’s orbit, and to explain the precession of the equinoxes, as concrete examples of Newton’s use of Consilience (since these distinct phenomena were at the time not apparently connected). It is remarkable that Mill focuses on the Method of Difference to the exclusion of Consilience and that Whewell does the reverse. That is to say, each was aware of the other’s preferred evidentiary mode, but was unwilling to give it serious consideration. This is especially puzzling given that each thought of his respective evidentiary mode as providing an operational characterization of Newton’s notion of a vera causa. In that regard, the view suggested here is that they each overstated the case;5 an argument to that effect is developed elsewhere (Kronz 2011).
4. A PROVISIONAL FORMALIZATION OF THE ENHANCED HYPOTHETICAL METHOD In light of the discussion above, it is reasonable to regard the Hypothetical Method as a general methodological rule that is complementary to the notion of Induction characterized by Achinstein. Following Achinstein’s inspiring claim that methodological rules are general and formalizable, a provisional formulation of the Enhanced Hypothetical Method is presented below that synthesizes key insights of Mill, Whewell, Charles S. Peirce, Karl Popper, Thomas Kuhn, and others concerning the effective use of hypotheses in science. The Hypothetical Method is often presented in a simple form. It is named here the Naïve Hypothetical Method.
Naïve Hypothetical Method 1. Conjecture a hypothesis as a solution to a problem or as a possible explanation of some puzzling phenomena. 2. Deduce empirical consequences from the hypothesis together with other empirically justifiable assumptions. 3. Conduct observations or experiments to determine whether the empirical consequences obtain. 4. Make an inductive inference to the hypothesis, provided that the empirical consequences obtain.
104
Philosophy of Science Matters
The Naïve Hypothetical Method does not bring into play any of the subtleties discussed above. As a result, it is susceptible to simple-minded counterexamples, a case in point being Russell’s involving the hypothesis “Pigs have wings.” That hypothesis together with the well-known fact that some pigs are good to eat entails the empirical consequence that some winged things are good to eat. But, it is unreasonable to regard that consequence as providing even a modicum of evidence for the hypothesis. Clearly, other considerations must be involved before step 4 is epistemically justified. As noted above, both Mill and Whewell require the satisfaction of an additional condition for the effective use of the Hypothetical Method. Consequently, it is reasonable to suppose that they would advocate replacing step 4 with the following: 4'. Make an inductive inference to the hypothesis, provided that the empirical consequences obtain and provided that there is X.
For Mill, X = a demonstrated difference (an effective use of his Method of Difference, broadly construed). For Whewell, X = a Consilience of Inductions. It is not necessary to determine whether either is correct. A better approach (the one adopted in this essay) is to suppose each is partially so, the upshot being that X = a demonstrated difference or a Consilience of Inductions or . . . (the ellipsis here corresponds to a disjunction of some other suitable set of conditions). Newton’s use of the Hypothetical Method supports this view; he sometimes used the Method of Difference and sometimes a Consilience of Inductions, as noted above. In light of the considerations in this section and related considerations in previous sections, a more sophisticated but provisional formalization of the Hypothetical Method is put forth. It is referred to here as the
Enhanced Hypothetical Method 1. Conjecture a hypothesis as a solution to a problem or as a possible explanation of some puzzling phenomena. 2. Deduce empirical consequences from the hypothesis together with other empirically justifiable assumptions. 3. Conduct observations or experiments to determine whether the empirical consequences obtain. 4. Make an inductive inference to the hypothesis, provided that the empirical consequences occur and provided that there is a. a consilience of inductions (in Whewell’s sense), or b. a demonstrated difference (in Mill’s sense), or c. a novel prediction (in Peirce’s sense),6 or d. a new organization of facts that facilitates solving a problem,7 or
Mill on the Hypothetical Method
105
e. a suitable condition (other than those already mentioned) is satisfied.8 5. If the empirical consequences do not obtain, then revise the hypothesis or revise an auxiliary assumption used to derive the empirical consequences, and then proceed as before (starting from step 2).9 The formal characterization above should be regarded as provisional and partial; a brief consideration of each of stage will suffice to show why. With regard to the first, the form of the hypothesis requires elaboration. Both general and statistical hypotheses are used in science, so it would be useful to characterize a range of formal features of such hypotheses. With regard to the reason for the conjecture, although it is often characterized in terms of its serving to solve a problem or to explain some puzzling phenomena, there may be other scientific reasons for making conjectures, and these should be included in the formulation. The nature of the act of conjecture also requires further characterization. Although some (such as Popper) have maintained that this stage is not inferential and merely involves making an educated guess or an intuitive leap, others (such as Peirce) maintain that this step involves a type of explanatory inference. In that case, the form of the inference and the underlying notion of explanation need to be explicated. In addition, it is possible that other formal modes are sometimes involved in the act of conjecture, such as an underlying process of pattern recognition; such modes would need to be appropriately formalized. The second step should be supplemented by an explication of what constitutes an empirically justifiable assumption. The third might appear to involve a straightforward comparison, and sometimes it does, but it is just as often a delicate matter especially when statistical hypotheses are involved. In some cases it might be better to use Neyman-Pearson statistical methods and in others Bayesian methods or something else; it would be worth providing formal criteria for distinguishing such cases to the extent that is possible. The fourth step should be elaborated with regard to the nature of the inductive inference, such as whether it constitutes an increase in the probability that the hypothesis is true (or approximately true), or perhaps crosses a threshold making it more likely than not that it is (approximately) true; moreover, the nature of that inference very likely depends on which conditions in step 4 are met. There is also the very interesting question as to what specific conditions alluded to in 4e could serve to “validate the induction” (notably weaker than what Mill intended in his desire to “fulfill the conditions of a complete induction”). With regard to step 5, one representative issue that needs to be addressed is the well-known Duhem-Quine thesis. Imre Lakatos (1970) has presented some provocative suggestions involving the notions of progressive
106
Philosophy of Science Matters
and degenerative problem shifts that counter that thesis to some extent, but his overarching view is not regarded as viable. Most of the issues mentioned in connection with each of the five stages identified above have been discussed in the literature (in many cases rather extensively). They are far from settled and further discussion of them is beyond the scope of this essay.
5. CONCLUDING REMARKS Achinstein interprets the Whewell–Mill debates as a dispute over which of two incompatible modes of inference, hypothetico-deduction and induction, is right for science. He sides with Mill and defends the view that there are universal rules of induction. He regards the rules put forth by Newton and by Mill as exemplary, and he maintains that they exclude the hypothetico-deductive method. But Achinstein fails to appreciate that their rules really do not exclude that method, and that neither Mill nor Newton (at least on Mill’s interpretation of Newton, which is based on Newton’s actual practice of science) is truly averse to its use in science. Furthermore, the Whewell–Mill debate is really a dispute about which of two compatible modes of inference, hypothetico-deduction and induction, is subsidiary to the other. Rather than side with Whewell or Mill, an alternative approach is advocated above: They are regarded as mutually complementary rather than hierarchically related, and it is the scientific context that determines which of the two (if either) is appropriate. Consonant with this approach and with Achinstein’s view that there are general methodological rules that govern the practice of science, a new version of the hypothetico-deductive method, the Enhanced Hypothetical Method, is provisionally formulated. It synthesizes key insights (put forth by Whewell, Mill, Peirce, Popper, Kuhn, and others) concerning the effective use of hypotheses in science.
ACKNOWLEDGMENTS The views expressed are those of the author. They do not represent the position of the National Science Foundation (NSF). No endorsement by NSF should be inferred. The author wishes to acknowledge the helpful comments and suggestions by Leslie Kronz and Melissa Jacquart, which served to improve an earlier draft of this essay, and the substantial editorial assistance of Greg Morgan in producing the final version.
Mill on the Hypothetical Method
107
REFERENCES Achinstein, P. 2010. Evidence, Explanation, and Realism. New York: Oxford University Press. Dewey, J. 1991. The Later Works, 1925–1953: Logic: The Theory of Inquiry, ed. J. A. Boydston. Carbondale: Southern Illinois University Press. Ducheyne, S. 2008. J. S. Mill’s Canons of Induction: From True Causes to Provisional Ones. History and Philosophy of Logic 29 (4): 361–76. Kronz, F. M. 2011. Scientific Method. In Leadership in Science and Technology, ed. W. S. Bainbridge. Thousand Oaks, CA: Sage. Kuhn, T. 1977. The Essential Tension: Selected Studies in Scientific Tradition and Change. Chicago: University of Chicago Press. Lakatos, I. 1970. Methodology of Scientific Research Programs. In Criticism and the Growth of Knowledge, ed. I. Lakatos and A. Musgrave. Cambridge: Cambridge University Press. Mackie, J. L. 1967. Mill’s Methods of Induction. In The Encyclopedia of Philosophy, vol. 5, ed. P. Edwards. New York: MacMillan. Mill, J. S. 1879. A System of Logic, 8th edition. New York: Harper & Brothers. Newton, I. 1999. Philosophiae Naturalis Principia Mathematica, 3rd edition. Trans. I. B. Cohen and A. Whitman. Berkeley: University of California Press. Peirce, C. S. 1998. The First Rule of Logic. In The Essential Peirce, vol. 2, ed. N. Houser, A. De Tienne, C. L. Clark, D. B. Davis, J. R. Eller and A. C. Lewis. Bloomington: Indiana University Press. Popper, K. 1968. The Logic of Scientific Discovery. New York: Harper Torchbooks. Russell, B. 1939. “Dewey’s New ‘Logic’” in Paul Arthur Schilpp ed. The Philosophy of John Dewey. New York: Tudor. Skyrms, B. 2000. Choice and Chance: An Introduction to Inductive Logic, 4th edition. Belmont, Calif.: Wadsworth Publishing Company. Whewell, W. 1860. On the Philosophy of Discovery. London: John W. Parker and Son. ——— . 1858. The Philosophy of the Inductive Sciences, 3rd edition. London: John W. Parker and Son. ——— . 1849. On Induction, with Especial Reference to Mr. J. Stuart Mill’s System of Logic. London: John W. Parker, West Strand.
NOTES 1. Section 3.8.1 denotes Book 3, Chapter 8, Section 1. 2. For further discussion of these methods, see Skyrms 2000, Chapter 5 or Ducheyne 2008. For further elaboration including additional variants and hybrids, see Mackie 1967. 3. Whewell has little to say about Mill’s Deductive Method per se, though he does argue that Mill’s Methods play only a secondary role in science (Whewell 1849, 43–54). He also notes other inductive methods aside from Mill’s (Whewell 1858, 202–32). 4. In each case, the negative instances are inferred rather than observed, which means (as Mill suggests) that Newton actually used a generalized form of the Indirect Method of Differences.
108
Philosophy of Science Matters
5. As noted by Peirce (1998, 56), it is best to regard any scientific hypothesis as provisionally accepted rather than as true no matter how well supported it might be (in other words, despite the strongest modes of evidentiary support that might be brought to bear in support of its truth), since it is always possible that an alternative hypothesis will be developed at some later time that will supersede the first. 6. That is, the predictions need not be future events, rather they need only to be derived antecedently to the investigator’s knowledge of their truth. 7. This criterion is suggested by Dewey (1991, 60). 8. Other criteria might include the five mentioned by Kuhn (1977) in connection with competing hypotheses; they are discussed briefly below. 9. A good starting place for developing criteria along these lines is Lakatos’s (1970) elaboration of Popper’s views on falsification (Popper 1968).
9 Waves, Particles, Independent Tests, and the Limits of Inductivism* Larry Laudan
In this paper, I shall be taking exception to a few of the ideas in Particles and Waves by Peter Achinstein (1991). In some of these cases, I think he is flat wrong. But that does not diminish in the least my admiration for his book, which is, in my judgment, the best extended piece of work we have on the epistemological problems posed by nineteenth-century physics.
1. INTRODUCTION Three decades ago, I told a story about the development of nineteenthcentury optics and empiricist epistemology (Laudan 1981). It went roughly as follows: the inductivist epistemology that became popular in the philosophical aftermath of Isaac Newton made it methodologically precarious to postulate theoretical entities, especially if those entities had properties unlike those of observed objects. This, in turn, meant that a variety of theories—including the wave theory of light—were to receive a hostile reception at the hands of many eighteenth-century empiricists and those natural philosophers heavily influenced by them. After all, theories that postulated the existence of elastic, imperceptible, and imponderable fluids were not the sorts of beliefs that an eighteenth-century empiricist could happily countenance. It was the moral of my story that, before such theories could begin to enjoy a wide acceptance, changes had to occur in prevailing methodological standards. Specifically, as I put it then, a shift was needed away from a narrow inductivism and toward a recognition of the merits of the method of hypothesis. Such a shift would enable fluid theorists to argue for their theories by pointing to their explanatory and predictive resources, even if the entities they postulated were quite
109
110
Philosophy of Science Matters
beyond the reach of ordinary observation and inductive generalization from the observable. As I showed, this hypothetico-deductive sort of inference (which I will subsequently call H-D), from a confirmation of consequences to the probability of the theory itself, carried no weight among traditional empiricists such as Bacon, Newton, Hume, or Reid. When Newton said “hypotheses non fingo,” it was this sort of inference he was repudiating. I suggested that it was thus no accident that the revivified wave theory of light and the method of hypothetico-deduction gained ascendancy at about the same point in the nineteenth century. The wave theorists needed the method of hypothesis to justify their approach, and the successes of the wave theory managed, in turn, to constitute vivid examples of the scientific fruits of hypothetico-deduction. I claimed that this linkage between the advocacy of ethereal fluids and anti-inductivism explains how, for instance, that principal advocate of the method of hypothesis in the first half of the nineteenth century, William Whewell, became one of the leading spokesmen for the wave theory. At about the same time that I was doing this research, a British historian of physics—Geoffrey Cantor—was coming to a complementary conclusion from a different direction (Cantor 1983). He had been studying the writings of such corpuscularians as Brougham and was struck by how heavily their criticism of the wave theory was imbued with inductivist language. Like me, he came to the conclusion that there was a close connection in the Enlightenment and early nineteenth-century science between where one stood on the wave/particle question and what theory of scientific method one espoused. Such a happy consilience of perspectives convinced Cantor and me that we were right, of course, especially as there were few demurrals to be heard from other scholars for more than a decade. Or that was this case until Peter Achinstein published his extremely interesting book, Particles and Waves.1 In that book, he has a different story to tell about this episode. According to Achinstein, there was no major methodological divide separating the corpuscularians from the undulationists. Methodological consensus happily prevailed among the physical scientists of the period. Still worse, at least as far as Cantor and I were concerned, the consensus that Achinstein detects was an agreement that induction, not the method of hypothesis, is the appropriate epistemology for science. Achinstein devotes two lengthy chapters of his important book to developing a probabilistic, quasi-Bayesian analysis of the episode, purporting to show that optical theorists in the early nineteenth century were all, at least implicitly, Bayesian conditionalizers. Now, I cannot speak for Cantor, but I want to say for my part that I think that Achinstein’s analysis has—on this particular point—got both the philosophy and the history wrong. It will be the purpose of my remarks today to
Waves, Particles, Independent Tests, and the Limits of Inductivism
111
motivate that reaction. For those of you who are saying to yourselves, “Who cares how the nineteenth-century light debates went?” I will try in passing, although this obviously cannot be my main concern today, to draw out some lessons from this episode for debates in contemporary philosophy of science. Several points about the historical record are uncontested. Let me begin with a summary of those: through much of the eighteenth century, Huygens’ wave theory of light was eclipsed by Newton’s corpuscular theory, not least because it seemed that Huygens could not explain the rectilinear propagation of light. At the turn of the nineteenth century, Thomas Young attempted to revive the wave theory using it to explain phenomena of diffraction and optical interference such as the colors of thin films and diffraction. Young’s theory in turn failed to be able to account for polarization. Then Fresnel came up with a kinematic model that conceived light as a transverse vibration transmitted in an elastic ethereal fluid. This enabled him to explain polarization and double refraction and to predict a number of surprising phenomena, including the famous bright spot at the center of a shadow cast by a disk. During the early 1830s, Cauchy developed a dynamical wave theoretic model that explained dispersion as well. After intense debate among physicists in the 1820s and early 1830s, most scientists had come to accept the superiority of the wave theory by the late 1830s, although a few hold-outs persisted for another generation. So much for the common ground. What is in dispute here, to put it in its most general terms, is this: what sorts of epistemic virtues led to the triumph of the wave theory? In very brief compass, the Laudan reply was this: the wave theory made a series of surprising predictions that turned out to be right and for which there were no counterparts in the corpuscular theory. It also explained a broader range of phenomena of diverse types without resorting to ad hoc adaptations.2 In sum, it solved more empirical problems than its corpuscularian rival and did so with less ad hocery. These, however, are virtues from an H-D perspective, not from an inductivist one. Achinstein’s answer—to which I shall turn in a moment—is, in brief: the wave theorists managed to show that the corpuscular theory had a vanishingly small probability and this created, by a kind of method of exclusion, a presumption that the probability of the wave theory was close to 1. Such positive confirmations and surprising successful predictions as the wave theory enjoyed merely reinforced this conclusion; they were, Achinstein insists, insufficient to motivate it. In other words, Achinstein denies that the ability of the wave theory to explain and predict a broad range of empirical phenomena, many of them surprising, did, or even in principle could have done, much to enhance its credibility. In what follows, I shall sketch out the two stories in more detail and indicate why I remain skeptical about Achinstein’s version.
112
Philosophy of Science Matters
2. THE ACHINSTEIN ACCOUNT It is important to note at the outset that Achinstein’s analysis is simultaneously working at two levels, the normative and the descriptive. What drives his normative analysis is a conviction that a Bayesian theory of evidence and testing is the only sound one. Descriptively, Achinstein is concerned to show that the participants in the light debates of the early nineteenth century were in fact making appraisals in accordance with Bayesian recipes. My principal concern here will be with the descriptive adequacy of Achinstein’s rational reconstruction rather than with its normative underpinnings. But if, as I expect to show, the Bayesian story falls short of being able to capture the reasoning of the agents involved, then it will be appropriate to ask, time allowing toward the end of my comments (or perhaps in the discussion to follow), whether Bayesianism could possibly capture the character of the scientific use of evidence in cases like this one. Achinstein’s rational reconstruction of the episode goes as follows. The wave theorists, he says, adopted a four-step strategy: 1. Start with the assumption that light is either a wave phenomenon or a stream of particles. 2. Show how each theory explains various optical phenomena. 3. Show that the particle theory, in explaining one or more of the observed phenomena, introduces improbable hypotheses while the wave theory does not. 4. Conclude that the wave theory is (very probably) true, because the particle theory is (very probably) false. Achinstein then proceeds to offer slightly more formal characterizations of these four steps. Step 1, he says, is tantamount to asserting that, relative to certain observations O and background knowledge b, (1) p(T1 or T2 / O & b) ~ 1(~ means ²is close to²)
where T1 is the wave theory and T2 is the particle theory. Step 3 above amounts to the claim that the particle theorists had recourse to certain auxiliary assumptions, h, such that although the auxiliaries are very plausible given the corpuscular theory, namely, although (3a) p(h /T2 & O & b) = 1,
the fact is that there is strong evidence against the truth of those auxiliaries, namely,
Waves, Particles, Independent Tests, and the Limits of Inductivism
113
(3b) p(h / O & b) = 0.
A quick application of Bayes’s theorem to (3a) and (3b) yields the result that (3c) p(T2 / O & b) = 0.
Combining (3c) with (1), Achinstein infers that the wave theory is very probably true, namely, (4) p(T1 / O & b) = 0.
Such then, in schematic form, is Achinstein’s proposed reconstruction of the arguments of the wave theorists. But what about step 2 and the comparison of the wave theory with optical phenomena? One might have thought that the single most important bit of evidence in appraising the wave theory was an examination of how it fared against the phenomena. But this process of checking the empirical consequences of the wave theory, according to Achinstein, can—even if the hypothesis stands up successfully to these tests—do little to enhance the probability of the wave theory. As he sees it, Bayesian inference insists that predictions and “explanations (no matter how numerous or varied) do not suffice to give an hypothesis high probability” (Achinstein 1991, 135). All such successes can do is to “ensure that the hypothesis retains whatever probability it has on other data” (135). This is quite a remarkable claim. It would certainly have come as a shock to many of the wave theorists who were impressed by the ability of the wave theory to make surprising predictions successfully and to explain many puzzling features of light. To someone like Whewell, who made a point of underscoring such successes of the wave theory, Achinstein’s rebuff is quick. An explanatory strategy of the sort advocated by Whewell and other supporters of the method of hypothesis will not be enough to guarantee high probability for h, no matter how many phenomena h explains, even if consilience and coherence . . . are satisfied. (137)
On Achinstein’s view, the only thing that can give the wave theory, or any other theory, a high probability is the demonstration that the probability of the disjunction of all its rivals is vanishingly small. Discrediting rivals is the only significant way of enhancing the probability of one’s pet hypothesis. A good offense it seems is not only the best defense; it is the only defense. The eliminative refutation of rivals is, for Achinstein, the only significant way of enhancing the credibility of a theory.
114
Philosophy of Science Matters
I think that his analysis is flawed both conceptually and contextually. I will now try to show why. First, and more briefly, philosophically:
(1) The conceptual problem There is a crucial equivocation at the beginning of Achinstein’s characterization of the problem facing theorists of light in the early nineteenth century. The wave theorists, on his reconstruction, began with the assumption that light is either a wave or a particle. That was itself fairly controversial since, even if we ignore possible theories and limit ourselves to then extant theories, there were theories that saw light as a fluid—very like heat was conceived in the 1820s—a hydrodynamic conception that is, strictly speaking, neither particulate nor undular. But leave that reservation to one side. Let us suppose that they and Achinstein were right in thinking that light almost certainly was either a wave or a particle. The equivocation I have in mind comes in the move from the claim that (0) p(light is a wave or light is a particle) = 1
to the claim that Achinstein needs for his reconstruction, namely, that the probability of a specific theory of light is close to 1 (viz., thesis [4]). The fact is that, even if it could be settled with certainty that light is not a corpuscle, and even if it could be inferred therefrom that light is almost certainly a wave-like phenomenon, it manifestly would not follow that we could thereby assign any particular probability—let alone a probability close to 1—to any specific wave theory of light. To establish by disjunctive elimination the high probability of a particular theory of light, one must not only discredit corpuscular approaches but, equally and obviously, one must show that rival wave conceptions to the one in question have a vanishingly small aggregated probability. For that reason, even if it is virtually certain that light is a wave, it does not follow without further ado that any particular theory of light is highly probable. This is more than a quibble since there were several different versions of wave theory on offer in the first half of the nineteenth century.3 To remind you of but two of them, recall that Young’s wave theory of light did not involve transverse vibrations, while Fresnel’s theory did. Other alternatives involved translational motion of the ether while some supposed only vibratory motion. For which of these many alternative wave theories is Achinstein claiming a probability close to 1? And how can he possibly get that claim from the refutation of a particular version of the
Waves, Particles, Independent Tests, and the Limits of Inductivism
115
particle theory, or even from the refutation of every known version of the corpuscular theory? Insofar as the undulationists were generally arguing for one specific version or other of the wave theory, Achinstein’s machinery is of no avail. Take the case of Fresnel during the 1820s. He did not see himself as addressing the ontological question “Is light a wave?” so much as he was attempting to ascertain the credibility of specific wave models or theories of light. Refutations of the corpuscularian hypothesis were powerless to guide him with respect to choices among rival wave conceptions. But perhaps, when Achinstein tells us that the probability of the wave theory is close to 1, he has in mind no particular, full-blown version of the wave theory but rather some generic Ur-wave theory that contains only those assumptions held in common between the various wave theories. Let us give him the benefit of the doubt here and suppose that it is that body of assumptions that he means by the term “wave theory.” What would such a theory look like? Absolutely essential to any attempt to characterize the common or generic elements of early nineteenth-century wave theories is the idea of a luminiferous ether. Although the student of twentieth-century physics, in contemplating a wave theory of light, has been trained to resist asking the question, “In what medium are the waves propagated?” no early nineteenth-century physicist could be so ontologically blasé. According to both its opponents and its detractors, the wave theory—in all its known versions—was committed to the existence of an all pervading, highly elastic, uniformly dense ether whose constituent parts were imponderable, that is, without weight. To accept the wave theory (in any sense stronger than as a useful predictive instrument) was, in this epoch, to claim to have a warrant for postulating such a medium. Much of the debate between the wave theory and its critics—a side of the debate that Achinstein largely ignores—is about the appropriateness of this commitment to a highly theoretical entity. Indeed, once one realizes that c is committed to this principle, we can recast Achinstein’s earlier prima facie eliminative disjunction into this form: (1¢) p(light is a particle or light is propagated through an elastic, homogeneous, imponderable fluid/ O & b) = 1
When thus recast, the core premise of the Achinstein reconstruction suddenly becomes, I submit, a great deal less plausible. What might have looked initially as a virtually exhaustive disjunction now comes to seem much less so. We can readily imagine that both disjuncts may be false and therefore that the initial assignment of a probability close to 1 to their disjunction is no longer compelling. And in that case, wave theorists are going to have to do a great deal more than argue negatively for a low probability that light is
116
Philosophy of Science Matters
a corpuscle. Of course, Achinstein is right that the wave theorists tried to show that light is not corpuscular, but if that was all that they had done, or the principal thing they had done, then they would have had no license whatever for supposing themselves to have provided a warrant for accepting the wave theory, not even in its generic version, given the implausibility of (1¢). And argue they did. Over and again, the wave theorists claimed that the strongest bit of their case rested on the ability of wave theories to solve a large range of empirical problems, including many that had been anomalous both for corpuscularian and for earlier undular theories.4 The philosophical point here about the precariousness of eliminative induction is a familiar one but it continues to be often ignored by Bayesians in our time, as it was ignored by Mill and his followers in the nineteenth century. Any account of evidence and theory evaluation that requires the enumeration of all the possible hypotheses for explaining some phenomenon, or in more modem probabilistic parlance, any approach that requires the enunciation of a set of hypotheses that are mutually exhaustive and pairwise exclusive, is, when applied to real scientific choices, almost guaranteed to fail. Scientists are rarely in a position to assert that they have canvassed all the relevant possibilities for explaining any body of phenomena. Indeed, properly viewed, the history of science is a record of refutations of such claims, whenever scientists have been so cheeky as to imagine that their extant theoretical options exhaust the available conceptual space. If Achinstein is right in claiming that hypotheses can acquire high credibility only by the exhaustive elimination of rivals, then we have to conclude that few if any scientific theories ever become credible. The alternative, of course, is to suggest that it is a reduction of the Bayesian position if it insists that credibility can be achieved by a theory only when all possible rivals to that theory have been both enumerated and vanquished. Before I move on to discuss what I earlier called the contextual problem, there is another conceptual problem that I want to mention. Full treatment of it would require another essay, but I can briefly summarize my worries in this fashion: as should already be clear, Achinstein believes that the accumulation of positive instances of a theory or hypothesis, however numerous, makes, at best, only a marginal difference to its probability. With that in mind, let us review Achinstein’s reconstruction of the wave theorists’ argument. Recall that the crucial step 3 in Achinstein’s reconstruction of the wave theorists’ argument involves their showing that auxiliaries introduced by the corpuscularians are highly improbable. In particular, their argument goes as follows: the corpuscular theory requires that the deflecting force is independent of the mass and shape of the deflecting aperture. This, say the wave theorists, is very unlikely, given that in the other cases of forces acting at a distance of which we are aware, the force in question is related to both the
Waves, Particles, Independent Tests, and the Limits of Inductivism
117
mass and the shape of the acting body. Now, how do the wave theorists know this? Well, one has to suppose they know it by virtue of an enumeration of known cases of bodies acting at a distance, combined with the knowledge that the force in question has thus far always depended on the mass and shape of the body exerting the force.5 In sum, the wave theorists are supposing that we have lots of instances of the hypothesis that forces exerted are dependent on mass and shape. It is that generalization that makes the corpuscularian hypothesis unacceptable. But, by Achinstein’s lights, such information can do nothing whatever to make probable the hypothesis that force depends on mass and shape. Instances of a generalization cannot—in his view—make that generalization probable. Only an eliminative argument can do that. As Achinstein himself points out, Thomas Young’s argument that the shape of bodies determines the kind of force they exert is based upon the fact that this is what we observe to be the case “with other known forces acting at a distance” (Achinstein 1991, 87). But, as I have said, on Achinstein’s own theory, such observations cannot possibly establish with high probability the claim that “all distance forces are dependent on the shape of the body exerting the force.” Yet that latter hypothesis is precisely the one that, on Achinstein’s reconstruction, the wave theorist needs. A similar argument could be made about the first premise of the wave theorists, namely, that all cases of motion involve either the transmission of a particle or of a disturbance in a medium. If you deny to the wave theorist the possibility of making a hypothesis credible by citing positive instances of it, then the wave theorist cannot begin to get (Achinstein’s version of) his argument against the corpuscular theory off the ground.
(2) The contextual problem I want now to turn away from the eliminationist issue in order to focus on what seems to me to be the central issue at stake in the debates between early nineteenth-century corpuscularians and undulationists. But we do not need to move too far afield, since (1¢) already allows me to direct attention to what I think was the core methodological divide between the wave theorists and the corpuscularians. Ever since Newton, corpuscularians had insisted that any theory about natural phenomena must not only be sufficient to explain the appearances, it must also involve postulating only true causes, or verae causae. This requirement, sometimes called the vera causa rule, is close to the core of late eighteenth-century empiricism; it was generally understood to mean that any entities postulated by theory must be ones to which we have independent access. Independent of what? Independent of the phenomena that the theory would be used to explain.
118
Philosophy of Science Matters
Between the time of Newton and Whewell, there was extensive discussion and refinement of this principle. Reid, Stewart, Priestley, Lyell, and Herschel were among its most ardent proponents. By the early nineteenth century, the vera causa demand had generally come to mean that any properties attributed to theoretical entities or processes must be a subset of the properties of observable bodies. The vera causa requirement, in other words, forbade attributing properties to unseen objects that were not exhibited broadly (perhaps universally) by objects accessible to inspection. Such a methodological demand was satisfied by the corpuscular theory of light; that indeed was one reason for the popularity of the corpuscular theory in the late eighteenth century. It postulated particles of light that, although obviously too small to be seen, behaved very like macroscopic objects, subject to Newton’s laws and to familiar forces of attraction and repulsion. Within the wave theory, however, the requirement of independent access or vera causa was apparently violated by the luminiferous ether.6 That ether consisted of particles that, being imponderable, had no weight. Corpuscularian critics of the wave theory like Brougham and Brewster claimed that no responsible empiricist had a license for propounding theories that, whatever their predictive or explanatory successes, involved entities whose properties were not drawn from common experience.7 When the corpuscularians demanded that there should be independent warrant for theories, this was what they had in mind.8 This was not a demand in which the wave theorists could acquiesce. A ponderable ether, which might have passed the vera causa test, would not do the jobs they required of their ether. Nor could they point to imponderable bodies in ordinary experience. It is for this reason, in my view, that the wave theorists found the method of hypothesis congenial, for what it offered was a way of freeing oneself from the vera causa requirement. The method of hypothesis allowed that a theory could be made plausible simply by examining its consequences (especially if they were of a broad and surprising character), without imposing any specific constraints on the sorts of entities postulated by the theories. Reading Whewell on this matter is instructive. A keen advocate of the wave theory, he goes to considerable lengths to castigate Newton and his followers for advocating the vera causa principle. Whewell sees that rule as an impediment to discovery and innovation and a gratuitous demand to make of a theory, especially if its consequential confirmation is impressive. Whewell saw clearly that, so long as the vera causa principle of independent warrant for a theory persisted, the wave theory of light would have tough sledding. Achinstein acknowledges that early nineteenth-century methodological standards required that there be independent support for theoretical
Waves, Particles, Independent Tests, and the Limits of Inductivism
119
entities. But, having acknowledged that, he proceeds to construe that requirement, when applied to the wave theory, as being satisfiable, by evidence that the corpuscular theory is erroneous! It is via that construal that Achinstein is able to act as if methodological consensus prevailed. The corpuscularians and the undulationists, he says, all accepted the principle that there must be independent empirical warrant for theories.9 Indeed, he characterizes the undulationists’ procedures as described in his steps (1) to (3) as a principle of independent warrant. What Achinstein fails to note is that the wave theorists’ form of independent warrant—if that is what it is—is completely unlike the traditional empiricist requirement of independent warrant.10 How the wave theorists established independent warrant, according to Achinstein, was by showing the implausibility of the auxiliaries used by the corpuscularians. But that has nothing whatever to do with satisfying the requirement of independent warrant as inductivists and corpuscularians then understood it. To the eighteenth-century empiricists and their successors in optics like Brougham and Brewster, independent confirmation of a theory T consisted in showing that ordinary bodies exhibited all the properties that T attributes to the entities it postulates. By contrast, Achinstein’s version of the independent support requirement dispenses with any constraint on the sorts of permissible entities. Rather, all it demands is evidence that the rivals to T are false or unsupported. I submit that no eighteenth-century empiricist, no advocate of the vera causa requirement and few if any corpuscularians would have accepted Achinstein’s characterization of the independent warrant requirement as an explication of what they were about. If the wave theorists’ strategy consists of the four steps that Achinstein attributes to them, then it automatically follows that—far from being inductivists in the then accepted sense of that phrase—they were entirely abandoning the inductivists’ project for subjecting theory to the vera causa requirement. Whewell saw clearly that the wave theory could not satisfy the traditional demand for being a vera causa; that is why he argued at length against the legitimacy of that requirement in scientific methodology. But even if Achinstein has got the corpuscularians wrong, it remains to ask whether his analysis of the case is one that wave theorists would have found congenial or close to the spirit of their project. I have my doubts. For reasons already indicated, the discrediting of known rivals—and that is all Achinstein’s independent confirmation requirement requires—is not sufficient grounds for asserting a theory. The wave theorists understood that and therefore spent much ink arguing that the principal virtue of the wave theory consisted in its ability to predict and explain a large range of phenomena, including many surprising phenomena. Achinstein’s philosophi-
120
Philosophy of Science Matters
cally-motivated conviction that this particular virtue cannot confer high probabilities on theories leads him to give less than its due to the prominent role accorded to positive evidence by the wave theorists. Convinced that positive confirmation, of whatever sort, cannot confer high probability on a theory, Achinstein supposes that nineteenth-century wave theorists must have accepted this point and acted accordingly. But I can find no evidence whatever, either direct or circumstantial, that they believed that positive confirmation was as impotent as Achinstein thinks it is. Let me put the challenge directly: where is the evidence that the wave theorists believed, as Achinstein does, that confirmation of salient instances cannot confer high credibility?11 Where is the evidence that they regarded the low probability of corpuscular theories as the principal ground of credibility for their own views? And if they did believe that, why were they so concerned with finding impressive corroborations of the wave theory?12 Indeed, if they really believed—as Achinstein suggests—that the wave theory acquires virtual certainty simply from the discrediting of the corpuscular theory, why give pride of place, in assessments of the wave theory, to its successful positive instances?13 John Stuart Mill, himself no friend of the wave theory, believed that the ability of theories to make surprising predictions successfully was of no epistemic moment. Such phenomena are, he said, designed only to impress “the ignorant vulgar.” I trust that it goes without saying that Achinstein is a Millian on these matters, even if his language is less figurative than Mill’s. But I see no historical basis for claiming that the wave theorists shared this dismissiveness about positive evidence in general, or about surprising instances in particular. For that reason, I doubt that eliminationism was the dominant methodological strategy of nineteenth-century theorists of light. Had it been so, nineteenth-century optics—both on its theoretical and on its experimental side—would look radically different from the way it actually does. REFERENCES Achinstein, P. 1992. Waves and Scientific Method. PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association Volume Two: Symposia and Invited Papers (1992), pp. 193–204. ——— . 1991. Particles and Waves. Oxford: Oxford University Press. Airy, G. 1831. Mathematical Tracts on the Lunar and Planetary Theories, 2nd edition. Cambridge: Cambridge University Press. Brewster, D. 1838. Review of Comte’s Cours de Philosophie Positive. Edinburgh Review 67: 271–308. Buchwald, J. 1981. The Quantitative Ether in the First Half of the 19th Century. In Conceptions of Ether: Studies in the History of Ether Theories, 1740–1900, ed. G. Cantor and M. Hodge. Cambridge: Cambridge University Press.
Waves, Particles, Independent Tests, and the Limits of Inductivism
121
Cantor, G. 1983. Optics after Newton. Manchester, England: University of Manchester Press. Herschel, J. 1830. Preliminary Discourse on the Study of Natural Philosophy. London: Longman. ——— . 1827. Light. In Encyclopaedia Metropolitana vol. 4, ed. P. Barlow. London: Griffin. Laudan, L. 1981. The Medium and Its Message: A Study of Some Philosophical Controversies about Ether. In Conceptions of Ether: Studies in the History of Ether Theories, 1740–1900, ed. G. Cantor and M. Hodge. Cambridge: Cambridge University Press. Powell, B. 1837. Recent Progress of Optical Science. British Annual and Epitome of the Progress of Science 1: 162–210. ——— . 1835. Remarks on the Nature of Evidence in Support of a Theory of Light. Edinburgh New Philosophical Journal 18: 275–85.
NOTES * Reprinted with permission from PSA: Proceedings of the Biennial Meeting of the Philosophy of Science Association Volume Two: Symposia and Invited Papers (1992), pp. 212–223. 1. Except where otherwise noted, all references to Achinstein will be to Particles and Waves (Achinstein 1991). 2. I claim no originality here. Whewell (1840) gave exactly the same analysis of the case. 3. To mention only a few: there was Young’s wave theory (without transverse vibrations), Fresnel’s wave theory of 1818, Fresnel’s 1821 ether (consisting of molecules acting at a distance) and Cauchy’s ether of the early 1830s. There were divergences among wave theorists about such matters as: the range of the molecular force associated with a particle of the ether (does it extend only to the next particle or fall off according to a l/r4 law, as Cauchy thought?); and how do ether particles and ordinary matter interact? These were not idle questions, as answers to them determined what sorts of empirical consequences a wave theory of light would have. Buchwald (1981) has convincingly argued that divergences among wave theorists about the nature of matterether interactions were “extremely important” in debates about the wave theory. 4. Thus, Herschel in his important monograph on light (Herschel 1827, 538) claims that “nothing stronger can be said in favour of an hypothesis, than that it enables us to anticipate the results of . . . experiment, and to predict facts opposed to received notions.” He was to make a similar point three years later in his 1830 classic: The surest and best characteristic of a well-founded and extensive induction, however, is when verifications of it spring up, as it were, spontaneously into notice, from quarters where they might be least expected, or even among instances of the very kind which were at first considered hostile to them. Evidence of this kind is irresistible, and compels assent with a weight which scarcely any other possesses (Herschel 1830, 170). These and like sentiments to be found in much of the writings of the wave theorists are scarcely the views of folks who think that the refutation of a rival hypothesis is the primary vehicle for establishing the credibility of a theory.
122
Philosophy of Science Matters
5. Achinstein himself describes the reasoning of the wave theorists here as a form of “inductive generalization” (Achinstein 1992). I fail to see how, by his lights, such inductive generalizations are capable of establishing the high probability of the claim that forces depend on masses and shapes. 6. I might ask in passing how, since Achinstein believes that it establishes a low probability for the corpuscular theory because it requires that the diffracting effect of an aperture is independent of its mass and shape, he can ignore the corpuscularians’ argument that the wave theory must have a low probability because of its commitment to imponderable particles of the ether? If light-deflecting apertures whose deflection depends on neither their shape nor mass is contrary to previous experience, it is surely as contrary to experience to postulate particles that have no weight. Although Achinstein claims that “no auxiliary assumption is introduced [by the wave theorists] whose probability given the theory is very high but whose probability on the phenomena alone is low” (Achinstein 1992), it seems to me beyond dispute that the hypothesis of the existence of an imponderable ether—although very probable given the wave theory—is a hypothesis whose probability “on the phenomena alone is low.” 7. David Brewster inveighed against the luminiferous ether because it was “invisible, intangible, imponderable [and] inseparable from all bodies” (Brewster 1838). On those grounds, he held that it could not be postulated as a legitimate causal agent. 8. The widespread acceptance of the vera causa demand shows up not only in the writings of the corpuscularians. During the 1820s and 1830s, there was a sizable group of optical theorists who, while opportunistic about using the mathematical and analytic tools of the wave theory, balked at the full acceptance of the wave theory because they did not see a warrant for the postulation of the optical ether. George Airy, for instance, claimed that the positive evidence as to the composition of the ether was too incomplete to enable one to judge which of the various ether models was correct (Airy 1831, vi). Baden Powell, happy to use the principle of interference, drew the line at accepting the ether precisely because it was not a vera causa whose existence had been independently established (Powell, 1835; 1837). 9. Three decades ago, I pointed out that the wave theorists’ idea of independent support “ought not be confused with the earlier empiricist requirement that theories must involve verae causae” (Laudan 1981, 175). 10. Achinstein writes that “both methodologies [those of the wave theorists and the particle theorists] stress the need for independent empirical warrant” (Achinstein 1991, 108). That may be so, but the fact remains that the two camps construed the demand for independent warrant in wholly different ways. 11. I am not here asking the normative question of whether they were correct in believing that positive confirmation can confer high probability. It is the descriptive issue that is at stake here. 12. Discussion of the confirmation of surprising predictions made by the wave theory was commonplace in this period. In 1833, for instance, Hamilton used Fresnel’s biaxial wave surface to predict (what was previously unknown) conical refraction. Within a year, Humphrey Lloyd, another partisan of the wave theory, had confirmed this result, triumphantly announcing the confirmation to the
Waves, Particles, Independent Tests, and the Limits of Inductivism
123
British Association meeting in 1834. Why, on Achinstein’s account, make such a fuss over results that could at best only marginally increase the credibility of the theory? 13. The only answer that I can find Achinstein offering to this question is that “the wave theorist wants to show that his theory is probable not just given some limited selection of optical phenomena but given all known optical phenomena” (Achinstein 1992,). That, I think, is not how the wave theorists express themselves. They are not saying: “Look, see how our theory retains its high probability even when it is extended to new phenomena.” Rather, they are saying: “The ability of the wave theory to be successfully extended to new phenomena vastly enhances its credibility.” Achinstein does not want them saying the latter, since his epistemic apparatus has no resources for making sense of such claims. But I think there can be no doubt that that is what they were claiming, justifiably or not.
10 What’s So Great about an Objective Concept of Evidence? Helen Longino
1. INTRODUCTION Peter Achinstein begins The Book of Evidence with an anecdote. At a faculty meeting, a university dean replies to a remark of Achinstein’s with “Peter, you have never made a contribution of interest to scientists” (Achinstein, 2001, 1). Wisely interpreting this rebuke as directed to philosophers of science in general, Achinstein decides to engage in a philosophical project that will be of value to scientists: a philosophical analysis of evidence, that, unlike other philosophical accounts, will address the questions scientists actually have about evidence, and is sensitive to the ways in which the concept of evidence works in actual research contexts. Achinstein’s exposition covers many questions of interest to philosophers, including interpretations of probability and their relevance to the analysis of evidence, the role of explanation in scientific inference, and holism. Along the way, he deploys many of his hallmark strategies, including the devastating counterexample and the absurd consequence. Most of Achinstein’s scientific examples are from its history, however, not from its present. I propose to pull out from the book-length treatment the central ideas about evidence that Achinstein develops, as well as some of the philosophical claims he makes on their basis, and put the ideas to work in thinking about issues in a contemporary research context. I will argue (1) that the account does have some utility for scientists, but (2) that it requires supplementation to achieve that utility, and (3) that the philosophical claims (in which I am interested) are not really supported by the analysis.
2. ACHINSTEIN’S CONCEPT(S) The basic idea Achinstein sets out to elaborate is that evidence is that which offers a good reason to believe that something else is true. One
124
What’s So Great about an Objective Concept of Evidence?
125
starting point is that whatever serves as evidence is a fact. The question about this fact is whether it counts as a good reason to believe that something else is true. It doesn’t matter whether anyone actually believes this something else. The question at issue is the relationship between this fact and some further (supposed) fact. The ambition is to provide a robust account of evidence that entitles us to say that the evidence relation is objective, non-trivial, and, ideally, non-contextual or nonrelative. Achinstein first distinguishes four concepts of evidence. ES evidence is a fact that counts as evidence for someone in what he calls a specific epistemic situation. An epistemic situation consists in beliefs or knowledge that certain propositions are true, the epistemic inaccessibility of the truth or falsity of certain other propositions, and knowledge (or the absence of knowledge) of how to reason from the proposition one takes to be true and some hypothesis of interest. In such a case, which characterizes all those cases in the history of science in which we take an individual or community to have had good reason to believe some hypothesis on the basis of the facts they had available to them, although those facts do not constitute good reasons for us, we can say that Sn has evidence e for hypothesis h relative to E-Sn. But what is evidence? Three additional definitions are provided to answer this question. The most important is that of potential evidence. Potential evidence is, for Achinstein, an objective concept, not relativized to an epistemic situation. e is potential evidence that h if i ii iii iv
p (h/e) > ½ p (there is an explanatory connection between e and h/h & e) > ½1 e is true e does not entail h
Potential evidence statements can be incomplete, in the sense that whether the fact that e does raise the probability of h above the threshold of ½ may depend on other facts about the situation, for example, that the experimental setup that produces the fact e is working as assumed. Potential evidence is distinguished from ES evidence in that the relativization to an epistemic situation is removed, although potential evidence can be subject to some relativizations (either to those just-mentioned assumed facts about a situation or to one of three types of “harmless” relativizations involving temporality, the assumption of non-interference by outside conditions, and the assumption that microconditions that may affect the precise value of measurements can be disregarded). Thus a more correct characterization of potential evidence includes reference to such assumptions: e is potential evidence that h, given b,2 if
126
i ii iii iv
Philosophy of Science Matters
p(h/e & b) > ½ p(there is an explanatory connection between e and h/e & b) > ½ e and b and true e does not entail h
Why is the second condition requiring an explanatory connection required? Why is it not sufficient to require that the probability of some hypothesis, on some fact e and assuming the truth of some set of background beliefs, be greater than the probability of its negation in order that e count as evidence for the hypothesis? Formally speaking, the following is possible. A set of statements can be assembled that satisfy the conditions set out in the definition of potential evidence, but the evidential work, one might say, is performed by the facts reported in the background assumption b, rather than by e which adds nothing to the probability conferred on h by b. Furthermore, attempts to overcome such a counterexample by requiring a deductive relationship will rule out some important cases in which we want to say that some fact, given background information b, is evidence.3 The concept of explanatory connection is not directional. There can be an explanatory connection if h, if true, would explain e, if e, if true, would explain h, or if there is a common explanation for both h and e. For simplicity’s sake, I will just use the former. Achinstein draws on his prior work on explanation to explicate the kind of relation that is involved. Some p1 explains p2 if p1 is a complete content-giving proposition with respect to a question whose complete presuppositions are given in p2. Such explanations can take the form: “The reason that p2 is that p1.” Suppose our question is, “Why did our colleague C remain in Europe until April 26?” The presuppositions of this question are that C was in Europe and remained in Europe until April 26. The answer to the question of why this was so is that European air traffic was shut down during the passage of a (very large) cloud of volcanic ash. Of course, there is additional information that could be included, but is presupposed: that C intended to return via air, that C was in one of the cities affected by the ash cloud, and so on. So the account of evidence is one according to which some e is evidence for some h if e makes it more reasonable to believe h than to believe not-h, glossed as p(h/e) > ½ or p(h/e & b) > ½,4 and that there is an explanatory connection (understood as just described) between h and e, given h and e has a probability greater than one half. These probabilities are understood as objective probabilities, and the reasonableness of belief is understood as supervenient on physical facts in the world. This is what makes the concept an objective concept: claims of evidential relevance are understood as true in virtue of facts in the world, whether these are causal facts, correlational facts, or other associational facts. The probability statements involved
What’s So Great about an Objective Concept of Evidence?
127
in this analysis are, as a consequence, empirical statements. They might be false, and anyone believing them would be wrong in their belief. It is the facts, not what anyone believes about the facts, that make beliefs reasonable. For any h, e pair, there may be an epistemic situation that would make it reasonable for someone in that epistemic situation to believe h given e, but, Achinstein emphasizes, “I am denying that a nonrelativized abstract evidential claim must always be understood as implicitly relativized to some particular epistemic situation” (Achinstein 2001, 97). Achinstein has a great many subtle observations to make about concepts and theories of probability and about alternatives to his exposition of reasonableness of belief in terms of probabilities. Among such alternatives are the Royall “likelihood” analysis and the Mayo error-statistical approach. My own concerns have to do not with these more technical issues, but with, granting the broad outlines of his analysis, understanding the reach of the claim that a nonrelativized evidential need not be understood as implicitly relativized to some particular epistemic situation.
3. EVIDENTIAL HOLISM AND CONTEXTUALISM Achinstein’s chapter called “Old-Age and New-Age Holism” in The Book of Evidence engages most directly with this question. “Old-age” holism is the holism attributed to Duhem and elaborated by Quine. It is expressed as, “Nothing can be evidence for or against an individual, isolated hypothesis, but only for or against some theoretical group of hypotheses” (Achinstein 2001, 231). Because Duhem uses a modus tollens setup in arguing that a piece of data is never evidence against a hypothesis considered in isolation but only against a hypothesis together with a body of theory that establishes the relevance of the datum to the hypothesis. Achinstein reads him as committed to a hypothetico-deductive account of evidence, which he has argued previously provides neither sufficient, nor necessary, conditions for evidence (146–8). But, of course, he realizes that one can express the relevant heart of the holist view without being committed to the hypothetico-deductive account. This is what he calls “new-age” holism, which is also recognizable as contextualism with respect to evidential relations.5 This he expresses as “where h is some isolated hypothesis and e is some isolated fact, it is never true that e is evidence for h [full stop]. Rather e is evidence for h, relative to some set of assumptions b.” Achinstein’s argument against this claim has several stages. First, he notes that one can always take those assumptions b and place them in the statement of e. As an example, consider a lottery. That Sylvia owns 95 tickets is evidence that Sylvia will win the lottery against the background
128
Philosophy of Science Matters
assumption/information that there are 100 tickets and the lottery is fair. But one can transform this into an unrelativized statement: that Sylvia owns 95 of 100 tickets in a fair lottery is evidence that Sylvia will win the lottery. But (1) this seems trivial and (2) it turns an empirical evidence statement into an a priori one. Lotteries play a large role in Achinstein’s explications of probability. One might argue that they are a misleading category of example to use for the explication of a probability concept relevant to cases in science, since in a lottery, there is a determinate (and known to someone) number of tickets, which makes probabilities of winning, given various distributions of tickets, quite objective. In this case, it is trivial to transform the relativized empirical statement into an unrelativized one. How might the “new-age” holist, or contextualist, object? One way might be to focus on the statements in b that are absorbed into the evidence statement. Achinstein puts a somewhat different objection into the mouth of the new-ager: that the unrelativized statement becomes a priori. This generates a restatement of the holist position: if an unrelativized evidential claim is empirical, then, on pain of incompleteness and/or lack of perspicuity, it must be understood as relativized to a set of assumptions that are necessary and sufficient for the truth of the claim. Achinstein’s response to this restatement is that if the argument is that the claim is somehow incomplete, the same can be said for any empirical claim. But this is unreasonable. Is the claim that a muon decays to an electron, an electron-antineutrino, and a muon-neutrino incomplete without a statement of the assumptions that are necessary and sufficient for its truth? If we were to demand this, we would not be able to distinguish a claim from what might be offered in its defense, what might be advanced as evidence for the claim. But surely this misrepresents the concern of the (let us now say) evidential contextualist. The contextualist is not motivated by distress over evidence statements interpreted to be a priori, nor is the contextualist saying that only if relativized to a statement of conditions necessary and sufficient for its truth is an evidential claim complete and perspicuous. Nor is the contextualist making a fuss about the “harmless relativizations” Achinstein allows. The contextualist is motivated by arguments about the underdetermination of hypotheses by evidence. The concern is not that evidence is not deductively determinative, but that the relevance of some state of affairs to a hypothesis requires background assumptions. Evidence statements are not categorical statements about the world, but statements about the relevance of facts to statements. Another way to put this is by noting, as Achinstein himself emphasizes, that an evidential claim is an empirical claim. This means that reasons not included in the evidential claim itself are relevant
What’s So Great about an Objective Concept of Evidence?
129
to its assessment. Some of these reasons will constitute evidence; others will constitute assumptions about the equipment and conditions of observation and experimentation; others may rule out alternatives; while still further assumptions may be general theoretical frameworks within which facts of certain types are connected with putative facts of other types. The contextualist wants to emphasize not the incompleteness of an evidence claim, but the dependence of its truth on a context of background assumptions that are themselves empirical in nature, not known to be true, and that might be false. These assumptions provide reasons to think not that some categorical assertion, like the one about the products of muon decay, is true, but that some fact is relevant to the truth of a claim about another fact. Where it matters most, the hypothetical facts for which some d is affirmed to be evidence are not directly accessible and we don’t have direct access to the associations between the hypothetical fact and the datum. Some theoretical background assumption(s) (whose truth we presuppose, but are not in a position to know) tell(s) us that there is some connection such that the factivity of one counts as a reason to believe the factivity of the other. The contextualist, then, is denying that unrelativized statements can ever (well, at least in the cases that matter) be independent of an epistemic context. Achinstein might say, of course, that the relevance of fact to hypothesis is established by the explanatory connection clause in the definition of evidence. Only if we focus on the first clause, requiring that the probability of h on e be greater than one half, does the relevance problem arise. Once the second clause is added, that the probability that there is an explanatory connection between h and e, on h and e, then the relevance of e to h is secured. Does this put the contextualist’s worries to rest? Let us look at an example. While this is still a somewhat fabricated example, it is closer to actual scientific practice than are lotteries. Consider a question about the causes of depression in young men. Suppose we want to consider how much this might be genetically influenced. One strategy would be to see if there is a higher frequency of a given gene in young men who are depressed than in young men who are not depressed. Following this strategy, we would first identify a population of young men who are depressed, probably generating our sample population from young men receiving treatment for depression, but then administering some standard test for depression. Then we would create a control sample, matched with our original population on a specified set of traits, but not depressed (using the same test as used on our first sample). We take blood samples from all our subjects, subject the samples to molecular analysis, and find that a higher percentage (imagine 20%) of those in the first group have one or
130
Philosophy of Science Matters
two short alleles of the serotonin transporter gene (5-HTTLPR) than do those in the second group (say 3%). For what and under what circumstances could this result (let’s call it d) be evidence for a hypothesis? There are several hypotheses for which the result could be evidence, and accordingly several claims in which d could figure. 1. d is evidence that one or two short alleles at 5-HTTLPR causes depression. (H1c) 2. d is evidence that one or two short alleles at 5-HTTLPR influences depression. (H1i) 3. d is evidence that one or two short alleles at 5-HTTLPR increases the likelihood of depression. (H1l) Remembering the analysis of evidence claims, these are restateable as 1. i. p(H1c (5-HTTLPR causes depression)/d) > ½ ii. p(that there is an explanatory connection between H1c (5-HTTLPR causes depression) and d/d) > ½ 2. i. p(H1i (5-HTTLPR causes depression)/d) > ½ ii. p (that there is an explanatory connection between H1i (5-HTTLPR influences depression) and d/d) > ½ 3. i. p(H1l (5-HTTLPR causes depression)/d) > ½ ii. p(that there is an explanatory connection between H1l (carrying 5-HTTLPR increases the likelihood of depression) and d/d) > ½ If any of these evidence claim pairs is true, this would entitle us to take our result d as evidence for whichever of these claims in which we are interested. The question is: what entitles us to assert any of these probabilities? In the lottery case, we know the total number of tickets and the number held by Sylvia. This makes calculating the probability a relatively simple matter. In the gene–depression case, this is much harder. The first two hypotheses affirm a causal relation between a genetic configuration and a phenotypic trait. The explanatory connection between h and d in those cases is based on a connection between the genetic structure and the phenotype. The explanatory connection in the third case could be understood as that of a common cause of the genetic structure and the phenotype, through linkage of the genetic structure with another genetic structure, which has the causal relation to the phenotype. It would enable prediction, but not intervention. Surely knowing that the probability that a hypothesis is true is greater than one half if some experimental result has a certain value is more than half the struggle in the process of inquiry. One can say the same for knowing that the probability that there is an explanatory connection between the hypothesis and the data is greater than one half. Carrying
What’s So Great about an Objective Concept of Evidence?
131
out the study is relatively simple compared to ascertaining its evidential import. We need to know the base rates both of the three types of allelic pairing and of the base rate of depression in the population in order that the associations in the study be taken to indicate anything about a potential explanatory connection. Moreover, we must be able to rule out that there are other genes that might be more highly represented in the sample population and are the genes influencing the phenotypic phenomenon. We must also be assuming that there is a causal chain from genetic structures through anatomy and physiology to temperamental phenotypes, that depression constitutes an identifiable phenotype related to endophenotypic (e.g., neurophysiological) structures and processes, and that the incidence of short alleles in the non-depressed control population can be accounted for. These kinds of assumptions are what concern the “new-age relativized holist,” who is not a holist but a contextualist. It’s not the entire theory that’s implicated, but a set of quite specific assumptions. To point out the web of assumptions against the background of which a datum will count as evidence is not to dismiss the analysis or the objectivity of the concept. One might say that the value of Achinstein’s analysis of evidence is that it forces one (the researcher, the community of scientists who assess the researcher, the lay public who wish to know whether to act on the basis of some hypothesis) who is in the position of having to evaluate the evidential support for some hypothesis to consider whether the datum confers the requisite probabilities on the hypothesis. And this consideration requires knowing what additional information to include in order to raise the probability to something greater than one half. In some cases, this will be empirical information, like information about base rates that can be obtained relatively readily. In other cases, what is missing will not be empirical information that helps to complete the evidence statement, but assurance of the truth of some fundamental framing assumptions, like assumptions that there is a causal pathway from the genetic structures in question to high-level behavioral and/or temperamental phenotypes or to the endophenotypes that underlie them. From the contextualist’s point of view, these assumptions are part of the Epistemic Situation of contemporary geneticists. They proceed on the assumption that there exist such causal pathways, and see it as their task to identify endpoints of a complex process: genetic structure and phenotype. The assumption that there is a causal pathway between the end points is not part of the evidence (how could it be when we do not know whether it is true?), but the background against which it makes sense to pursue genetic investigation at all. Achinstein’s positive relevance account shows that it is necessary to spell out both the additional empirical information
132
Philosophy of Science Matters
and the background assumptions. To that extent it is of potential value to scientists. But it does not give us or them a principled way to determine what should go into the evidence statement and what should remain among the assumptions in light of which the results in question would count as evidence at all. To that extent, the account is incomplete. For an additional example, consider the case to which Achinstein devotes an entire chapter, Perrin’s experiments with Brownian motion that were rewarded by a Nobel prize. There can be some disagreement as what precisely these experiments were relevant to: the reality of molecules or that the structure of matter is particulate rather than continuous. The former interpretation is stressed by philosophers using this example as part of an argument for realism. The latter arguably makes more sense of the historical situation and Perrin’s own statements about his accomplishments. The point, however, is that the observed behavior of resin droplets (painstakingly produced as analogs to unobservable molecules just over the threshold of visibility) in different gases (that could slow their motion sufficiently that it could be measured) and the comparisons of measurements of droplets in different gases yields, with a lot of additional assumptions, a value for the number of molecules in a gram weight of the gas close enough to Avogadro’s number to be thought to coincide with it. Those assumptions, hard won during the physics of the nineteenth century, were necessary in order that Perrin even conceive his experiments. Once he was able to generate an empirical measurement that yielded Avogadro’s number, many other parametric measurements fell into place, entitling physicists and philosophers to consider the kinetic theory of gases as empirically grounded. It’s quite possible to reconstruct Perrin’s reasoning to see how he took the behavior of his analogs as evidence (of a kind that satisfies Achinstein’s criteria), but he was in a particular epistemic situation, one in which we remain.6
4. CONCLUSION So, how is the positive relevance account of evidence of use to scientists? What good is an objective account of evidence? To say that it is objective is to say that the connection affirmed by the attribution to some data of evidential relevance to a hypothesis is a real connection. Thus, it is, in an ontological sense, independent of what any individual may believe about it. As a criterion of evidential status, the positive relevance account offers a standard of reference by which to evaluate any claim of the evidential relevance of a datum, or set of data, to a hypothesis. But ontological objectivity does not entail epistemological a-contextualism. Applying
What’s So Great about an Objective Concept of Evidence?
133
the standard in any case more interesting and complicated than closed lottery situations is likely to reveal the need for additional information. Depending on how widely shared assumptions are in a given scientific community, application will also reveal the dependence of evidential status on assumptions framing the general research approach within which the datum has or data have been generated. If fully shared, the dependence is likely to remain unnoticed. But unnoticed does not equal nonexistent. In any case, adoption of such a standard is useful in guiding scientists to further research, research that might lead to discharging some of the assumptions. Certainly, if there were a principled way to determine which assumptions could or should be discharged and which should remain as assumptions, the account would be more useful, but that fact does not diminish the utility I have identified. Of course, the scientific community has to agree that this is the standard to which they wish to hold themselves. Does this account lay holism, or contextualism, to rest? I don’t see that Achinstein has offered a satisfactory argument against the contextualists’ claim. The contextualists are not claiming that whether some fact has evidential relevance is relative to individuals’ beliefs nor are they claiming that any empirical claim must include its grounds in order to be perspicuous or meaningful. The contextualists are claiming that empirical claims about evidential import (generally, and in the kinds of cases scientists are interested in) presuppose the truth of assumptions we are generally not in a position to ascertain. This assertion is supported by analyzing cases that are now or have in the past been of interest. The contextualists claim that we are always in an epistemic situation, and that, in the cases worth worrying about, we always have, at best, ES evidence. Since we don’t know what we don’t know, the trick lies in identifying just what our epistemic situation is. REFERENCES Achinstein, P. 2001. The Book of Evidence. New York: Oxford University Press. Longino, H. 2009. Perilous Thoughts: Comment on van Fraassen. Philosophical Studies 143 (1): 25–32. ——— . 1990. Science as Social Knowledge. Princeton: Princeton University Press. Van Fraassen, B. 2009. Perils of Perrin. Philosophical Studies 143 (1): 1–24.
NOTES 1. Provision i is absorbable into ii, as the truth of ii requires the truth of i. But, although the shorter definition is more elegant, it is more perspicuous for my purposes to articulate both probability provisions.
134
Philosophy of Science Matters
2. There are alternative ways of expressing this relation that can transform i into an a priori statement, but these can be disregarded. 3. Achinstein’s example concerns taking frequencies in a sample of a population as evidence of frequency in the general population. 4. Achinstein does not consider cases where there might be, instead of two hypotheses, h and not-h, three or four, say, hi, hii,, and hiii. Presumably such cases can be addressed by altering the probability threshold. 5. Thus bearing at least a family resemblance to the contextual empiricism I have elsewhere defended (Longino 1990). 6. See van Fraassen 2009, and my comment (Longino 2009) for an elaboration.
11 The Objective Epistemic Probabilist and the Severe Tester Deborah G. Mayo
1. INTRODUCTION In 1998, Peter Achinstein participated in a PSA Symposium I organized, the goal of which was to promote philosophical accounts of evidence as relevant to scientific practice.1 I was therefore somewhat jarred that Achinstein titled his presentation “Why Philosophical Theories of Evidence Are (and Ought To Be) Ignored by Scientists” (Achinstein 2000). But it turns out we were entirely in sync as regards the reasons for his lament: the problem with philosophical accounts is (1) they are far too weak to give scientists what they want from evidence, and (2) they make the evidential relationship a priori whereas establishing claims of evidence requires empirical investigation. From this agreement it became clear that we share fundamental theses about evidence. As Achinstein has recently noted, we concur “that whether e, if true, is evidence that h, in the most important sense of ‘evidence,’ is an objective fact, not a subjective one of the sort many Bayesians have in mind. We agree further that it is an empirical fact, not an a priori one of the sort Carnap has in mind” (Achinstein 2010, 170). In fact, Achinstein is to be credited as being one of the only philosophers of science to explicitly incorporate the need for empirical checks of evidence in his account. In addition, we are of like mind in insisting on a “threshold” concept for warranted evidence—if data x do little to warrant H, then to infer H is unwarranted, by dint of x. (Although he would put this in terms of beliefs, the idea is the same.) Nevertheless, Achinstein alleges that he and I disagree on the fundamental role of probability in an adequate account of evidence. Achinstein’s (objective) epistemic probabilist holds a variation on the view that probability enters to quantify how reasonable it is to believe in H, given data x; in the view I advance, probability arises to quantify how well, or how severely, H
135
136
Philosophy of Science Matters
is tested by means of x. Where Achinstein’s threshold for evidence for H requires the posterior (objective epistemic) probability for H to be sufficiently high (at least greater than 0.5), mine requires H to have passed a test that is sufficiently probative or severe. Not only is severity not based on a Bayesian calculation of posterior probabilities in hypotheses, it can even happen, Achinstein argues, that H passes with high severity while the posterior probability of H is low. If, as Achinstein holds, high posterior probability is necessary (though not sufficient) for warranted belief, then it appears the severe tester would countenance warranting H in cases where Achinstein advises withholding warranted belief. Conversely, it can happen that Achinstein’s epistemic probabilist assigns a high degree of reasonableness to belief in H when the severe tester withholds inferring H. This “Achinstein–Mayo conflict” may be dubbed the “highly probable vs. highly probed” conflict (Mayo 2005), and it is the focus of this paper. Whether the conflict constitutes a problem for the severe tester or for the epistemic probabilist turns on which measure more adequately captures the evidential warrant for the hypotheses in the “counterexamples” raised. Achinstein and I have had several exchanges over the years revolving around this conflict (Achinstein 2000, 2005, 2010, and Mayo 2000, 2005, 2010), and I am grateful to have the opportunity to revisit the issue once more. For I am increasingly convinced, especially given our most recent (2010) exchange, that the severity account is actually in sync with the goals and special features of Achinstein’s objective epistemic probabilist (although I will only be able to discuss some of the reasons for this here). When it comes to objective evidence—the only account I will be considering—Achinstein breaks ranks with the typical Bayesian account of confirmation in several respects: first, the prior probabilities (in an exhaustive set of hypotheses) are to be neither measures of subjective degree of belief nor a priori logical measures (either of the Carnapian or more modern “reference” or information-theoretic Bayesian varieties); second, he denies that it is necessary or sufficient for confirmation that x increase the probability of H, although the posterior probability of H must reach a threshold for reasonable belief (at least 0.5); and third, it is required that there be some kind of (“non-Bayesian”?) explanatory connection between x and H (either x explains H, H explains x, or there is a third factor explanatorily connected to both). According to Achinstein, objective epistemic probabilists “are committed to saying the inference is justified only if the objective epistemic probability (the posterior probability) of the inductive conclusion” is sufficiently high (Achinstein 2010, 179). This assertion is uncontroversial if he is merely playing on the fact that, in ordinary English, we may use “H is probably true” to express that there is good evidence for claim H. Indeed I had
The Objective Epistemic Probabilist and the Severe Tester
137
earlier assumed that Achinstein intended “high epistemic probability” as a formal way to abbreviate something like “high inductive evidential warrant.” Consider, too, Achinstein’s additional requirement that: P(there is an explanatory connection between H and e|e) > 0.5.
Is this just an abbreviation for something like: “there is good evidence for an explanatory connection between H and e”? If it is left as a qualitative sum-up of empirical background information, then it would seem we already must be in possession of some non-Bayesian way to appraise evidence for such explanatory claims. And if “probability” can live here merely as an abbreviation for how good the evidence is for an explanatory connection, then why not also in the after-data sum-up of the warrant for H? Achinstein (2010) makes it clear that he intends high objective epistemic probability in hypothesis H to be a posterior probability in H, as computed using conditional probability (or Bayes’s theorem). True, this assumption is often regarded (by philosophers at least) as obvious or at least innocuous. I argue that it is neither, even when probability is used merely as a metaconcept for philosophical discussion.2 To begin with, there are the obstacles to arriving at the ingredients required for the Bayesian computation along with the challenge to show why the numbers thereby obtained may be interpreted as some kind of objective weight of evidence or belief. Moreover, probability logic seems to inadequately capture the reasoning appropriate to inductive inference in science, or so I argue (Mayo 2010). The examples that arise in illustrating posterior probability-severity conflicts involve dichotomous hypotheses that I label as H0 and H1, so I restrict my discussion to these.
2. EVIDENCE AS PASSING A SEVERE TEST Even though I intend to give a very general account of evidence, I use the notion of “testing” to emphasize a kind of standpoint or “burden of proof” that the severe tester demands. Clearly, evidence is not being taken seriously in appraising hypothesis H if it is predetermined that a way would be found to either obtain or interpret data as in agreement with (or as “passing”) hypothesis H, regardless of the evidence. Here is one of many ways to state this: Severity Requirement (weakest): An agreement between data x and H fails to count as evidence for a hypothesis or claim H if the test would (almost certainly) yield so good an agreement even if H is false.
Because such a test procedure had little or no ability to find flaws in H, finding none scarcely counts in H’s favor.
138
Philosophy of Science Matters
The weak severity requirement can be substantiated in terms of the goals of learning. To flout it would be tantamount to ignoring evidence and would permit being wrong with maximal probability. The onus is on the person claiming to have evidence for H to show their procedure is not guilty of so egregious a lack of severity. Although one can get considerable mileage even stopping with the weak severity requirement, I am prepared to accept the converse as well: Severity Principle (full): Data x provide a good indication of or evidence for hypothesis H (just) to the extent that test T severely passes H with x.
(1) Examples: blowout preventers and college readiness Example 1: testing the BOP on the Deepwater Horizon rig Although the blowout preventer (BOP) on the Deepwater Horizon drilling rig passed the required government tests, the question arises: did those passing results provide ample evidence that H1: the BOP on the Deepwater Horizon rig would perform adequately (to prevent a blowout in the Macondo well)?
Not if the government tests, conducted approximately every 2 weeks, are performed under conditions that render it easy for H1 to pass, even if H1 is false, that is, even if H0: the BOP would not perform adequately (either the “blind sheer ram” would be unable to cut the actual pipe, and/or it would malfunction in the conditions of extreme pressure and temperature that would be encountered).
Passing the government tests shows H1 “agrees with” data x (H1 might even logically entail passing results x). But if there is only a small probability, say 0.1, that the rig fails the government tests even if H1 is false (H0 is true), then H1 is very poorly corroborated; that is, the severity is ~ 0.1. So the error probability associated with the inference to H1 would be 0.9—clearly high.3 Using this rule would, in the long run, very often erroneously affirm H1—that is one thing entailed by the high error probability. But that is not the reason we deny that x warrants H1. We deny this because of what the data fail to indicate in the case at hand. As with scientific instruments more generally, the reported error probabilities (when computed correctly) inform us of the general capability of tools (in this case, it is a testing tool). The capability we care about here is the tests’ abilities to alert us to errors in reasoning. For example, hypothesis H1 would more realistically be framed in terms of quantities or parameters: the minimum thickness of
The Objective Epistemic Probabilist and the Severe Tester
139
the pipe the BOP would need to cut, and at what minimal pressure. High error probabilities in the government tests, when well specified,4 inform us if the simulated conditions are inadequately demanding as regards either requirement needed for H1. Severity assessments may arise from formal statistical models, but in the strongest cases, they are built on entirely informal grounds. The next example specifically arose in relation to the Achinstein–Mayo conflict (Howson 1997a, 1997, Mayo 1997b, 1997c, 2003a, 2005, 2010, 192–5).
Example 2: Isaac and his college readiness A student, Isaac, is given a wide variety of tests designed to check that he has sufficient mastery of high school material in order to be ready for work in a four-year college, that is, to check what is called “college readiness.” We are told that Isaac has scored nearly perfect grades on rigorous standardized tests covering science, history, literature, mathematics, and so on. We are further told that obtaining such high scores would be extremely improbable were Isaac not college ready. (We would prefer to consider degrees of readiness, but the critic’s example requires just the dichotomous hypotheses.) Reasoning that it is practically impossible for a student who lacked readiness to have consistently scored so well on a wide variety of tests—barring cheating—we infer the test results are evidence of his college readiness. I come back to “barring cheating” shortly. I first make a few points about severity that emerge from these examples.
(2) Features of severity reasoning For this it helps to have an abbreviation: We can abbreviate “the severity with which test T with outcome x passes hypothesis H1” as SEV(H1, test T, x),
where it should be understood in what follows that the test T would delineate the kind of experiment, the mode of data collection and data model, and the possible outcomes and hypotheses. The “logic” of SEV is not probability logic. In example 1, we had SEV(H1, test T, x) = 0.1
Although the severity for H1 is low, it does not follow that its denial H0 has passed a severe test. In fact, apparently the passing results of the Deepwater Horizon on April 20, 2010, warranted neither H1 nor H0. This is one of the
140
Philosophy of Science Matters
reasons that an error probability (associated with a test and an inference) is not well captured by the logic of probability. Neither hypothesis has reached the threshold for evidence. At the same time there is no danger of falling into the situation Achinstein rightly wants to avoid, namely, allowing both a hypothesis and its denial to be warranted by x. That is because if SEV(H, test T, x ) is high, then SEV(~H, test T, x) is low.
However, the converse does not hold: If the SEV(H, test T, x ) is low, the severity for ~H will in some cases be low, in other cases high. Severity is relative to a hypothesis and a set of data. It is crucially important to recognize that a severity assessment is always relative to (i) a hypothesis (or set of hypotheses) being considered for inference, and (ii) a specific set of data. Since it was so easy to pass H1: BOP erroneously, the passing result x fails to provide good evidence for H1. But suppose instead that a different result occurred, call it y, and that with result y test T passes H0: the BOP is inadequate. The very fact that test T sets such a low bar for the BOP being declared adequate (H1) gives all the more grounds for inferring its inadequacy (H0) when the results are y, that is, SEV(H0, test T, y) = high. So, in this account, it is not possible to assess a test’s severity without specifying the data and the specific inference being entertained in claiming to have evidence for H. The relativity of severity enables the account to smoothly handle familiar problems. In example 2, for instance, Isaac’s scores pass with severity the hypothesis of college readiness: H1(I) Isaac is college ready,
but they do not severely pass claims about the cause of his readiness (e.g., that Isaac was first-born). Nor would his scores severely pass claims resulting from “tacking on” irrelevant conjuncts to hypothesis H1(I) (e.g., the BOP on the Deepwater Horizon is adequate). His college exam results did nothing to probe or rule out errors regarding the cause of his readiness, or the inadequacy of the BOP. Any test rule that would regard Isaac’s test scores as evidence for these other hypotheses would violate the weak severity principle. This is the basis for getting around classic problems with both probabilistic and hypothetico-deductive accounts, discussed elsewhere (Mayo 1996; Mayo and Spanos 2010).
3. WHAT’S IN A TEST? Without restricting the account to formal contexts, statistical tests provide some general elements that help to illuminate the posterior probability-severity conflict. In those examples, there are two (exhaustive) hypotheses H0 and H1; so, while over-simple, I stick with such cases.
The Objective Epistemic Probabilist and the Severe Tester
141
(A) Hypotheses: H is typically a claim about some aspect of the process that generated the data, data x0 = (x1, . . . , xn): so the two (exhaustive) hypotheses H0 and H1 make rival claims about the aspect of interest. Probabilities of various outcomes are to be “computed under the supposition that Hi is correct” (about the generating mechanism), and that is the way to read: P(x; Hi).
Note that this is not a conditional probability, since that would assume there is a prior probability for Hi. Since it is less restrictive, we can use this notation and still entertain contexts where a Bayesian calculation is being considered. (B) Distance Function: a function of the data, d(X ), the test statistic, reflects how well or poorly the data x0 = (x1, . . . , xn) fit the hypothesis H0—the larger the value of d(x0), the farther the outcome is from what is expected under H0 in the direction of alternative H1, with respect to the particular question being asked. (Note: X is a random variable, and x0 is a fixed value of X; bolding the random variable indicates it is a vector.)
Suppose in our college-ready example, the data d(X ) is the average of 6 different standardized college-readiness tests, each with 800 as its maximum score; and say the test infers evidence of readiness so long as d(X ) is at least 750. (C) Test Rule T: infer that x is evidence that H1(x): x is college ready iff {d(X) > 750}.
This refers to a generic rule. Once x0 is in hand, the rule infers H1(x0). For example, if the student is Isaac, the inference would be H1(I): Isaac is college ready. (D) Error Probability: Applications of an inductive test rule can be “in error” as regards the data generation mechanism, so long as the data are limited. The probability of {d(X ) > 750} under the assumption that H0, is the probability of erroneously inferring a student is college ready. That is, P({d(X ) > d*};H0(x)-not college ready)
is an error probability. It is given by the probability distribution of d(X )—called its sampling distribution (computed under one or another hypothesis). The sampling distribution characterizes the capability of the inferential rule to unearth flaws and distinguish hypotheses. What makes an account “error statistical” is its consideration of these error probabilities. I am belaboring this because the confusion between the distribution of d(X) and the distribution of hypotheses is ubiquitous. Suppose Isaac’s average yields 770, so x agrees with H1 (and disagrees with H0). From the givens of the example, the probability in (D) equals some very low value p.
142
Philosophy of Science Matters
Therefore, H1 passes with high severity, 1 – p, and we infer that x is evidence that Isaac is college ready. (E) Empirical Assumptions: Quite a lot of empirical background knowledge goes into implementing these computations. We can place them into two groups of questions:
1 How probative would the test be if its assumptions were approximately satisfied? 2 Are its assumptions approximately satisfied? Answering question #1 turns on a yes answer to question #2. For instance, in the college-readiness test, the judgment (1) that it is practically impossible for a student to do so well if she were not college ready is based on assuming (2) that the tests are “working” in the case at hand, for example, that the students are not cheating, but achieving their scores as a result of their academic knowledge.5 The task of checking such assumptions calls for its own discussion, which I do not have space to consider here. The main thing to note is that both the severity and the posterior probability calculation require satisfying the assumptions in question #2, and any alleged posterior probability-severity conflict must already assume a yes answer to #2.
4. POSTERIOR PROBABILITY-SEVERITY CONFLICTS Note that the test ingredients in Section 3 do not include any assessment of the probabilities of the hypotheses themselves. The severe tester is keen to assess quantitatively “how false” or “how true” hypotheses are, but that is very different from assigning them a probability. The claims Hi(x0) are either correct or incorrect as regards the mechanism generating the data x0, and I take it Achinstein would agree. Insofar as we are interested in using data to make inferences about what generated this data, in this world, and insofar as we are keeping to a frequentist account of probability, Achinstein rightly observes that it would make no sense to speak of the probability of Hi, as if universes were as plenty as blackberries from which we randomly selected this one universe (as Peirce would say).6 By contrast, Achinstein is prepared to obtain his epistemic probabilities by a kind of straight rule: Achinstein’s Straight Rule for Objective Epistemic Probabilities: If xi is randomly selected from a population where p% have property C, then the objective epistemic probability that xi has C equals p.
The Objective Epistemic Probabilist and the Severe Tester
143
(1) Modifications. Before turning to the Bayesian computation, two of the modifications called for in order to bring the Bayesian calculation into the frequentist error-statistical fold should be noted, to avoid misunderstanding how the conflict can even arise. Turning error probabilities into likelihoods There is often confusion about likelihoods, so let me explain by means of a particular case, example 2. A single application of the six exams gives one outcome x0 (where x0 = the observed six scores) leading to one average score d(x0). Suppose Isaac’s average scores are d(x0) = 770. Once x0 is fixed, we can consider the probability of x0 under the various Hi. This is the likelihood of Hi given data x0.7 But the error probability in Section 3(D) is not a likelihood for the experiment being considered. It refers to average test scores other than the particular d(x0) observed, in particular, to all scores as great as or greater than 750. But Bayesians use likelihoods, not error probabilities. To bring the error probability into the Bayesian realm, the critic lumps together outcomes so that there are only two, call them success: s, or failure: ~s. An outcome is called a success (s), in relation to our college readiness example, if the observed scores are at least 750, else it is a failure (~s). By condensing outcomes, the likelihood that enters the Bayesian computation—for example, P(s|Hi)—is the same as the error probability, for example, P(d(X ) > d(x0); Hi). Turning events into hypotheses Second, to make out a posterior probability-severity conflict the critic considers “hypotheses” to which a frequentist might seem willing to assign a probability, in other words, the hypotheses are what we would normally consider specific events: that a sample possesses a characteristic such as “being able to prevent a blowout” or “being college ready.” (Hi(x) is like a one-place predicate, which does not have a truth value until x is replaced by a constant x0.) But we also want hypotheses to assign probabilities to outcomes. So, to help the critic’s example, let us stipulate that, once a particular name, Deepwater Horizon or Isaac, replaces x, the resulting hypotheses assign probabilities to the outcomes s and ~s. But what shall we say about the priors?
(2) Ready or not? The posterior probability-severity conflict in the case of poor Isaac goes like this: The severe tester takes the outcome s “success” as evidence of
144
Philosophy of Science Matters
Isaac’s readiness H1(I). Certainly she would deny s is evidence of his unreadiness (H0(I) doesn’t even “fit” the very high scores, s). But now imagine, says our critic, that Isaac was randomly selected from a population—call it Fewready Town—where the relative frequency of college-ready students is some very small number, say one out of ten thousand. The critic, applying Achinstein’s straight rule for priors, infers that the prior probability P(H1(I)) is small (e.g., 0.0001). But then, the critic continues, the posterior probability that Isaac is college-ready H1(I), given his high test results, would be very low (though the posterior probability has increased from the prior).8 There are two questions that arise: First, would frequentists accept the straight rule for assigning probabilities to hypotheses? Second, even if a frequentist would regard such a rule as fallacious, does the posterior thereby obtained yield the more intuitively adequate degree of evidential warrant? We say no to both. Fallacy of probabilistic instantiation Although the probability of randomly selecting a particular student from the population of high schoolers from Fewready Town is small, say 0.0001, it does not follow that Isaac, the one we happened to select, has a probability of 0.0001 of being college ready. To suppose it does is to commit what may be called a fallacy of probabilistic instantiation (Mayo 1997a, 1997b, 2003b, 2005, 117). Suppose the experiment is to randomly select a member of Fewready Town. To infer from the premises P(H1(x)) = very low (e.g., 0.0001) and x 0 = I (i.e., Isaac) to the inference:(*) P(H1(I)) = 0.0001.
is a fallacious instantiation that can easily give rise to contradictions. But Achinstein, in his recent update, grants this, so we turn to the second question that might be raised.9 The charge of committing a fallacy of probabilistic instantiation, he now allows, would be true if the probabilities in question were construed as relative frequencies. However, . . . I am concerned with epistemic probability. If all we know is that Isaac was chosen at random from a very disadvantaged population, very few of whose members are college ready . . . then we would be justified in believing that it is very [improbable] that Isaac is college ready [i.e., p(H1) is very low and hence p(H1/s) is very low; I replace his symbols with mine]. (Achinstein 2010, 187)
The Objective Epistemic Probabilist and the Severe Tester
145
Hence, (*) would give a legitimate objective epistemic prior. But does it yield a relevant degree of warrant for H1? Even confronted with Isaac’s high test scores, s, Achinstein’s probabilist is justified in denying that the scores are good evidence for H1(I). Rather, they are evidence for believing H0(I)—that Isaac is not college ready. Whether Achinstein merely denies evidence for readiness or also goes on to make the stronger claim, that the high scores s are evidence for H0 non-readiness, there is still the conflict he wishes to describe. (In the Stand and Deliver clarification he seems clearly to be making the stronger claim. See Section 3 of this essay.) But is this a problem for the epistemic probabilist or the severe tester? Severe testers always evaluate an inferential report, regardless of the procedure that generated it, by considering its relevant error probabilities, without which they cannot assess severity. Given a Bayesian report of a high posterior degree of belief, say 0.95, in a hypothesis H1(I), a severe tester asks: how often would such a high assignment occur even if H is false? The example has only two outcomes: success and failure—s or ~s. Clearly getting ~s, the lower average grades, gives even less evidence of readiness; that is, P(H0(I)|~s) > P(H0(I)|s). Therefore, whether Isaac scored as high as s or lower, it appears that Achinstein’s epistemic probabilist denies there is evidence of readiness. The probability of Achinstein finding evidence of Isaac’s readiness even if in fact he is ready is minimal, if not zero, thereby violating the weak severity requirement! If a procedure would very often instruct us to infer or believe in H, even if H is false, then its inferring H does not seem to provide an objective warrant for the truth of H. To take this example further: Suppose in fact Isaac was selected randomly, not from Fewready Town, but from a population where college readiness is common, Fewdeficient Town. The same average score s now warrants a strong objective epistemic belief in Isaac’s readiness. So students from Fewready High would need to score quite a bit higher on these same tests to have their scores be regarded as evidence for college readiness (than would students selected from Fewdeficient High). The concern here, of course, is not the political implications of such “reverse discrimination” but the way this consequence brings out the shortcomings of using the recommended posterior probability as a measure of what Isaac’s scores tell us about Isaac.
(3) Achinstein and Stand and Deliver Achinstein illuminates his analysis by reminding us of the movie Stand and Deliver where students in a poor area with a lot of student dropouts
146
Philosophy of Science Matters
pass the advanced placement tests in math with flying colors, having been drilled by a caring high school math teacher (I never saw the movie). Officials from the Educational Testing Service are suspicious, and insist that the students retake the test “under the watchful eyes of their own representatives.” Says Achinstein, [G]iven that all they knew were the test results and the background of the students, were the officials from the E.T.S. justified in believing that, despite the high test results, it is unlikely10 that the students have a good knowledge of basic calculus? I think that this is a reasonable epistemic attitude. They assigned a very low prior epistemic probability to the hypothesis that these students know calculus, and hence a low posterior probability to this hypothesis given the first test results. We may suppose that after the results of the second test, conducted under more stringent testing conditions, the posterior probability of the hypothesis was significantly high. (Achinstein 2010, 187–8, my emphasis)
There are two key points that Achinstein obviously intends as relevant to his position on our college-readiness exchange, even though here H1 refers to the (math) ability of a group of students. Yet they seem deeply puzzling. First, to raise the possibility of cheating would be to question the assumption on which the posterior probability-severity conflict is based. It is given in the college-readiness example that “the test is severe in the sense that passing it is very difficult to do if one is not ready for college” (Achinstein 2010, 186). The example we are supposed to entertain just is the second test (where assumptions hold). Second, to suppose that the results of the second test now yield a high posterior probability to the group’s ability sidesteps the issue at hand. For however small the likelihood for readiness is with the second test’s super high scores, s*, the critic adduces a small enough prior for readiness so as to yield a low posterior. Then, the original question that Achinstein asks the severe tester would be re-asked of him: Are you prepared to take s* as evidence of H1 even though the posterior epistemic probability of H1 (given s*) is low? Is Achinstein now intimating that he is leaning against the low posterior probability to H1(I) given the very high scores s*, even given that Isaac was randomly selected from what we might call Veryfewready Town? Might he even agree, further, that the very high scores s* are evidence for rejecting the prior as a relevant measure of evidence? When we move from hypotheses like “Isaac is college ready” to more realistic generalizations, the problem of obtaining priors by means of something like Achinstein’s straight rule is even more pronounced, but there is no space to describe them here (see Mayo 1997a, 1997b, 2010, 194–99).
The Objective Epistemic Probabilist and the Severe Tester
147
5. MILL AS SEVERE TESTER Perhaps the most interesting feature of Achinstein’s most recent discussion is to be found in his defense of Mill’s account of induction: Although “one of Mill’s definitions of an inductive inference [is] one in which ‘we conclude that what is true of certain individuals of a class is true of the whole class’ . . . he does not say, and indeed he explicitly denies, that any induction so defined is justified. . . .” Achinstein (2010, 173). He quotes Mill: When a fact has been observed a certain number of times to be true, and is not in any instance known to be false; if we at once affirm that fact as an universal truth or law of nature, without ever testing it by any of the four methods of induction [Mill’s famous methods for establishing causal laws], or deducing it from other known laws, we shall in general err grossly. (Mill 1888, 373)
To avoid erring grossly, it is necessary to run empirical checks of “different sorts of ‘fallacies’ or mistakes in reasoning, including failure to look for negative instances and making the generalization broader than the evidence allows” (Mill 1888, 373). Whether such flaws and fallacies are avoided “depends on non-formal empirical facts regarding the sample, the sampling, and the properties being generalized. I would take Mill to be espousing at least some version of the idea of ‘severe testing’ ” (Achinstein 2010, 174). So would I. But what about the central point on which Achinstein and I are said to disagree? To my surprise, Achinstein declares, speaking about both Mill and Newton, that Neither in their abstract formulations of inductive generalizations . . . nor in their examples of particular inductions to general conclusions of the form "all A’s are B’s," does the term "probability" occur. Both write that from certain specific facts we can conclude general ones—not that we can conclude general propositions with probability, or that general propositions have a probability, or that they have a probability conditional on the specific facts. From the inductive premises we simply conclude that the generalization is true, or . . . "very nearly true," by which [Newton] appears to mean not "probably true," but "approximately true." (Achinstein 2010, 176)
Now I am certainly no Mill scholar, and I am glad Achinstein has set out “to give Mill a better run for his money.”11 Adhering to Achinstein’s description of Mill, there is no inclination to foist a posterior probability on inductive generalization H. Background considerations enter not to assign prior probabilities to an exhaustive list of hypotheses (e.g., about the proportion of A’s that are B’s in a population). Rather, there is an
148
Philosophy of Science Matters
appeal to empirical considerations both to avoid as well as to check, a set of classic flaws and foibles that, if committed and not compensated for, would invalidate the inductive move. Far from wishing to justify the familiar inductive “straight rule,” Mill appears to be saying that an induction following this pattern would be warranted just when the test did a good job of ruling out the ways we can “err grossly” in going from the sample correlation to the population. Contemporary experimenters have a more sophisticated set of tools at their disposal, but the overarching reasoning is the same. From Achinstein’s own enlightening depiction, I too “would take Mill to be espousing at least some version of the idea of ‘severe testing.’ ” Having done a plum job giving evidence of this, however, Achinstein promptly discards the evidence, and instead converts Mill into an epistemic probabilist of the Bayesian kind. I will not get into the acrobatics required for Achinstein to convert Mill’s remarks about probabilities of events so that Mill may be seen as assigning probabilities to inductive hypotheses. What reason is there to convert Mill thus, aside from insisting that a way would be found to interpret Mill as fitting a certain Bayesian philosophy (thereby violating the minimal severity principle)?
6. CONCLUDING REMARKS Achinstein is critical of the severity account because it is possible for H to pass severely even though the posterior probability of H might be low. I argue that the priors involved in such “posterior probability-severity” conflicts are not kosher for a frequentist. Further, in no case where this conflict occurs does the posterior probability seem to get the intuitively correct appraisal of reasonableness to believe. Where severity disagrees with a posterior, it is more reasonable to regard the evidence as grounds to reject the prior probability (or one of the test assumptions). In defending Mill, Achinstein gives further evidence that I am not entirely alone in viewing inductive inference as nonBayesian.
REFERENCES Achinstein, P. 2010. Mill’s Sins or Mayo’s Errors? In Error and Inference: Recent Exchanges on Experimental Reasoning, Reliability, and the Objectivity and Rationality of Science, ed. D. Mayo and A. Spanos. Cambridge: Cambridge University Press.
The Objective Epistemic Probabilist and the Severe Tester
149
——— , ed. 2005. Scientific Evidence: Philosophical Theories and Applications. Baltimore, Md.: Johns Hopkins University Press. ——— . 2000. Why Philosophical Theories of Evidence Are (and Ought To Be) Ignored by Scientists. Philosophy of Science 67 (3): 180–92. Howson, P. 1997a. A Logic of Induction. Philosophy of Science 64 (2): 268–90. ——— . 1997b. Error Probabilities in Error. Philosophy of Science 64 (4): 185–94. Kass, R. and L. Wasserman. 1996. The Selection of Prior Distributions by Formal Rules. Journal of the American Statistical Association 91 (435): 1343–70. Mayo, D. G. 2010. Sins of the Epistemic Probabilist: Exchanges with Peter Achinstein. In Error and Inference: Recent Exchanges on Experimental Reasoning, Reliability, and the Objectivity and Rationality of Science, ed. D. Mayo and A. Spanos. Cambridge: Cambridge University Press. ———. 2008. How to Discount Double-Counting when It Counts: Some Clarifications. British Journal for the Philosophy of Science 59 (4): 857–79. ——— . 2005. Evidence as Passing Severe Tests: Highly Probed vs. Highly Proved. In Scientific Evidence: Philosophical Theories and Applications, ed. P. Achinstein. Baltimore, Md.: Johns Hopkins University Press. ——— . 2003a. Could Fisher, Jeffreys, and Neyman Have Agreed on Testing? Commentary on J. Berger’s Fisher Address. Statistical Science 18 (203): 19–24. ——— . 2003b. Severe Testing as a Guide for Inductive Learning. In Probability Is the Very Guide in Life, ed. H. Kyburg. Chicago: Open Court. ——— . 2000. Experimental Practice and an Error Statistical Account of Evidence. Philosophy of Science 67 (3): 193–207. ——— . 1997a. Duhem’s Problem, The Bayesian Way, and Error Statistics, or “What’s Belief Got To Do with It?” Philosophy of Science 64 (2): 223–44. ——— . 1997b. Error Statistics and Learning from Error: Making a Virtue of Necessity. In Philosophy of Science 64 (PSA Symposia Proceedings), ed. L. Darden. Chicago: University of Chicago Press. ——— . 1997c. Response to Howson and Laudan. Philosophy of Science 64 (2): 323–33. ——— . 1996. Error and the Growth of Experimental Knowledge. Chicago: University of Chicago Press. Mayo, D. G. and D. R. Cox. 2010. Objectivity and Conditionality in Frequentist Inference. In Error and Inference: Recent Exchanges on Experimental Reasoning, Reliability, and the Objectivity and Rationality of Science, ed. D. Mayo and A. Spanos. Cambridge: Cambridge University Press. Mayo, D. G. and A. Spanos. 2006. Severe Testing as a Basic Concept in a Neyman– Pearson Philosophy of Induction. British Journal for the Philosophy of Science 57 (2): 323–57. ——— , eds. 2010. Error and Inference: Recent Exchanges on Experimental Reasoning, Reliability, and the Objectivity and Rationality of Science. Cambridge: Cambridge University Press. Mill, J. S. 1888. A System of Logic, 8th edition. New York: Harper and Bros. Peirce, C. S. 1931–1935. Collected Papers, ed. C. Hartshorne and P. Weiss. Cambridge, Mass: Harvard University Press. Woodward, J. 2000. Data, Phenomena, and Reliability. Philosophy of Science 67 (3): 163–97.
150
Philosophy of Science Matters
NOTES 1. The other participant in this Symposium was Jim Woodward (see Woodward 2000). 2. In the most sophisticated accounts now in use, the prior probability assignments are based on information-theoretical systems, and are specifically not intended to represent degrees of belief, or even probabilities. See Kass and Wasserman 1996 and Mayo and Cox 2010. 3. It does not matter which we make H0 and which H1 in this account. I choose to make H0 the hypothesis of “inadequacy” or “unreadiness” in both examples 1 and 2, respectively, for ease in reading. They must, of course, be denials of each other. 4. For a discussion of the relevance of error probabilities, see Mayo and Cox 2010 and Mayo 2008. 5. The testers might qualify “college-readiness” to mean simply mastery of high school material, to put to one side any controversy about whether such mastery is a valid indication of “college-readiness.” 6. Except for very special cases. 7. Likelihoods do not obey the probability axioms; for example, the sum of the likelihoods of a hypothesis and its denial is not one. 8. There will not always be such an increase in the examples used in this type of criticism! In classic examples, statistically significant results against a null hypothesis H0 are shown to correspond to results accorded a high Bayesian posterior; moreover, the posterior exceeds the prior. 9. To clarify, consider a case where a legitimate frequentist prior might be possible. Conceivably, various genetic factors might allow computing that the (frequentist) probability that Isaac would have brown hair is, say, 0.6. Compare that to an experiment of randomly selecting from a population where 10% have brown hair, and drawing Isaac. The probability is 0.1 of drawing a student with brown hair. But the (frequentist) probability that the student we drew— Isaac—has brown hair is 0.6. 10. The use of “unlikely” here means “improbable.” 11. It would follow that C. S. Peirce was incorrect to accuse Mill of omitting the two key rules of inductive inference: predesignation and random sampling (Peirce 1931–1935, 1.95 N).
12 Achinstein and Whewell on Theoretical Coherence Gregory J. Morgan
1. INTRODUCTION In Particles and Waves, Peter Achinstein gives a precise probabilistic version of theoretical coherence inspired by William Whewell’s somewhat vague notion of coherence (Achinstein 2010). Whewell believed that as theoretical science proceeds, it becomes more coherent and rejects false incoherent theories. Achinstein offers a challenge: try to make Whewell’s idea more precise while maintaining the properties that Whewell claimed coherence to have. I will argue (1) that Achinstein’s probabilistic rendition of coherence fails to capture Whewell’s notion, since the probabilistic rendition of coherence is not an a priori sign of truth, and (2) that Achinstein’s approach is better seen as a critique of Whewell’s central methodological claims than as an interpretation of Whewell’s ideas.
2. WHEWELL ON COHERENT COHERENCE William Whewell, in his classic Philosophy of the Inductive Sciences, analyzes the progress of science: . . . we have to notice a distinction which is found to prevail in the progress of true and false theories. In the former class all the additional suppositions tend to simplicity and harmony; the new suppositions resolve themselves into the old ones, or at least require only some easy modification of the hypothesis first assumed: the system becomes more coherent as it is further extended. The elements which we require for explaining a new class of facts are already contained in our system. Different members of the theory run together, and we have thus a constant convergence to unity. In false theories, the contrary is the case. (Whewell 1847, 233, my emphasis)
151
152
Philosophy of Science Matters
While Whewell illustrates his claim with numerous episodes in the history of science, he believes that one can know a priori that coherence is a sign of truth. “Truth may give rise to such a coincidence; falsehood cannot” (71). For Whewell, truth is coherent and simple (72). False hypotheses cannot explain all phenomena, whereas true hypotheses can (62). For him, the true philosopher, if successful, “binds together” otherwise incoherent facts like stringing a set of pearls, to use his simile. In a claim that arguably conflates psychological inability and logical impossibility, Whewell goes so far as to say that once a system of facts have been bound together, it is often impossible to see the facts as “incoherent and detached” (52). Robert Butts claims that Whewell views facts thus bound together as necessary and known by intuition (Butts 1968, 16). Indeed, consistent with his neo-Kantianism, Whewell sees the connection between coherence and truth as necessary and a knowable a priori. Whewell’s notion of coherence is closely related to his better-known notion of consilience. For Whewell, a consilient theory correctly predicts unobserved phenomena of a kind different than it was designed to explain. In so doing, a consilient theory unifies different classes of phenomena (Whewell 1847, 230). Let us make his idea more precise. Assume there are two classes of observable phenomena, O1, . . . On and On+1, . . . On+m. Imagine scientist S constructs a theory T1 that entails O1, . . . On and On+1, . . . On+m, but does not believe that On+1, . . . On+m are relevant in formulating her theory. In this case, T1 would be a consilient theory. Whewell’s notion of consilience has both an objective dimension and a subjective dimension. If observable phenomena O1, . . . On and On+1, . . . On+m are different kinds of phenomena, then this fact does not depend on whether anyone believes that this is the case. On the other hand, whether S considers or contemplates certain phenomena in formulating her theory depends trivially on the propositional attitudes of S, and thus Whewell incorporates a subjective component into his notion of consilience. A theory is coherent if, as new kinds of phenomena are observed, the theory explains them without having to introduce ad hoc assumptions, that is, if it repetitively unifies new kinds of phenomena with little or no modification. Thus coherence, like consilience, depends on the historical development of the theory in question. If a theory has a history of successful consilience, then it is coherent.
3. ACHINSTEIN’S SANITIZED VERSION OF COHERENCE Achinstein (1991), in attempting to make Whewell’s notion more precise, removes the subjective component from coherence giving what I call
Achinstein and Whewell on Theoretical Coherence
153
a sanitized version of coherence. First, he stipulates that a theory T consists of a set of hypotheses, h1, h2, . . . , hm. He then defines what it is for a hypothesis h1 to be coherent with the remaining hypotheses. h1 is coherent with h2, . . . , hm on B, if and only if p(h1/h2, . . . hm & B) > k, and p(h1/h2, . . . , hm & B) > p(h1/B) (Achinstein 1991, 130).
Then he introduces the notion of coherence for the set of hypotheses. A set of hypotheses h1, . . . , hm is coherent, on B, if and only if each hypothesis is coherent with the other members of the set on B (130).
Finally, Achinstein defines coherence for a theory: theory T is coherent if and only if the set of hypotheses, h1, . . . , hm is coherent. Let us call the person who claims, like Whewell, that coherence is a sign of truth a coherentist. This central question of this chapter is whether one can adopt Achinstein’s sanitized version of coherence and also be a coherentist.
4. TWO CONCEPTIONS OF SIGN OF TRUTH As it is commonly used, the notion “sign of truth” is ambiguous and vague. Restricting our focus to usage in theoretical science, let me distinguish two distinct senses: one usage of the term suggests that possession of a sign of truth at least makes a theory (T) likely and another usage suggests that possession of a sign of truth makes a theory more likely. First there is an absolute notion or threshold sign of truth that I will call a sign1 of truth: If coherence is a sign1 of truth, then, ∀T, p(T/T is coherent) > k, where k ³ 0.5.
The consequent is not a sufficient condition since it could be the case that p(T/~(T is coherent)) = p(T/T is coherent) > k, in which case coherence is irrelevant to the probability of a theory, but T is nonetheless likely. Presumably, adding the condition p(T/~(T is coherent)) £ k to the above necessary condition would generate necessary and sufficient conditions for coherence being a sign1 of truth. One weakness of the necessary condition is that it is undefined when p(T is coherent) = 0. Notice that the existence of refuted coherent theories do not necessarily undermine the coherentist’s claim as in that case one is often considering a different conditional probability, that is, p(T/T is coherent and there is signification evidence against T) < k. If coherence were an infallible sign1 of truth, the value of k would be 1. To be charitable, I will consider cases where coherence is not an infallible sign of truth, that is, cases where k<1. Since Achinstein is concerned with
154
Philosophy of Science Matters
what one should believe, I will consider this type of sign when k ³ 0.5. If k < 0.5, then it is more reasonable to believe ~T than to believe T. There is also a weaker notion of a sign of truth. While sign1 of truth is absolute, the second notion of sign of truth, which I call a sign2 of truth, is relative: Coherence is a sign2 of truth if and only if, given two theories that are identical in all epistemic properties except that T1 is coherent and T2 is not, then p(T1) > p(T2).
The idea here is that, by possessing coherence, a theory’s probability is increased. There can be relative signs that are not threshold signs. For example, a sign could increase the probability of finding the truth without reaching the threshold k.
5. FOUR DISTINCT POSITIONS One can use the empirical/ a priori distinction to create four distinct positions that coherentists might occupy. To illustrate these positions, let us consider a sign1 of truth, although similar considerations also apply to a sign2 of truth. The core inference that a coherentist endorses for a particular coherent theory T1 is the following: Premise A: ∀T, If T is coherent, then there is a reason to believe that it is true. Premise B: Theory T1 is coherent. Conclusion: Therefore, there is a reason to believe that T1 is true.
For each of the premises, one may ask the question: is it an empirical or an a priori claim? A given coherentist’s answers to these questions classifies her into one of four positions. Position 1: Both premises A and B can be known a priori. A particular sanitized interpretation of William Whewell puts him in the rationalist coherentist camp. On this view of Whewell, it is an a priori truth that there is a connection between coherence and truth and one can tell a priori whether a given theory is coherent. Position 2: Premise A expresses an a priori truth, but whether a particular theory is coherent or not is an empirical matter. A more orthodox interpretation of William Whewell puts him in this camp. Position 3: Premise A is empirical, but it is an a priori matter whether a particular theory is coherent. As with Position 1, on this view one need not perform any worldly investigation to determine if a given theory is coherent.
Achinstein and Whewell on Theoretical Coherence
155
Position 4: Both premises A and B are empirical claims. Those who hold that the general principle is empirical (Positions 3 and 4) often justify their belief inductively. For example, if coherent theories in the past have been likely, then they argue we have some reason to think that current coherent theories are likely.
6. IS COHERENCE A SIGN 1 OF TRUTH? The ardent coherentist claims that coherence is a sign1 of truth, a necessary condition of which is that p(T/T is coherent) > 0.5. If coherence were a sign1 of truth and a theory’s coherence were determinable without empirical inquiry, then it would be possible to justify a belief in a coherent theory without testing it experimentally. This would be a significant achievement, but I will argue it is a mirage. Satisfying the necessary condition of a sign1 of truth is a tall order. Consider a competing theory T*, such that ~(T & T*). If T* is also coherent, then it cannot be the case that p(T/T & T* are coherent) > 0.5 and p(T*/T & T* are coherent) > 0.5 since the probability of mutually exclusive theories cannot sum to greater than 1. In some cases, we can construct a competing T* from the components of a coherent T. Assume that T (= {h1, h2. . . . hm}) is coherent. From Achinstein’s definitions it follows that p(h1/h2, . . . , hm & B) > p(h1/B). Whewell claims that falsity cannot exhibit coherence. However, consider the theory T* that consists of {~h1, ~(h2 &, . . . , hm)}. Given plausible assumptions, T* is also coherent. (See Appendix 1.) At least one of T and T* must be false. In particular, even if T is as coherent as possible, that is, k = 1, then an incompatible competing theory is coherent also. The coherentist might argue that useful judgments of a theory’s probability are not made in a vacuum, but rather are made in the context of additional evidence and consequently one should consider the effect of coherence and the additional evidence. This complaint voices a legitimate point. The goal of using signs of truth is to learn what one should believe; any such judgment would likely be made on the basis of more evidence than merely the coherence of T. No one, to my knowledge, advocates that scientists use the internal coherence of a theory as the sole criterion of theory acceptance. To make the coherentist’s position as cogent as possible, let us consider the role of coherence in the belief in empirically adequate theories. I will assume that T is empirically adequate with respect to evidence E if and only if T entails E. It follows that, for any T, if T is empirically adequate with respect to E, then p(T/E & T is coherent) ³ p(T/T is coherent). If one can show that even in cases of empirically adequate theories, p(T/E & T is coherent) ³ 0.5
156
Philosophy of Science Matters
for some T and E, then it follows that there exists a T such that p(T/T is coherent) < 0.5, that is, that coherence is not a sign1 of truth. This result should not be too surprising, since if coherence does not render an empirically adequate theory rationally believable, then it is implausible that it renders an empirically inadequate theory rationally believable.
(1) The problem of coherent rival theories The coherentist faces an uphill battle to show that coherence is a sign1 of truth. John Earman has presented what might be called the problem of rival theories (Earman 1994). He shows that if we assume that there exists a rival theory T* that has at least as high prior probability as T, p(T*/B) ³ p(T/B), then no matter how many more confirming observations we make (in the sense that both T and T* individually entail O1 & O2, . . . ,On), then p(T/O1 & O2, . . . ,On & B) £ 0.5 (where B stands for “background knowledge,” and Oi stands for the ith observation). Thus, we cannot infer a likely theory from the observable phenomena unless we “load the dice” against rival theories, to use Earman’s phrase. Actually, the situation is even worse. One can generalize Earman’s result: if we assume in addition to T, m-1 mutually incompatible theories that exhaust all the remaining possibilities, and if these theories also save all the phenomena, then p(T/O1 & O2, . . . ,On & B) ≤ 1/m (See Appendix 2). Are the dice loaded against incoherent theories? It is a popular proposal to claim that if non-empirical factors play a role in theory confirmation, they influence the prior probability. Wesley Salmon, for example, claims that simplicity and symmetry have a “significant bearing” on the prior probability of a theory (Salmon 1996, 283). A coherentist might suggest that an empirically adequate theory with the most coherence has a higher probability than its less coherent rivals. Does this suggestion work? Unfortunately one can construct a variant of Earman’s argument that I call the problem of coherent rival theories. There is no reason to think that a theory with the most coherence will be unique. Nothing about coherence requires that there is only one most coherent theory. There could be multiple theories with the maximal amount of coherence. Given this possibility, there is good reason to think that an a priori proof of uniqueness is impossible. If there is a rival theory T* that also saves the phenomena and is equally coherent, then by analogous argument, p(T/B) = p(T*/B) and p(T/B) £ 0.5, since T and T* are incompatible. At this point the coherentist might shift gears. Perhaps she concedes that Earman’s theorem suggests that in the long run the coherence of a theory cannot guarantee that its probability will not converge to a number above 0.5. However, echoing John Maynard Keynes, she might suggest
Achinstein and Whewell on Theoretical Coherence
157
that what happens in the long run is irrelevant to what one should believe now, given current evidence. It still could be the case that current evidence is such that it is now rational to believe that coherent theories are likely. It is to this question that I now turn. In its simplest form, Bayes’s Theorem shows the three components of a theory’s probability: p(T/E) = p(T)× p(E/T)/p(E) = prior probability× likelihood/expectedness
A slightly more complicated version of the theorem includes a term for background knowledge B. To consider the role of the coherence of T, consider background knowledge B that includes “T is coherent.” p(T/E&B) = p(T/B) p(E/T&B)/p(E/B)
The probability of a theory depends on three things: the prior probability, the probability of the evidence given the theory (that is, the likelihood), and the probability of the evidence. The denominator, the probability of the evidence, p(E/B), sometimes called the expectedness of the evidence, depends upon E and B and is the same for all theories competing to account for the same phenomena regardless of whether they are coherent. If the coherence of a theory has any effect upon the theory’s plausibility, it must either affect the prior probability or the likelihood.
(2) Does coherence influence the likelihood of a theory? What then of likelihood, the probability of the evidence given background knowledge and that the theory is true, that is, p(E/T&B)? Often this expression is taken as one measure of the explanatory power of the theory. The problem with linking coherence to likelihood is that we can conceive of incoherent theories having a high likelihood. An incoherent theory may even entail the evidence E, in which case the likelihood p(E/T&B) = 1. Indeed, if we frame the problem confronting the theoretician to be a choice among empirically adequate theories as we did earlier, then the likelihood is set to 1 for all candidates, coherent or not.
(3) Does coherence influence the prior probability of a theory? Given their name, one might think that all prior probabilities are determined a priori. However, we should distinguish between two types of priors: (1) empirical priors and (2) a priori priors. A priori priors, symbolized p(T), are not conditional on any contingent background knowledge.
158
Philosophy of Science Matters
Because of the lack of empirical constraint, some, with good reason, argue that a priori priors are often undefined. The coherentist could argue that coherence provides the needed constraint. I am suspicious of this suggestion. Empirical prior probabilities, symbolized p(T/B), are typically conditional upon the background knowledge B and consequently will depend upon the nature of the background knowledge. As the saying goes, today’s priors were yesterday’s posteriors. Given the empirical nature of these prior probabilities, it is difficult to assign values from the philosopher’s armchair. Here I think the burden is on the coherentist to show that there are features of the background knowledge that would determine that the prior probabilities of coherent theories are high enough to guarantee that the posterior probabilities are greater than 0.5. The underlying intuition behind the coherentist’s assignment of probability is that a coherent theory has a higher probability than an incoherent theory. I have serious reservations about this intuition. Consider a logically weakened version of T1, called T1°, that is, assume T1 entails T1°, but T1° does not entail T1. If T1 is coherent or more coherent than T2 it does not follow that T1° is coherent or more coherent than T2. For example, T1° might contain only the probabilistically independent components of T1. The coherentist might be inclined to argue that if T1 is more coherent than T1°, then p(T1) > p(T1°). However, this claim violates the probability calculus since if T1 entails T1°, then by the axioms of probability p(T1) £ p(T1°). Is the procedure of assigning coherent theories (or more coherent theories) higher prior probabilities rational? It is not clear that any standards of rationality apply here. Normal ways of constraining rational belief, such as Dutch book arguments, do not dictate that coherent theories have higher probabilities than incoherent theories. Practically any assignment of prior probability, as long as the probabilities of mutually exclusive and exhaustive theories sum to one, is allowed. Perhaps one might claim that the coherentist’s strategy is rational in the sense that it violates neither deductive logic nor the probability calculus. But this claim is weak. There is no good sense in which the coherentist’s strategy of assigning prior probability is more rational than an innumerable number of alternative assignments. To see this point, consider a competing strategy, that of the anti-coherentist. The anti-coherentist thinks that less coherent theories should have higher prior probabilities than more coherent ones. He repudiates every relative assignment of prior probability that the coherentist makes. If T1 is more coherent than T2 and the coherentist asserts that p(T1) > p(T2), then the anti-coherentist asserts that p(T1) > p(T2). There is no a priori reason to privilege one approach over the other. The ardent
Achinstein and Whewell on Theoretical Coherence
159
coherentists disagree, but on what basis can they argue for a difference between the two symmetrical positions? It cannot be from experience, since we are considering a priori priors. This leaves some non-empirical means. There are additional reasons to think that coherence is irrelevant to the prior probability of a theory. Consider the following: let T = {h1, h2, h3}, p(h1) = p(h2) = p(h3) = c < k <1, and p(T) = p(h1 & h2 & h3) = c3. Let us assume that h1, h2, and h3 are mutually statistically independent of each other and consequently T is not coherent. Consider another theory T* = {h1*, h2*, h3*} where h1* = h1 & h2, h2*= h2 & h3, and h3* = h3 & h1. In this case p(h1*/h2* & h3*) = p(h2*/h1* & h3*) = p(h3*/h1* & h2*) = 1 > k. Therefore, T* is coherent. Is p(T*) > p(T)? No, the two theories actually express the same propositions—the only real difference between them is that T* is coherent and T is not. This type of counterexample can be extended to more complicated cases in which each hypothesis is not entailed by the conjunction of the rest. To generalize, coherence appears to be a property of the way the theory is represented, not a property of the content of a theory, and so can vary when we vary the way of representing the same facts. If the coherentists could provide an a priori proof of their principle, they would be convincing, but no proof is forthcoming. To bias our a priori probabilities in favor of coherence is at best a metaphysical article of faith, which either unjustifiably biases our beliefs in favor of coherence or divorces probability assignments from belief. That we must privilege the a priori priors of coherent theories is not a methodological necessity either. On could argue that science progresses as efficiently, or perhaps even more efficiently, without a bias toward coherence. If I am correct, this leaves the coherentist’s defense of the epistemic role of coherence in a precarious position and no more rational than competing approaches. Whether the coherentist strategy works for empirical prior probabilities is an open question.
7. IS COHERENCE A SIGN 2 OF TRUTH? If the previous argument was successful, then the coherentist cannot successfully defend the view that the coherence of a theory is a sign1 of truth without appeal to empirical evidence that philosophers are ill suited to garner. At this point in the debate, the coherentist might retreat to the weaker position that coherence is a sign2 of truth. I will argue that the coherentist cannot successfully defend this weaker position on a priori grounds either.
160
Philosophy of Science Matters
Consider two competing mutually exclusive and empirically adequate theories T1 and T2 that are comparable on all epistemic respects, except that T1 is coherent and T2 is not. If coherence is a sign2 of truth, then p(T1/E&B) > p(T2/E&B). If one can show that the inequality does not hold then by modus tollens, coherence is not a sign2 of truth. One advantage of this comparative approach is that we no longer need to consider the expectedness of the evidence p(E/B). From Bayes’s theorem it follows from the consequent that p(T1/B) p(E/T1&B)/p(E/B) > p(T2/B) p(E/T2&B)/p(E/B). The denominators cancel, leaving the following inequality that the coherentist must prove: p(T1/B) p(E/T1&B) > p(T2/B) p(E/T2&B). Rearranging we get P(T1/B)/P(T2/B) > p(E/T2&B)/p(E/T1&B).
Thus it is the relative value of the ratios of priors and likelihoods of the two theories that determine the truth of the coherentist’s claim. There are four ways in which this inequality could be true. (1) Driven by the priors: The coherence of T1 increases the prior probability of T1 over T2 to the degree that differences in the likelihood are irrelevant. (2) Driven by the likelihoods: The coherence of T1 increases the likelihood of T1 over T2 to the degree that differences in the priors are irrelevant. (3) Driven by the priors and likelihoods: the coherence of the theory increases both the prior and likelihoods. Possession of coherence increases both the priors and the likelihoods. (4) Coherence has no systematic effect on the priors or likelihoods individually, but nonetheless p(T1/B) p(E/T1&B) > p(T2/B) p(E/T2&B). I will deal with the four cases in reverse order, ending with the most plausible case. The fourth case is the least likely to occur. For it to obtain, when the prior probability T1 is less than T2, the likelihood of T1 must be greater than T2 by a greater amount than T1 is less than T2. Furthermore, where the prior probability of T1 is greater than T2, the likelihood of T1 cannot be even greater than the likelihood of T2. That this is the case is implausible and I can think of no mechanism that would guarantee that these relations exist. The mechanism would have to increase a coherent theory’s priors when it decreases the likelihoods and vice versa. Arguments against case 3 will be similar to arguments against cases 1 and 2. If any of my arguments against case 1 and 2 are successful, they will also be successful arguments against case 3.
Achinstein and Whewell on Theoretical Coherence
161
Case 2 concerns signs2 of truth that influence the posterior probability of a theory by increasing the likelihood of the coherent theory, p(E/T1&B), over a rival, p(E/T2&B), without the alleged sign of truth. Similar remarks apply to consideration of likelihoods as those mentioned in Section 5.2. First, if we are considering empirically adequate theories, T1 and T2, then the likelihoods are equal to 1 and consequently they are equal for the theory with the alleged sign of truth and the theory without it. Second, if we have a conception of coherence that necessarily involves making the evidence less surprising, then the notion of coherence risks becoming relativized to a set of evidence. Theories can be coherent without accounting for the evidence and judgments of a theory’s coherence, and do not depend upon which evidence one considers. I see no good reason to assume that coherence systematically increases the likelihoods of a theory. Case 1 is the most interesting of the four cases. As mentioned above, the prior probability of the theory is the least implausible component of a theory’s probability where coherence is alleged to have an effect on the probability of a theory. Does a coherent theory have a higher prior probability than a non-coherent one? If so, this fact is either a priori or empirical. In Section 6.3, I have argued that it is not a priori that coherent theories have higher prior probabilities. Not all is lost for the coherentist, however. A more plausible position is to argue that whether a sign2 of truth exists is an a posteriori question to be settled with empirical evidence. If this is the case, then scientists and historians of science would more easily provide the appropriate evidence, if it exists, than philosophers of science.
8. CONCLUSION What should one say about the relation between coherence and truth? On my view, a general probabilistic connection between coherence and truth (or falsity) would be a contingent connection. Here I disagree with Whewell. Discovering this connection would be an empirical discovery. If this general fact were part of the background knowledge, and if T1 is a coherent theory, then the general connection is relevant to the probability of T1. If it is knowable a priori that T1 is coherent, then adding that T1 is coherent to the background knowledge does not increase the probability of T1. In an important sense, the background knowledge already contains this (and all other) a priori facts. However, even in this case, it does not follow that coherence is irrelevant to the believability of theories since whether there is a general empirical connection between coherence and truth is relevant. Consider the general claim that most coherent theories
162
Philosophy of Science Matters
are true. Whether this is part of the background knowledge does affect the probability of the theory: p(T1/B & most coherent theories are true) > p(T1/B & most coherent theories are false). If the coherence of a theory is determinable a priori, coherence is not a sign2 of truth, but it is possible that coherence is relevant to a theory’s probability. Whether it is relevant, that is, whether there is a general connection between coherence and truth, is an empirical question—one better answered by an empirical survey than conceptual analysis. Achinstein’s sanitized version of coherence fails to capture Whewell’s notion since the probabilistic rendition of coherence is not an a priori sign of truth. His approach is better seen as a critique of Whewell’s central methodological claims than as an interpretation of Whewell’s ideas.
APPENDIX 1 Part of the proof is straightforward: we can show that ~h1 is coherent with ~(h2, . . . hm) on B. Here is the proof: p(h1/h2, . . . , hm & B) > p(h1/B) by assumption p(h2, . . . , & hm/h1 & B) > p(h2, . . . , & hm/B) by Bayes’s theorem 1 - p(h2, . . . , & hm/h1 & B) > 1 - p(h2, . . . , & hm/B) p(~(h2, . . . , & hm)/h1 & B) > p(~(h2, . . . , & hm)/B) p(h1/~(h2, . . . , & hm) & B) > p (h1/B) by Bayes’s theorem 1 - p(h1/~(h2, . . . , & hm) & B) > 1 - p (h1/B) p(~h1/~(h2, . . . , & hm) & B) > p (~h1/B) To complete the proof of ~h1 is coherent with ~(h2, . . . , & hm) on B, we need to show that p(~h1/~(h2, . . . , & hm) & B) > k. I cannot offer a proof of this claim, but using Bayes’s theorem again one can show that p(~h1/~(h2, . . . , & hm) &B) > [p(~h1/B) – p(~(h2, . . . , & hm)/B) - kp (~(h2, . . . , & hm)/B)]/p(~h1/B) follows from the coherence of T. As k ® 1, then [p(~h1/B) – p(~(h2, . . . , & hm)/B) + kp(~(h2, . . . , & hm)/B)]/p(~h1/B) ® 1 and p(~h1/~(h2, . . . , & hm) &B) ® 1. Thus it is possible that, if T is coherent on B, then ~h1 is coherent with ~(h2, . . . , & hm) on B. Next we must consider the second part of Achinstein’s definition. One needs to show that ~(h2, . . . ,&hm) is coherent with ~h1 on B. Using an analogous argument it is possible that p(~(h2, . . . , &hm)/~h1) >k. Finally, one must show that p(~(h2, . . . , &hm)/~h1 & B) > p(~(h2, . . . , hm)/B) and thus show that ~(h2, . . . ,&hm) is coherent with ~h1 satisfying the second part of Achinstein’s definition. The last step follows deductively:
Achinstein and Whewell on Theoretical Coherence
163
p(~h1/~(h2, . . . ,&hm) &B) > p(~h1/B) p(~h1/B)p(~(h2, . . . , &hm)/~h1 & B)/p(~(h2, . . . , hm/B) > p(~h1/B) by Bayes’s Theorem p(~(h2, . . . , &hm)/~h1 & B) > p(~(h2, . . . , hm)/B)
APPENDIX 2 Proof: let a theory T exist, which in conjunction with background knowledge B entail O1, O2, . . . , On. Claim: assume that p(T/B) < 1/m. If (i) T1, . . . ,Tm–1 also entail O1, . . . On, and (ii) B entails ¬ (Ti & Tj) where i ¹ j, and (iii) ∀i (p(Ti/B) ³ p(T/B)) and (iv) p(T Ú T1 Ú T2 Ú . . . Ú Tm–1/B) = 1, then for any n, no matter how large p(T/O1 & O2, . . . , & On & B) £ 1/m. From Bayes’ theorem and (i) one can derive the following equality p(T / O1 & On & B) p(T / B) = p(T1 ∨ T2 ∨ Tm −1 / O1 & On & B) p(T1 ∨ T2 ∨ Tm −1 / B)
Now we know p(T/B) < 1/m, by assumption, and by (iv) p(T1 Ú T2 Ú ...Ú Tm–1/B) ³ (m–1)/m. Assume that the claim is false, i.e., p(T/O1 & O2, . . . & On & B) > 1/m, then p(T1 Ú T2 Ú . . . Ú Tm–1/O1 & O2, . . . & On & B) < (m–1)/m, and the above equality cannot hold. REFERENCES Achinstein P. 1991. Particles and Waves: Historical Essays in the Philosophy of Science. New York: Oxford University Press. Butts, R. 1968. William Whewell’s Theory of Scientific Method. Pittsburgh: University of Pittsburgh Press. Earman, J. 1994. Concepts of Projectibility and the Problem of Induction. In Grue!, ed. D. Stalker. Chicago: Open Court. Salmon, W. C. 1996. Tom Kuhn meets Tom Bayes. In The Philosophy of Science, ed. D. Papineau. Oxford: Oxford University Press. Whewell, W. 1847. The Philosophy of the Inductive Sciences. London: John W. Parker.
13 Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist John D. Norton
1. INTRODUCTION For several years, through the “material theory of induction,” I have urged that inductive inferences are not licensed by universal schemas, but by material facts that hold only locally (Norton 2003; 2005). My goal has been to defend inductive inference against inductive skeptics by demonstrating how inductive inferences are properly made. Since I have always admired Peter Achinstein as a staunch defender of induction, it was a surprise when Peter identified me as one of the skeptical enemies in “The War on Induction” (Achinstein 2010). Peter reproached me for “taking on” his inductive heroes, Newton and Mill, and their celebrated rules of inductive inference. That my project could lead me to become a foe of induction was unimaginable. Or it was, until I began an analysis of a problem in philosophy of physics, whose elaboration is the purpose of this note. I wanted to endorse certain inductive inferences whose cogency seems unassailable: we have never seen certain pathologies in spacetime, so inductive inference should assure us that we never will. However I was unable to display the facts that license these inferences. After some reflection, I believe this problem shows that induction has less reach than we thought. There are problems we expect inductive inference to solve but that it cannot solve. This admission is no failure of the material theory of induction. No other approach to inductive inference fares any better with them. I will argue below that attempts to ground the inductive inferences in universal inductive principles founder on both the vagueness of the principles and the tendency of the applicable principles to contradict each other. While
164
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
165
Peter has identified me as a foe of induction, I doubt that his analysis will help these principled approaches. Contrary to my expectations, Peter’s recently published book chapter (Enter John Norton) does not mount a full-blown defense of Newton’s and Mill’s rules as a priori, even though he finds the rules to codify “a type of reasoning that is crucial to empirical science.” (Achinstein 2010) Rather, he agrees with me that empirical considerations do determine whether an inductive inference is good. An inference that merely has the familiar inductive form “that A is B, so all A’s are B” may fail to be good if the underlying facts are inhospitable. Indeed, the analysis of the problem presented here is a success for the material theory of induction, provided one is prepared to limit the reach of inductive inference in this and similar cases. For the material theory enables us to delineate which inductive problems are tractable and which are not. That decision reduces to whether we can identify appropriate warranting facts. Theories of inductive inference based on universal principles are unable to make the corresponding discrimination, for their universal principles must hold everywhere. A failure of inductive inference is for them inexplicable. The problem at issue concerns observationally indistinguishable spacetimes, described in the following section. In them, deductive inference cannot determine which spacetime is ours, no matter how extensive a portion of the spacetime is observed. In Section 3, I will argue that these results do not illustrate an underdetermination of theory by evidence, since they make no decision between competing theories, and they make little contact with the inductive considerations that must ground such a decision. Rather, in Section 4, I will describe how they exemplify a different sort of failure manifested by physical theories, a form of generic indeterminism in general relativity. In it, a specification of the observable past always fails to fix the remainder of a spacetime. While we may have no means to distinguish deductively among different cosmic futures in the cases considered in this literature, I will urge in Section 5 that we can pick among them with quite familiar sorts of inductive arguments whose cogency seems unassailable. Nonetheless, in Section 6, I will urge that these inductions are troubling in that they are what I shall call “opaque.” That is, we cannot see through the inductive inferences to an unproblematic warrant, whether it be in matters of principle or fact.
2. OBSERVATIONALLY INDISTINGUISHABLE SPACETIMES The existence of observationally indistinguishable spacetimes in general relativity was brought to the attention of philosophy of science by Clark Glymour (1977) and David Malament (1977). An observer at any event
166
Philosophy of Science Matters
in a spacetime is depicted as having full knowledge of everything in the temporal past of that event. The temporal past is that set of events from which ordinary processes, propagating at less than the speed of light, may reach the observer’s event. It may happen that exactly the same observable past arises somewhere in a second spacetime. The first spacetime of the observer is observationally indistinguishable from the second, if this finding is assured no matter where the observer may be in the first spacetime. Manchak (2009) proved that any well-behaved1 spacetime always has many geometrically distinct, nemesis spacetimes from which it is observationally indistinguishable. Moreover the nemesis spacetimes will be “locally” the same as the observer’s spacetime. In the first spacetime, one might have a condition that holds at each event, such as the Einstein gravitational field equations; or, more simply, a different condition that just asserts everywhere vanishing geometric curvature. The locality clause of the theorem assures us that the nemesis spacetimes will satisfy these same conditions. The theorem and its proof involve some sophisticated spacetime geometry. But the basic idea is simple. A loose analogy, shown in Figure 13.1, illustrates it. Imagine an ant on an infinite, flat (Euclidean) sheet of paper who can survey only the surrounding 10,000 square foot patch. No matter where the ant may be, it cannot distinguish its sheet from a nemesis sheet, which consists of a copy of the original sheet of paper rolled into a cylinder with a circumference of one mile.
Fig. 13.1. Ant on a sheet of paper
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
167
3. WHAT THE SIGNIFICANCE IS NOT: UNDERDETERMINATION OF THEORY The thesis of the underdetermination of theory by evidence, in its strong and interesting form, asserts that all evidence necessarily fails inductively to fix theory.2 The similarity in the terms “observationally equivalent theories” and “observationally indistinguishable spacetimes” requires some thought.3 Are the observationally indistinguishable spacetimes illustrations of the thesis of the underdetermination of theory by evidence? I will argue that they are not. I agree with the opening remarks of Manchak (2009). He notes the distinctness of his result from the skeptical thesis that acceptance of some particular scientific claim can be resisted in the face of all evidence by revision to background theories. The assurance of observationally indistinguishable spacetimes in general relativity fails to bear on the thesis on the underdetermination of theory by evidence in two ways. First, the indistinguishability does not pertain to theory. We are not presented, for example, with general relativity and some competitor theory, indistinguishable from it. Rather, what we cannot distinguish is whether this spacetime is the one of our observations or whether it is that one.4 Second, the indistinguishability asserts a failure of deductive inference, whereas the thesis of the underdetermination of theory by evidence asserts a failure of inductive inference. Many spacetimes are logically compatible with the fragment of spacetime that we observe. So deductive inference does not allow us to fix which of them is our spacetime. This failure is no failure of inductive inference, which routinely fares better with such problems. Deductive inference cannot assure us that our spacetime will continue to manifest the conservation of electric charge, even though it has been observed to do so without exception. The simplest inductive inferences can give us that assurance. Manchak’s theorem, however, is stronger. It does preclude particular sorts of inductive inferences from distinguishing the spacetimes. Our observable spacetime is four-dimensional and has a Lorentz signature metrical structure. We are allowed the inductive inference that this will persist in the unobserved part. More generally, we are allowed to infer inductively to the persistence of any local condition, such as the obtaining of the Einstein gravitational field equations, in both the observer’s and the nemesis spacetimes. These inductive inferences, the theorem shows, will still not enable us to separate the spacetimes, for both will agree on them. What is not shown, however, is whether other inductive inferences would enable us to separate the two spacetimes. It is essential to the
168
Philosophy of Science Matters
theorem that the observer’s spacetime and its nemesis are factually distinct. What needs to be shown is that these factual differences cannot be exploited by an inductive inference that can separate the two spacetimes. I will suggest below in Section 5 that apparently routine inductive inferences are capable of exploiting these factual differences to discriminate a spacetime from its nemesis. In Section 6, however, I will urge that this routine appearance is deceptive in that the warrants of these inductive inferences are unclear.
4. WHAT THE SIGNIFICANCE IS: A FORM OF INDETERMINISM The results on observationally indistinguishable spacetimes amount to this: we fix some part of the spacetime and, within the context of some physical theory like general relativity, the rest of the spacetime remains undetermined. This result is a form of indeterminism. Indeterminism in physical theories arises whenever the full specification of the present fails to fix the future. Indeterminism is routine in standard, collapsed versions of quantum theory. The full specification of the present state of a radioactive atom does not determine when it will decay in the future. Determinism arises commonly in the spacetimes of general relativity. For example, in the Robertson-Walker spacetimes of relativistic cosmology, selecting a single moment of cosmic time identifies a three dimensional surface in the four-dimensional spacetime that has the special property of being a "Cauchy surface." Having this property entails that, if we fix the spacetime geometry and matter fields of the universe on this surface, they are fixed by the theory for all times. The failure of determinism in quantum theory in the 1920s was shocking, since it implied a profound limit on what we could know about the future. It told us that no matter how much we knew about the present state of some suitably chosen quantum system, we could not know assuredly what it would do in the future. This kind of principled epistemic limit on what we can know is not captured well by seeking to implement determinism in terms of the Cauchy surface “nows” of cosmic time in relativistic spacetimes. For no observer can observe the entirety of one of these surfaces. Rather, what an observer can know at one moment is delimited better by the observer’s temporal past, even though it represents an excessively optimistic assessment of our observational abilities. Then the results on observationally indistinguishable spacetimes place powerful constraints on just
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
169
what can be inferred directly from our spacetime theories about the remaining unobserved parts of our spacetime. They tell us that, even with the assistance of local spacetime theories, we can never assuredly fix a unique extension to the portion we have observed. In this regard these results are the appropriate analog of the indeterminism of quantum theory.5 However, there is a strong disanalogy to the indeterminism of quantum theory. Both forms of indeterminism express an impossibility of the past deductively determining the future. They differ markedly when we consider the possibility of inductive determination of the future. While inductive discrimination is possible in both cases, as we shall see below, they employ different sorts of inductive inferences.
5. SOME COSMIC INDUCTIONS Inductive inferences can discriminate a spacetime from an observationally indistinguishable nemesis arising in the results on observationally indistinguishable spacetimes. A simple example illustrates the types of induction required. Consider a Minkowski spacetime. It is observationally indistinguishable from a “half Minkowski spacetime”; that is a Minkowski spacetime in which half has simply been excised. This excised half is the “t = 0” hypersurface, in a standard coordinate system, and all events to its future. The observational indistinguishability depends on the fact that every observer’s past in either spacetime is identical to every other observer’s past in either spacetime; they are all geometric clones of one another, as illustrated in Figure 13.2. Consider the timelike curves of any inertial observer in either spacetime. No such observer would ever detect a failure of the observer’s world line to extend by a millisecond of proper time.6 Every observer would have repeatedly done the experiment of waiting a millisecond and found always that their worldline was extended by a millisecond, as shown in Figure 13.3. The natural inductive inference is that all future terminated inertial worldlines can be extended by one millisecond of proper time. But that condition can only be met in the full Minkowski spacetime. Hence, even though the two spacetimes are observationally indistinguishable as far as deductive discriminations are concerned, this induction indicates in favor of the full Minkowski spacetime. This last example uses a peculiar spacetime, an extendable half Minkowski spacetime. These extendable spacetimes are only a little more peculiar than the constructions used to generate indistinguishable spacetimes. Where the nemesis of a Minkowski spacetime was created by
170
Philosophy of Science Matters t=0
observer
observer time
observer’s temporal past
observer’s temporal past
space
Minkowski spacetime
Half Minkowski spacetime
Fig.13.2. Minkowski and half Minkowski spacetimes
t=0
1 millisecond
1 millisecond time
space
Minkowski spacetime
Half Minkowski spacetime
Fig. 13.3. Millisecond extensions of inertial observers’ worldlines
subtracting spacetime structure, more common examples in the literature create nemeses by adding. The ingenious chain construction of Manchak’s proof requires us to build an infinity of duplicate spacetimes and then stitch them together in an infinite chain by what amounts to wormholes. In the case of a full Minkowski spacetime, observers would never detect the wormholes in the portions of spacetime they observe. Thus must remain forever unsure of whether such a wormhole link to the duplicated Minkowski spacetimes will eventually enter the growing portion that they can observe. Deduction cannot rule out the possibility. Induction can: these odd structures have never been seen, so one expects never to see them. There are more examples in which spacetime structure is added. A familiar case is a two-dimensional de Sitter spacetime and versions of it
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
171
time Original de Sitter spacetime
Doubled de Sitter spacetime
Tripled de Sitter spacetime
space
Fig. 13.4. Two dimensional de Sitter spacetimes
spatially unrolled into larger spacetimes of twice, thrice, etc. the spatial size of the original spacetime.7 This de Sitter spacetime can be pictured as a two-dimensional hyperboloid in a three-dimensional Minkowski spacetime. Its spatial slices are circles and the unrolling just consists of replacing them with larger circles of twice, thrice, etc., the circumference. The original and unrolled versions are depicted in Figure 13.4.8 The unrolled versions have the curious property of harboring space-time regions that are duplicated twice, thrice, etc. according to the extent of the unrolling. This property is illustrated in the figure by the presence of a single observer’s temporal past in the original de Sitter spacetime, then two copies of it in the doubled de Sitter spacetime, followed by three copies in the tripled de Sitter spacetime. A spacetime with no duplications and a spacetime with 27 duplications of the observer’s past will be observationally indistinguishable by deductive means. However, Occam’s razor motivates an inductive inference to the first spacetime.
6. THE OPACITY OF COSMIC INDUCTIONS While we can discriminate inductively among possible futures in both cases, the indeterminism arising through observationally indistinguishable spacetimes is more troubling than the indeterminism of quantum theory. In the case of quantum theory, the warrant for the inductive inferences is quite clear and unproblematic. In the spacetime case, it is hard to see through the inductions to the warrant that lies behind them. Insofar as warrants can be found, they are suspect. I will call these latter inductions “opaque.” In the case of quantum theory, the theory supplies physical chances to help us pick among the possible futures. Take the radioactive decay of an atom. We are equally sure that the atom will or will not decay over a
172
Philosophy of Science Matters
single half-life; both outcomes have the same physical chance of 1/2. We can be very sure, inductively, that decay will have happened if we wait ten half-lives; the physical chance of decay is then 1– (1/2)10 = 0.999. These inferences from the known past to the unknown future are grounded by the physical chances of the quantum theory itself. We can see through these inductions to the physical theory that grounds them; in this sense, they are “transparent.” The inductions arising in observationally indistinguishable spacetimes are of a different kind.9 Relativity theory provides no physical chances to weight the different spacetime extensions that it allows for our temporal past. The theory itself merely declares that the various alternatives are possible and nothing more. It leaves to us the task of determining if one or other of them is somehow preferred. We must look outside the physical principles of cosmology to decide this question. This is a natural project for inductive inference. However, the examples of Section 5 above reveal no single, principled inductive approach that can be applied across the many cases of indeterminism. Rather, we must consider each on a case-by-case basis and hope that we can find principled grounds in each. Take the extrapolation of the extendability of observed spacetime curves to all spacetime curves. Can it be grounded in an inductive principle that asserts that what has always been so, will always be so? Such a universal principle is untenable. It can only apply to some things that have been so, otherwise we rule out the possibility of any novel changes in the future. It must be modified so some future novelty is possible. But what could these modifications be, if we are to secure a universal inductive principle applicable beyond the one case? The danger is that we reduce the principle to a blatant circularity, in which we solemnly declare that it applies except when it does not. Worse, we must also be able to overrule the principle if further facts demand it. In a cosmology with a future “big crunch” singularity, we will have the same records of assured millisecond extensions, yet our inertial trajectories will not be indefinitely extendable. That failure can be deduced from present facts through the singularity theorems of general relativity. We face similar problems when we try to rule out the funhouse mirror duplications of the unrolled de Sitter spacetimes or the extravagantly duplicated spacetimes connected by wormholes. We would like to ground the inductive inference in a principle like Occam’s razor. However, the idea behind it, that simplicity is often accompanied by truth, is more a convenient and fallible rule of thumb than a guarantee. These problems are deepened by an absence of any clear rules as to just what counts as simple.10 I have long harbored dissatisfaction with the evident failure of any universal inductive principle such as the ones just listed. My solution has
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
173
been to propose that we abandon the search for universal, formal approaches to inductive inference. In a material theory of induction, I urge (Norton 2003; 2005) that inductive inferences are not warranted by general principles, but by facts. A fitting application of this material approach is to the inductive inferences just seen on radioactive decay. The laws of quantum theory are the facts that warrant them. What is troublesome from the material perspective is the absence of warranting facts for the inductions in the spacetime case. It seems natural to infer inductively to the fully extended Minkowski spacetime rather than the extendable half Minkowski spacetime, or to avoid admitting holed spacetimes that are created from other spacetimes by excising even quite small parts. However it is very hard to specify just what facts ground the inference. That we have never seen holes in spacetime does not settle the matter. By their construction, there cannot be an observable trace of holes. That remains true even if our world tubes pass directly through the hole. We would cease to be for the portion of our world tubes coinciding with the excision. However, the portion of our world tubes in the future of the hole would be reconstituted with all the memories and other traces of the excised spacetime. If observed facts do not ground the inductive inference, what of physical laws? We could cite the common postulate in relativity texts that spacetimes are inextendable. However, that postulate is merely the supposition of precisely what is at issue, and is distinctive in that it is dispensable from a physical perspective. It is present for mathematical convenience. In his unpublished manuscript, Manchak reports the justifications usually given for assuming inextendability (Manchak 2009a). They amount to invoking Leibniz’s principle of plenitude. Manchak quotes from the writings of the mathematical physicist Robert Geroch as a typical justification: “Why, after all, would Nature stop building our universe . . . when she could just as well have carried on?” One cannot help but be struck by how tenuous the grounding has become. We are now to secure our inductions in abstract metaphysics. The principle of plenitude itself is sufficiently implausible that we need to prop it up with anthropomorphic metaphors. We are to imagine a personified Nature in the act of creating spacetime, much as I might be painting my fence on the weekend. Just as I might not want to stop when there is one board remaining unpainted, so Nature is supposedly loath to halt with a cubic mile-year of spacetime still uncreated. If the complete arbitrariness of the principle of plenitude is not already clear, we might pause to apply it elsewhere. We are supposed to prefer spacetimes without duplications by virtue of a metaphysics of simplicity. Yet surely the metaphysics of plenitude would direct the opposite result.
174
Philosophy of Science Matters
Why would Nature, guided by the slogan “make all you can make,” eschew yet another duplication of the spacetime if it is possible? All these inductive inferences are opaque in that we cannot see through them to their warrants. If we seek to justify them by means of general inductive principles, we resort to principles that are clearly false, or, if not, so hedged as to be inapplicable. If we seek to justify them materially in facts, we arrive almost immediately in the dubious, abstract metaphysics of plenitude and simplicity. This circumstance is to be contrasted with the transparent inductive inferences in the quantum context. Their grounding is found directly in the laws of quantum theory; we can in turn satisfy ourselves of those laws by tracing back further warrants in the experimental and theoretical foundations of quantum theory. In sum, we have what appears to me to be an intractable problem.11 On the one hand, it seems completely unjustified to presume that wormholes we have never seen in our past spacetime will appear in the future. It seems completely unjustified to presume that processes we observe here are duplicated many times over in an unrolled spacetime, when those duplications are by construction, necessarily invisible to us. It seems completely unjustified to assume that there are holes in spacetime, when the spacetime would, by construction, look identical to us if there were no holes. Indeed, even if our world tubes had no past, we would have memories of a past that never was. The inductive inference from those memories to the reality of the past seems irresistible, as do the inductive inferences that reject spatial duplications and future wormholes to new universes. To deny these inductive inferences would, in other contexts, be denounced as delusional. We routinely dismiss as desperate zealots those who tell us our universe was created last Wednesday complete with all records of an ancient past. Yet, when we try to display the proper warrant of those inductive inferences we favor, whether the warrant is in general principles or material facts, the ground crumbles around our feet.
ACKNOWLEDGMENTS It is a pleasure to present this essay in honor of Peter Achinstein, with gratitude for his scholarship and dedication to philosophy of science. I thank him for his thoughtful discussion in Achinstein (2010) and also for correspondence, in which he clarified his views and made suggestions for editing this note. I also thank Claus Beisbart and John Manchak for helpful discussion on an earlier draft.
Observationally Indistinguishable Spacetimes: A Challenge for Any Inductivist
175
REFERENCES Achinstein, P. 2010. Evidence, Explanation, and Realism. New York, Oxford University Press. Beisbart, C. and T. Jung. 2006. Privileged, Typical, or Not Even That? Our Place in the World According to the Copernican and the Cosmological Principles. Journal for General Philosophy of Science 37 (2): 225–56. Glymour, C. 1977. Indistinguishable Space-Times and the Fundamental Group. In Foundations of Space-Time Theories: Minnesota Studies in the Philosophy of Science, vol. 7, ed. J. Earman, C. Glymour, and J. Stachel. Minneapolis: University of Minnesota Press. Magnus, P. D. 2005. Reckoning the Shape of Everything: Underdetermination and Cosmotopology. British Journal for the Philosophy of Science 56 (3): 541–57. Malament, D. 1977. Observationally Indistinguishable Spacetimes. In Foundations of Space-Time Theories: Minnesota Studies in the Philosophy of Science, vol. 7, ed. J. Earman, C. Glymour and J. Stachel. Minneapolis: University of Minnesota Press. Manchak, J. 2009. Can We Know the Global Structure of Spacetime? Studies in History and Philosophy of Modern Physics 40 (1): 53–6. ——— . (2009a) “What is a ‘Physically Reasonable’ Spacetime?” URL: http:// philsci-archive.pitt.edu/id/eprint/4506Spacetime? Norton, J. D. 2008. Must Theory Underdetermine Evidence? In The Challenge of the Social and the Pressure of Practice: Science and Values Revisited, ed. M. Carrier, D. Howard, and J. Kourany. Pittsburgh, Pa.: University of Pittsburgh Press. ——— . 2006. The Formal Equivalence of Grue and Green and How It Undoes the New Riddle of Induction. Synthese 150 (2): 185–207. ——— . 2005. A Little Survey of Induction. In Scientific Evidence: Philosophical Theories and Applications, ed. P. Achinstein. Baltimore, Md.: Johns Hopkins University Press. ——— . 2003. A Material Theory of Induction. Philosophy of Science 70 (4): 647–70.
NOTES 1. The theorem excludes spacetimes in which the entire spacetime is observable from one event. They are “bizarre,” because they include closed timelike curves, which permit time travel. 2. In a recent paper, I urge that the thesis is groundless (Norton 2008). For further analysis of the relation of these examples to the underdetermination thesis and the possibility that the observationally indistinguishable spacetimes may just be variant presentations of the same facts, see Magnus 2005. 3. It requires more thought than I gave it in writing footnote 13 of Norton 2008. 4. There is an ambiguity in the use of the term “theory.” One might conceive an individual spacetime as a theory in its own right. The geometry of Minkowski spacetime is the special theory of relativity. However this use is unnatural in general relativity, in which the particular spacetimes are models of the general theory.
176
Philosophy of Science Matters
5. Claus Beisbart has pointed out to me that, aside from Manchak’s result, there is a familiar expectation of this sort of indeterminism. Fixing one’s temporal past leaves open the possibility of influences propagating to one’s future from the region of spacetime outside one’s past light cone. 6. In the half Minkowski spacetime, some worldlines will not extend by a millisecond, when the observer’s worldline runs into the non-existent t = 0 excision. That observer ceases to exist and there is no detection or record of the failure. 7. This spacetime has only one spatial dimension. The spatial duplications described are harder to implement with three-dimensional spaces. The simplest case arises with a topologically toroidal Euclidean space. It is created by taking a cube of Euclidean space and identifying opposite faces. The space can be unrolled by connecting its faces to duplicate cubes of Euclidean space. 8. The figures are misleading insofar as it appears that the doubling is achieved by a uniform expansion of the spacetime. That would alter the spacetime curvature at every point of the spacetime, so that the temporal pasts in the two spacetimes would no longer be isometric. Rather the doubling is effected by a cutting and pasting that leaves local spacetime structure unaffected. Take a de Sitter spacetime “1” and copy of it, de Sitter spacetime “2.” Cut each spacetime along a timelike geodesic that then exposes edges “1L” and “1R” in spacetime 1 and “2L” and “2R” in a spacetime 2. Glue 1L to 2R and 1R to 2L to form the doubled de Sitter spacetime. 9. The problem has been long discussed in the context of justifying the cosmological principle. Its justification requires an inductive inference from the large scale, spatial homogeneity, and isotropy of the observed part of spacetime to all spacetime. For a recent discussion, see Beisbart and Jung 2006. 10. I set aside Bayesian analyses. All they will do is take something like these principles and use them to determine prior probabilities and likelihoods. The resulting analysis will be no more secure than the principles used to set these quantities, although this will be harder to see because of the complications of the computational machinery. 11. Peter Achinstein has urged me to explain how this problem differs from another notorious intractability in induction, “grue.” In Norton 2006, I argue that grue only provides a novel inductive challenge if we grue-ify our total science. However, then the standard and grue-ified alternatives are isomorphic, so we cannot rule out the possibility that they are merely notational variants of the same facts. Hence, we should not expect an inductive logic to separate them. A variation on this approach may assist us in the case of spacetimes with excisions. Since no experience will ever enable us to learn whether ours is the fully extended or mutilated spacetime, strict empiricist leanings may incline us to say that the two do not really differ factually. However, this sort of thinking cannot help us if we are choosing among spacetimes that may have wormholes in our future. These wormholes will eventually manifest in our observations.
14 Making Contact with Molecules: On Perrin and Achinstein Stathis Psillos
1. INTRODUCTION In his essay, “Philosophy in France in 1912,” André Lalande made the following observation. M. Perrin, professor of physics at the Sorbonne, has described in Les Atomes, with his usual lucidity and vigor, the recent experiments (in which he has taken so considerable a part) which prove conclusively that the atoms are physical realities and not symbolical conceptions as people have for a long time been fond of calling them. By giving precise and concordant measures for their weights and dimensions, it is proved that bodies actually exist which, though invisible, are analogous at all points to those which we see and touch. An old philosophical question thus receives a positive solution. (Lalande 1913, 366–7)
This brief and matter-of-fact announcement expressed a rather widely shared sentiment on the European continent that Jean Perrin’s experimental work had clinched the issue of the reality of atoms. Indeed, it is now obvious that between roughly 1908 and 1912, there was a massive shift in the scientific community in favor of the atomic hypothesis. It is also obvious that Perrin’s experimental work on the causes of Brownian motion played a major role in this shift. When Perrin received the Nobel Prize for physics in 1926, it was noted in the presentation speech by Professor C. W. Oseen that he “put a definite end to the long struggle regarding the real existence of molecules.” Peter Achinstein has offered one of the most systematic expositions and reconstructions of Perrin’s argument, aiming (a) to show how his own theory of evidence best accounts for the significance of Perrin’s results; and (b) how Perrin has offered a local and experimental argument for scientific realism. After some detailed presentation of Perrin’s
177
178
Philosophy of Science Matters
argument, I will offer my own reconstruction of it and will show why it is superior to Achinstein’s. Finally, I will try to draw some lessons for scientific realism.1
2. ENTER PERRIN Over time, Perrin seems to have shifted from a neutral position, with friendly gestures to atomism, to a full endorsement of the atomic theory. In his textbook of 1903, he contrasted two methods of doing science: the inductive, which proceeds by analogy and generalization, and the intuitivedeductive, which consists “in imagining a priori matter to have a structure that still escapes our imperfect senses, such that its knowledge permits deductive predictions about the sensible properties of the universe” (Perrin 1903, viii). The latter method fosters “the illusion of a satisfactory explanation . . . [of] the visible in terms of the invisible, even when [it does not] lead to the discovery of new facts” (viii). Though he notes that in that book he will adopt the inductive method, he nonetheless claims he will not aim to condemn “en bloc” the molecular theories, but rather to submit them to a critical scrutiny in such a way that their essential elements are preserved. Perrin was sensitive to the fact that for many (notably Duhem and perhaps Ostwald and Mach), the atomic hypothesis was a venture into metaphysics. Surprisingly, he added: “I do not forget that the sensation is the only reality.” This would seem to condemn the atomic hypothesis from the start. Yet, Perrin added two important caveats. The first is that “[sensation] is the only reality, on the condition that to the actual sensations all possible sensations are added.” This is important because he thought that the atomic hypothesis could, in the end, be rooted in sensations. How this could be is illustrated by his second caveat, in which he drew an analogy between molecules and microbes—the latter did become the object of “possible sensation” via the microscope. Here is how he put it. One would certainly have been able, without the aid of the microscope, to arrive at the thought that contagious diseases were due to the multiplication of very small living beings. One, guided by these ideas a priori, would have been able to discover almost all of the Pasteurian technique. One would have thus followed deductive science and cured the contagious diseases, but following a way condemned by the supporters solely of the inductive method, until the very day in which the microscope had proved that the microbe hypothesis expressed several possible sensations. Here then is an indisputable example of a structure which could escape our senses and the knowledge of which allows anticipation of certain properties which are [to our senses] directly accessible. (Perrin 1903, ix–x)
Making Contact with Molecules: On Perrin and Achinstein
179
The point is that a hypothetico-deductive account of scientific method won’t provide strong grounds for accepting the reality of the explanatory posits—more is needed, and this more comes, in the end, from experimental confirmation and, in particular, from placing the hypothesized entities into a causal network that ends up in certain observational trails. By the time he wrote his Les Atomes, he had become an ardent defender of the intuitive-deductive method. In the preface, he noted, To divine in this way the existence and properties of objects that still lie outside our ken, to explain the complications of the visible in terms of invisible simplicity is the function of the intuitive intelligence which, thanks to men such as Dalton and Boltzmann, has given us the doctrine of Atoms. This book aims at giving an exposition of that doctrine. (Perrin 1916, vii)
However, even then, he very much hoped that there would be some day in which atoms would be “as easy to observe as are microbes today,” though for him the use of a microscope is within the “realm of experimental reality” (1916, x). The point that needs to be appreciated is that for Perrin science should proceed by refusing to limit itself “to the part of the universe we actually see,” and that the best way to achieve this is to aim at explanation-by-postulation, that is by aiming to explain the visible in terms of the invisible (1916, xii). Perrin’s more technical work is collected in his Brownian Movement and Molecular Reality, which appeared in French in 1909 and was translated into English in 1910. In this book, Perrin makes almost no methodological remarks, but I shall attempt to reconstruct the structure of his argument for the reality of molecules in a way that his methodology is clearly brought out. The key point of his strategy is this: “Instead of taking this hypothesis [the atomic hypothesis] ready made and seeing how it renders account of the Brownian movement, it appears preferable to me to show that, possibly, it is logically suggested by this phenomenon alone, and this is what I propose to try” (1910, 7). Perrin takes it that the atomic hypothesis is an already given plausible hypothesis, its plausibility being grounded in the fact that it remains the only serious admissible explanation of Brownian movement. Reviewing the work of Léon Gouy and others, Perrin suggests that several potential causes of the movement can be safely eliminated and that, in particular, the cause of the movement is internal and not external (cf. 1901, 6). This kind of eliminative approach paves the way for rendering the standard atomic explanation of Brownian movement “by the incessant movements of the molecules of the fluid” the only serious admissible explanation. This is not enough to render it true or probable; and yet, by the end of his reasoning, Perrin does think that it is probable and true. This happens
180
Philosophy of Science Matters
because Perrin succeeds in showing that Brownian movement is itself an instance of molecular movement and hence that it obeys the laws of the molecular movement. Hence, it can be used to (a) determine Avogadro’s number and (b) specify the individuating properties of atoms. To see all this, let us follow his steps in some detail. Perrin’s theoretical schema proceeds as follows. Let us suppose we have a uniform emulsion (all granules are identical) in equilibrium that fills a vertical cylinder of cross section s. Consider a horizontal slice contained between the levels, where this slice is enclosed between two semi-permeable pistons—they are permeable to the molecules of water but impermeable to the granules. Each piston is subjected to osmotic pressure. This slice of granules does not fall; hence there must be an equilibrium between the force that tends to move it upward (viz., the difference of the osmotic pressures) and the force that tends to move it downward (viz., the total weight of the granules less the buoyancy of the liquid). Having estimated both forces, Perrin arrives at the equation of the distribution of the emulsion 2/3Wlog(n0/n) = j(D−d )gh
(1)
where W is the mean granular energy, j the volume of each granule, D its density, d the density of the intergranular liquid and n and n0 respectively the concentrations of the granules at the two levels separated by height h. The task then is to measure all magnitudes other than W; hence, to determine W (cf. 1910, 24). The important assumption that Perrin makes is that the mean granular energy W of the particles in Brownian motion is equal to mean molecular energy W¢. In other words, he argues that the Brownian particles behave as large molecules and hence obey the laws of the gases (see also 1916, 89, 92). Indeed, the first few sections of his 1910 work aim to motivate this claim. The mean kinetic energy W¢ of the molecules of a gram-molecule of a gas is a function of Avogadro’s number N. It is equal to (3R/2N)T, where T is the absolute temperature of the gas and R is the constant of the perfect gases (cf. 1910, 19). Hence, W¢ = (3R/2N)T
(2)
Perrin relies on van’t Hoff’s proof that the invariability of energy (viz., that the mean kinetic energy is the same for all gases at the same temperature) holds also for the molecules of dilute solutions and generalized it further to all fluids, including emulsions. The claim that “the mean energy of translation of a molecule [is] equal to that possessed by the granules of an emulsion”—that is that W = W¢—is crucial. It paved the way for an experimentum crucis: either W = W¢ or W
Making Contact with Molecules: On Perrin and Achinstein
181
¹ W¢ and given that both W and W¢ could be calculated, we might have “the right to regard the molecular theory of this movement as established” (1910, 21). Being an extremely skillful experimenter, Perrin managed to prepare suitable emulsions of gamboge and mastic, with spherical granules of radius a. (1) thus becomes 2/3Wlog(n0/n) = 4/3pα3(D−d )gh.
(1¢)
Here again, all magnitudes but W are measurable. Determining the ratio (n0/n) was quite demanding, but Perrin used the microscope to take instantaneous snapshots of the emulsion. Determining the value a of the radius was even more demanding, but Perrin used three distinct methods to achieve this, one relying on Stokes’s equation (capturing the movement of a sphere in a viscous fluid), and two without applying this equation (using, instead, a camera lucida). These calculations were in impressive agreement, which led Perrin to conclude, among other things, that the otherwise controversial application of Stokes’s equation (because it was meant to apply to continuous fluids) was indeed legitimate. When all was said and done, Perrin was able to calculate the granular energy W (which is independent of the emulsion chosen). If W = W’ (if, that is, the Brownian particles do behave as heavy molecules and hence if the laws of the gases do hold for them too), there is a direct prediction of Avogadro’s number N from (1¢) and (2), that is, (RT/N)log(n0/n) = 4/3pa3(D−d)gh
and N = 3RTlog(n0/n)/4pa3(D−d)gh.
(1²)
This prediction could then be compared with known calculations of N based on the kinetic theory of gases, for example, that by van der Waals’s (N = 6×1023) (cf. 1910, 44). Perrin made a number of experiments and concomitant calculations and the agreement was always impressive. As he put it, “It is manifest that these values agree with that which we have foreseen for the molecular energy. The mean departure does not exceed 15 percent and the number given by the equation of van der Waals does not allow for this degree of accuracy” (Perrin 1910, 46). Perrin became immediately convinced that “this agreement can leave no doubt as to the origin of Brownian movement” (1910, 46). “[A]t the same time,” he said, “it becomes very difficult to deny the objective reality of molecules” (1916, 105). What convinced him, he says, was that on any other hypothesis (better, on the negation of the atomic hypothesis), the expected
182
Philosophy of Science Matters
value of N from the study of the movement of granules suspended in a liquid would be either infinite or zero—it would be infinite if all granules actually fell to the bottom of the vessel, and zero if the fall of the granules was negligible. Hence, on the hypothesis that matter has not molecular structure, the probability that the predicted value of N would be the specific one observed would be zero; on the contrary, this probability is high given the atomic hypothesis. This, Perrin noted, “cannot be considered as the result of chance.” Perrin takes another step. He stresses that the determination of Avogadro’s number by (1²) affords a determination of the properties of molecules that can be calculated on its basis. Moreover, this determination of N is now “capable of unlimited precision,” since all magnitudes in (1²) other than N can be determined “to whatever degree of precision desired.” Hence, Perrin went on to calculate N and to conclude that its value is N=7×1023. From this, he calculated the weight and the dimensions of molecules. He also reported on a number of other calculations of Avogadro’s number, including: the measurement of the coefficient of diffusion; the mobility of ions; the blue color of the sky (the diffraction of the sunlight by the atmospheric molecules); the charge of ions; radioactive bodies; and the infrared part of the spectrum of the black-body radiation. Though all these calculations were less accurate than his own, Perrin took them to prove molecular reality (cf. 1910, 90), since they are in considerable agreement, showing that this number is “essentially invariant” (1910, 74). Here is his conclusion: I think it impossible that a mind, free from all preconception, can reflect upon the extreme diversity of the phenomena which thus converge to the same result, without experiencing a very strong impression, and I think it will henceforth be difficult to defend by rational arguments a hostile attitude to molecular hypotheses, which, one after another, carry conviction, and to which at least as much confidence will be accorded as to the principles of energetics (1910, 91).
What then is the logical structure of Perrin’s argument? Recall his claim that he was after a crucial experiment for the reality of atoms. Of course, there are no crucial experiments in the strict sense of the expression, viz., in the sense of disproving a hypothesis or of proving a hypothesis. But as Poincaré has put it, an experiment can condemn a hypothesis, even if it does not—strictly speaking—falsify it. Perrin’s argument was precisely meant to condemn the denial of the atomic hypothesis—which, of course, is not to say that he intended to condemn energetics. As we have just seen, he did think (and he had already noted this in his 1903 work) that energetics need not imply the denial of the atomic hypothesis, namely, that matter is continuous.
Making Contact with Molecules: On Perrin and Achinstein
183
The way, then, I think Perrin’s argument should be reconstructed is as follows. With the argument sketched above, Perrin has made available two important probabilities, namely, Prob (n=N/AH) = very high Prob(n=N/-AH) = very low
That is, the probability that the number of molecules in a gram-molecule of a gas (including an emulsion, which does behave as a gas) is equal to Avogadro’s number given the atomic hypothesis is very high, while the probability that the number of molecules is equal to Avogadro’s number given the denial of the atomic hypothesis is very low. These two likelihoods can be used to specify the so called Bayes factor f. f= prob(n=N/-AH)/prob(n=N/AH)
Bayes’s theorem states prob(AH/n=N) = prob(n=N/AH)prob(AH)/prob(n=N)
where: prob(n=N) = prob(n=N/AH)prob(AH)+prob(n=N/−AH)prob(−AH).
Using the Bayes factor, Bayses’s theorem becomes: prob(AH/n=N) = prob(AH)/(prob(AH) + f prob(-AH)).
Perrin’s argument then can be put thus: 1. f is very small. 2. N = n is the case. 3. prob(AH) is not very low. Therefore, prob(AH/n=N) is high. Now, premise 1 (that f is very small) is established by the body of Perrin’s demonstration, which shows that given the denial of the atomic hypothesis, it is extremely unlikely that Avogadro’s number has the specific value it does. Premise 2 is established by a series of experiments involving different methods and different domains. Premise 3 is crucial, since it is required for the probabilistic validity of Perrin’s argument. It specifies the prior probability of the atomic hypothesis and without the prior probability the argument noted above would commit the base-rate fallacy. Perrin’s preparatory eliminative work has aimed to show that, by eliminating several alternative potential explanations of Brownian movement,
184
Philosophy of Science Matters
the atomic hypothesis has gained at least some initial plausibility which is reflected in its having some prior probability of being true. Actually, the following might be added. There is a rare case in which the prior probability of a hypothesis does not matter, and this is when the Bayes factor is zero. This happens when just one theory can explain the evidence. Then, we can dispense with the priors. If the Bayes factor is zero, no matter what prob(AH) is, the posterior probability prob(AH/ n=N) is unity. And the Bayes factor is zero if prob(n=N/-AH) is zero. Recall Perrin’s wording: “That, in the immense interval [0, infinity] which a priori seems possible for N, the number should fall precisely on a value so near to the value predicted, certainly cannot be considered as the result of chance” (1910, 46; cf. 1916, 105). This is almost tantamount to saying that his experiments established that prob(n=N/-AH) = 0. This kind of claim would (and does) explain Perrin’s confidence that the atomic hypothesis has been “established”; that he has offered “a decisive proof” of it (1916, 104). Admittedly, there is room for manoeuver here, since it might be argued that prob(n=N/-AH) has, after all, a small finite value. In that case, some reliance on the prior probability prob(AH) is inevitable and the usual philosophical dialogue would kick off: How are the priors fixed? Are they objective? If not, is the case for the reality of atoms strong? I do not want to follow this dialogue now (except to note that I agree with Achinstein that prior probabilities need not be subjective or idiosyncratic degrees of belief). I want to stress, however, that it seems to me that the major role Perrin’s work has had in persuading scientists to adopt the atomic hypothesis lies mostly in its presenting a rare but very important case in which the posterior probability of the atomic hypothesis becomes (almost) unity—given, of course, that it is assigned a non-zero prior, which it seems everybody but Duhem did. A chief point that Perrin makes is precisely that size does not matter, but causal role does! Like microbes, molecules do end up being the objects of possible sensation—in the broad sense in which Perrin understands this, namely, to include detection through the microscope. Hence Perrin, like Pasteur before him, places the atoms firmly within the laboratory, grounding their causal role and offering experimental means for their detection and the specification of their properties. This is of great significance because it becomes clear that Perrin’s argument should be compelling for anyone who does not take it that strict naked-eye observability is a necessary condition for accepting the reality of an entity. It should also be compelling for anyone who thinks that continuity of causal role is a criterion for accepting the reality of an entity—irrespective of whether some instances of this entity are observable, while others are not.
Making Contact with Molecules: On Perrin and Achinstein
185
By the same token, it becomes clear that the real issue about the so-called theoretical entities is not their unobservability, but rather their accessibility. In this sense, what Ostwald aptly called “the scientific horizon” is not fixed and immovable; claims that are once below it can move above it. What facilitates this change is not that some suspicious entities become observable, but rather that some suspicious entities enhance their explanatory role: claims about them are highly confirmed by ordinary scientific methods; their causal role is established experimentally; they become the locus of unification of disparate phenomena. Perrin’s case is instructive because it shows vividly that there are points after which resistance to accepting the reality of certain entities becomes dogmatic and mostly motivated by philosophical prejudice (cf. Krips 1986).
3. ENTER ACHINSTEIN The core of Achinstein’s claim is that the calculation of Avogadro’s number by Perrin’s experiments using (a notational variant of) equation (1²) above confirmed Perrin’s core hypothesis, namely that molecules exist and that Avogadro’s number is 6×1023. More specifically, Achinstein takes proposition T to express the core hypothesis of the atomic theory: T = Chemical substances are composed of molecules, the number N of which in a gram molecular weight of any substance is (approximately) 6 ×1023.
He takes it that this proposition already has had some support from background knowledge b and other evidence. In particular, he rightly claims that T’s plausibility (and in fact its non-zero probability) was based on the application of “causal eliminative” reasoning (Achinstein 2001, 255). Actually, the initial probability that Achinstein assigns (or claims that Perrin assigned) to T, given background knowledge, is Prob(T/b)>1/2. He then claims that Perrin’s experimental result led him to accept the following proposition: C = The calculation of N done by means of Perrin’s experiments on Brownian particles using equation [(1²)] is 6×1023, and this number remains constant [when several parameters in equation (1²) are varied].
Achinstein goes on to claim that C is more probable given the truth of T than without T and to make his crucial point that C confirms T. This is because given (i) prob(C/T&b) > prob(C/b) (ii) prob(T/b)>0 (iii) prob(C/b)>0
186
Philosophy of Science Matters
it follows from an application of Bayes’s theorem that (iv) prob(T/C&b)>prob(T/b). Moreover, since he has assumed that prob(T/b)>1/2, it follows that (v) prob(T/C&b)>1/2. Achinstein put this in the service of his own theory of evidence. In broad outline, two statements are the main features of Achinstein’s theory. The first is that for something e to be evidence for a hypothesis H, it must be the case that the probability of H given e should be higher than ½. That is, prob(H/e)>1/2. So, Achinstein works with an absolute concept of evidence: e is evidence for H only if e is not evidence for the denial of H. This is meant to capture the view that evidence should provide a good reason to believe. But, second, this absolute conception of evidence (though necessary) is not sufficient for reasonable belief. What is added is that the probability that there is an explanatory connection between H and e, given H and the evidence e, should be more than ½. Call E(H/e) the claim that there is an explanatory connection between H and e. Achinstein’s second feature is that prob(E(H/e)/H&e)>1/2. More accurately, e is evidence (a good reason) for H only if the product of prob(E(H/e)/e&H) with prob(H/e) should be greater than ½. Given this conception, (v) is far more important than (iv) above. Besides, the foregoing requirement that there is an explanatory connection between the hypothesis and the evidence is satisfied in Perrin’s case, and Achinstein argues that (vi) prob(E(T/C&b)/T&C&b)>1/2. Of course, there is no guarantee that the product of prob(E(T/C&b)/T&C&b) with prob(T/C&b) is great than ½. The values of the two factors should be chosen by hand such that their product is greater than ½. Achinstein argues that they can be plausibly taken to be sufficiently high in Perrin’s case. But this is certainly an extra and defeasible assumption. In any case, Achinstein’s conclusion is that not only did Perrin provide evidence for the reality of molecules, but also that this is best captured by his own theory of evidence. It seems to me this is not right. Achinstein’s reconstruction leads to a weak conclusion vis-à-vis the case at hand. If Perrin just succeeded in establishing that prob(T/C&b)>1/2, it seems there is no explanation of why his achievement was taken (by himself and almost everybody else) to be decisive in establishing the reality of atoms. On Achinstein’s reconstruction, all Perrin achieved was to show that the atomic hypothesis is more likely than not. This is not a mean feat, of course. But it is hardly sufficient to sway the balance in favor of the atomic hypothesis in the way it actually did. There
Making Contact with Molecules: On Perrin and Achinstein
187
is no natural way to increase the posterior probability of T in Achinstein’s account, unless T is given a very high prior probability and the product prob(E(T/C&b)/T&C&b) × prob(T/C&b) is fiddled with. My account, on the contrary, does show that AH (which is roughly equivalent to Achinstein’s T) becomes very probable, given Perrin’s experimental results. Besides, my account, unlike Achinstein’s, captures the strength of the evidence. More specifically, Achinstein notes that his own account of evidence cannot easily explain why some qualities of some piece of evidence (e.g., precision and directness) provide stronger support for a hypothesis than pieces of evidence that lack these qualities (Achinstein 2001, 262). (Achinstein ends up adding these qualities by hand into his theory.) In my account, the precision of the determination of Avogadro’s number and the diversity of means by which this precise determination was achieved makes it all the more improbable that this will be the right number (that is, that n will be equal to N) given the negation of AH. Defending his own theory of evidence against other attempts to reconstruct and explain Perrin’s achievements, Achinstein (2001, 259) notes that his own account (i) is better than Salmon’s (which was based on the common cause principle) because on his own account the molecular hypothesis does get confirmed by the evidence; and (ii) is better than an ordinary hypothetico-deductive reconstruction, since it does not suppose a deductive link between the molecular hypothesis and Perrin’s results. It’s patently the case that my own account fares at least as well as Achinstein’s vis-à-vis the other two stories. Achinstein is very sensitive to the charge that Perrin’s reasoning might be circular, since Perrin seems to assume the reality of the molecules before he tries to prove it (Achinstein 2001, 259). His answer to this charge is that Perrin does not start with an unquestioned assumption that molecules exist, but that he does take this assumption to have some initial probability, based on the causal-eliminative reasoning that preceded his own strategy. I think this is right and is actually brought out by my own account too. Hence, my account offered in the previous section has all the strengths and none of the weaknesses of Achinstein’s.
4. LESSONS FOR SCIENTIFIC REALISM Perrin’s case highlights a claim that lately I tend to find all the more forceful, namely, that commitment to the reality of specific explanatory posits is a matter that depends on the context. This is so because, as I have argued in Knowing the Structure of Nature (Psillos 2009), there are two types of evidence that are brought to bear on the truth of scientific hypotheses
188
Philosophy of Science Matters
(and which inform judgments of prior probability and of confirmation). The first type is first-order evidence and is related to whatever evidence scientists have in favor of a hypothesis. In Perrin’s case, this evidence includes the several methods of determination of Avogadro’s number, the evidence that goes into fixing a non-zero prior probability to the atomic hypothesis (e.g., the evidence that the cause of Brownian movement is internal to the fluid), and so on. The second type of evidence, what I call second-order evidence, comes from the track record of scientific theories and/or meta-theoretical (philosophical) considerations that have to do with the reliability of scientific methodology. This, for instance, is evidence that many past explanatory hypotheses have been abandoned, that there have been alternative potential explanations of some phenomena that came to be accepted later on, and so on. This kind of (historical-philosophical) evidence does not concern particular scientific theories but science as a whole. It is the kind of evidence that, for instance, motivates the pessimistic induction. I have argued that the proper philosophical task is to balance these two kinds of evidence and that this balancing is context-dependent (Psillos 2009). Perrin’s case is very instructive precisely because it shows that the context can settle issues of balance. For instance, it is clear that Perrin’s case is so strong that the first-order evidence for the reality of molecules takes precedent over the second-order evidence there might be for being skeptical about explanatory posits. The fact that other explanatory hypotheses have failed in the past is trumped—in this context—by the strength of the first-order evidence. It would be folly, however, to think that considerations concerning the second-order evidence should be totally wiped out—or worse, that these are considerations to which working scientists are blind. These are meta-theoretical or philosophical considerations that do get into the evidential balance sheet nonetheless. Achinstein seems to imply that these considerations are almost irrelevant to the issue of the reality of explanatory posits (Achinstein 2002, 495). They are not. Achinstein (2002) is right in stressing that the proper battleground for scientific realism is made of specific arguments in favor of the reality of specific unobservable entities. Given the key role that explanatory considerations play in specifying the prior probabilities of certain hypotheses (e.g., the atomic hypothesis), it is an exaggeration to call the proper argument for realism “experimental.” Better put, the proper argument for realism is explanatory-experimental, the latter component meaning to stress that causal contact with the explanatory posits enhances the claim to their reality. But Achinstein (2002) seems to want to draw the further conclusion that the realism debate as it has been conducted so far is independent of the kind of argument for realism you get from Perrin. This is wrong.
Making Contact with Molecules: On Perrin and Achinstein
189
I will confine myself to two points. First, Perrin already works within what I have elsewhere (Psillos forthcoming [a]) called “the realist framework.” Simply put, Perrin already works within the framework that seeks explanation of the manifest behavior of bodies while positing typically unobservable entities; hence he adopts a framework that allows for the assignment of prior probabilities to “invisible” entities. This is not something that evidence or a priori reasoning forces on anyone. To see this, just think of die-hard opponents of realism (Duhem? van Fraassen?) who refuse to adopt this framework; and hence, who refuse to assign nonzero (Duhem) or anything-but-vague (van Fraassen) prior probabilities to specific explanatory posits—such as the molecules. To put the point somewhat crudely, Perrin’s argument does not amount to an argument for scientific realism in general (as opposed to an argument for the reality of certain entities) because it is launched within the realist framework. Hence, the debate about the realist framework itself is alive and well. My second point concerns the relation between Perrin’s argument and the so-called “no miracles” argument (NMA) for realism. Achinstein intends to distance Perrin’s argument from NMA (Achinstein 2002, 486). But he does not have to do so. The relation between Perrin’s argument and NMA is precisely the one anticipated by realists like Boyd and myself (Psillos 1999), namely, Perrin’s argument is one of the very many first-order instances of inference to the best explanation (IBE), which feed the premises of the realist argument that IBE is reliable. And this is precisely what the NMA aims to do, namely, to defend the reliability of IBE. I have defended all this in a forthcoming paper (Psillos forthcoming [b]). What Perrin’s case has taught me, among other things, is that the first-order IBEtype of reasoning that leads to commitment to the reality of certain explanatory posits has a fine structure and a strength that is shaped, by and large, by the context.
REFERENCES Achinstein, P. 2002. Is There a Valid Experimental Argument for Scientific Realism? Journal of Philosophy 99 (9): 470–95. ——— . 2001. The Book of Evidence. New York: Oxford University Press. Krips, H. 1986. Atomism, Poincaré and Planck. Studies in History and Philosophy of Science 17 (1): 43–63. Lalande, A. 1913. Philosophy in France in 1912. Philosophical Review 22 (4): 357–74. Perrin, J. 1916. Atoms, trans. D. L. Hammick. London: Constable & Company Ltd. ——— . 1910. Brownian Movement and Molecular Reality, trans. F. Soddy. London: Taylor and Francis. ——— . 1903. Traité de Chimie Physique: Les Principes. Paris: Gauthier-Villars.
190
Philosophy of Science Matters
Psillos, S. Forthcoming (a). Choosing the Realist Framework. Synthese, DOI 10.1007/s11229-009-9606-9. ——— . Forthcoming (b). The Scope and Limits of the No-Miracles Argument. In The Philosophy of Science in a European Perspective, vol. II, ed. F. Stadler et al. Springer. ——— . 2009. Knowing the Sructure of Nature. London: MacMillan-Palgrave. ——— . 1999. Scientific Realism: How Science Tracks Truth. London & New York: Routledge.
NOTES 1. I dedicate this essay to Peter, who has taught me (and us, I hope) how important it is to combine philosophical rigor with historical sensitivity.
15 Achinstein and the Evidence for Evolution Richard A. Richards
1. INTRODUCTION Peter Achinstein begins his Book of Evidence with the striking claim that “standard philosophical theories about evidence are (and ought to be) ignored by scientists” (Achinstein 2001, 3). The problem with standard theories of evidence, according to Achinstein, is that they “are based on assumptions incompatible with ones scientists make when they speak of, and offer, evidence for hypotheses” (3). Achinstein is right as far as he goes, but he could have gone even further. For historians of science, as well as scientists, the standard philosophical ways of conceiving evidence are of little use, and may in fact be counterproductive in the efforts to think clearly about what counts as evidence for a scientific hypothesis. In his Book of Evidence, Achinstein lays out what would be required for a concept of evidence to be useful for scientists. First, it must be strong enough to warrant belief in a hypothesis—not merely raise its probability (Achinstein 2001, 8–9). Second, it must be empirical—not a priori, semantic, or merely mathematical (Achinstein 2001, 5, 26, 28). In other words, scientists assume that evidence provides a good reason to believe a hypothesis—not just some reason, and on the basis of the empirical facts, and not just logical or formal relations either. I endorse Achinstein’s criticism of standard philosophical theories of evidence and both of his conditions of adequacy, relative to the assumptions of scientists. But we can also ask about the adequacy of a concept of evidence for historians of science, who try to understand the actual acceptance and rejection of hypotheses on the basis of evidence. For historians, a concept of evidence must help us understand not just what should count as evidence for a hypothesis in general, that is, what constitutes a good reason for anyone to believe, but what in fact did count as evidence, that is, what caused belief in a hypothesis. This is not to say that historians never care about what should count as evidence. But sometimes this concern is contextual, and is
191
192
Philosophy of Science Matters
about what a scientist should believe, given what he or she knows, or is in a position to know. Historians might, for instance, ask if Galileo was justified in believing the heliocentric model of the cosmos rather than the geocentric, given what he knew at the time or was in a position to know. There are then two main tasks for a historically adequate theory of evidence: first, a prescriptive task to provide conditions for having a good reason to believe a hypothesis; second, a descriptive task to provide conditions for having an actual reason to believe. And for historians both tasks must be sensitive to context. As we shall see, while standard philosophical theories of evidence are inadequate relative to these tasks, Achinstein’s framework provides the conceptual resources to carry out both tasks and with sufficient sensitivity to context. To see more precisely what a historically adequate theory of evidence requires, we shall first look at a puzzling case in the history of science: Charles Darwin’s belief in branching evolution prior to his acceptance of a mechanism (natural selection), and his seemingly inconsistent views about whether he was justified in doing so. Second, we will look at standard philosophical theories of evidence, to see how they are inadequate for understanding the details of Darwin’s ambivalence. Third, we shall see how Achinstein’s framework can help us understand the complexities here. In particular, his concepts of subjective and epistemic situation evidence will be of value in understanding the details of Darwin’s views about evidence and evolution. Subjective evidence gives us a descriptive framework, while the idea of an epistemic situation gives us the resources to think about the prescriptive question relative to a variety of contexts.
2. A DARWINIAN PUZZLE After Darwin’s return from the five-year Beagle voyage at the end of 1836, he turned his attention to the “mystery of mysteries,” the origin of species. During this time, and before his discovery in 1838 of natural selection, he began speculating about the possibility of branching evolution: the formation of new species by branching and modification. (His first branching diagram appeared in the middle of 1837 in his “B” notebook.) Two decades later in the Origin, he told us that his belief in branching evolution was on the basis of what the historian of science Mary Winsor calls the “taxonomic facts,” which include the hierarchical groupings of organisms generated by similarities and differences (Winsor 2009). It is a truly wonderful fact—the wonder of which we are apt to overlook from familiarity—that all animals and all plants throughout all time and space should be related to each other in group subordinate to group, in the
Achinstein and the Evidence for Evolution
193
manner which we everywhere behold—namely, varieties of the same species most closely related together, species of the same genus less closely and unequally related together, forming sections and sub-genera, families, orders, sub-classes, and classes. (Darwin 1964, 128)
At the end of the Origin, Darwin first claimed that he believed in branching evolution on the basis of these taxonomic facts. Then he claimed that he “should without hesitation” adopt this hypothesis about branching evolution on the basis of just these facts. . . .The several classes of facts in this chapter seem to me to proclaim so plainly that the innumerable species, genera, and families of organic beings, with which this world is peopled, have all descended, each within its own class or group, from common parents, and have all been modified in the course of descent, that I should without hesitation adopt this view, even if it were unsupported by other facts or arguments. (Darwin 1964, 457–8, my emphasis)
But earlier in the Origin, Darwin seemed inclined in another direction, toward the view that the taxonomic facts were not sufficient to justify belief in branching evolution. . . . [A] naturalist . . . might come to the conclusion that each species . . . had descended from other species. Nevertheless, such a conclusion, even if well founded, would be unsatisfactory, until it could be shown how the innumerable species inhabiting this world had been modified. (Darwin 1964, 3, my emphasis)
Here he seems to have claimed instead that the taxonomic facts were by themselves insufficient. Also required was a suitable mechanism of change, natural selection. Darwin’s views about whether the taxonomic facts were evidence for branching evolution are complicated and puzzling in several ways. First, he seems to have claimed that he in fact came to believe in branching evolution based on the taxonomic facts. Second, he seems to have claimed that he should believe in branching evolution just on the basis of the taxonomic facts. But third, he also seems to have claimed that something more was required—mechanism. The first claim seems to be descriptive, about what actually caused him to believe in branching evolution. The second claim seems to be prescriptive, about what he would be justified in believing. The third claim seems to be prescriptive, not about his own belief, but the belief of “a naturalist”—some abstract naturalist, or perhaps his fellow naturalists of the time. But why would he be justified in believing in branching evolution on the basis of taxonomic facts, if others were not? Shouldn’t the same facts that justified his belief also justify the beliefs of his fellow naturalists or some abstract naturalist? To fully understand
194
Philosophy of Science Matters
Darwin’s views here, we need some way of assessing both descriptive and prescriptive claims about evidence relative to Darwin, and relative to either some abstract naturalist, or his contemporary naturalists. An adequate concept of evidence should render situations like this explicable. But standard philosophical theories of evidence do not seem up to this task.
3. STANDARD PHILOSOPHICAL THEORIES OF EVIDENCE There have been theories of scientific evidence as long as there has been something we can reasonably call science. Some of the standard philosophical theories go back to the beginnings of modern science in the views of Francis Bacon, René Descartes, and Isaac Newton. More recent theories originated in the nineteenth century in the work of John Herschel, William Whewell, John Stuart Mill, and Charles Peirce. The twentieth century has given us the theories of Carl Hempel, Rudolf Carnap, Karl Popper, and more. I cannot address the full variety and complexity of these theories here, but we can think about them in more general and schematic terms. John Norton has done so, arguing that there are really three main families of theories of evidence: inductive generalization, hypothetical induction, and probabilistic. In each of these families, the concept of evidence is prescriptive, and the evidential relation is formalized in distinctive ways, relative to a distinctive inductive schemata. The first family of theories of evidence is inductive generalization, and is based on the principle that “an instance confirms a generalization” (Norton 2003, 652). This can be from the very simple enumerative induction where “an instance of an A that is B confirms the generalization that all A’s are B’s” (652). But it also includes more complex principles such as Mill’s joint method of agreement and disagreement. The important idea here is that there is some formal schema that gives the evidential requirements for whether an observation should count as evidence for a hypothesis. For instance, is the observation of some black ravens evidence for the hypothesis that all ravens are black? If it satisfies the formal schema, the answer is “yes.” The second family of theories of evidence, according to Norton, is hypothetical induction, and is based on the idea “that the ability of a theory, possibly with auxiliary assumptions, to entail the evidence is a mark of its truth” (Norton 2003, 653). This is the familiar hypothetico-deductive method, whereby a hypothesis is confirmed to some degree when its predictions are verified by observation. The abstract, formal schema here goes something like: “if hypothesis h, then some observation o; if o, therefore
Achinstein and the Evidence for Evolution
195
probably h.” Because this requirement is widely seen as too weak, a variety of restrictions have been proposed based on the requirements that the prediction be unlikely if the hypothesis is false, that the hypothesis be simple or parsimonious, or that the hypothesis be the best explanation. Here, as in the case of inductive generalization, these are prescriptive theories, giving the conditions of adequacy for when we are justified in taking some observation to be evidence for a hypothesis, and based on some schematic formalization that lays out the proper form of reasoning. The third family of theories of evidence is probabilistic and based on the idea that the evidential relation conforms to the demands of some “definite calculus, usually the probability calculus” (Norton 2003, 659). This family is most at home when stochastic processes are involved. These theories sometimes also assert that “degrees of belief should correspond to the same calculus as governs stochastic processes in the physical world” (660). Here, as in the two other families of theories, the claim is prescriptive, about what should count as evidence, and it is based on some formal schematization, typically the probability calculus and Bayes’s theorem. There are well-known problems with each of these theories of evidence, as we have seen in the historical debates about Hume’s problem of induction, the Raven paradox, and the strange predicate “grue.” My focus here, however, is not on these problems with induction, but with the inadequacy of these standard theories for understanding the history of science. The first and most obvious problem is that all three families of theories of evidence are prescriptive rather than descriptive; they give the conditions of adequacy for what should count as evidence, not what in fact has counted as evidence. Newton gave us his rules of reason to tell us how we should reason, not to describe actual reasoning—even if he thought his own reasoning conformed to his Rules. Bacon, Descartes, Herschel, Whewell, Mill, and Peirce similarly gave guidance on what should count as evidence, not what has in fact counted as evidence. Contemporary philosophical accounts of evidence follow in this tradition. Elliott Sober, for instance, at the beginning of his Evidence and Evolution, seems to endorse this tradition, telling us that “historians and sociologists study science as it is, whereas philosophers of science study science as it ought to be. Philosophy of science is a normative discipline” (Sober 2008, xv). Sober then gives an extended analysis of prescriptive thinking about evidence. But to understand the Darwin case, we need to know whether or not Darwin did in fact believe in branching evolution on the basis of the taxonomic facts, not just about whether he should have believed. If philosophical theories of evidence are going to be useful to historians of science, they must provide the resources suitable for thinking about the
196
Philosophy of Science Matters
historical facts—what actually counted as evidence for scientists in context, not just the prescriptive ideal. But even the prescriptive question poses problems for standard theories of evidence. How do we determine what Darwin or anyone else should have believed? The second problem with standard philosophical theories of evidence, that they are not empirical, arises because these theories typically give formal, a priori criteria for what should count as evidence. On the simplest enumerative version of inductive generalization, something counts as evidence for a hypothesis because it satisfies the hypothesis. Any observed A that is also B satisfies the hypothesis “all A’s are B’s,” and can therefore serve as evidence for that hypothesis. Since this is a formal theory, it applies whenever the formal relations are satisfied— whenever the schema is instantiated. Because it is formal, it is also a priori, in that the “satisfaction” requirement does not depend on empirical facts, but on the logical relation between “A is a B” and “all A’s are B’s.” We see a similar formalization in Mill’s methods, and those theories of evidence based on the probability calculus and Bayes’s theorem. In each case, since the prescriptive criteria are formal, they are a priori. As Achinstein argues, however, the concept of evidence used by scientists is not a priori. It is empirical in that it relies on empirical assumptions. For example, whether an instance of an “A that is a B” counts as evidence for the generalization that “all A’s are B’s” depends on empirical facts about the A’s and B’s. On an a priori theory of evidence, by contrast, the fact that a sample of water has a particular melting point is evidence for the hypothesis that all samples of water have that melting point, and the fact that a piece of wax has a particular melting point is evidence that all samples of wax have that melting point. This is because both inferences satisfy the general, formal schema. But we know that water and wax are different in important ways; instances of water are uniform in ways that instances of wax are not. We can therefore make generalizations about water that we cannot make about wax. In particular, the melting points of water are uniform in the ways that the melting points of wax are not. Consequently, the empirical facts justify a generalization about the melting point of water, but not one about the melting point of wax. Knowledge of these differences will surely be relevant to the justification of the belief in a particular hypothesis about melting points (Norton 2003, 649). Because standard philosophical theories of evidence are formal and a priori, they render the evidential relation insensitive to the empirical facts. This has implications for the history of science. When historians ask what Darwin or anyone else should have believed, they may want to know what he should have believed irrespective of anything else. But they may
Achinstein and the Evidence for Evolution
197
also want to know what he should have believed given his historical context—what he was in position to know or believe about the empirical facts. To answer such contextual questions, an adequate theory of prescriptive evidence must be able to take context into account. As we shall see, Achinstein’s framework provides the necessary resources.
4. ACHINSTEIN ON EVIDENCE The most basic, and perhaps most important feature of Achinstein’s theory of evidence is its recognition of multiple concepts of evidence. Achinstein introduces these concepts by reference to Heinrich Hertz’s famous experiments with cathode rays in 1883. Hertz wanted to know if cathode rays were electrically charged—if they “gave rise to electrostatic forces” (Achinstein 2001, 14). To answer this question, Hertz constructed an experimental apparatus whereby cathode rays would be sent through a glass tube that had an electrometer attached to measure any electrostatic effect. The results of this experiment were negative, and Hertz concluded that “cathode rays are electrically indifferent” (15). A later experiment by J. J. Thomson in 1897, however, arrived at a different outcome, suggesting that cathode rays were in fact electrically charged. Thomson achieved this outcome, in part, by better evacuating the air from the glass tube (17). Were the negative results from the 1883 experiment evidence for the hypothesis that cathode rays are electrically indifferent? Achinstein argues that we can answer this question in different ways. We could answer by saying that given what Hertz knew in 1883, the experiment did indeed provide evidence for the hypothesis. Or we could answer that between 1883 and 1897 the experiment provided evidence, but not after 1897. Or we could answer that the experiment never was evidence for the hypothesis, even though Hertz believed it was. These three ways of answering the question about evidence, according to Achinstein, reflect the use of different concepts. Using one concept we might answer in one way, but using another concept we might answer in another way. Achinstein identifies four fundamental concepts of evidence: subjective, epistemic situation, potential, and veridical. The subjective concept of evidence, according to Achinstein, would be focused on what Hertz actually believed, based on everything else he believed or knew. The basic idea is that the outcome of the experiment e is evidence for the hypothesis h that cathode rays are electrically inert, if and only if Hertz believed that e was evidence for h, he believed that h was true, and his reason for believing that h was true is e (Achinstein 2001, 25). This is a subjective concept in that it depends on what a
198
Philosophy of Science Matters
particular subject, Hertz, believed and why he believed it. It is also a descriptive concept in that it tells us about what actually served as evidence, not what should have served as evidence. Epistemic situation evidence, according to Achinstein, is objective in that it does not depend on the actual beliefs of individuals, but it is relativized to an epistemic situation at a particular time. [An epistemic situation] is an abstract type of situation in which, among other things, one knows or believes certain propositions are true, one is not in a position to know or believe that others are, and one knows(or does not know) how to reason from the former to the hypothesis of interest, even if such a situation does not in fact obtain for any persons. (Achinstein 2001, 20)
Using this concept of evidence, we could ask what Hertz was in a position to know or believe, or not know or believe in 1883 (regardless of what he actually knew or believed), and how that might provide a good reason to believe that cathode rays were electrically inert. This concept of evidence is prescriptive in that it focuses on what can justifiably be inferred, not what was actually inferred. Alternatively, we could ask instead whether Hertz’s experiment in 1883 was evidence for the hypothesis that cathode rays are electrically inert, regardless of what anyone knows or believes. And we could ask whether Thomson’s experiment in 1897 was evidence for this hypothesis, regardless of what anyone knows or believes. This way of framing the evidential question does not require reference to any actual beliefs, nor does it invoke an epistemic situation. It is, in effect, asking whether the experiments were really evidence. Or, as Achinstein puts it, whether they were a “sign” of the truth of the hypothesis. There are two versions of this way of thinking about evidence. The first, weaker version, potential evidence, does not presuppose the truth of the hypothesis. “Some fact e can be evidence that h . . . and hence be a good reason to believe h . . . even if h is false” (Achinstein 2001, 28). On the stronger version, veridical evidence, “if e is a good reason to believe h, then h is true” (26). According to Achinstein, it is this last version of evidence that scientists want. They want e to really be evidence for h, regardless of what anybody knows or believes, and they want h to really be true—and they sometimes think they have this kind of strong evidence. I suspect that Achinstein is right about what scientists want, and I agree that standard philosophical theories of evidence are too weak—they don’t give good reasons to believe. But my concern here is with the value of philosophical theories of evidence to understanding the history of science, and in particular in understanding the Darwin case. As we shall see, the first two concepts of evidence outlined here, subjective and epistemic situation evidence, will be most helpful.
Achinstein and the Evidence for Evolution
199
5. THE DARWIN CASE As we have seen, Darwin had seemingly conflicting views about whether the taxonomic facts were sufficient evidence to believe in branching evolution. First, he believed that the taxonomic facts were sufficient for him to believe in branching evolution and that he came to believe on the basis of these facts. Second, he seemed to think that he should believe on just these grounds—that he would be justified in his belief. But third, he seemed to think that these same taxonomic facts were insufficient for others to believe; something else, a mechanism, was required. Standard philosophical theories of evidence, as we have also seen, are inadequate to help us understand these three historical facts because, first, they are prescriptive and to understand the first fact, we need a descriptive concept. Second, they are abstract and a priori, and don’t seem to be able to take into account differences in what individuals believe or know. What counts as evidence for one person should therefore count as evidence for another. On this way of thinking about evidence, it is paradoxical that Darwin would think something would justifiably be evidence for himself but not for others. Darwin’s descriptive claim about evidence, that he did in fact come to believe in branching evolution on the basis of the taxonomic facts, is easily explained and understandable in terms of Achinstein’s subjective evidence. There are, as outlined by Achinstein, three conditions to be met for something to count as subjective evidence. First, Darwin must have believed that the taxonomic facts were evidence for branching evolution; second, he must have believed that branching evolution were true; and third, his reason for believing in branching evolution must have been the taxonomic facts. If we want to know the historical fact—about whether the taxonomic facts were evidence for Darwin for branching evolution, we simply answer these questions using the standard methods of historical reconstruction. So while standard theories of evidence don’t even seem to be applicable to this descriptive task, Achinstein’s framework gives straightforward guidance about how to proceed. Darwin’s prescriptive claims about whether belief in branching evolution was justified are more complicated, but can be easily understood by reference to epistemic situation evidence. Darwin thought that he should believe in branching evolution on the basis of the taxonomic facts, but he seemed to suggest that others should not believe, because for them something else would be required—a mechanism. This fact is inexplicable by any philosophical theory of evidence that is abstract and a priori. On the inductive generalization theories, for instance, either the facts satisfy the hypothesis or they don’t. If an “A is a B,” then it is evidence for “all A’s
200
Philosophy of Science Matters
are B’s.” Similarly, on the hypothetical induction theories, if an observed fact is a deductive consequence of the hypothesis, then it is evidence (given the appropriate additional assumptions about simplicity, etc.). If not, it is not. In these cases, what someone knew or believed, or was in a position to know or believe, is irrelevant. But we can see how the epistemic situation concept of evidence can straightforwardly take individual differences into account. On this concept, as we have already seen, evidence is relativized to an epistemic situation—what a person or persons is in the position to know or believe and how to reason from what is known or believed. In this formulation, the epistemic situation is an abstraction, but it can be abstracted in various ways. First, we can abstract an epistemic situation relative to a single, concrete individual. We can conceptualize Darwin’s epistemic situation at a particular time in terms of what he actually knew or believed and how he thought he could reason from what he knew and believed. In this case, we evaluate his actual beliefs on the basis of his other beliefs and knowledge and his views about what constitutes correct reasoning. And we can evaluate his actual reasoning on the basis of his set of actual beliefs and knowledge, relative to objective standards of reasoning. In some sense then, this abstracted epistemic situation has a concrete basis, in that it contains only the actual beliefs and knowledge of an actual person at a particular time. But it is also abstract in that we are then considering the epistemic situation itself, as a set of beliefs, and independently of the person whose views constitute the epistemic situation. Because it is abstract, we can evaluate the reasoning and beliefs of the person who formed the basis of the epistemic situation. This, however, is only one way to think about evidence in prescriptive terms. Second, we could adopt a more abstract conception of epistemic situation as what some particular person is in a position to know or believe and how he or she could reason from this. This is abstract because it doesn’t rely on actual knowledge and beliefs, or views about correct reasoning. It is also prescriptive in the sense that we can identify what conclusion can be legitimately drawn, including conclusions about a particular hypothesis. Just as we can decide whether Hertz was justified in believing that cathode rays were electrically inert in 1883, we can decide whether Darwin was justified in believing in branching evolution in 1837. In both cases, this would be on the basis of what they were in a position to know or believe, not in terms of what they actually knew or believed. Third, we can adopt a yet more abstract epistemic situation in that it is not relativized to what a particular person would be in a position to believe or know, but to what some person or other was in the position to know at a particular time and in a particular cultural or social context.
Achinstein and the Evidence for Evolution
201
So just as we could ask what a German physicist might be in a position to know or believe in 1883, and an English physicist in 1897, we could ask what a naturalist in England might be in a position to know or believe in 1837. This sense of epistemic situation evidence is relativized not to a particular person, but what someone of a particular group might be in a position to know or believe. It might include more than an epistemic situation relativized to a particular person, because more of what is known and believed is typically available to some person or other of a group than to any single person in the group. But it might also include less. A particular person may be in the position to know more than the group of which he or she is a member. It might be the case, for instance, that in 1837 and because of his Beagle voyage, Darwin was in a position to know and believe things that his fellow English naturalists were not. We could continue this process of abstraction by considering epistemic situations that are not localized geographically or culturally, but only temporally. What beliefs and knowledge were available in 1837, independent of location? This epistemic situation would include beliefs and knowledge not available everywhere, so a person in Great Britain may not, strictly speaking, have access to what was available in Germany, and vice versa. Furthermore, we can relativize the epistemic situation to various counterfactuals. What would someone be in a position to know and believe in 1837, if they also knew something about the mechanisms of heredity? The value of this epistemic situation will depend on what questions are being asked. It may not be of broad interest, but nonetheless it is potentially a useful abstraction for thinking about prescriptive questions of evidence within various historical contexts. Here then is how we might understand Darwin’s seemingly conflicting views about what was necessary for justified belief in branching evolution. When he claimed that he was justified in his own belief in branching evidence, and on the mere basis of the taxonomic facts, he can be understood in terms of his own concrete epistemic situation. Given his concrete set of beliefs, knowledge, and views about correct reasoning, he thought he was justified in believing in branching evolution. He did not need anything more. But when he was considering the beliefs and knowledge of his fellow naturalists, he was thinking about a different epistemic situation— one that was not constituted in precisely the same way. His fellow naturalists, for instance, were not acquainted with the biogeography of the Galapagos and Falkland islands, and hence did not have that as part of their epistemic situation. Furthermore, they may not have had the same views about what constitutes good and correct reasoning. If this is correct, then Darwin’s seemingly contradictory views are easily explained by his
202
Philosophy of Science Matters
consideration of the different epistemic situations. What was justified in one epistemic situation was not necessarily justified in another.
6. CONCLUSION An idea implicit in Achinstein’s thinking about evidence is that an adequate theory of evidence must help us understand the actual reasoning of scientists. Standard philosophical theories of evidence fail here, according to Achinstein, because they don’t reflect the fact that for scientists, evidence is strong—it must provide a good reason to believe. A philosophical concept of evidence that is too weak will not help us understand the actual reasoning of scientists. Similarly, I argue here, a theory of evidence must make sense of the reasoning of historians of science. To do so, it must have the resources to answer descriptive questions about what individuals actually took to be evidence. Achinstein’s subjective evidence provides the resources to do this. An adequate theory must also be able to help us think prescriptively about what was justified, and in a way that can account for differences in context, individual knowledge, and belief. Otherwise, it is not clear how we can understand situations like Darwin’s, where he was seemingly contradicting himself about the evidential adequacy of the taxonomic facts. Achinstein’s epistemic situation evidence provides the resources to make sense of this and similar cases. In a passage quoted earlier in this essay, Elliott Sober claims that the philosophy of science is a normative discipline, telling us what science should be like. That may be true, but that fact does not require that philosophy of science ignore the descriptive and contextual. An adequate philosophy of science must be able to help us understand the complexities of the history of science, as scientists came to believe their hypotheses on the basis of empirical facts, and in light of what they were in a position to know or believe. REFERENCES Achinstein, P. 2001. The Book of Evidence. New York: Oxford University Press. Norton, J. 2003. A Material Theory of Induction. Philosophy of Science 70 (4): 647–70. Sober, E. 2008. Evidence and Evolution: The Logic Behind the Science. Cambridge: Cambridge University Press. Winsor, M. P. 2009. Taxonomy was the Foundation of Darwin’s Evolution. Taxon 58 (1): 43–9.
16 The Place of Artificial Selection in Charles Darwin’s Theory of Evolution through Natural Selection Michael Ruse
1. INTRODUCTION Anyone taking up the philosophy of science around 1960, as I did, was bound—in Britain and America at least—to be immersed in the school known as “logical empiricism.” Headed then by people like Richard Braithwaite (1953) in England and Carl Hempel (1965) and Ernest Nagel (1961) in America, there was a great emphasis on explicating the underlying logical structure of scientific arguments and how they relate to evidence. Thus, in the case of theories, the popular analysis was in terms of so-called “hypothetico-deductive systems,” where one had axiom systems with high-level hypotheses (often referring to unseen, so-called “theoretical entities”) and then from these, lower-level empirical laws could be deduced. Physics of course was the usual science of choice from which illustrative examples could be drawn. Because this philosophy was firmly committed to the distinction between the context of discovery and the context of justification, with only the latter really of concern to the philosopher, there was not much interest in heuristic analogies or like things. However, there was a theory of so-called “models.” These were things that were linked by being different interpretations of the same formal system, the same “calculus.” You have a theory and then you have a different interpretation, a model. It was okay, we learned, to use models, particularly for the cause of understanding—a solar system for the hydrogen atom, for instance—but you had to be very careful. Don’t think that what is going on in the model necessarily holds in the theory. I still remember the dire warning: “The price of the employment of models is eternal vigilance” (Braithwaite 1953, 93).
203
204
Philosophy of Science Matters
2. THE IMPORTANCE OF PETER ACHINSTEIN By the mid 1960s, this whole philosophy was under attack. Mainly it came from those who had turned to science and its history, finding that things are very different—in many respects much messier—than the logical empiricists supposed. People like Stephen Toulmin (1972), Norwood Russell Hanson (1958), and then above all Thomas Kuhn (1962) started to sweep things very much away. I myself was deeply affected by this movement and it has led to a lifetime’s fascination with the history of science, as much as with its philosophy. But it was not only from without but from within that logical empiricism was under attack, and it was from here that I set off on the path that I still follow. I still remember the moment when I realized that there was something wrong with the philosophy I had embraced. It was when reading Peter Achinstein’s first book, Concepts of Science (1968), where he pointed out with devastating clarity that the logical empiricist analysis of models just won’t do. You can have a formal similarity, and yet no one would think of models in such a case. For instance, if you have a flow of liquid through a tube, it is governed by Poiseulle’s law. V=
c (p1−p2) L
(V is the volume of liquid flowing through the tube across a fixed cross section per unit time; c is a constant; L is the length of the pipe; and p1 – p2 is the pressure difference between the ends of the tube.) Formally this law is the same as that governing the expansion of gases, being heated under a constant pressure. W = k (V1−V2) L
(W is work done; P is the pressure of the gas; m is the mass; and V2 – V1 is the volume change.) No one would think of the liquid as a model for the gases. However, the liquid case is also formally the same as the transmission of electricity through a wire, Ohm’s law. I = (v1−v2)
(I is the quantity of charge passing through a fixed cross section of the wire in a unit time; L is the length of the wire; k is a constant; and v1 – v2 the difference in potential between the ends of the wire.) Here we do think in terms of models. But as Achinstein pointed out, it is because we go beyond the formal similarity to physical similarity. In both we are dealing with a long narrow object of length L; in both we are concerned with the quantity of something that passes through a
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
205
special cross-section of this narrow object in unit time; and in both we are concerned with the difference in value of certain quantities between the ends of this narrow object. To draw the analogy one considers not only the formal structure of the equations but also similarities in the designate of the symbols they contain (Achinstein 1968, 245). I cannot begin to tell you what a thrilling shock that was. The logical empiricists were not God. They could be wrong! It took me some time to get a real handle on this—writing your dissertation and your first book is not the moment to change underlying philosophies—but as I got more and more seduced by Kuhn and his fellows, that liberating moment from Achinstein stayed with me and I have grown ever more appreciative. I am sure it also had a major role in what has become a lifelong interest in analogical and like reasoning—especially models and metaphors—in science. Appropriately, therefore, it is this interest that fuels this little essay. We are just finished with the “Darwin Year”—the 200th anniversary celebrations of the birth of Charles Darwin and the related 150th anniversary celebrations of his great book, On the Origin of Species by Means of Natural Selection. Naturally we have been led to think again about the work and its structure and contents, as well as its influence. In my case this has been particularly so because I wrote the introduction to a new, good book on the Origin, written by (and from the perspective of) one of today’s leading evolutionary biologists (Reznick 2009). As always, I have been struck by the importance in that work of analogies and metaphors, and above all by the example (call it what you will) of artificial selection, which is the topic of the first chapter and the prolegomenon of what is to come after. I want to turn to this topic here, a move made particularly pertinent by the fact that a new book has just appeared, co-authored by the well-known philosopher Jerry Fodor, that takes Darwin to task for the artificial selection model and argues that this is an irredeemable flaw that ruins the whole work (Fodor and Piattelli-Palmarini 2010). The book built upon an earlier argument made by Fodor: Selection theory is rotten at its core, because of the place of artificial selection. The present worry is that the explication of natural selection by appeal to selective breeding is seriously misleading, and that it thoroughly misled Darwin. Because breeders have minds, there’s a fact of the matter about what traits they breed for; if you want to know, just ask them. Natural selection, by contrast, is mindless; it acts without malice aforethought. That strains the analogy between natural selection and breeding, perhaps to the breaking point. What, then, is the intended interpretation when one speaks of natural selection? The question is wide open as of this writing. (Fodor 2007)
206
Philosophy of Science Matters
3. FROM ARTIFICE TO NATURE The discussion in the Origin of artificial selection, the picking of and breeding from desirable plants and animals by humans, is straightforward and follows Darwin’s usual path of mixing the fact of change with the mechanism of change. Picking the rock pigeon as his example, Darwin argues that all of the different varieties and forms, wonderful though they may be, are in fact descended from the same root stock. Altogether at least a score of pigeons might be chosen, which if shown to an ornithologist, and he were told that they were wild birds, would certainly, I think, be ranked by him as well-defined species. Moreover, I do not believe that any ornithologist would place the English carrier, the short-faced tumbler, the runt, the barb, pouter, and fantail in the same genus; more especially as in each of these breeds several truly inherited sub-breeds, or species as he might have called them, could be shown him. Great as the differences are between the breeds of pigeons, I am fully convinced that the common opinion of naturalists is correct, namely, that all have descended from the rock pigeon (Columba livia), including under this term several geographical races or sub-species, which differ from each other in the most trifling respects. (Darwin 1859, 23)
Darwin argues to this conclusion on the basis of several interrelated facts, particularly that the pigeons of today are all completely interfertile, the strongest mark that we are dealing with one species rather than many. He then goes on to talk about selection. Breeders of animals and plants decide what they want and then they proceed to produce it, picking carefully from the stock and developing new features and variations. Breeders habitually speak of an animal’s organisation as something quite plastic, which they can model almost as they please. If I had space I could quote numerous passages to this effect from highly competent authorities. Youatt, who was probably better acquainted with the works of agriculturalists than almost any other individual, and who was himself a very good judge of an animal, speaks of the principle of selection as “that which enables the agriculturist, not only to modify the character of his flock, but to change it altogether. It is the magician’s wand, by means of which he may summon into life whatever form and mould he pleases.” Lord Somerville, speaking of what breeders have done for sheep, says, “It would seem as if they had chalked out upon a wall a form perfect in itself, and then had given it existence.” That most skilful breeder, Sir John Sebright, used to say, with respect to pigeons, that “he would produce any given feather in three years, but it would take him six years to obtain head and beak.” (Darwin 1859, 31).
He now moves on to the natural world, and having established that there is abundant variation in nature and that there is an ongoing struggle
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
207
for existence (and more importantly for reproduction), he is ready to bring on his mechanism of natural selection. How will the struggle for existence, discussed too briefly in the last chapter, act in regard to variation? Can the principle of selection, which we have seen is so potent in the hands of man, apply in nature? I think we shall see that it can act most effectually. Let it be borne in mind in what an endless number of strange peculiarities our domestic productions, and, in a lesser degree, those under nature, vary; and how strong the hereditary tendency is. Under domestication, it may be truly said that the whole organisation becomes in some degree plastic. Let it be borne in mind how infinitely complex and close-fitting are the mutual relations of all organic beings to each other and to their physical conditions of life. Can it, then, be thought improbable, seeing that variations useful to man have undoubtedly occurred, that other variations, useful in some way to each being in the great and complex battle of life, should sometimes occur in the course of thousands of generations? If such do occur, can we doubt (remembering that many more individuals are born than can possibly survive) that individuals having any advantage, however slight, over others, would have the best chance of surviving and of procreating their kind? On the other hand, we may feel sure that any variation in the least degree injurious would be rigidly destroyed. This preservation of favourable variations and the rejection of injurious variations, I call Natural Selection. (Darwin 1859, 80–1)
Darwin also makes use of the artificial selection analogy a moment later when he introduces his secondary mechanism, sexual selection, which is the differentiation that comes within a species as organisms compete for mates. He distinguishes two kinds of such selection and in support makes mention explicitly of the human world. First there is male combat, where males fight each other for mates. This brings on adaptations designed to defeat rivals. Sexual selection by always allowing the victor to breed might surely give indomitable courage, length to the spur, and strength to the wing to strike in the spurred leg, as well as the brutal cock-fighter, who knows well that he can improve his breed by careful selection of the best cocks. (Darwin 1859, 88)
Then there is female choice, where beautiful features appear in the males in order to attract the females. It may appear childish to attribute any effect to such apparently weak means: I cannot here enter on the details necessary to support this view; but if man can in a short time give elegant carriage and beauty to his bantams, according to his standard of beauty, I can see no good reason to doubt that female birds, by selecting, during thousands of generations, the most melodious or beautiful males, according to their standard of beauty, might produce a marked effect (89).
208
Philosophy of Science Matters
One interesting fact about the Origin is that there are absolutely no examples of natural selection in action today causing change. Darwin gives a pretend example of wolves pursuing two strategies—short and stocky and slow versus light and fast—and suggests that there may be some truth to this, but generally there is no big search or apology. Even when Henry Walter Bates (1862) came up with his brilliant (and experiment-supported) explanation of butterfly mimicry, although Darwin was very pleased with the work—praising it generously and getting Bates a job—he only introduced it into later editions of the Origin and then almost at the end of the book. (The fourth edition of 1866 and in the penultimate chapter.) Basically, Darwin did not think that selection could act all that quickly, and so thought the search for evidence of it in action and having an effect was doomed to failure. Ironically, in 1878 an entomologist sent Darwin a letter documenting industrial melanism, the change in color of a moth’s wings toward a darker form due to the camouflage advantages brought on by concealment against an ever dirtier background. Apparently he did not reply and there is no evidence that he thought it that significant (Hart et al. 2010). However, Darwin was quite willing and eager to use the artificial selection model throughout the Origin when he wanted to make a point. The best example occurs in the discussion of embryology, where Darwin is arguing that the reason why the embryos of very different adults are often very similar is that the struggle for existence and consequent selection only kicks in as the organisms move towards maturity. In the egg or the womb, they are protected. In support, Darwin turns to the world of the breeder, suggesting that we should find something similar. And we do! Some authors who have written on Dogs, maintain that the greyhound and bulldog, though appearing so different, are really varieties most closely allied, and have probably descended from the same wild stock; hence I was curious to see how far their puppies differed from each other: I was told by breeders that they differed just as much as their parents, and this, judging by the eye, seemed almost to be the case; but on actually measuring the old dogs and their six-days old puppies, I found that the puppies had not nearly acquired their full amount of proportional difference. So, again, I was told that the foals of cart and race-horses differed as much as the full-grown animals; and this surprised me greatly, as I think it probable that the difference between these two breeds has been wholly caused by selection under domestication; but having had careful measurements made of the dam and of a three-days old colt of a race and heavy cart-horse, I find that the colts have by no means acquired their full amount of proportional difference. (Darwin 1859, 444–5)
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
209
4. PART OF THE THEORY? The question I want to ask now as a philosopher (rather than as a historian) is whether the artificial selection model/analogy is part of Darwin’s theory proper. Is it an essential piece of the theory of evolution through natural selection as given by Charles Darwin? It occurs between the covers of the Origin of Species, but so then does the title page. Is it really a part of the theory? I am assuming that the background to this question is that a logical empiricist would argue that it is not. It would be agreed that it is there, but it would be dismissed as not central in some way. And actually, going back to history, there is a good reason why one might feel confident in arguing in this way. The co-discoverer of natural selection, Alfred Russel Wallace, did not introduce the analogy into his little essay that he sent to Darwin. In fact, he argued against the analogy! Dealing with the objection that breeders never make new species—reproductively isolated groups— Wallace wrote, It will be observed that this argument rests entirely on the assumption, that varieties occurring in a state of nature are in all respects analogous to or even identical with those of domestic animals, and are governed by the same laws as regards their permanence or further variation. But it is the object of the present paper to show that this assumption is altogether false, that there is a general principle in nature which will cause many varieties to survive the parent species, and to give rise to successive variations departing further and further from the original type, and which also produces, in domesticated animals, the tendency of varieties to return to the parent form. (Wallace 1858, 54)
Having said this, although they were basically the same, Darwin’s theory was not exactly the same as Wallace’s theory. Darwin always focused on selection working at the individual level, whereas Wallace was always more inclined to see selection working at multi-levels, including producing benefits for the group at the expense of the individual (Ruse 1980). So the history is hardly decisive. Moving back to philosophy, let us ask what it was that Darwin was doing with artificial selection. Essentially part of his theory or not, what role was it playing in his thinking? A case can be made that there were at least three roles. First, there was the heuristic role. Darwin always claimed that he got to natural selection through the analogy from artificial selection. My first note-book was opened in July 1837. I worked on true Baconian principles, and without any theory collected facts on a wholesale scale, more especially with respect to domesticated productions, by printed enquiries, by conversation with skilful breeders and gardeners, and by extensive reading. When I see the list of books of all kinds which I read and abstracted,
210
Philosophy of Science Matters
including whole series of Journals and Transactions, I am surprised at my industry. I soon perceived that selection was the keystone of man’s success in making useful races of animals and plants. (Darwin 1958, 119)
He then tells us that reading Malthus showed exactly how selection could work in nature. In fact, if you look carefully at the selection notebooks to which Darwin is referring, you find that the artificial selection analogy is not as clear as he later suggests. However, there seems little reason to doubt the general recollection that it was the world of the breeders that put Darwin on the route to the big mechanism. So let us agree that there was a heuristic role for Darwin. Invoking the distinction between the context of discovery and that of justification, you might complain that this shows precisely that the analogy was not part of Darwin’s theory. Following Kuhn and many others, I am not sure that this distinction is quite as clean cut as the logical empiricists implied. But for the sake of argument, let us for the moment agree that the heuristic role is ancillary to the theory. This however leads us straight on to the second role, namely the pedagogical one. The analogy is used to teach us, to introduce us to the notion of selection in general and to lead us to natural selection. Presenting the analogy and its heuristic role enables the rest of us to follow along and (Darwin hopes) to accept the central mechanism. Remember the already-quoted passage where Darwin does introduce natural selection (taken from the beginning of Chapter Four). We are invited to think about artificial selection and then go on to natural selection. “Can the principle of selection, which we have seen is so potent in the hands of man, apply in nature? I think we shall see that it can act most effectually” (Darwin 1859, 80).
5. A “VERA CAUSA” Again I suspect that the logical empiricist is going to complain or whine that pedagogy is not part of the theory proper. No one wants to downplay the teaching of the young or uninitiated, but that is not part of real science. Again I suspect that the follower of Kuhn will differ, for one of the most important roles that he saw paradigms as playing was precisely that of bringing students into the field and teaching or indoctrinating them. But, again, for the sake of argument, let us give the logical empiricist the case. For now we come to the third role of the analogy, namely that of support. Here I think we can claim definitely that we have part of the theory proper. Darwin always looked upon the world of the breeders as part of the justification for his theory, something made particularly pressing in the light of the (earlier discussed) failure to offer any direct
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
211
justification from the natural world. Note that I say “direct” justification, for Darwin clearly thought that something like biogeographical distribution justified his case. In fact, he was very self-conscious about the role that the analogy was playing. In the language of the day, it was for him a proof that natural selection was a vera causa. Following the physicistphilosopher John F. W. Herschel, Darwin looked for a cause that we control as an analogy of a cause that exists naturally. If the analogy of two phenomena be very close and striking, while, at the same time, the cause of one is very obvious, it becomes scarcely possible to refuse to admit the action of an analogous cause in the other, though not so obvious in itself. For instance, when we see a stone whirled around in a sling, describing a circular orbit round the hand, keeping the string stretched, and flying away the moment it breaks, we never hesitate to regard it as retained in its orbit by the tension of the string, that is, by a force directed to the centre; for we feel that we do really exert such a force. We have here the direct perception of the cause. (Herschel 1830, 149)
Note that artificial selection exactly fits this mold, for here we have a human-caused force doing exactly what a natural-caused force is presumed to do. (See Ruse 1975 for a full discussion of the vera causa principle and the role it played in Darwin’s thinking.) Darwin always stressed the role of artificial selection when in the years after the Origin he was challenged on his theory. For instance, writing to the botanist George Bentham he says, In fact the belief in natural selection must at present be grounded entirely on general considerations. (1) on its being a vera causa, from the struggle for existence; & the certain geological fact that species do somehow change (2) from the analogy of change under domestication by man’s selection. (3) & chiefly from this view connecting under an intelligible point of view a host of facts. (Darwin 1985, vol. 11, 433)
Going back to the embryological discussion for a moment, we see that the breeders’ work is being used quite openly as support for his claims about what happens in nature and why different adults have similar young. “Fanciers select their horses, dogs, and pigeons, for breeding, when they are nearly grown up: they are indifferent whether the desired qualities and structures have been acquired earlier or later in life, if the fullgrown animal possesses them” (Darwin 1859, 446). This is then applied to nature. Now let us apply these facts . . . to species in a state of nature. Let us take a genus of birds, descended on my theory from some one parent-species, and of which the several new species have become modified through natural selection in accordance with their diverse habits. Then, from the many slight
212
Philosophy of Science Matters
successive steps of variation having supervened at a rather late age, and having been inherited at a corresponding age, the young of the new species of our supposed genus will manifestly tend to resemble each other much more closely than do the adults, just as we have seen in the case of pigeons.” (446–7)
I don’t see any way that you could, or really would, want to argue that this is not part of Darwin’s theory. So at some real level, the analogy does belong. And this point I suspect that the logical empiricist will pull the old ploy: “We’ve known this all along!” Inasmuch as the analogy is evidence for the main claims of the theory, it is part of the theory, just as the evidence for Kepler’s laws is part of the Newtonian theory. Or if it is not part of the theory, then it is fully acknowledged support for the theory. Either way, it has a place in the logical empiricist worldview.
The design metaphor Again, I think you have to agree to this, but I think there is (in this case at least) something going on that the logical empiricist picture does rather ignore or downplay. In both the artificial and the natural selection cases you have a differential reproduction, and one could—indeed population genetics from the 1930s on have—set everything out formally and show we have the same structure, the same calculus. But sensitized by Achinstein we start to look for something more. And there is something more. The artificial and natural are linked by more than formal similarity. There is if you like a metaphor or physical analogy at the heart of the matter, namely design (Ruse 2003). As Darwin noted, the breeders design the animal or plant that they want and then set about producing it. The same process is supposedly going on in nature. Natural selection does not just bring about change, but change of a particular kind, namely in the direction of designlike features—the eye and the hand and bark and leaves—that are adaptations. This is what makes the whole thing so powerful, and controversial. We have real design in the human world, metaphorical design in nature. In an early version of his theory, Darwin went even further, supposing in imagination that there is a designer. Let us now suppose a Being with penetration sufficient to perceive differences in the outer and innermost organization quite imperceptible to man, and with forethought extending over future centuries to watch with unerring care and select for any object the offspring of an organism produced under the foregoing circumstances; I can see no conceivable reason why he could not form a new race (or several were he to separate the stock of the original organism and work on several islands) adapted to new ends. As we assume his discrimination, and his forethought, and his steadiness of
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
213
object, to be incomparably greater that those qualities in man, so we may suppose the beauty and complications of the adaptations of the new races and their differences from the original stock to be greater than in the domestic races produced by man’s agency: the ground-work of his labours we may aid by supposing that the external conditions of the volcanic island, from its continued emergence and the occasional introduction of new immigrants, vary; and thus to act on the reproductive system of the organism, on which he is at work, and so keep its organization somewhat plastic. With time enough, such a Being might rationally (without some unknown law opposed him) aim at almost any result. (Darwin and Wallace 1958, 114)
It cannot be overly stressed how important this metaphor of design is within the Darwinian system. By thinking of the organic world as if designed, then we can do science. Why does the honeybee build a hexagonal comb? Because that is the strongest and most efficient. Why does the bird have feathers? (Archaeopteryx is introduced into the fourth edition of the Origin.) In order to fly. Why are forelimbs of man, mole, horse, porpoise, and bat so different? To grasp, to dig, to run, to swim, and to fly. It is this (that is to say, things of this sort) that is missed in the logical empiricist account of science, and it is clearly this that has got under the skin of Jerry Fodor. In particular, though there is no end of it in popular accounts of adaptationism, it is a Very Bad Idea to try and save the bacon by indulging in metaphorical anthropomorphisms. It couldn’t, for example, be literally true that the traits selected for are the ones Mother Nature has in mind when she does the selecting; nor can it be literally true that they are the traits one’s selfish genes have in mind when they undertake to reproduce themselves. There is, after all, no Mother Nature, and genes don’t have, or lack, personality defects. Metaphors are fine things; science probably couldn’t be done without them. But they are supposed to be the sort of things that can, in a pinch, be cashed. Lacking a serious and literal construal of “selection for,” adaptationism founders on this methodological truism. (Fodor 2007)
But here’s the rub! Should metaphors be cashed? Some would say that you cannot cash metaphors, even in principle. I won’t go there but I will say that there are very good reasons why you should not always rush to cash metaphors. The fact that not only are there heuristics that might lead you to a theory, but within the theory itself thanks to metaphor there are strengths that lead to new discoveries. Take the stegosaurus, that ungainly dino with funny triangular plates down its back. Why does it have them? This s surely a legitimate question that is quite unanswerable unless you think in terms of design. What is their point? Today, incidentally, this question is usually answered in terms of heat transfer. The plates are just like those found in cooling stations and apparently have the same function (Farlow, Thompson, and Rosner 1976). In Kuhn’s language, because of its
214
Philosophy of Science Matters
metaphors, a good paradigm can go on solving new puzzles. (Presumably the example of Wallace shows that you could get to the metaphor of design without the move from artificial selection, although looking again at the paper that Wallace sent to Darwin, I am not at all sure that he grasped fully the ways in which selection—a term he did not use— could help us to understand the design-like nature of adaptation. Remember that after the paper Wallace never got to develop his thinking independently of Darwin, whose greater importance Wallace always proclaimed.) Finally, is there something particularly offensive about this particular example? Was Darwin wrong to think in terms of design? I don’t see why. It cannot be overemphasized that it was a metaphor and not literal. As it happens, at the time of writing the Origin, Darwin did still believe in a Designer, but one who works at a distance and does not involve Himself in the day-to-day details of evolution. The whole point from that day to this—witness the scorn of contemporary evolutionists for so-called Intelligent Design Theorists—consciousness is kept out of the evolutionary process (qualifications being made for such things as sexual selection in humans).
6. CONCLUSION The world of the breeders provides a crucial piece of Darwin’s thinking. It offers a model or analogy leading to the central evolutionary force of natural (and its sister sexual) selection. Can it be given a full and adequate analysis by the logical empiricist approach to science? If you stretch and pull and lop—particularly if you lop—you can probably go a long way. But right at the end you miss the most important part of the analogy, the metaphor of design that is at the heart of selection-based evolutionary thinking. What a pity that Jerry Fodor never read Peter Achinstein at the same formative stage as I did! REFERENCES Achinstein, P. 1968. Concepts of Science. Baltimore: Johns Hopkins University Press. Bates, H. W. 1862. Contributions to an Insect Fauna of the Amazon Valley. Transactions of the Linnaean Society of London 23: 495–515. Braithwaite, R. 1953. Scientific Explanation. Cambridge: Cambridge University Press. Darwin, C. 1985. The Correspondence of Charles Darwin. Cambridge: Cambridge University Press.
The Place of Artificial Selection in Charles Darwin's Theory of Evolution
215
——— . 1958. The Autobiography of Charles Darwin, 1809–1882, ed. N. Barlow. London: Collins. ———. 1859. The Origin of Species. London: John Murray. Darwin, C., and A. R. Wallace. 1958. Evolution by Natural Selection. Cambridge: Cambridge University Press. Farlow, J. O., C. V. Thompson, and D. E. Rosner. 1976. Plates of the Dinosaur Stegosaurus: Forced Convection Heat Loss Fins? Science 192 (4244): 1123–25. Fodor, J. 2007. Why Pigs Don’t Have Wings: The Case against Natural Selection. London Review of Books, October 18. Fodor, J., and M. Piattelli-Palmarini. 2010. What Darwin Got Wrong. New York: Farrar, Straus, and Giroux. Hanson, N. R. 1958. Patterns of Discovery. Cambridge: Cambridge University Press. Hart, A. G., R. Stafford, A. L. Smith, and A. E. Goodenough. 2010. Evidence for Contemporary Evolution during Darwin’s Lifetime. Current Biology 20 (3): R95. Hempel, C. G. 1965. Aspects of Scientific Explanation. New York: Free Press. Herschel, J. F. W. 1830. Preliminary Discourse on the Study of Natural Philosophy. London: Longman, Rees, Orme, Brown, Green, and Longman. Kuhn, T. 1962. The Structure of Scientific Revolutions. Chicago: University of Chicago Press. Nagel, E. 1961. The Structure of Science: Problems in the Logic of Scientific Explanation. New York: Harcourt, Brace, and World. Reznick, D. N. 2009. The “Origin” Then and Now: An Interpretive Guide to the “Origin of Species.” Princeton: Princeton University Press. Ruse, M. 2003. Darwin and Design: Does Evolution Have a Purpose? Cambridge, Mass.: Harvard University Press. ——— . 1980. Charles Darwin and Group Selection. Annals of Science 37 (6): 615–30. ——— . 1975. Darwin’s Debt to Philosophy: An Examination of the Influence of the Philosophical Ideas of John F.W. Herschel and William Whewell on the Development of Charles Darwin’s Theory of Evolution. Studies in History and Philosophy of Science 6 (2): 159–81. Toulmin, S. 1972. Human Understanding. Oxford: Clarendon Press. Wallace, A. R. 1858. On the Tendency of Varieties to Depart Indefinitely from the Original Type. Journal of the Proceedings of the Linnean Society, Zoology 3 (9): 53–62.
17 Evidence and Justification Kent Staley
1. INTRODUCTION Peter Achinstein has been a persistent advocate of a kind of pluralism about evidence. In different scientific contexts, distinct evidence concepts come into play. The purpose of the present paper is to advocate a further extension of Achinstein’s pluralism. I will argue that his account should be supplemented with a further concept to more adequately account for the justification of scientific inferences. My proposal does not introduce any new notion of evidence, but articulates instead a relativized, nonprobabilistic notion (the securing of evidence) that helps make sense of certain strategies by which scientists justify their claims about what evidence they have and their inferences from that evidence. That Achinstein’s framework would benefit from such a supplement is a claim needing defense, because that framework already has a concept— called “ES-evidence”—that he says is “based on the idea of providing an epistemic justification for belief” (Achinstein 2001, 19), and is thus relativized to particular epistemic situations. In what follows, then, I will be at pains to explain Achinstein’s view of ES-evidence, the relationship ES-evidence bears to Achinstein’s other evidence concepts, and its relationship to this new notion: the securing of evidence.
2. ES-EVIDENCE AND VERIDICAL EVIDENCE Crucially, Achinstein conceives of some kinds of statements that scientists make about evidential relationships between facts and hypotheses as both (1) objective in the sense that statements of those kinds are true or false independently of what anyone believes about the hypotheses and facts in question; and (2) empirical in the sense that ascertaining the truth or falsehood of such statements is (at least sometimes) a matter for investi-
216
Evidence and Justification
217
gation by means of experiment and observation rather than a priori calculation. These features are incorporated into the concepts of “potential evidence” and “veridical evidence.” He proposes the following as necessary and sufficient conditions for the former: (PE) e is potential evidence that h, given b if and only if: 1. p(there is an explanatory connection between h and e/e and b) > ½ 2. e and b are true 3. e does not entail h. (Achinstein 2001, 170). Here e refers to some fact, h is a hypothesis, and b is background information. The probability statement in the first condition should be understood in terms of objective epistemic probability, according to which a statement of the form “p(h/e) = r” should be interpreted as asserting that “the degree of reasonableness of believing h, on the assumption of e, is r” (106). Veridical evidence is defined by adding to the conditions specified in (PE) the further requirement that h is true. Both potential evidence and veridical evidence share a feature that is central to Achinstein’s concerns: if e is evidence that h in either the potential or veridical sense, then e is a good reason to believe that h is true. Moreover, it is a good reason to believe h in a sense that is completely independent of any epistemic situation. An epistemic situation, according to Achinstein, is . . . an abstract type of situation in which, among other things, one knows or believes that certain propositions are true, one is not in a position to know or believe that others are, and one knows (or does not know) how to reason from the former to the hypothesis of interest, even if such a situation does not in fact obtain for any person. (Achinstein 2001, 20)
It is in this sense that the term “epistemic situation” will be used throughout this paper. Potential and veridical evidence are distinguished from ES-evidence, which provides “a justification for believing h for anyone in epistemic situation ES” (Achinstein 2001, 174).1 Achinstein defines the notion thus: (ES) e is ES-evidence that h with respect to an epistemic situation ES if and only if: 1. e is true 2. anyone in ES is justified in believing that e is (probably) veridical evidence that h. (174) Because veridical evidence satisfies the requirement of constituting a good reason to believe the relevant hypothesis and moreover entails that the
218
Philosophy of Science Matters
hypothesis in question is true, it captures, according to Achinstein, the aim of scientific investigation. Scientists, he writes, “are not satisfied with providing only a justification of belief for those in certain situations, even their own, since the belief might turn out to be false.” Hence ES-evidence does not constitute the aim of scientific inquiry. And because they “want their hypotheses to be true,” potential evidence also falls short of capturing the aim of scientific inquiry (34). Achinstein does not deny that scientists seek, in addition to veridical evidence, justification in drawing the conclusions that they do. But he treats this as at most a secondary concern. Indeed, when discussing the contexts in which the various evidence concepts come into play, he associates ES-evidence primarily with historical investigations that seek to answer questions about whether a particular scientist was justified in believing certain hypotheses, given her epistemic situation (Achinstein 2001, 37). This does not mean that a concern with epistemic justification (and hence ES-evidence) arises only in historical investigations. “To be sure, when a scientist claims that e is evidence that h, he believes and hopes that, given his knowledge, he is justified in believing h on the basis of e. But he believes and hopes for something much more”: a good reason for believing h that is independent of any epistemic situation, even his own (Achinstein 2001, 37). Perhaps because he primarily thinks of ES-evidence as coming into play in historical contexts, Achinstein rarely addresses directly the ways in which investigators seek to alter their own epistemic situations with regard to given evidence claims. Even where he does discuss the improvement of an epistemic situation, he considers not the improvement of the investigator’s own situation, but that of his audience. Writing about J. J. Thomson’s experiments on cathode rays yielding evidence that they carry an electric charge, Achinstein notes that, because others might not be able to recognize that his results are a good reason to believe that cathode rays are electrically charged, “What Thomson may need to do, and what a good physics text does, is to set the stage by including sufficient information so that others can be in an appropriate epistemic situation, so that others can become justified in believing the hypothesis for that reason” (Achinstein 2001, 35). True enough, but how did Thomson himself become justified in believing that his results are evidence that cathode rays carry an electric charge? I propose that the notion of ES-evidence as Achinstein defines it does not much help us to answer questions like this. Although ES-evidence acknowledges a role for the notion of epistemic justification relative to one’s epistemic situation, it cannot tell us how to understand that con-
Evidence and Justification
219
cept. Recall from the above definition of ES-evidence that for e to be ES-evidence that h relative to some epistemic situation ES, it must be the case that anyone in ES is justified in believing that e is (probably) veridical evidence that h. Clearly Achinstein here does not even intend, in defining ES-evidence, to give us an account of what it is to be justified in believing something relative to an epistemic situation, since he has simply moved that concept to the other side of the biconditional. I conclude that his discussion of ES-evidence is meant to acknowledge the role of epistemic justification in some uses of evidence-talk in science, and to clarify how such justification relates to other evidence concepts, but is not meant to give an account of epistemic justification as such. Although there is already a vast literature in general epistemology that seeks to analyze the notion of justification, I would contend that such accounts do not help us much to understand justification in science. More precisely, I propose shifting the focus of the discussion of justification away from analytic epistemology’s concern with specifying necessary and sufficient conditions for the truth of “S is justified in believing p,” and toward specifying a conceptual framework that serves to explicate justificatory practices in the sciences. Such a conceptual framework centers on an ideal of justification, at which justificatory practices aim (Staley and Cobb, forthcoming), and thus approaches justification by asking what it is that such practices accomplish, and how.
3. SECURING EXPERIMENTAL CONCLUSIONS Consider the following kind of situation: A researcher presents a conclusion from experimental data at a meeting of specialists.2 The decision to present a conclusion indicates the conviction of the researcher and her collaborators that they are prepared to justify their inference in response to whatever challenges they expect to encounter. Their confidence will result from their having already posed many such challenges to themselves. New challenges will emerge from the community of researchers with which they communicate. Such challenges take many forms, depending on the nature of the experiment and conclusions: Are there biases in the sampling procedure? Have confounding variables been taken into account? To what extent have alternative explanations been considered? Are estimates of background reliable? Can the conclusion be reconciled with the results of other experiments? Have instruments been adequately shielded, calibrated, and maintained? To a large extent, such challenges present possible scenarios in which the experimenters have gone wrong in drawing the conclusions that
220
Philosophy of Science Matters
they do. Such challenges are not posed arbitrarily. Being logically possible does not suffice, for example, to constitute a challenge that the experimenter is responsible for addressing. Rather, such scenarios are judged significant by those in a certain kind of epistemic situation, incorporating relevant disciplinary knowledge; and an appropriate response needs to provide a basis for concluding that the scenario in question is not actual. I propose thinking of practices of justifying an inference as the securing of that inference against scenarios under which it would be invalid (“error scenarios”), where the concept of security is defined as follows: SEC: Let Ω0 be the set of all scenarios that are epistemically possible relative to an epistemic situation K. Suppose that Ω1 Í Ω0. Proposition P is secure throughout Ω1 relative to K iff for every scenario ω Î Ω1, P is true in ω. If P is secure throughout Ω0, then P is fully secure relative to K.
Before proceeding, some explanation of terminology is in order. This definition employs the notion of epistemic possibility, which can be thought of as the modality employed in such expressions as “For all I know, there might be a third-generation leptoquark with a rest mass of 250 GeV/c2” and “For all I know, I might have left my sunglasses on the train.” Hintikka’s seminal work (1962) takes expressions of the form, “It is possible, for all that S knows, that P,” to have the same meaning as, “It does not follow from what S knows that not-P.”3 I have borrowed the notion of a scenario from David Chalmers for heuristic purposes. He describes a scenario as a “maximally specific way things might be” (Chalmers, 2011). If there is, relative to one’s epistemic situation, an epistemically possible scenario in which a proposition P is true, that means that, for all one knows, a complete and maximally specific description of the world entails P. In practice, no one ever considers scenarios as such, of course, but rather focuses on salient differences between one scenario and another. To put this notion more intuitively, then, a proposition is secure for an epistemic agent just insofar as, whatever might be the case for all that the agent knows, that proposition remains true. Applied to an inference from fact e to hypothesis h, an inference from e to h is secure relative to K insofar as the proposition “e is good evidence for h” is secure relative to K. For Achinstein’s account, the relevant way to explicate this might be to say that the conditions for e being potential evidence for h are secure relative to K. Note that the notion of a fully secure inference functions as an ideal for use in articulating an account of justification. Second, this account does not suppose that investigators can or should attempt to determine some degree of security of any of their inferences.
Evidence and Justification
221
Rather, the value of the concept of security lies in its capacity to conceptualize methods of justification encountered in scientific practice in a systematic way. Indeed, the methodologically significant notion is not security per se, but the securing of inferences, that is, the use of methods that serve to increase the relative security of an inference, either by expanding the range of validity of an inference across a fixed space of possible scenarios, or by decreasing the range of possible scenarios in which the inference would be invalid. Using the notion of security to complement Achinstein’s theory of evidence, I propose the following as an ideal of justification: JE (Justified Evidence): An assertion of h as a conclusion inferred from observed fact(s) e is fully justified relative to epistemic situation K if:
(1) e is potential evidence for h; and (2) the proposition “e is potential evidence for h” is secure throughout all scenarios that are epistemically possible relative to K.4 Note that these conditions are stronger than those for either veridical evidence or ES-evidence. This account articulates a notion of full justification as an ideal. The point is that methods of justification serve two distinct purposes. First, they aim (fallibly) to create conditions that will render (1) true for the inference at which the investigators arrive. Second, they aim to facilitate the pursuit of (2) by providing investigators with the resources to respond to the challenge of possible error-scenarios and, thus, serve to secure the inference proposed. Though full security may remain an unachieved ideal, the increase in relative security puts investigators in a better epistemic situation, and it is in this sense that methods aimed at securing evidence claims provide justification. Two general strategies for the securing of evidence pervade experimental science: In a weakening strategy one replaces a conclusion h with a weaker conclusion h’ that is true across a broader range of epistemically possible scenarios. A strengthening strategy calls for changing one’s epistemic situation ES into a stronger situation ES’ such that error scenarios epistemically possible relative to ES are not possible relative to ES’. To illustrate these strategies and see how the concept of security might complement Achinstein’s account of evidence, let us consider experiments undertaken by Heinrich Hertz, which Achinstein uses to introduce the notion of ES-evidence. As he sees it, Hertz’s experiments provided ES-evidence, relative to Hertz’s epistemic situation, that cathode rays are electrically neutral. However, later experiments by J. J. Thomson showed that Hertz’s results were not veridical or even potential evidence for the neutrality of cathode rays.
222
Philosophy of Science Matters
3. HEINRICH HERTZ AND THE JUSTIFICATION OF EVIDENCE FOR THE ELECTRICAL NEUTRALITY OF CATHODE RAYS In his 1883 paper, “Experiments on the Cathode Discharge,” Heinrich Hertz describes a series of experiments carried out in the Physical Institute of the University of Berlin on cathode phenomena. The conclusions that he claims to have “proved” via these experiments include that “cathode rays are only a phenomenon accompanying the discharge, and have nothing directly to do with the path of the current” and that “the electrostatic and electromagnetic properties of the cathode rays are either nil or very feeble” (Hertz 1896, 254).5 Hertz’s research should be understood in the context of the work already undertaken by his Berlin colleague Eugen Goldstein, who regarded the rays as (1) an entirely novel production not readily assimilable to known categories of electrical phenomena, but as (2) consisting of “open currents.” The latter view treated the rays as involving a kind of motion in the ether that propagated through the tube without transferring any material particle; any movement of charge through the tube did not traverse the length of the tube. Instead, rays originate as longitudinal waves at the cathode that terminate when they strike a particle, which in turn becomes the origin of another ray. As Buchwald notes, Goldstein “conflated longitudinal waves in the ether with unterminated currents, conceiving that the ether might possess conductivity. For him cathode rays were simultaneously a kind of current and a kind of wave” (Buchwald 1994, 137).
(1) Rays are not currents Hertz found an ingenious way to simultaneously challenge and endorse Goldstein’s claims by attacking (2) in a way that strengthened the support for (1). According to Buchwald’s account, the conception of rays in Geissler tubes as a novel state was central to Goldstein’s view, while the claim that they are open currents was secondary, an interpretation that distinguished the rays from any known electrical phenomenon. By showing that the rays were not any kind of current at all, Hertz could enhance the claim to their ontological distinctness while also establishing his own experimental acumen (Buchwald 1994, 137–41). In one experiment seeking to distinguish between the current in the tube and the rays, he set out to trace the current’s path in the tube by means of its magnetic effect, and to compare that path to the path of the rays through the tube. This involved constructing a tube that was in fact a rectangular frame holding two panes of glass enclosing an evacuated space
Evidence and Justification
223
(Figure 17.1). The cathode and anode were inserted just into that space through adjacent sides of the rectangle. A magnetic needle was suspended above the frame such that its deflection could be recorded as the frame was moved about. Hertz’s results clearly indicated a distinction between the paths of the current lines as recorded by the magnetic deflection and the paths of the cathode rays (Figure 17.2). But, Hertz noted, to justifiably regard these results as evidence that the current and rays were distinct one needed to rule out a possibility. If the cathode rays exerted some non-electromagnetic effect on the needle, then one could not use the deflection of the needle to map the current, and thus could draw no conclusion about the relationship between the current paths and the ray paths. Hertz performed another experiment to show that “no such effect occurs” (Hertz 1896, 239). In this experiment, Hertz devised a radially symmetric cathode-anode construction. The cathode consisted of a brass disk with a diameter equal to the opening of the tube into which it was inserted. Through a hole in the center of that disk protruded a thermometer tube, through which in turn protruded the anode. Hertz positioned the tube “as near as possible to the magnet, first in a position that the magnet would indicate a force tangential to the tube, then radial, and lastly, parallel to the tube.” But he found no deflection of the magnet. By contrast, if the anode was placed further down the length of the tube, so that the path from cathode to anode ran parallel to the rays, he did observe “deflections of thirty to forty scale divisions” (240). Hertz concluded that the rays had no detectable non-electromagnetic effect on the magnet, since eliminating the electromagnetic effect via a symmetric configuration eliminated any detectable deflection. Hertz’s use of a subsidiary experiment here exemplifies what I call a “strengthening strategy” for securing his conclusions from the mapping experiment. That is, he added to his knowledge, strengthening his epistemic situation so that some scenarios (those in which cathode rays exert detectable non-electromagnetic effects on the magnetic needle) that were previously epistemically possible and would potentially invalidate any
Fig. 17.1. Hertz’s “tube” turned into a rectangular frame. From Hertz (1896).
224
Philosophy of Science Matters
b
a
Fig. 17.2. The lines are equipotential lines (roughly indicating current paths), which are clearly distinct from the cathode light (a) and the positive striae (b). From Hertz (1896).
conclusions drawn from the mapping experiment, became no longer epistemically possible.
(2) Rays lack electrostatic properties Hertz next turned to the question “Have the cathode rays electrostatic properties?” He divides this question into two parts: “Firstly: do the cathode rays give rise to electrostatic forces in their neighbourhood? Secondly: In their course are they affected by external electrostatic forces?” (Hertz 1896, 249). He then described two experiments directed at these two questions respectively. In the first, as depicted in Figure 17.3, he used a glass tube containing the cathode (α). “All of the parts marked β are in good metallic connection
Evidence and Justification
225
with each other, and such of them as lie inside the tube form the anode” (Hertz 1896, 249). This included a brass tube surrounding the cathode, such that the only opening through which the cathode rays may pass is a 10mm hole opposite the cathode. They must then pass through a wire mesh, also forming part of the anode. The anode also connected to an external mantel connected to a “delicate electrometer.” Also connected to the electrometer was a metal case “which completely surrounds the greater part of the tube and screens that part of the gas-space which lies beyond the wire-gauze from any electrostatic forces which might be produced by induction from without, e.g., from the cathode” (250). On the basis of the previously described experiment, Hertz regarded the rays that pass through the mesh into the space of the tube as “pure” in the sense of not including any of the current that flows from cathode to anode. Thus, should any electrostatic forces be exerted in the vicinity of the rays, or should they be found to respond to any such forces, he could ascribe these effects to the rays themselves. Here again, Hertz took care to secure his conclusions against a threatening scenario: that in which electrostatic effects are manifested, but only because of a commingling of cathode-anode current with the rays which he had established as a distinct phenomenon. But Hertz took a further step, the consideration of which shows how studying the error scenarios deemed relevant by an investigator might shed light on his understanding of the conclusions he draws. To show that his apparatus was adequate to discriminate between the presence and absence of the phenomenon in question, Hertz considered how to determine the magnitude of the effect to be expected if the rays were to exert electrostatic forces. This he did by simulating the effect in question: he replaced the glass tube inside the mantle with a metal rod “which had about the same size and position as the cathode rays” and was in contact with the cathode. This arrangement produced deflections of the electrometer “too great to be measured” but estimated at “two to three thousand scale divisions” which ceased when the current stopped. Note his next comment: “Now if the cathode rays consisted of a stream of particles charged to the potential of the cathode, they would produce effects quantitatively similar to the above, or qualitatively similar if they produced any electrostatic forces whatever in their neighbourhood” (Hertz 1896, 250). When the experiment was performed, Hertz reported that the electrometer exhibited vibrations through “ten or twenty scale divisions from its position of rest.” He infers: “As far as the accuracy of the experiment allows, we can conclude with certainty that no electrostatic effect due to the cathode rays can be perceived; and that if they consist of streams
226
Philosophy of Science Matters
A
b ∝ b
b
b
b
b g
g
B
Fig. 17.3. Cathode tube used to “purify” rays from current. From Hertz (1896).
of electrified particles, the potential on their outer surface is at most one-hundredth that of the cathode” (Hertz 1896, 251). The last point about the “potential on their outer surface” deserves attention. The metal rod served as a satisfactory simulation of the effect being investigated because Hertz assumed that cathode rays bearing electrostatic properties would be “particles charged to the potential of the cathode.” Buchwald comments that “Hertz was . . . thinking of small bits of metal literally torn away from the cathode” and retaining the property of conductivity such that they can be “charged to a potential.” To secure his conclusion that cathode rays did not exert electrostatic forces, Hertz needed to rule out the scenario in which, although they did exert such forces, his apparatus was unable to detect them. But his treatment of this scenario indicates that he was in fact thinking of the “charged ray” hypothesis according to a particular model quite distinct from that later supported by Thomson’s experiments, for example.
Evidence and Justification
227
Of course, to judge Hertz’s accomplishments in his cathode ray research in light of ideas that were not available to him would reek of anachronism. Nonetheless, by following Buchwald’s lead in paying close attention to the error scenarios that did concern Hertz, we can engage in a non-vicious form of anachronism. In particular, we can use a weakening strategy on Hertz’s behalf to clarify in what sense Hertz’s conclusion from this experiment was justified, but in a way that is informed by our knowledge of later developments. The fact that the developments that make our reconstruction possible resulted from later work can form the basis of understanding why Hertz himself did not employ this strategy. The following weakening strategy seems applicable: replace h: “the electrostatic and electromagnetic properties of the cathode rays are either nil or very feeble” with the weaker claim that “either h or else the cathode rays bear those properties in some manner distinct from that in which bodies with the property of conductivity bear them.” The latter possibility did not occur to Hertz, and in that sense he justifiably drew his conclusion. Yet his justification did not meet the ideal set forth in JE. Ideally, experimenters anticipate all of the error scenarios relevant to their inferences and acquire the resources to rule them out. In truth, the ideal is rarely met, and Hertz did not meet it. Employing the framework of security in understanding the ideal, along with additional empirical knowledge about the phenomena Hertz was investigating, allows us both to acknowledge the justifications available to him, and to note those that eluded him.
(3) Cathode rays are not affected by electrostatic forces Finally, Hertz addresses the question of whether electrostatic forces affect cathode rays. Hertz used a “ray purifying” tube similar to that in Figure 17.3, but this time placed between “two strongly and oppositely electrified plates” (Hertz 1896, 252). In the path of the rays Hertz placed a “fine wire” that cast a shadow on the phosphorescent patch at the end of the tube. Hertz proposed that a movement of the shadow would indicate any deflection of the rays. This effect he did not observe. In another instance of anticipating an error scenario, Hertz notes, “But here there was a doubt whether the large electrostatic force to which the tube was subjected might not be compensated by an electrical distribution produced inside it” (Hertz 1896, 252). As Buchwald notes, the concern here was directed at the idea that the gas inside the tube may have been rendered conductive in the manner of “an attenuated metal,” not in the sense of the gas becoming ionized—the effect later identified by Thomson (Buchwald 1994, 166).
228
Philosophy of Science Matters
To rule out this possibility, Hertz moved the plates to the inside of the tube. Under these conditions, Buchwald notes, conductivity in the gas as Hertz conceived it would result in discharge between the plates, an effect that was not observed. The shadow showed no displacement. To investigate the effect of electromotive rather than electrostatic forces on the cathode rays, Hertz then connected the plates to a battery sufficient to induce arc-discharges between them. Hertz observed that under these conditions the phosphorescence appeared “distorted through deflection in the neighbourhood of the negative strip; but the part of the shadow in the middle between the two strips was not visibly displaced” (Hertz 1896, 252). This final series of observations acquired particular importance later insofar as Thomson’s series of experiments included a similar arrangement but with a superior evacuation of the tube to rule out precisely the kind of ionization of the gas (and resultant shielding of the rays from the electric field) that was not the target of Hertz’s concern about the gas being rendered conductive. Buchwald notes that, although many believe that Hertz’s tube suffered from an ionization effect, such a scenario “would not even have occurred to Hertz in 1883, because he understood conductivity as an extensive property of matter in bulk . . . from Hertz’s point of view the gas in the tube was either conducting, in which case it behaved like an attenuated metallic mass, or else it was not, in which case it behaved like an insulator. He had to examine whether the rays were or were not deflected under both circumstances, which he did” (Buchwald 1994, 168). Here again we see how meticulously Hertz ruled out error scenarios that he deemed relevant against the background of the theoretical possibilities within his ken. Nonetheless, possibilities did emerge later that would have been relevant had they occurred to him.
4. CONCLUSION Under Achinstein’s interpretation, the claim that Hertz’s results are evidence that “cathode rays are not electrically charged” is correct insofar as we take the relevant evidence concept to be ES-evidence, but Thomson’s later experiments show that it is incorrect in terms of potential or veridical evidence. Here I take no issue with these claims. The purpose of the above discussion has been to show how we can regard much of what Hertz did in the course of his experimental activities as serving two purposes simultaneously. First, he sought to arrive at results that stand as a matter of
Evidence and Justification
229
objective fact in a relationship of evidential support for a hypothesis of interest. Second, he sought to acquire the epistemic resources to justify the claim that those results are such evidence for that hypothesis, particularly by being able to rule out scenarios that, if actual, would invalidate such a claim. Furthermore, although Achinstein’s evidence concepts— particularly the notion of veridical evidence—might serve us well in understanding how Hertz’s efforts were directed at the first aim, he has no account of justification that will shed substantive light on the second aim. I propose that the securing of evidence, as here discussed, can illuminate this justificatory aspect of science.
REFERENCES Achinstein, P. 2001. The Book of Evidence. New York: Oxford University Press. Buchwald, J. Z. 1994. The Creation of Scientific Effects. Chicago: University of Chicago Press. Chalmers, D. 2011. The Nature of Epistemic Space. In Epistemic Modality, ed. A. Egan and B. Weatherson. Oxford: Oxford University Press. DeRose, K. 1991. Epistemic Possibilities. Philosophical Review 100 (4): 581–605. Hertz, H. 1896. Miscellaneous Papers, trans. D. E. Jones and G. A. Schott. London: MacMillan and Co. Hintikka, J. 1962. Knowledge and Belief: An Introduction to the Logic of the Two Notions. Ithaca, N.Y.: Cornell University Press. Mayo, D. 1996. Error and the Growth of Experimental Knowledge. Chicago: University of Chicago Press. Mayo, D. and A. Spanos. 2009. Error and Inference: Recent Exchanges on Experimental Reasoning, Reliability, and the Objectivity and Rationality of Science. New York: Cambridge University Press. Staley, K. and A. Cobb. Forthcoming. Internalist and Externalist Aspects of Justification in Scientific Inquiry. Synthese. DOI: 10.1007/s11229-010-9754-y
NOTES 1. Here Achinstein refers specifically to epistemic justification, and one should construe all references to justification in this essay likewise. 2. The discussion in this section parallels that of Staley and Cobb (forthcoming), in which we articulate a similar suggestion in the context of Deborah Mayo’s error-statistical account of evidence (Mayo 1996; Mayo and Spanos 2009). 3. Just how to formulate the semantics of such statements is, however, contested (see, e.g., DeRose 1991 and Chalmers 2011). The central claims of the present proposal are independent of disputed issues regarding the semantics of epistemic possibility. 4. One might entertain as well an alternative formulation that replaces potential with veridical evidence. Such a formulation entails that the ideal is met only
230
Philosophy of Science Matters
if one already knows h to be true (lest it be the case that for all one knows h might be false, in which scenario e is not veridical evidence for h). Because I do not wish to claim that JE is the only justificatory ideal that might be relevant to science, I offer (1) and (2) as sufficient but not necessary conditions. Someone who wishes to embrace a single ideal of justification might reformulate these conditions as both sufficient and necessary. 5. For details regarding Hertz’s research, see Jed Buchwald’s fascinating and careful discussion (Buchwald 1994, 131–74), on which I draw heavily in what follows.
18 What Was Perrin’s Real Achievement? Bas C. van Fraassen
1. PHILOSOPHY LORE ABOUT THE PERRIN EPISODE There is a bit of conventional wisdom often recounted by Peter Achinstein and other Scientific Realists concerning the history of science: LORE: until the early twentieth century, there was insufficient evidence to establish the reality of atoms and molecules; but then Perrin’s experimental results on Brownian motion convinced the scientific community to believe that they are real.
There are two views regarding the rationale for the nineteenth century disputes over, and in opposition to, the atomic theory. Steven Brush and John Nyhof, for example, argued that the opponents held positivist philosophical presumptions against admitting the unobservable1. Penelope Maddy holds, on the contrary, that the dispute was purely scientific. Either way, once the philosophers’ lore is accepted, the question becomes only how we can understand Perrin’s work as epistemically legitimating the conclusion drawn from it, that is, the reality of atoms and molecules. This question of legitimation (with its presupposition intact) is addressed by Wesley Salmon, Clark Glymour, and Peter Achinstein, with different answers (Salmon 1984, 221; Glymour 1975, 409, n. 12; Achinstein 2001, 243–65). Achinstein criticizes earlier accounts of Perrin’s reasoning, and offers his own “legitimation”—surprisingly, the premises he lists include a prior probability of at least ½ for the atomic hypothesis, and the conclusion is only that it is “greater” (Achinstein 2001). Maddy takes for granted that Perrin’s reasoning and results do legitimate the conclusion, and does not offer a competing account to these (Maddy 2001). Of course, this question of legitimation arises on the assumption that the story presents both historical events that happened and what was scientifically at stake, and therefore it presents the real significance of this scientific advance. One presumption involved is that success of a theory means that
231
232
Philosophy of Science Matters
it comes to be believed to be true, and that the work done to that end was significant precisely in the way and to the extent that it produced evidence and arguments to justify that belief. This presumption is supported by a plethora of quotes from eminent scientists of the time, including erstwhile opponents of the atomic theory who changed their minds, that show that the advance consisted in demonstrating, beyond all reasonable doubt, that the atomic hypothesis was true. But do scientists, in practice, make the distinction so familiar to philosophers, between what is true and what is good for the future of their enterprise? Do they make the distinction between, on the one hand, counsel to doubt that there are atoms and, on the other, counsel to doubt that the atomic hypothesis points in a good direction for the advance of physics? When scientists describe the acceptance of a scientific theory, do they think in terms of such distinctions as those between truth, empirical adequacy, and pragmatic value? Even if particular scientists do so, should we take their judgments as free of interpretation? Or should we take them as unconditioned by social, cultural, or educational factors? Whether on the lips of scientists or of philosophers, it remains that LORE is an interpretation, though unacknowledged as interpretation. It can be challenged—or perhaps I should say, exposed as such—by presenting an alternative interpretation, and scrutinizing the credentials these rival interpretations may have. Only if alternative interpretations are considered can we see whether there are ambiguities in the story, and whether there are interpretative leaps.
2. DIFFICULTIES BESETTING THIS PHILOSOPHICAL LORE When the story is told in terms current in philosophy of science, we must be especially critical. Thus Maddy says simply: . . . in a case like the post-Einstein/Perrin atomic theorist, it seems incorrect to interpret the claim “there are atoms” to mean that the assertion of the existence of atoms is empirically adequate: it was considered empirically adequate before Einstein and Perrin; afterwards it graduated to another status. (Maddy 2001, 59)
But “empirically adequate” is a philosophical term of art; the scientists did not have that term. If they had had it, they certainly could not have thought that the evidence established that the atomic theory was empirically adequate, for that claim would extend far beyond the evidence. The history is anyway badly portrayed! If the reference is instead to empirical success conceived in a broader sense (taking “empirically adequate”
What Was Perrin’s Real Achievement?
233
in a less philosophically technical sense), then Maddy does follow Perrin’s presentation, but that portrayal looks Pollyannic given the severe problems of the atomic theory in the two decades preceding Perrin’s work.2 As always, the bottom line in the empirical sciences is to meet the criteria of success that relate directly to test and experiment. So let us leave this philosophical lore and the realism/empiricism debate behind and look into some actual empirical constraints on theories and models.
3. THE EMPIRICAL GROUNDING CRITERION The understanding in practice of what is required for a good empirical theory was in evidence throughout the development of modern science, but was not explicitly formulated until recently. I will begin with an early example to guide us, and then present a contemporary formulation, before entering on its application to the development of the atomic theory.
(1) Newtonians and the Cartesian critique The Cartesians’ critique of Newton was that, with his introduction of non-kinematic parameters such as mass and force, he had brought back “occult qualities.” The Newtonian response was, in effect, that admittedly what is measured directly are lengths and durations, but that they could show nevertheless how to measure mass and force. The rationale of this response was thoroughly reinvestigated in the nineteenth and early twentieth century, by Mach, Duhem, and Poincaré.3 As they showed, the measurement of those dynamic parameters on a body is an operation that counts as such a measurement relative to Newtonian theory. To say that the operation measures mass, for example, is to presuppose the applicability of Newton’s second and/or third law. So, for example, measurements of mass ratios with the Atwood machine,4 or by contracting springs, presuppose that the setup as a whole is a Newtonian system. The values of the masses are indeed calculated from the observations of kinematic quantities, but via Newton’s laws. The Newtonian response was precisely to the point, and it reveals quite clearly the norms concerning empirical constraint accepted in modern scientific practice. All the parameters in the theoretical models must admit of such empirical grounding.5 If not, they are empirically superfluous, and provide an obstacle to the acceptability of the theory. The baseline criteria for science are empirical. That explains why hidden variable theories do not get any attention among the scientists
234
Philosophy of Science Matters
T
M T
Mg m
mg
Fig. 18.1. The machine of Rev. George Atwood (1746–1807). The accelerations are equal but opposite in direction, and proportional to (M−m)/(M+m), which determines M/m.
SPRING-CONNECTED MASSES
M m
CM
Fig. 18.2. The system’s center of mass (CM) is unchanged as the two carts approach one another.
themselves, as opposed to philosophers, until and unless there is some suggestion of a possibility of empirical testing. It is not relevant to object that all the evidence is as much in accord with the hidden variable variant as with the original. Parameters introduced into modeling must not be empirically superfluous—there must be, in some way, even if at some distance, coordination with empirically differentiating phenomena. Sometimes the parameters that appear to be empirically superfluous can be simply removed without imperiling either the coherence of the theory or its empirical strength and credentials. The “grounding” requirement turns into a salient problem only when elimination is not possible, that is, when
What Was Perrin’s Real Achievement?
235
there are no theoretically specifiable conditions in which their values can be determined, relative to the theory, on the basis of measurement results. The appropriate, and typical, response in that case is to start enriching the theory so that it becomes more informative, informative enough to allow the design of experiments in which this empirical determination of the values does become possible.6 But meanwhile, can we imagine the Cartesians’ feelings? Those measurements of mass or force make sense only in the context of the assumption that the setup or target is itself a Newtonian system—something that the Newtonian postulates. So how, in what sense, is this evidence that bears out Newton’s theory? How can the evidence, taken in a way that is neutral between the Cartesians and the Newtonians, legitimate the conclusion to the truth of the Newtonian theory? We can imagine the Cartesian asking these questions, and the dissatisfaction with this on the Cartesian side, especially since Cartesian general epistemology is paradigmatically foundational. But in this (uncharitable? anachronistic?) imagined response, the Cartesian is barking up the wrong tree—as are philosophers today if they remain fixated on the question of evidence for the truth of a theory or for the reality of the entities a theory postulates.
Weyl and Glymour: The empirical constraints on science. The relevant methodological insight was, as I said, formulated much later; some of the current philosophical “conventional wisdom” seems never to have assimilated it. As initial formulation, here is Hermann Weyl in his Philosophy of Mathematics and Natural Science:7 1. Concordance. The definite value which a quantity occurring in the theory assumes in a certain individual case will be determined from the empirical data on the basis of the theoretically posited connections. Every such determination has to yield the same result. . . . Not infrequently a (relatively) direct observation of the quantity in question . . . is compared with a computation on the basis of other observations . . . 2. It must in principle always be possible to determine on the basis of observational data the definite value which a quantity occurring in the theory will have in a given individual case. (Weyl 1963, 121–2)
It is easier to read these points in reverse order. Given that one is called Concordance let us call the other Determinability. This deserves detailed discussion, but for now I just want to put the following empirical grounding requirement on center stage: • Determinability: any theoretically significant parameter must be such that there are conditions under which its value can be determined on the basis of measurement.
236
Philosophy of Science Matters
• Concordance has two aspects: ° Theory-Relativity: this determination can, may, and generally must be made on the basis of the theoretically posited connections Uniqueness: the quantities must be “uniquely coordinated”; there ° needs to be concordance in the values thus determined by different means.8 There is, at first blush, a glaring possible objection to the completeness of this formula, if viewed as putatively sufficient. If the theory’s being thus borne out by experimental and measurement results is on the basis of the theoretically posited connections, why does that fact not trivialize the putative evidence? This concern was addressed explicitly by Clark Glymour in his account of relevant evidence and testing. Glymour was explicitly following Weyl here, but saw the need for the additional constraint to prevent self-fulfilling prophecy in science. I will adapt the following to our present purpose, from Glymour’s Theory and Evidence—adapt and amend, since his presentation of the “bootstrapping method” was confusingly conflated with what was then called “confirmation theory.”9 For simplicity let’s take theory T to be presented as a set of equations, involving certain parameters, some directly measurable and some theoretical, and take relevant evidence to consist similarly in a set of equations that simply assign values to some of the measurable parameters. Then Glymour imposes the constraint that there must be an alternative possible outcome for the same measurements that would have refuted the hypothesis on the basis of the same theoretically posited connections. His conception may be presented initially as follows: E provides relevant evidence for H relative to theory T exactly if E has some alternative E’ and T some subtheory T’ such that: (1) (2) (3) (4)
T ∪ E ∪ {H} has a solution. T’ ∪ E’ has a solution. All solutions of T’ ∪ E are solutions of H. No solutions of T’ ∪ E’ are solutions of H.
For example, if T consists simply of the equation P(t)V(t) = RT(t), with R a theoretical constant, then we can take H to be just T itself, and E could be E = {P(1) = 2, V(1) = 3, T(1) = 30; P(2) = 3, V(2) = 1, T(2) = 15}
which satisfies T while determining the value of R to be 5. It has the requisite possible alternative that could have been found instead; for example: E¢ = {P(1) = 2, V(1) = 3, T(1) = 30; P(2) = 3, V(2) = 1, T(2) = 11}
What Was Perrin’s Real Achievement?
237
which does not satisfy T for any possible value of R. (Here the subtheory T’ is trivial, empty, or tautologous, which is what makes the example very simple.) The threat of trivializing circularity or vacuity may not be entirely eliminated, in logical principle, by Glymour’s additional requirement. It would be surprising if we could find complete sufficient conditions for having an empirically solid theory so quickly. But satisfaction of the above requirements characterizes well and clearly what can be offered on behalf of the significance of a particular empirical grounding of the theoretical parameters in any specific case.
4. THE PROBLEM OF EMPIRICAL GROUNDING IN THE NINETEENTH CENTURY Now we have come to the besetting problem of the atomic theory that Dalton introduced early in the nineteenth century, and that was extended into the kinetic theory of heat, and finally into the statistical mechanics that rivaled phenomenological thermodynamics. I’ll use the term “kinetic theory” to refer to all of that, for short. This methodological demand for empirical grounding, that we see so clearly operative throughout the modern history of science, applies to the kinetic theory as well. The attitude toward the atomic and molecular structure postulated in the nineteenth century was precisely that the models provided by the atomic theory must be thoroughly coordinated with measurement procedures. Let’s make the demand explicit in general terms: (I). If two such models of a given phenomenon differ only in the values of certain parameters, there must be in-principle measurement results that will differentiate between them. (II). Similarly, for any distinct states in the theory’s state-space, in which the model locates the systems’ trajectories, there must be in-principle measurable quantities that differentiate them. The term “in-principle” refers here not just to the idealization that measurements have unlimited precision, but also to Weyl’s observation that the differentiation is not crudely theory-neutral, but on the contrary, relative to the theory itself (and perhaps to additions from background theory). If these demands are satisfied, let us call those parameters, or the theory, empirically well-grounded. In a kinetic model of a gas, there are many parameters that pertain to the individual molecules. The empirical success of such models is related to the measurement of “gross” quantities such as mean kinetic energy. If two such
238
Philosophy of Science Matters
models of a gas agreed on those quantities that were meaningfully measurable in phenomenological thermodynamics, but differed in the values of such parameters as individual masses, sizes, momenta, or number of molecules, could there be measurements to differentiate those, in principle? Philosophers’ history of the scientific research directed to this question has largely seen it displaying philosophical rather than scientific motivations. But if we look at the texts with new eyes we see that the objections and challenges concerned precisely the question of whether the parameters in the atomic theory could have their values determined by measurement relative to the theory—the question of empirical grounding.
5. PERRIN BEGINS To report on Perrin’s work here I will rely on the general presentation he gave of his work in 1909, just a year after the publication of his epochmaking experimental results.10
(1) How and where empirical grounding is needed Early on in his account, Perrin lists the parameters that have resisted satisfactory or sufficient empirical grounding to date. Throughout the nineteenth century, hypotheses had been added to enrich the basic kinetic theory, and this meant that more and more procedures could be designed that would count as measurements, to determine values or at least to relate values of various quantities to each other. The prime example, very early on, was the addition of Avogadro’s hypothesis that allowed deduction of molecular mass ratios. In Perrin’s formulation, that hypothesis is that any two gram-molecules of a substance contains the same number of molecules, Avogadro’s number N.11 There is a similar theoretical relation between N and the mean kinetic energy of the molecules, via the ideal gas law; and this can in the same way be used to yield an equation connecting N and the mean square molecular speed. The perfect gas law is the well known equation PV = RT, where R is the perfect gas constant and the temperature T was proved to be proportional by the factor 3R/2N to the mean kinetic energy (Perrin 2005, 11–2). So we have as resultant the equation pV = (1/3)Nm<s2>, where • N = the number of molecules • m = the mass of each molecule • <s2> = the mean square speed of the molecules
What Was Perrin’s Real Achievement?
239
Pressure p and volume V can be measured directly; but on the right we see three theoretical parameters. The number of unknowns can be reduced still more if we bring in more such equations. For example, Perrin points out, when the theory of electricity is also brought into the fold, a relation can be deduced between the minimal electric charge e and this number N. On the kinetic theory, electrolysis is explained by postulating that in electrolysis the molecules are dissociated into ions carrying a fixed electric charge. A Faraday is the quantity F of electricity that passes in the decomposition of 1 gram-molecule of hydrochloric acid; it is at the same time equated to the charge carried by one gram-molecule of hydrogen ions. It is known empirically that the decomposition by current of a gram-molecule of an electrolyte displays always the passing of the same quantity of electricity, and it is always a whole number of Faradays. This number must be the product of the number of ions taken with the number of minimal electric charges e that they carry. Putting these equations together and noting that by hypothesis one gram-molecule of hydrogen consists of N hydrogen atoms, we have Ne = F
where F is an empirically known quantity, and we have two theoretical parameters. Of course the two above equations can be combined, so as to place an equivalent constraint on just three of the theoretical parameters, with the fourth defined in terms of the other three. It is easily seen that these theoretical developments consist only partly in calculations, and partly in the introduction of further hypotheses to elaborate on the basic kinetic model. At this point, measuring an electric charge, a mass, and a volume places on the numerical relations between parameters pertaining to the molecules and their motion some quite definite constraints relative to the theory as developed so far. In his exposition prior to the statement of his results, Perrin continues these points by adding in a similar hypothesis due to Maxwell, on the statistical independence of the spatial components of molecular speeds. To be specific, Maxwell derived a law of the distribution of molecular velocities in a gas, but on the special assumption that the distribution along spatial direction x is statistically independent of the distribution along the orthogonal y and z axes. Adding then a special hypothesis relating the internal friction between two parallel layers of a gas that are moving at different speeds to exchange of molecules between these layers, Maxwell found a further linkage to measurement, namely that:
240
Philosophy of Science Matters
the coefficient z of internal friction, or viscosity, which is experimentally measurable, should be very nearly equal to one-third of the product of the following three quantities: the absolute density δ of the gas . . . the mean molecular speed Ω . . . and the mean free path L which a molecule traverses in a straight line between the two successive impacts. (Perrin 2005, 14)
This gives information about the mean molecular speed, here designated as Ω. Adding to this a further hypothesis that provides a kinetic model for the internal friction between two parallel layers of gas, moving at different speeds, Maxwell arrived at the equation z = 0.31dWL
where: • z is the coefficient of internal friction (viscosity)—an experimentally measurable parameter • δ is the absolute density of the gas (also measurable) • W is the mean molecular speed (mentioned above) • L is the mean free path of the molecules Given the hypotheses and measurement results so far then, the mean free path is calculable. Then with the still further addition that the molecules are spheres— one of very few shapes the kinetic models had admitted—Clausius and Maxwell derived an equation that fixes the molecular diameter approximately as a function of the mean free path and the number n of molecules per cubic centimeter. The latter will bring us to Avogadro’s number N, but what is still needed to solve the equation in question is a second constraint on the relation between the molecular diameter and that number n. So now we have seen what the first part of the 1910 monograph was devoted to spelling out: that empirical research and theoretical development in tandem had progressed to the point where you could see that relative to the theory (taken sufficiently broadly) only one more parameter needed empirical grounding to finish the job. At this stage in history Perrin can point out that relative to the theory, the measurement of any of these so far undetermined, or only partially determined, parameters would fix the others as well (Perrin 2005, 12). Thus we see in principle a number of approaches to the empirical grounding of the so far remaining indispensable parameters in the models provided by the kinetic theory. By this we must mean of course: operations that will count as measurement, relative to the theory, that is, utilizing the above theoretically derived equations.
What Was Perrin’s Real Achievement?
241
Perrin’s research on Brownian motion is directed precisely to this end; and to begin it was quite independent of the new theoretical results due to Einstein. After his initial successes in determining values for those parameters, he continued by guiding further experiments in ways linked to Einstein’s work, and found good agreement with the previous results.12 To be realistic, we should note that the theoretical derivations that Perrin assumes are also largely dependent on assumptions added to the kinetic theory, in the construction of specific models. Most of the work proceeds with models in which the molecules are perfect spheres, for example, though Perrin notes that other hypotheses are needed in other contexts (Perrin 2005, 14). As long as the simple models work, to allow a transition from the empirically obtained results to values for the theoretical parameters, and as long as these values obtained in a number of different ways agree with each other and with what is theoretically allowed—to within appropriate margins of error—this counts as success. The addition Perrin made to this already almost century-old story follows the same pattern. As Achinstein emphasizes, Perrin also introduces an addition to the theory, a “crucial assumption, viz. that visible particles comprising a dilute emulsion will behave like molecules in a gas with respect to their vertical distribution.” (Achinstein 2001, 246) Not that this is a blithe addition: Perrin argues for its plausibility, but in terms that clearly appreciate the postulational status of this step in his reasoning. After a discussion of the range of sizes of molecules (according to results derived from measurements via such extensions of the theory as we have just been inspecting) he writes, Let us now consider a particle a little larger still, itself formed of several molecules, in a word a dust. Will it proceed to react towards the impact of the molecules encompassing it according to a new law? Will it not comport itself simply as a very large molecule, in the sense that its mean energy has still the same value as that of an isolated molecule? This cannot be averred without hesitation, but the hypothesis at least is sufficiently plausible to make it worthwhile to discuss its consequences. (Perrin 2005, 20)
On this basis, the results of measurements made on collections of particles in Brownian motion give direct information about the molecular motions in the fluid, always of course within the kinetic theory model of this situation. But that is just what was needed for empirical grounding of those remaining theoretical parameters. This was not the end of the story for Perrin. What Weyl calls the requirement of concordance and unique coordination was apparently very much on his mind. Perrin begins Part III of this 1910 work with the remark
242
Philosophy of Science Matters
that his experiments have allowed “the various molecular magnitudes to be determined” but then adds, But another experimental advance was possible, and has been suggested by Einstein at the conclusion of the very beautiful theoretical investigation of which I must now speak. (Perrin 2005, 51)
Perrin notes that Einstein obtained his results in part “by the aid of hypotheses which are not necessarily implied by the irregularity of the Brownian movement” and details two of them. These include his own main hypothesis, namely that “the mean energy of a granule is equal to the molecular energy” (2005, 53). After discussing a rather large amount of experimental work bearing on Einstein’s results, and its nevertheless inconclusive outcome, Perrin himself set about (to use his own words) an experimental confirmation of Einstein’s theory. In this he was very successful as well. Not only that: in his experimental work related to Einstein’s theory, he arrived at the same values for the theoretical quantities as he had found in his own previous research. Logically speaking, the outcomes could have been at odds with each other, since no matter how tightly the theory is constructed, the actual results of measurement are after all “up to nature”. So we can read this part of the story as not simply a further inquiry but a demonstration that Weyl’s concordance requirement is taken into account and the credentials of the theory with respect to this empirical constraint are demonstrably provided.
(2) How Perrin ends his 1910 monograph Finally, although Perrin’s text is such a boon to scientific realist writing, I think we should attend to his own emphasis on how thoroughly empirical his work was. His explanation is precisely in line with what I have here displayed as the project of empirical grounding. This comes in his final section, headed “43. Molecular reality,” and begins with the telling sentence Lastly, although with the existence of molecules or atoms the various realities of number, mass, or charge, of which we have been able to fix the magnitude, obtrude themselves forcibly, it is manifest that we ought always to be in a position to express all the visible realities without making any appeal to elements still invisible. But it is very easy to show how this may be done for all the phenomena referred to in the course of this Memoir. (Perrin 2005 [1910], 91)
He then explains how to isolate the empirical content (still, although he does not say so, relative to the theory!) of the theoretical results. For example, he suggests comparing two laws in which Avogadro’s constant enters:
What Was Perrin’s Real Achievement?
243
The one expresses this constant in terms of certain variables, a, a¢, a², …, N = f[a,a¢,a²....];
the other expresses it in terms of other variables, b, b¢, b²,…, N = g[b,b¢,b²….].
Equating these two expressions we have a relation f[a,a¢, a²….] = g[b,b¢,b²….]
where only evident realities enter, and which expresses a profound connection between two phenomena at first sight completely independent, such as the transmutation of radium and the Brownian movement (Perrin 2005, 91–2). Once again, it seems to me mistaken to read this in a philosophical vein. I do not offer this as a case of an apparent scientific realist contributing grist for the empiricist’s mill. Rather, this passage is important because of how it illustrates how the factors of Determinability and Concordance function in empirical grounding.
6. CONCLUSION In sum then, I propose we see the century-long story of research to establish the credentials of the kinetic theory as a truly empirical enterprise— not as a century-long search for independent evidence for the truth of a well-defined hypothesis about what nature is like, but in a quite different light! Perrin aimed to develop the theory itself, and to enrich it so as to allow construction of models for special cases in its domain—all so as to make empirical grounding possible for its theoretical quantities. That enterprise essentially involves the concurrent development of measurement procedures to implement the grounding thus made possible. It is neither purely theoretical nor purely empirical, the theoretical and empirical are indissolubly entangled, but what is achieved is an empirical success. One greatly gratifying aspect of Perrin’s work was that when he followed up his own research on Brownian motion with an experimental inquiry into Einstein’s new theoretical development, he found a satisfactory concordance in the results obtained. It is still possible, of course, to also read these results as providing evidence for the reality of molecules. But it is in retrospect rather a strange reading—however much encouraged by Perrin’s own later prose and by the commentaries on his work in the scientific and philosophical
244
Philosophy of Science Matters
community. For Perrin’s research was entirely in the framework of the classical kinetic theory in which atoms and molecules were mainly represented as hard but elastic spheres of definite diameter, position, and velocity, at a time when Rutherford was earning the Nobel prize for subatomic research. Moreover, the monograph begins with the conviction on Perrin’s part that there is no need at this late date to give evidence for the general belief in the particulate character of gases and fluids. On the contrary (as Achinstein saw), Perrin begins his theoretical work in a context where the postulate of atomic structure is taken for granted. What can we make of all those triumphant remarks that suggest otherwise? I submit that we can interpret them as follows: that Perrin’s work laid to rest the idea that it might be good for physics to opt for a different way of modeling nature, one that rivaled atomic theories of matter. Precisely that is what was, in retrospect, well vindicated—an outcome as welcome to empiricists as to scientific realists, I would say. But for the methodology and epistemology of science the most salient conclusion to draw is, it seems to me, that evidence can be had only relative to the theories themselves (the “bootstrapping” moral) and that this is so because a theory needs to be informative enough to make testing possible at all. Thus the extent to which we can have evidence that bears out a theory is a function of two factors: first, of how logically strong and informative a theory is, sufficiently informative to design experiments that can test the different parts of the theory relative to assumptions that the theory applies to the experimental set-up; and second, of how well the measurement results in situations of this sort are in concord with each other. But third, the testing involved cannot be adequately or properly portrayed as just checking on implied consequences (along such lines as suggested by the “hypothetico-deductive method” or Hempel’s “confirmation theory”). To properly credential the theory, the procedures that count as tests and measurements in the eyes of the theory must provide an empirical grounding for all its significant parameters. The completion of this task, which made the kinetic theory into, at least in principle, a truly empirical theory, was Perrin’s real achievement.
ACKNOWLEDGMENTS This is a somewhat modified version of my 2009 paper (van Fraassen 2009), offered here gratefully to Peter Achinstein for all that I learned from his work over the years. My thanks to the commentary by Helen Longino and the discussion at the 2008 Oberlin symposium, and for Greg Morgan’s help with revision. Research for this essay was supported by NSF Senior Scholar Award SES-0549002 and Award 1026183.
What Was Perrin’s Real Achievement?
245
REFERENCES Achinstein, P. 2001. The Book of Evidence. Oxford: Oxford University Press. Brush, S. G. 1976. The Kind of Motion We Call Heat. Amsterdam: North-Holland Publishing Company. Clark, P. 1976. Atomism Versus Thermodynamics. In Howson ed. Method and appraisal in the physical sciences. Cambridge: Cambridge University Press. De Regt, H. W. 1996. Philosophy and the Kinetic Theory of Gases. British Journal for the Philosophy of Science 47 (1): 31–62. Duhem, P. 1996. Essays in the History and Philosophy of Science, trans. R. Ariew and P. Barker. Indianapolis: Hackett. Earman, John, ed. 1983. Testing Scientific Theories. Minnesota Studies in the Philosophy of Science, vol. X. Minneapolis: University of Minnesota Press. Glymour, C. 1980. Theory and Evidence. Princeton: Princeton University Press. ——— . 1975. Relevant Evidence. Journal of Philosophy 72 (14): 403–26. Hanson, N. R. 1958. Patterns of Discovery. Cambridge: Cambridge University Press. Kuhn, T. S. 1961. The Function of Measurement in Modern Physical Science. Isis 52 (2): 161–93. Mach, E. 1960. The Science of Mechanics: A Critical and Historical Account of Its Development, 6th English edition. LaSalle, Ill.: Open Court. Maddy, P. 2007. Second Philosophy: A Naturalistic Method. Oxford: Oxford University Press. ——— . 2001. Naturalism: Friends and Foes. Philosophical Perspectives 15: 37–67. ——— . 2000. Naturalism in Mathematics. Oxford: Oxford University Press. Nyhof, J. 1988. Philosophical Objections to the Kinetic Theory. British Journal for the Philosophy of Science 39 (1): 81–109. Perrin, J. 2005. Brownian Movement and Molecular Reality, trans. F. Soddy. New York: Dover. Poincaré. 1905. Science and Hypothesis. London: Walter Scott Publishing. Salmon, W. 1984. Scientific Explanation and the Causal Structure of the World. Princeton: Princeton University Press. Van Fraassen, B. C. 2009. The Perils of Perrin, in the Hands of Philosophers. Philosophical Studies 143 (1): 5–24. ——— . 2008. Scientific Representation: Paradoxes of Perspective. Oxford: Oxford University Press. ——— . 1983a. Glymour on Evidence and Explanation. In Testing Scientific Theories, ed. J. Earman. Minneapolis: University of Minnesota Press. ——— . 1983b. Theory Comparison and Relevant Evidence. In Testing Scientific Theories, ed. J. Earman. Minneapolis: University of Minnesota Press. Weyl, H. 1963. Philosophy of Mathematics and Natural Science. New York: Atheneum.
NOTES 1. “In retrospect it seems clear that the criticisms of the kinetic theory in this period were motivated not primarily by technical problems, such as specific heats of polyatomic molecules but, rather by a general philosophical reaction against mechanistic or “materialistic” science and a preference for empirical or phenomenological theories. as opposed to atomic models” (Brush 1976, 245). See further Nyhof 1988.
246
Philosophy of Science Matters
2. Perrin took a definite side in a live controversy, though one that was already being overtaken by the radical conceptual departures from classical mechanics at that time (in which Perrin took no part). The difficulties of the kinetic theory between 1880 and 1905 are graphically described by Clark (1976, 82–8), and while his philosophical take on this period is challenged by Nyhoff (1988), and their dispute evaluated critically by De Regt (1996), the role and impact of those difficulties on scientific research in that period are not challenged. 3. See Mach 1960, 264–6 and discussions in Duhem 1996, 121–2 and Poincaré 1905, 89–110. 4. I will make this more precise below. It is appropriate, I think, to relate this to the older term “coordination” of Mach, Schlick, and Reichenbach; see van Fraassen 2008, 115–40. 5. It is not the case that a logically weaker theory is automatically better confirmed or better supported by the evidence. The weaker theory may not be informative enough to allow for the design of a test. So, for example, the mere hypothesis of atomic structure, taken by itself, is not a well-tested theory (or tested at all!), though it is part of well-tested theories. 6. This is quoted in Glymour 1980, 50 and discussed there, though it is not clear just how Glymour connects what Weyl writes with Glymour’s own central concern, which was confirmation of hypotheses relative to theories. 7. This third point emphasizes here what Schlick and Reichenbach insisted on in the phrase “unique coordination,” though this was obscured by the lack of clarity, at the time, on the distinctions and relations between truth and verification. 8. In fact Glymour’s conception, to replace the then (and still!) current notions of confirmation, was a measure that was a function of both confirmation and information. For an analysis see van Fraassen 1983a and 1983b. 9. This was translated into English within the year as the book Brownian Movement and Molecular Reality. Although less historically and technically informative on one hand, and on the other hand less explicit with respect to Perrin’s own interpretation of his results, it is much closer to the actual work than his later book Atoms. 10. A gram-molecule of a substance is the mass of this substance which in the gaseous state occupies the same volume as two grams of hydrogen measured at the same temperature and pressure. 11. Here the philosophical literature is not always in accord with Perrin’s own account. As I shall discuss below, Perrin (2005, originally published in 1910) presents his own research entirely before beginning part III with, “But another experimental advance was possible, and has been suggested by Einstein at the conclusion of the very beautiful theoretical investigations of which I must now speak” (2005, 51). Compare to this the order of presentation in Maddy 2000, 139–40 or Maddy 2007, 72, noting also the omitted background of initial experimental setbacks for Einstein’s work (Clark 1976, 97). 12. For the machine’s uses in measurement procedures cf. Hanson 1958, 100–2; Kuhn 1961, 169–72.
19 Causes, Conditions, and the Pragmatics of Causal Explanation Jim Woodward
1. INTRODUCTION Many standard accounts of causation, whether framed in terms of regularities involving necessary and sufficient conditions, statistical relevance, or counterfactual dependence, fail to distinguish factors that (at least in ordinary language) are described as “causes” from those regarded as mere “enabling conditions.” Suppose, to adapt an example from Judith Thomson (2003), a bridge is constructed (B), providing, for the first time, access from the mainland to an island. X crosses the bridge to the island and commits a robbery R there. Suppose B is an INUS condition for R, that it raises the probability of R, and that without B, R would not have occurred. Despite the satisfaction of these standard “objective” conditions for causal relevance, most people are reluctant to judge that B “caused” R. Instead we regard B as merely an “enabling condition” for R. What is the basis for this distinction? Although Peter Achinstein has not, to my knowledge, directly addressed this question, many attempts to answer it appeal to an idea to which (in related contexts) he has been sympathetic: that explanation, including causal explanation, has an important pragmatic dimension, where this means, roughly, that the goodness of an explanation (and perhaps other distinctions we may wish to make among explanations) depends in part on facts about the interests, attitudes, or background knowledge of the explainer or her audience. In particular, many philosophers have claimed there is no “objective” basis for the distinction between causes and conditions, and that the distinction instead has to do with what speakers or their audiences find most more salient, interesting, important, or satisfying. Going further, the very fact that we distinguish between causes and condition is often thought to support the view that important aspects of
247
248
Philosophy of Science Matters
causal explanation are pragmatic in character, since (the argument goes) it hard to see what other basis this distinction might have. This essay is organized as follows. Section 2 focuses on the cause/ condition distinction, arguing that in some substantial range of cases there are objective, structural, “out there in the world” differences between the way that causes and conditions are related to effects, and that these features influence judgments about the cause/condition distinction. Section 3 explores some consequences for pragmatic approaches to causal explanation like Achinstein’s.
2. CAUSAL SELECTION To set the stage for more detailed discussion, it will be useful to begin with the general notion of one factor, event etc. being causally relevant to another. One natural way of characterizing this is in terms of counterfactuals. Think of C and E as variables taking more specific values, corresponding, for example, to the presence or absence of some causal factor. In the simplest case (and the only one that will matter to us) C is causally relevant to E if, were the value of C to be different in some appropriate way, then the value of E would be different. This counterfactual should be understood as non-backtracking or “interventionist” counterfactual, in the sense of Woodward 2003. According to this notion of relevance, often many factors will be causally relevant to outcomes, including many that we would not readily describe as causes in ordinary language: the short circuit is causally relevant to the fire, but so is the presence of oxygen and (arguably) the absence of a sprinkler system; X’s desire for Y’s money is causally relevant to the robbery X commits, but so is the existence of the bridge. In both cases, the first factor cited is typically described as a cause of the outcome in question, while the other factors may be described as mere conditions. To avoid confusion I will write “causen” for this narrow notion of cause that contrasts with “condition.” The problem of “causal selection” is that of understanding the basis on which the causen/condition distinction is made. A widespread philosophical view is that causal selection is capricious and arbitrary, reflecting idiosyncratic facts about explainer and audience, but having no other more objective basis. John Stuart Mill writes, “Nothing can better show the absence of any scientific ground for the distinction between the cause of a phenomenon and its conditions, than the capricious manner in which we select from among the conditions that which we choose to denominate the cause” (Mill 1846, 198). David Lewis expresses a similar sentiment.
Causes, Conditions, and the Pragmatics of Causal Explanation
249
We sometimes single out one among all the cause of some event and call it “the” cause, as if there were no others. Or we single out a few as the “causes,” calling the rest mere “causal factors” or “causal conditions” . . . We may select the abnormal or extraordinary causes, or those under human control, or those we deem good or bad, or just those we want to talk about. I have nothing to say about these principles of invidious discrimination. (Lewis 1986, 162)
Many philosophers (and psychologists) frame this claim about the arbitrariness of causal selection primarily as a claim about how we use (or ought to use) the word “cause,” but there are larger issues at stake. In many areas of scientific inquiry, researchers seem to distinguish among factors that are causally relevant in the broad sense to some explanandum of interest, focusing primarily on just a few of these, and relegating many others to the status of background. For example, many biologists focus on “genes” or “DNA” in contrast to other factors that are also causally relevant to phenotypic outcomes. This in turn prompts critics (e.g., Oyama 2000) to respond that this selective focus is arbitrary and without any principled basis. According to the critics, if two factors C1 (e.g., DNA, genes) and C2 (other cellular machinery, environmental factors) are both causally relevant to some outcome E, considerations of “causal parity” or “causal democracy” favor giving each equal attention. In effect, the critics agree with Mill and Lewis that any distinction among these factors must be without “any scientific ground” or “invidious” and take this to argue for a more egalitarian approach. One obvious response to Mill and Lewis is this: even if the conceptions of causation they favor (causation as the instantiation of a regularity of some kind in Mill’s case, causation as the ancestral of counterfactual dependence in Lewis’s case) provide no resources for distinguishing among the factors that are causally relevant to E (all such factors seem symmetrically related to E) perhaps there are other considerations that, even if influenced by pragmatic considerations, also draw on more objective facts about causal structure that contribute to the basis for the causen/condition distinction. This is the position I favor, but before turning to details, some further explanation is in order. The considerations I will discuss, having to do with (what I call) stability and specificity, should be understood as considerations that often or typically bear on causal selection; I do not claim these give us either strictly necessary or sufficient conditions for some factor to be regarded as a causen (or as causally important, interesting, etc.) Moreover, these are certainly not the only considerations bearing on causal selection. In saying that these considerations bear on selection, I intend to make a broadly empirical claim: both ordinary people’s and scientist’s
250
Philosophy of Science Matters
judgments about the cause/condition distinction and related matters are influenced by these considerations. I also think that it makes normative sense that people’s judgments are so influenced, but defending this claim is beyond the scope of this essay1. Turning to details, I begin with stability. Suppose that we have a relationship of causal relevance (understood as counterfactual dependence) between C and E. The stability of this relationship has to with the extent to which it will continue to hold as various background factors/circumstances Bi change, where a background factor is simply any factor distinct from C and E. The “larger” the range of background factors over which the C ® E relationship continues to hold, or the more “important” these factors, the more stable is the C ® E relationship. Other things being equal, we tend to treat factors C1 that are causally relevant to E and that bear more stable relations to E as “causesn” of E (or more important causes) and those factors C2 that are causally relevant to E but bear less stable relationships to E as “conditions.” Here are two examples, both from Lewis 1986, illustrating the intuition behind stability. In the first, Lewis writes (R) a letter of recommendation with the result that X gets a job she would not otherwise have gotten, meets and marries someone she would not otherwise have met and married, and has children who would not otherwise exist, who then engage E in certain activities. R is causally relevant to the birth B of these children and to E. Nonetheless, we do not find it natural to describe R as a cause of either B or E; instead R is at best an enabling condition. My suggestion is that this judgment is influenced by the fact that the relationship between R and B (and E), although one of causal relevance, is highly unstable or “sensitive” in Lewis’ language. Change any one of a large number of background factors, and the relationship of causal relevance (counterfactual dependence) between R and B/E would no longer hold— this would happen if, for example, a stronger candidate had applied for the job X got, if X had lingered a little less long in the bar where she met her spouse and so on. Contrast this with an example in which Y is shot (S) through the heart by a large caliber bullet at point blank range and dies (D). S is certainly causally relevant to D, but in addition, in comparison with the previous example, the relation of counterfactual dependence between S and D is relatively stable. No doubt there are some possible background circumstances in which, were Y shot in this way, she would not die, but (we may assume) these are unlikely or far-fetched. In most realistic non-science fictitious scenarios in which Y is shot in this way, she dies. This is connected, I suggest, to our willingness to regard S as a cause, and not a mere condition for D.
Causes, Conditions, and the Pragmatics of Causal Explanation
251
I said above that the “larger” or “more important” the range of background factors over which the C ® E relation holds, the more stable this relationship. I lack both the space and the ability to satisfactorily unpack the quoted phrases, but the following considerations are relevant. First, sometimes a natural measure, supplied by subject-matter-specific considerations, over relevant ranges of background conditions will be available and this will allow us to sensibly talk about larger or smaller ranges of such conditions. A gas released into a larger container will expand for all but a set of measure zero initial conditions—this counts as a “large” set and makes the claim about expansion a stable one. Second, background conditions that are common or frequent or typical (around here, right now) count more for purposes of assessing stability than background conditions that are not common—this is part of what led us to regard the S ® D relationship as relatively stable. The limiting case is a situation in which there is a condition Bk, the presence of which is sufficient for E to be counterfactually dependent on C, and Bk is virtually always present—as, for example, oxygen is near the surface of the earth. Third, subject-matterspecific considerations can also tell us that, for certain causal relationships, stability under some background conditions is more important than stability under others. For example, if a gene G is causally relevant to phenotypic trait P in highly specific environmental conditions E, but the G ® P relationship would not continue to hold under environmental conditions different from E, then, ceteris paribus, biologists are likely to regard the G ® P relationship as unstable and to this extent a condition rather than a cause for P. As an illustration, suppose the expression of gene G is required for learning mathematics and let P be the acquisition M of some particular bit of mathematical knowledge. G is causally relevant to M but most biologists would be unlikely to describe G as a cause of M, again at least in part because the G ® M relationship is highly unstable under changes in environmental conditions—place the learner in a different pedagogical environment and she would not learn M. Expression of G is instead an enabling condition for acquiring M, with a better candidate for the causen of M being, for example, the learning regime involved in acquiring M. I believe a similar analysis applies to many other examples discussed above. The construction B of the bridge is causally relevant to R in Thomson’s example, but the B ® R relationship is naturally regarded as highly unstable. Change any one of a large number of background conditions, including X’s desire to commit robbery, the availability of a victim and so on, and the B ® R relationship would be disrupted. This influences us in the direction of thinking of B as a “condition for” rather a “causen” of R. A second set of considerations that is relevant to the cause/condition distinction has to do with specificity. Specificity has several different
252
Philosophy of Science Matters
dimensions but for reasons of space I focus on just one, which is related to Lewis’s (2001) notion of influence. Suppose that C is causally relevant to E. The relation of C to E will be causally specific (or C will be a specific cause of E) to the extent that C and E are variables that can be in any one of a number of different possible states or can take a number of different possible values (c1 . . . cn), (e1 . . . em) and there is a pattern of systematic dependence between these two sets of values. More specifically, there should be a mapping F from C to E such that for many different states of C each such state has a unique image under F in E (that is, F is a function or close to it); not too many different states of C are mapped onto the same state of E (the ideal being that F is 1–1); and most states of E are the image under F of some state of C (ideally F should be onto). F should describe patterns of counterfactual dependence between states of C and states of E, where the dependence in question is of the sort associated with non-backtracking or interventionist counterfactuals. To the extent that this condition is satisfied, then which particular state of C is realized provides for the possibility of a kind of fine-grained control over (or modulation of) which particular state of E is realized. The contrasting notion of a non-specific cause is that of a causally relevant factor, the operation of which is more nearly switch-like: in the limiting case, there are just two possible states of C, c1 and c2, and two possible states of E, e1 and e2. Changing from, for example, c1 to c2 changes which state of E is realized but no further; more fine-grained modulation of E by altering the state of C is possible. As an illustration, consider an ordinary copying machine. Any one of an indefinitely large number of different texts T = t1 … tk may be introduced into the machine and, as long as the machine is on, in each case a correspondingly different copy R is produced. The text inputted is a highly specific cause of the text that is outputted—by varying T, one may achieve a very fine-grained, modulated level of control or influence over R. The copying machine also has an on-off button B, capable of assuming just two states, also causally relevant to the output of the machine. For the machine to produce copies at all, B must be in the “on” rather than the “off” position, but it is not possible, by introducing more fine- grained variations in the value of B, to further modulate what the output of the machine is. B is a relatively non-specific cause of whatever output the machine produces. My suggestion is that, other things being equal, we are more likely to regard causally relevant factors that are relatively specific as “causes” in the narrow sense (causen) that contrasts with “condition” and more likely to regard causally relevant but non-specific causes as conditions. In the copying machine, both the text being copied and B are causally relevant
Causes, Conditions, and the Pragmatics of Causal Explanation
253
to the output, but, if we are engaged in causal selection, we are more likely to regard whether the machine is on as mere condition for the output and the text being copied as “the” causen of the output. Now consider a more scientifically interesting example. Molecular biologists sometimes argue that the state of the cellular DNA plays a different kind of role in determining which proteins are synthesized than other factors that are also causally relevant to protein synthesis, such as the presence of RNA polymerase (an enzyme involved in RNA synthesis) and other elements of the “cellular machinery” involved in protein synthesis, even though all are causally relevant to protein synthesis. Davidson (2001), for example, tells us that it makes sense to focus on DNA sequence in explaining protein synthesis and the generation of morphological diversity, because DNA sequence is a “more specific” cause of these outcomes than the other machinery involved in producing it. This fits with the idea that considerations of specificity play a role in causal selection. Variations in DNA sequence are causally specific in the sense described above while the presence of RNA polymerase is non-specific; if the latter is absent or diminished, protein synthesis will slow down or stop, but variations in the level of RNA polymerase do not modulate which proteins are synthesized in a fine-grained way. Let us take stock. I have drawn attention to two kinds of considerations— stability and specificity—that I claim are relevant to causal selection and to the cause/condition distinction. These considerations have an objective structure or core that I have tried to characterize. However, I do not claim that these considerations can be specified in a completely abstract, noncontextual way that is entirely independent of the interests and background knowledge of inquirers. Instead, contextual and interest-relative considerations influence the ways in which stability and specificity are understood or fleshed out or applied in particular situations, but they do so in a way that is mediated or structured by the content of these ideas. The stability of a relationship has to do with whether it holds under some range of background circumstances, but which such circumstances are regarded as most important for the purposes of assessing stability may depend on subject-matter-specific considerations, or the interests of inquirers. Similarly, judgments of specificity will be influenced in part by how fine- or coarse-grained the predicates are that we use to describe nature, and this will in part reflect our cognitive and practical interests. The presence of a short circuit is likely to strike us as a more specific cause of a fire than the presence of oxygen, in part because we find it natural to think of possible variations in the place or time at which the short circuit occurs, and hence to think in terms of the time and place of the fire as subject to modulation by means of these variations. By contrast, special circumstances
254
Philosophy of Science Matters
aside, the presence of oxygen is likely to be conceptualized in such a way that the only two possibilities are that it is either present everywhere in the situation of interest or not, which encourages us to think of it as non-specific. But while context and interests influence the ways in which stability and specificity are interpreted, and hence influence processes of causal selection, this influence does not, so to speak, operate in a completely indiscriminate and unstructured way, but rather through more structured considerations of a sort that I have attempted to describe. And while context and interests influence judgments of stability and specificity, so too do “objective” facts about the structure and behavior of the systems we are trying to describe—it is in part facts about what the world is like that make it the case that, for example, the relationship between Lewis’s letter of recommendation and the births of X’s children is relatively unstable. So far I have said nothing about the relationship between causal selection and another feature of causal claims—that they exhibit contrastive focus—that has been a topic of a considerable philosophical discussion. By contrastive focus, I mean that causal claims often (perhaps always) can be understood as exhibiting a “rather than” structure in both the effect and cause position: C caused E is perspicuously rendered as something like “C’s being in state c1 rather than state c2 caused E’s being in state e1 rather than state e2.” Often selection of a causally relevant factor as a causen rather than a condition can be naturally expressed in terms of contrastive focus—indeed thinking of a causally relevant factor as a causen rather than a condition is often tantamount to thinking of an effect in terms of one contrastive focus rather than another. If we think of the occurrence of a short circuit as a causen, and the presence of oxygen as a condition, then it will be natural to think of the effect as something like the occurrence of a fire at time t and place p, rather than the occurrence of a fire at some other time and place. The cause is then understood as the occurrence of a short circuit at some particular time and place, appropriately related to the fire, rather than the occurrence of a short circuit (or perhaps some other fire-causing agent) at some other time and place. Since oxygen is assumed to be ubiquitous, its presence does not cause (cannot be used to causally explain) the fire at one place or time rather than another. Or to put the matter the other way around, to think of the effect as the occurrence of a fire at t and p, rather than some other time and place is to think of the occurrence of the short circuit as a cause and the presence of the oxygen as a mere condition. Similarly in Thomson’s example, if, as would be natural, one thinks of the effect as X’s committing robbery rather than engaging in some other activity Z, such as sight-seeing on the island, then the presence of the bridge is not a causen of this effect since it is a precondition for both X and Z.
Causes, Conditions, and the Pragmatics of Causal Explanation
255
Does this mean that there is nothing more to the problem of causal selection than the choice of one contrastive focus rather another in causal inquiry? No. For one thing, problems of causal selection can remain even if the contrastive focus of the effect is fixed, as happens when one must select among causes that are more or less proximal to the effect or at different “levels” of analysis. But in addition, if one thinks of selection and contrastive focus as connected in the way described, then one can equally think of factors affecting selection like stability and specificity as also affecting choices about contrastive focus—stability and specificity lead us to conceive of the explanatory problem we face in terms of one contrastive focus rather than another. To the extent that we value identifying causes that are specific, this leads us to think of the effect we are trying to explain as, for example, the synthesis of one particular protein rather than another, and not as the synthesis of some protein or other (rather than none at all.) Similarly, to the extent that we value stability, it will be natural to think of the effect of interest as X’s robbing rather than X’s engaging in some other island-involving activity. In other words, although contrastive focus gives us a (partial) framework for representing problems of causal selection, considerations like stability and specificity remain in the picture, as factors influencing the choice of contrastive focus.
3. PRAGMATICS I now turn to some connections with Achinstein’s views on the pragmatics of explanation. According to Achinstein (1984), an explanation-sentence is “strongly pragmatic” if (i) it contains terms that refer to an explainer or audience and (ii) the truth-value of the explanation-sentence can vary with the person giving or receiving the explanation. Peter discusses the following notion of a “good explanation” that is strongly pragmatic: “E will be a good explanation for an explainer to give in explaining q to an audience if E is capable of rendering q understandable in an appropriate way to that audience by producing the knowledge of the answer to Q that it supplies that it is correct . . .” (Achinstein 1984, 284). In a striking discussion, Achinstein goes on to suggest there will be counterexamples to both the necessity and sufficiency of any purported universal condition for explanatory goodness (including those advanced in familiar philosophical theories of explanation) that is specified in a non-contextual, non-audience-relative way. For any such condition, there will be audiences for whom some explanations satisfying the condition do not produce the appropriate sort of understanding and explanations not satisfying the condition that do produce understanding.
256
Philosophy of Science Matters
I don’t want to directly challenge this claim, but rather to suggest that, even if it is correct, traditional non-pragmatic theories of explanation may remain valuable and illuminating. (I don’t think that Peter would disagree.) First, let us note that the notion of a “strongly pragmatic explanation” is consistent with a range of possibilities. One possibility is that (as an empirical matter) some common set of features of many explanations is judged as contributing to their explanatory goodness by many different audiences, even if not by all possible audiences. In other words, the assessment of these features as good may be relatively (even if not universally) robust across variations in audience. And the good-making features in question may admit of some interestingly general characterization in terms of shared structural features (as illustrated, perhaps, by stability and specificity.) Another contrasting possibility is that the assessment of some class of explanations is highly non-robust across variations in audience, changing in capricious and idiosyncratic ways depending on audience characteristics. Relatedly, there are no general features or patterns in these explanations that account in a non-trivial way for this variation—all that can be said is that some audiences find the explanations satisfying or appropriate and others do not. This second possibility might be characterized as a kind of “radical” (as opposed to merely “strong”) pragmatism about explanation. Something like this may be what Mill and Lewis have in mind in connection with causal selection. Their claim is not just that for any principle that might be proposed as governing causal selection, there will be some cases in which some audience does not find the selection of the cause recommended by the principle appropriate or satisfying, and some cases and audiences with the opposite profile. They endorse the much more radical view that there is lots and lots of audience-relative variation; causal selection varies widely and arbitrarily across different audiences, depending on the vagaries of their interests. Relatedly, they also hold there are no objective features out there in the causal structure of the world that play a role in guiding causal selection. My discussion above is intended to illustrate the possibility that causal selection (and causal explanation more generally) can have a “strongly pragmatic” aspect to it, in Peter’s sense, without being idiosyncratic to particular audiences in the way suggested by radical pragmatism. A second, related theme is that, supposing that we agree that context, interests, and background knowledge affect assessments of explanations, this can happen in different ways. One possibility—again perhaps illustrated by the Lewis/Mill view of causal selection—is that their influence is unstructured and conforms to no general patterns. Another possibility is that their influence is more structured, patterned and mediated: people’s interests do guide their causal selections but there are general patterns of a
Causes, Conditions, and the Pragmatics of Causal Explanation
257
non-idiosyncratic sort that in turn influence what they find interesting: for example, other things being equal, people are more interested in stable causal relationships, find them more explanatorily satisfying, are more likely to cite them in causal judgments, and so on. It is this possibility that is consistent with traditional accounts of explanation remaining valuable and interesting. It would be worthwhile to try to distinguish between these possibilities empirically. One could see whether for different groups of subjects, their causal selections, explanatory judgments, and so on are correlated with or independent of the presence of features like stability and specificity in presented causal scenarios.2 If the account sketched above is correct, there will be such correlations even if they fall well short of 1.0 (as will be the case if there are no strictly necessary or sufficient conditions for selection as cause rather than a condition). The Mill/Lewis view predicts no such correlations, at least when subject’s interests are manipulated independently of these structural features. REFERENCES Achinstein, P. 1984. The Pragmatic Character of Explanation. In Proceedings of the Biennial Meeting of the Philosophy of Science Association, ed. P. D. Asquith and P. Kitcher. Chicago: University of Chicago Press. Davidson, E. H. 2001. Genomic Regulatory Systems: Development and Evolution. San Diego: Academic Press. Lewis, D. 2000. Causation as Influence. Journal of Philosophy 97 (4): 182–97. ——— . 1986. Philosophical Papers. Oxford: Oxford University Press. Mill, J. S. 1846. A System of Logic. New York: Harper & Brothers. Oyama, S. 2000. Causal Contributions and Causal Democracy in Developmental Systems Theory. Philosophy of Science 67 (Proceedings): S332-S347. Thomson, J. J. 2003. Causation: Omissions. Philosophy and Phenomenological Research 66 (1): 81–103. Woodward, J. 2010. Causation in Biology: Stability, Specificity, and the Choice of Levels of Explanation. Biology and Philosophy 25 (3): 287–318. ——— . 2003. Making Things Happen: A Theory of Causal Explanation. New York: Oxford University Press.
NOTES 1. See Woodward 2010 for such a defense and for more details on stability and specificity. 2. Forthcoming experimental work by Tania Lambrozo explores these questions.
20 Achinstein’s Replies
I feel most honored by the contributors for their essays on my views in the philosophy of science. Below I respond to them individually. I want to express deep appreciation to Greg Morgan not just for getting the idea for such a volume, but for his wisdom and patience in organizing and editing it. He informed me that according to the agreement with Oxford University Press (and its classy editor Peter Ohlin), the volume was not to be the usual sort of festschrift in which contributors submit a paper on some topic or other that I might enjoy reading, but a paper about my views that I might not enjoy reading, to which I was to reply. Greg has accomplished his mission: the papers confront my work directly; and in fact I did enjoy reading them even when my philosophical foibles were exposed. What they did, which I fully appreciate, is force me to think in new ways about my views, to see implications that I have not previously appreciated, and in the space I have been allotted, to offer compact formulations that will enable readers to more easily identify and understand the main differences between the positions I take and those of others, including my critics. It is my good fortune to have had outstanding graduate students during my career, ten of whom are contributors to this volume. They are Victor Di Fate, Gerald Doppelt, Steve Gimbel, Adam Goldstein, Fred Kronz, Helen Longino, Jeff Maynes, Greg Morgan, Richard Richards, and Kent Staley. Some of them explain my views, some criticize them, some extend them, some do all three—for which I am very grateful. I am also very pleased to have such formidable philosophers as Nancy Cartwright, Jordi Cat, Philip Kitcher, Larry Laudan, Deborah Mayo, John Norton, Stathis Psillos, Michael Ruse, Bas van Fraassen, and Jim Woodward critically discuss my positions. In my responses below I hope that I have interpreted the account of each of the contributors correctly and that what I say will help clarify my positions and even successfully answer some of their challenges. Finally, I offer special thanks to Linda S. Brown for very perceptive comments and suggestions that helped improve my replies.
258
Achinstein’s Replies
259
GIMBEL AND MAYNES ON MY PROVENANCE Many thanks to Steve Gimbel and Jeff Maynes for their very perceptive account of early major influences on my thinking: logical positivism, particularly the views of Rudolf Carnap and Carl G. Hempel; ordinary language philosophy, especially as practiced by John Austin and Peter Strawson; and the views about logic and pragmatism of W. V. Quine. To Carnap and Hempel I am indebted for raising the question of what it means, or should mean, to talk about evidence and explanation in science, and for providing answers that I have found very stimulating, even though I suggest quite different ones of my own. Austin’s work on speech acts, particularly illocutionary ones, has been influential in the development of my own theory of scientific explanation. Strawson got me to think in new ways about induction and its justification. Finally, in developing my own views about evidence, I found that I needed to respond to Quine’s challenging doctrine of holism. I have been fortunate in having had personal contact with, and learning from, these major philosophers. Quine, Hempel, Austin, and Strawson were teachers of mine. Carnap was not, though we exchanged ideas in letters and in print.
CARTWRIGHT ON MY “MANTRA” It is most pleasing for me to know that I have an ally, especially such an important one, in the person of Nancy Cartwright. She is concerned with “effectiveness predictions” stating that a certain policy treatment T will result in an outcome O, where randomized controlled trials (RCTs) are standardly taken to be evidence for such predictions. Using the mantra “evidential relevance = explanatory relevance” that she ascribes to me, she questions whether RCTs really do provide evidence, and she explains what is generally missing in such studies in terms of that mantra. Cartwright and I use somewhat different terminologies, but she is right in claiming that evidence (which she also calls “evidential relevance”) requires correct explanation. Even if p(h/e) or p(e/h) are high, or even if p(h/e) > p(h), this is not sufficient. On my own view, e is (veridical) evidence that h (the kind that scientists in general seek), given background information b, if and only if (1) p(there is an explanatory connection between h and e/e&b) > ½; (2) e, b, and h are all true; (3) e does not entail h; and (4) (in the strongest and most interesting type of veridical evidence) there is an explanatory connection between h and e. By an explanatory connection between h and e I mean that the fact that e is true correctly explains why h is true; or the fact that h is true correctly explains why
260
Philosophy of Science Matters
e is true; or some hypothesis correctly explains why both e and h are true. Cartwright has a more nuanced way of expressing this idea, but I think we are in general agreement here. We are also in agreement that what I call an explanatory connection, and what she calls explanatory relevance, is an objective concept, not to be relativized to anyone’s epistemic situation. She is not at all worried if the concept of correct explanation is not further definable (though, unlike her, I do attempt to provide a general definition that I claim will work). On her view, as well as mine, one can have evidence for an evidential claim, since the latter is most usually an empirical claim. And she asserts that in randomized controlled trials, if we do want to claim that the policy treatment will result in a certain outcome, then we must have evidence that this explanatory connection holds. The latter is frequently not the case with RCTs—which is why she believes that such evidential claims are given insufficient justification by their proponents. For Cartwright, the most interesting cases are ones in which what she calls a “study conclusion” is taken to be evidence for a “target conclusion.” Here a policy treatment T is given to one group (the study group) and the results are then taken to be evidence that the treatment will have similar effects in the target group. Such a claim presupposes that there is a common explanation for effects in both groups (or, at least, this is the simple way I would put her point). Yet such a presupposition may well be false, as many examples show. That is why in such cases the evidential claims are faulty. I am sure Nancy and I can find aspects of evidence in which our views don’t coincide. But I accept the mantra she has given me with pleasure, and I welcome her claims about interesting cases in the social sciences in which “explanatory relevance” is necessary for evidence.
CAT ON MAXWELL Before turning to Cat’s discussion of my claims about James Clerk Maxwell’s “method of physical speculation,” it is important to note that throughout much of his career Maxwell was concerned with an important methodological question: is it scientifically legitimate to introduce physical hypotheses dealing with “unobservable” entities and causes, and if so, how can this be done? He provided different answers at different times. In 1855, in his paper “On Faraday’s Lines of Force,” when dealing with “electrical science,” Maxwell chose to avoid such hypotheses altogether in favor of what he called “physical analogies.” He introduced an analogy between the electromagnetic field and a purely imaginary
Achinstein’s Replies
261
incompressible fluid flowing through tubes of varying section. He believed that such an analogy would provide a “simplification and reduction of the results of previous investigation to a form in which the mind can grasp them” without postulating or inferring the real existence of the analog introduced or of any “unobservable” entities or causes. In 1860, in his first paper on kinetic-molecular theory, although Maxwell introduces “unobservable” molecules, he speaks of the theory as providing a “physical analogy,” as he did in the previous case. Nevertheless, this case is very different from the electrical one. In the latter, the properties of the imaginary incompressible fluid are analogs of, but not identical with, those of the electromagnetic field (e.g., the velocity of the fluid at a point in the fluid is the analog of the electrical force at a point in the field). But in the kinetic theory case, the properties attributed to the system of molecules postulated are the same as those attributed to a gas (e.g., pressure, temperature, volume, satisfying Boyle’s law). Rather the idea here seems to be to engage in pure speculation about the internal structure of gases, by supposing that a gas is identical with (not an analog of) a system of molecules; by supposing that such a system satisfies a set of simplifying assumptions Maxwell introduces; and by showing how to derive interesting theoretical and observational consequences from these assumptions. All of this is done without making inferences that the hypotheses introduced are true, or probable, or believable. At one point, Maxwell refers to this as an “exercise in mechanics.” (For a discussion of this paper, and a contrast with the previous one, see my Particles and Waves.) In 1875, Maxwell took yet a third approach to “unobservables” in his paper “On the Dynamical Evidence of the Molecular Constitution of Bodies” when he introduces what he calls a “method of physical speculation.” (It is this method that I discuss in my paper “What to Do If You Want to Defend a Theory You Cannot Prove,” Journal of Philosophy, Jan. 2010, reprinted in my Evidence, Explanation, and Realism.) In his 1875 paper, Maxwell claims that in dealing with unobservable molecules, he will employ a method that will yield less than decisive proof but more than mere “confirming instances” supplied by the hypothetico-deductive method. Such a method, if properly employed, can furnish justified belief, at least in the central and distinctive assumptions of a theory. The method, as I reconstruct it, has four components. The first component, independent warrant, contains experimental results arrived at independently of the theory in question from other domains that provide a causal-inductive basis for supposing that the postulated unobservables exist and have the properties attributed to them. It may include other sorts of reasons as well; for example, in defense of the dynamical principles introduced that
262
Philosophy of Science Matters
are supposed to govern molecules, Maxwell cites their generality and fundamental nature. But whatever reasons are given in favor of the basic assumptions, they are not of the hypothetico-deductive form “if we make these assumptions, then we can explain and predict such and such.” The second component consists of derivations and explanations of known phenomena. The third is a “theoretical development” of the theory, which involves raising and answering a range of questions about what properties and principles in addition to those introduced in components 1 and 2 the unobservables satisfy. This task usually requires introducing new theoretical assumptions for which there may not be independent warrant, and derivations of new results that may not be testable at the time (e.g., Maxwell’s distribution law for molecular velocities). Depending on how extensive the “theoretical development” is, it can provide an important basis for defending the theory non-epistemically on grounds of completeness and precision—criteria that Maxwell valued particularly highly. The final component consists of a listing of unsolved problems. In my paper I show how, especially with respect to components 1 and 3, this method is different from hypothetico-deductivism. With this brief summary of Maxwell’s “method of physical speculation,” let me respond to three of Cat’s claims about my understanding of this method. First, he claims that I examine Maxwell’s method “in pursuit of a theory-oriented argument for realism.” Although I am indeed a scientific realist, in the paper in question I explicitly reject the idea that Maxwell’s method must be understood in the context of, or as presupposing, realism (see fn. 2 in that paper). Anti-realists can raise the same question as Maxwell is by asking how to defend a theory when you cannot prove that it “saves the phenomena.” Maxwell’s method of physical speculation can be understood as a method for answering this question as well. My paper was not written with the aim of defending realism. Second, Cat claims that whereas I understand Maxwell’s method in a “disunified” way, he understands it rather simply as (A) “an inductive method linked to empirical warrant. The method of physical speculation, I propose, is the method of generalized hypothesis.” Whether my account is “disunified” because it doesn’t cover all of Maxwell’s various pronouncements about methodological strategies, or whether it is because my reconstruction of the method involves four distinct components, Cat does not say. If it is the latter, then I invite him to say why a method with four components is disunified. (This is not like postulating four distinct forces in nature.) If it is the former (which I suspect), then I would simply repeat what I say in the first part of my reply above: throughout his career Maxwell struggled with the question of how to deal with hypotheses involving “unobservables,” and he did in fact
Achinstein’s Replies
263
employ quite different methodological strategies for doing so in the three papers I mentioned. Cat’s characterization of the method of physical speculation in (A) is sufficiently broad and vague to apply to my reconstruction of that method, since the latter involves “generalized hypotheses” as well as induction. But it also applies to the method of hypothesis, which Maxwell rejects in his 1875 paper. Third, Cat objects to some of the examples I invoke in the “independent warrant” component of the method, although his grounds for doing so are not completely clear to me. In my paper I refer to experiments of Rumford and Davy at the end of the eighteenth century, showing the impossibility of treating heat as a fluid (caloric), and those of Joule in the 1840s on heat produced by friction of bodies establishing a relationship between mechanical work and heat. I claim that Maxwell may well have had these experiments in mind in his 1875 paper when he speaks of having “experimental proof that bodies may be divided into parts so small that we cannot perceive them.” I also cite Maxwell’s explicit claim regarding the success of dynamical principles in astronomy and electrical science as his basis for applying those principles to molecules. Cat’s objection seems to be that these experiments, as well as the appeal to dynamical principles, can be understood hypothetico-deductively. No doubt a hypothetico-deductivist can respond to Maxwell by showing how he can interpret the appeal to experiments and the success of dynamical theories in accordance with his viewpoint. But in the 1875 paper Maxwell rejects hypothetico-deductivism. And his appeal to the success of dynamical theories is pretty clearly an inductive argument to their success in the molecular realm. In his brief comment, Cat makes a reference to a different work of Maxwell’s, namely, an encyclopedia article entitled “Atom,” published just one year after his paper introducing the “method of physical speculation.” Here Maxwell cites various experiments, and Cat may be claiming that Maxwell is simply arguing hypothetico-deductively that because molecular theory can explain the experimental results, it is made probable by such results. In this article there is no mention of the method of physical speculation and no explicit rejection of hypothetico-deductivism. Nevertheless, Maxwell does begin his section on “Modern Molecular Science” by offering independent warrant for some of the basic assumptions of the theory, and as in the earlier paper he proceeds by introducing the Clausius virial equation, which is derived from classical mechanics. I believe the strategy employed by Maxwell in this encyclopedia article can be shown to conform reasonably well to his method of physical speculation, as I have interpreted it, although I will not pursue that here. Whatever the case, my interest is in the latter method, and what that commits one to, whether or not Maxwell
264
Philosophy of Science Matters
earlier or later failed to employ it. I regard this method as an important one for philosophers to examine, since scientists are, I think, frequently in the position outlined by Maxwell at the beginning of his 1875 paper. They are frequently able to defend a theory both epistemically and nonepistemically without having experimental proof, but having a sufficient basis to claim that the theory is a good one and a sufficient basis to be justified in the belief that the central ideas of the theory are correct.
DI FATE AND “JUSTIFICATORY FORCE” In a relatively short space, Victor Di Fate has characterized my views about induction and evidence better than I could have. For this, and for his help intellectually and practically in many matters, I am very grateful. His more critical comments and questions are reserved for his final section 4, and I will try to respond to them in the course of what follows. My claim is that whether an inductive argument of the sort championed by Newton and Mill is valid, and whether some fact is evidence for a given hypothesis, are usually, and in the most interesting cases, empirical and objective matters. The question then arises as to how my view is different from that of John Norton, who makes a similar claim, in arguing against the possibility and usefulness of universal formal rules governing induction and evidence such as those supplied by Newton and Mill. (Norton says all inductions are “local” and make no use of, and receive no justification from, universal formal rules.) Yet, in opposition to Norton, and I think, in the spirit of Newton and Mill, I claim that such universal rules are useful and do have a “justificatory force.” Di Fate asks, reasonably: how can this be? To begin with, Mill (unlike Newton) provides a “formal” definition of induction as “the process by which we conclude that what is true of certain individuals of a class is true of the whole class, or that what is true at certain times will be true in similar circumstances at all times.” He makes it clear in his discussion that some inductions are justified, some are not, and that whether they are justified is empirically decidable by examining how the instances were selected as well as by invoking other empirical background information. The mere fact that something is an induction from all observed A’s are B’s to all A’s are B’s does not mean that its conclusion is justified on the basis of its premise. I think Newton would have agreed. Of what value, then, is Mill’s definition, and for that matter, Newton’s inductive rules 3 and 4, or his causal rules 1 and 2? For one thing Newton’s rules help guide us through a long and rather complex argument of the sort that he gives for his law of universal gravitation. “Here is my empirical argument for universal gravity,” Newton
Achinstein’s Replies
265
in effect is saying, “and notice how this step involves an inductive generalization, and that one an inference to the same cause from similar effects.” (See my Evidence, Explanation, and Realism, Ch. 4, for specific examples.) Citing his universal rules, as he does, serves as a useful guide through rather difficult terrain. This is by no means all. Newton is also claiming that the only legitimate way to proceed in science is by “deductions from the phenomena,” that is, by producing arguments that are valid inductive or causal inferences, or by deriving propositions (as we would now say) deductively from others so established. And for Mill, in cases typical in science involving inferences to propositions invoking multiple causes, we are to use what he calls the “deductive method,” the first step of which requires inductions to the various causes involved. So the rules of Newton and Mill have a normative aspect to them. Proceeding to scientific conclusions by deriving such conclusions from a priori first principles, such as Descartes did in the case of his laws of motion, or by using the hypothetico-deductive method or “inference to the best explanation” (as suggested by Whewell), are forbidden strategies for both Newton and Mill. The rules tell us, in a general and abstract way, what sorts of arguments are to be used to support scientific conclusions. Finally, then, what about my claim—one that Di Fate focuses on critically—that these rules have a “justificatory” force or role? What could that mean? My answer is not that the mere fact that the inference has a form corresponding to Mill’s definition of “induction” (or to what Newton took an induction to be) justifies the particular inference in the sense of making the conclusion believable on the basis of the premise. There are good and bad inductions with such forms. My claim is only that, for Newton and Mill, having such a form is a necessary but not a sufficient condition for being the sort of inference (or one of them) that scientists should be making. (Unlike Di Fate and Norton, I don’t regard such a necessary condition as “trivial”; indeed, it is a pretty strong claim.) Having established that a particular inference has, or can be given, a formal structure of the sort Mill and Newton demand, determining whether that inference is a good scientific one does indeed require that empirical facts be invoked.
DOPPELT’S “RELEVANCE” Gerald Doppelt seeks “relevance” in any concept of evidence, and he thinks I do too. However, he claims that my concepts of potential and veridical evidence, in contrast to his own version of “inference to the best explanation,” fail to satisfy this criterion.
266
Philosophy of Science Matters
“Relevance” is not a term I use, but perhaps Doppelt is referring to my claim (in the opening pages of The Book of Evidence) that standard theories of evidence proposed by philosophers are, and ought to be, ignored by scientists, because such theories are based on assumptions incompatible with ones scientists make when they speak of, and offer, evidence for their hypotheses. One assumption many philosophers make is that evidence is a weak notion; you don’t need very much to have evidence for a hypothesis (e.g., on the standard Bayesian view, you just need something that increases the probability of the hypothesis). A second assumption (made, e.g., by Carnap’s logical theory, hypothetico-deductivists, Glymour’s bootstrap theory, and Hempel’s satisfaction theory) is that the evidential relation is logical, not empirical: whether e if true is evidence that h, and how strong it is, are logical, or mathematical questions whose answers can be determined by a priori “calculation.” Both assumptions I strongly reject. And my claim is that no concept based on these assumptions will be of interest to (or, to use Doppelt’s term, “relevant for”) scientists. Doppelt wants to make much stronger claims about “relevance.” First, he wants to say that “relevance” requires a concept of evidence that is tied to an epistemic situation. But unlike my concept of ES-evidence, Doppelt’s “evidence” refers to a judgment made by someone in a given epistemic situation—for example, our judgment that Thomson’s experimental results provide veridical evidence for the charge hypothesis, which rests, of course, on our own epistemic situation. (My ES-evidence is relativized to an epistemic situation, but requires no judgments to be made by anyone, or indeed that anyone be in that epistemic situation.) Second, he seems to be saying that unless a theory of evidence tells us how to make such a judgment by telling us how to determine whether, in any particular case, it is true that e is veridical evidence that h, it fails the “relevance” test. It may fail his test, but not mine. I distinguish between (a) the meaning of a statement of the form “e is veridical evidence that h” from (b) how to determine the truth of such a statement in a given case. For (a) I provide a general definition of veridical evidence. For (b) I provide no general account, since veridical evidence statements are empirical; and, as with empirical statements generally, how one determines their truth varies from case to case. (One can define “bachelor” as “an adult unmarried male,” but determining that a particular person satisfies this definition requires empirical investigation, the nature of which can vary from one case to another.) My “relevance” claim pertains to (a), not (b). My aim is to define a concept of evidence that is “relevant” to scientists in the sense that it reflects what scientists actually seek when they seek evidence. (See also my replies to Kitcher and Longino.)
Achinstein’s Replies
267
Perhaps Doppelt’s thought that I have a notion of “relevance” requiring that a judgment be made about whether e is veridical evidence that h arises because of the way he characterizes my concept of veridical evidence. On his characterization, e is veridical evidence that h only if both e and h are true, and “the reasoning from e to h is correct.” The quoted phrase is his, not mine, and I think it is misleading, since it suggests that someone is in fact doing some reasoning. My claim is that veridical, as well as potential, evidence are like “sign” and “symptom.” X’s rash may be a sign of a poison ivy contact, whatever (good or bad) reasoning someone in a given epistemic situation might use in inferring one from the other, indeed, even if no one at all is reasoning. Analogously, the rash is a good reason to believe that there was a poison ivy contact whatever reasoning (good or bad) someone in a given epistemic situation might use in inferring one from the other, and even if no one makes such an inference. Veridical evidence requires that e provide a good reason to believe h, not that someone in a given epistemic situation be reasoning in some particular way. Finally, I will respond briefly to Doppelt’s complaint that I provide no general account of correct explanation, or at least not one that plays any role in my account of veridical evidence. In The Nature of Explanation as well as in The Book of Evidence, I give general conditions for an explanation to be correct. It is an objective notion that is not tied to any particular context or explainer, in contrast to what, following Austin, I call “illocutionary” ways to evaluate explanations, which are contextual and perfectly legitimate as well. (See my reply to Goldstein.) But for the concept of veridical evidence, what is needed is a non-contextual concept of “correct explanation.” Contrary to what Doppelt seems to be suggesting in his essay, this concept, I claim, is sufficient for understanding “correct explanation” underlying the idea of veridical evidence.
GOLDSTEIN ON A PRAGMATIC ACCOUNT OF EXPLANATION In various writings I have defended what is called a pragmatic account of explanation. Where q is an indirect question, and Q the direct form of the question, I take sentences of the form “E is an explanation of q” to be true if and only if Q is what I call a content-question, and E is an ordered pair whose first member is what I call a complete content-giving proposition with respect to Q and whose second member is the act-type explaining q. We need not consider the details here. The main point is that on this view an explanation is something that provides an answer to certain types of questions—an answer that can be given in an act of explaining. (In my book The Nature of Explanation all of these technical ideas are defined.)
268
Philosophy of Science Matters
Several things make this view pragmatic. One is the introduction of the idea of a type of explaining act in which a speaker explains q by uttering something u with the intention that his utterance of u render q understandable by producing the knowledge that u expresses a correct answer to Q. The other is the idea that one important way to evaluate an explanation is by considering the needs, interests, and backgrounds of the potential explainer and audience. What may be a good explanation for one explainer and audience may not be so for another. Scientific textbooks illustrate this all the time. An elementary text may explain a given phenomenon in one way, a more advanced one in a very different way. Both may be perfectly good explanations for the respective audiences, and not good for others. In his paper Adam Goldstein is concerned with the importance of pragmatic evaluations, which he illustrates nicely in the case of population genetics. There is, however, another way to evaluate an explanation, namely, non-pragmatically, which is important for my theory of evidence. Since Goldstein focuses solely on pragmatic evaluations and does not mention non-pragmatic ones, and since some critics ask how my non-pragmatic theory of (veridical, as well as potential) evidence, which invokes a concept of explanation, squares with my pragmatic views on explanation, let me say just a few words. The explanation that the soldier died because he received a severe chest wound in Iraq may be perfectly correct, even if, “pragmatically” speaking, this might not be a good (enough) explanation for his commanding officer or the medical examiner to give in his report. This evaluation as “correct” is non-pragmatic since it is independent of the intentions and interests of any particular or type of explainer or audience. It can readily be explicated using the ordered pair theory sketched above. Where Q and E satisfy the ordered pair conditions given in the previous paragraph, and u is the complete content-giving proposition with respect to Q that is the first member of the ordered pair, E is a correct explanation of q if and only if u is true. In the soldier case, Q = why did the soldier die?; u = the soldier died because he received a severe chest wound in Iraq. The idea of a non-pragmatic “correct” explanation is built into my concept of veridical evidence. Finally, let me note that most standard theories of explanation, such as ones proposed by Hempel, Salmon, and Railton, are non-pragmatic. They too are meant to provide conditions for being a “correct” explanation. However, there are numerous problems I find with these theories. I’ll mention two. First, as I have shown elsewhere, there are counterexamples to each of them, that is, explanations satisfying the conditions they impose that, nevertheless, provide incorrect answers to the explanatory question raised. Second, these models propose conditions that are much too stringent.
Achinstein’s Replies
269
For example, all of them require explicit use of scientific laws, whether universal or probabilistic. But as the “soldier’s death” example above shows, we may give a perfectly “correct” explanation without invoking any laws whatever, and without even knowing what laws, if any, could be invoked. If the response is “Yes, but the explanation you gave above, while correct in some sense, is not very good,” I would reply by saying that now you are introducing pragmatic considerations. For certain contexts (e.g., one involving a medical autopsy) such an explanation may be inadequate. For another context (e.g., an explanation given to the soldier’s parents) this may be perfectly adequate. In The Nature of Explanation I argue that the same holds for “scientific” explanations, even ones involving laws of nature. Unlike what the usual non-pragmatic theories assume, there is no set of “universal instructions” by reference to which all explanations in science must be evaluated.
KITCHER’S CHALLENGE The first chapter of The Book of Evidence is entitled “The Dean’s Challenge.” I claim that the usual philosophical accounts of evidence are, and should be, of little scientific interest because they yield concepts that are much too weak, and because in general they assume (incorrectly, I claim) that the relationship between evidence and hypothesis is always a priori. The Dean’s Challenge is to provide one or more concepts of evidence that reflect what scientists mean when they speak of, and offer, evidence for their hypotheses, and that are not incompatible with assumptions scientists make when they do so. Philip Kitcher raises the meaty question: am I really giving scientists anything of use to them? He is skeptical that I am, and, indeed, he is skeptical that he himself, or perhaps any philosopher, could. What does Kitcher want in order to meet the Dean’s Challenge? His answer: the philosopher should be able to say when a scientist is “scientifically responsible” in making a judgment about the evidence (i.e., a judgment about whether to accept or reject a hypothesis on the basis of that evidence) that “accords with the evidence.” That, he claims, is what a scientist really needs to know about evidence, and that is what I don’t supply and he is not able to either. But is that really my aim, and should it be? To begin with, I don’t deal with the question of when and whether a scientist is “responsible” in his judgments about evidence. If, as Kitcher claims, this is really what the Dean (either Kitcher’s practical dean or his more theoretical one) wants, then I have not met the challenge. Kitcher claims that the closest I come to this is with my concept “e’s being ES-evidence that h,” which requires that e be true and that anyone in the
270
Philosophy of Science Matters
epistemic situation ES be justified in believing that e is veridical evidence that h. I am not convinced that this is close to what Kitcher wants, since e can be ES-evidence that h even if there are no persons in that epistemic situation; even if there are, such persons may not be making any judgments about evidence. ES-evidence has nothing to do with such judgments, if that means judgments about the truth of claims of the form “e is evidence that h.” (See my reply to Longino.) Kitcher is right in saying that I don’t provide any general account of “justification” needed for the concept of ES-evidence. My main concerns are (i) with “veridical evidence,” since (as Kitcher agrees) this is in fact the sort of evidence scientists seek, and (ii) with “potential evidence” needed for the definition of this concept. For these concepts, definitions are provided that do not appeal to, or presuppose, any judgments of the scientist or any notion of “responsibility.” My claim is that whether e is veridical (or potential) evidence that h is an objective, empirical fact (like that of something’s being a sign or symptom of some state of affairs) that is independent of judgments of particular scientists about the evidence, about their “responsibility,” or anything else. The results of J.J. Thomson’s experiments constituted (veridical) evidence that cathode rays are electrically charged; the results of Hertz’s experiments did not constitute (veridical) evidence that cathode rays are electrically neutral. The definitions I offer of veridical (and potential) evidence are meant to clarify what such claims mean. One of Kitcher’s complaints is that my definitions appeal to probabilities, and in real scientific life, where the situation gets messy (since different parts of the evidence may point in different directions), I offer no way of assigning probabilities. But my definitions don’t require that I do. Even in messy situations of the sort Kitcher likes to cite (the sort I deal with in a section of my book on different meta-analyses), all that is necessary is that the hypothesis be more probable than not. (More strictly, what is necessary is that the probability that there is an explanatory connection between h and e, given e, be greater than the probability of there being no such connection.) Although in the book I try to characterize an objective concept of probability of the sort required, and I do provide definitions of “explanation” and “explanatory connection,” these will not enable the scientist to churn out a priori an answer to the question of whether e is evidence that h. My claim is that whether the requisite probability and explanatory conditions hold is not an a priori matter for which armchair philosophers can crank out formulas (à la Carnap), but an empirical one, for which scientists themselves do, or ought to, provide reasons. Thomson himself gave strong empirical reasons for his conclusion that his experimental results are evidence that cathode rays are charged. These empirical reasons constitute the basis for saying that there is a high probability of an explanatory
Achinstein’s Replies
271
connection between his experimental results and the presence of negatively charged particles. With the highly evacuated cathode tubes Thomson was able to obtain, the fact that the cathode rays were deflected toward the positively charged electric plate makes it probable that the reason for this deflection is that the cathode rays are negatively charged. Of course, the latter claim can itself be defended by appeal to further empirical facts, for example, that oppositely charged bodies attract. On my view, when we provide such empirical reasons in defense of an evidential claim we are doing physics, not philosophy. What the philosopher of science ought to do is figure out what the original evidential statement means. This is what I am supposing (or at least hoping) that my dean had in mind. In addition, following in the footsteps of Newton and Mill, the philosopher of science can and should identify certain basic forms of inductive and causal reasoning that scientists do and ought to follow in defending an empirical claim, including an evidential one. (See my “The War on Induction,” reprinted in my Evidence, Explanation, and Realism; also my reply to Di Fate in this volume.) But, as is the case with evidential claims, whether any particular causal or inductive inference satisfying one of these forms is justified is an empirical issue. Perhaps Kitcher and his deans want more. If what they seek are empirical ways to defend or criticize particular evidential or other scientific claims, then neither Philip nor I, as philosophers of science, will be able to supply what is wanted.
KRONZ’S HYPOTHETICAL METHOD Frederick Kronz notes that on my view, as well as on the views I attribute to Newton and Mill, “induction and hypothetico-deduction are mutually exclusive.” He argues, on the contrary, that these methodologies are compatible and complementary, and indeed that “neither Mill nor Newton . . . is really averse to the use of hypotheses in science.” I have two responses. The first is to distinguish the “use of hypotheses” from “hypothetico-deduction” (or what is also called the hypothetical or hypothetico-deductive method). The latter constitutes a form of reasoning in which the hypothesis is inferred to be true, or probable, given the empirical data, on the grounds that the hypothesis entails, explains, or predicts the data. In more sophisticated versions, certain conditions may be imposed, such as Whewellian “consilience”—requiring that the hypothesis explain and predict data not only of the kind that generated the hypothesis in the first place, but of other different kinds as well. Whatever these additional conditions are, however, they do not include the requirement that there be inductive support of a sort demanded by Newton and
272
Philosophy of Science Matters
Mill. So, Newton and Mill would say, the fact that Newton’s law of gravity explains and predicts a wide range of observed phenomena is not sufficient to infer its truth. What is necessary is the sort of causal-inductive reasoning to the law from observed phenomena involving the planets and their satellites that Newton gives in Book 3 of the Principia. Now, contrast this with “the use of hypotheses” in science. Newton defines a hypothesis as a proposition that is not “deduced from the phenomena” using causal-inductive reasoning of the kind he requires. Although Newton famously wrote that “hypotheses . . . have no place in experimental philosophy,” he does introduce them from time to time. (For example, in the Principia the proposition that “the center of the system of the world is at rest” is labeled “hypothesis 1.”) So, despite his famous claim to the contrary, Newton is not averse to introducing and using hypotheses. What I take his position to be is this: you can introduce a hypothesis, but from the fact that it helps to explain and predict phenomena you cannot infer that it is true or probable, since a conflicting hypothesis may explain and predict the same phenomena. In other words, Newton is distinguishing between the use of the hypothetico-deductive method and the use of hypotheses. The same goes for Mill, who allows hypotheses to be introduced to suggest possible explanations of phenomena. But from the fact that a given hypothesis yields such an explanation one cannot infer its truth or probability. In the first part of his paper, Kronz does not sufficiently emphasize this distinction between the use of hypotheses and the use of the hypothetico-deductive method. My second response is to Kronz’s “enhanced hypothetical method,” which is designed to show, among other things, how the inductivism of Newton and Mill is not incompatible with, but complementary to, hypothetico-deductivism. In addition to stipulating that the empirical consequences of the hypothesis obtain, the enhanced hypothetical method requires “an inductive inference to the hypothesis” and the satisfaction of an incompletely formulated disjunctive set of conditions (including “consilience,” novel prediction, etc.) Now besides the incompleteness of Kronz’s set, there is a problem that I see with this proposal. What exactly does Kronz mean by an “inductive inference” here? Is it (1) any inference from the truth of the observed consequences of a hypothesis to the truth or probability of the hypothesis? Or is it (2) any inference that “what is true of certain individuals of a class is true of the whole class”? If it is (1), then Newton and Mill will object, and deny complementarity. For example, as mentioned, one of Kronz’s disjuncts is Whewellian consilience. But Mill explicitly rejects the idea that you can infer that a hypothesis is
Achinstein’s Replies
273
true or probable on the grounds that it explains or predicts a range of observable phenomena, even if they are of different kinds. As Kronz notes, Mill defends what he calls the “deductive method,” which he sharply distinguishes from the hypothetico-deductive method. The former has three components: first inductions (in Mill’s sense, viz. (2)) to a set of causal laws; second, “ratiocination,” which involves deductions from the set of inductively inferred causal laws; third, empirical verification of phenomena newly predicted from this set. For Mill, the hypothetico-deductive method omits the first step and contains only the second and third. It is clear from his discussion that adding any of the disjunctive requirements in step 4 of Kronz’s “enhanced hypothetical method” to his inductive requirement, understood in the sense of (1), will not suffice to generate a justified conclusion. So if (1) is how “inductive inference” is to be understood, then Mill’s inductivism and hypothetico-deductivism are mutually exclusive. On the other hand, if by an “inductive inference” Kronz means what Mill does, that is, (2), then we end up with Mill’s “deductive method,” perhaps enhanced with a disjunction of further conditions, the satisfaction of any of which is supposed to strengthen the inference to the hypothesis. I will not here discuss whether the additional conditions Kronz lists are indeed “inference enhancing.” I will note that some are vague, for example, “a new organization of facts that facilitates solving a problem,” and a “suitable condition (other than those already mentioned).” If this is the way “inductive inference” is to be understood then I would conclude that Kronz’s “hypothetico-deductivism” is pretty close to Mill’s inductivism (as expressed in his “deductive method”). Indeed, it includes it; it is not really a complementary methodology.
LAUDAN’S NINETEENTH-CENTURY WAVE-THEORIST I have long admired Larry Laudan’s historical-philosophical work, despite our different takes on the nineteenth century wave-particle debate. He understands that debate to be largely influenced by a difference in methodological viewpoints. On his view, the wave theorists, being proponents of the method of hypothesis, or hypothetico-deductivism, were free to introduce unobservable waves and an ether to support such waves, so long as the theory gave correct predictions for a variety of observable optical phenomena. The particle theorists, being proponents of an inductive methodology, required that any entities or causes introduced be observable or at least have properties like those found in observable cases. (The latter, Laudan claims, is how eighteenth- and nineteenth-century followers of Newton understood his “vera causa” rule 1 in the Principia.)
274
Philosophy of Science Matters
On my reconstruction of the debate, Laudan’s methodological schism is exaggerated. As he notes, I claim that a strategy wave theorists use is typically this: (i) start with the assumption that light is either a wave phenomenon (a wave motion or pulse transmitted through some medium), or a stream of particles (subject to forces obeying Newton’s laws of motion); (ii) show how each theory explains various observed optical phenomena (e.g., rectilinear propagation, reflection, refraction, diffraction); (iii) show that the particle theory, in explaining one or more of these phenomena, introduces improbable hypotheses, while the wave theory does not; (iv) conclude that the wave theory is very probably true, while the particle theory is not. This eliminative strategy is one adopted by wave theorists such as Young, Fresnel, Herschel, and Lloyd. So far it is an open question whether this should be given an inductive or a hypothetico-deductive, or some other, methodological interpretation. Laudan denies even this much. Wave theorists, he claims, did not, and could not, employ an eliminative strategy, since the number of different specific wave theories was too great to be exhaustively considered and rejected. Perhaps Laudan has difficulties here, but wave theorists such as Young, Lloyd, and Herschel did not. Here is Young. It is allowed on all sides, that light either consists in the emission of very minute particles from luminous substances, which are actually projected and continue to move with the velocity commonly attributed to light, or in the excitation of an undulatory motion . . . in a highly light and elastic medium pervading the universe; but the judgments of philosophers of all ages have been much divided with respect to the preference of one or the other of these opinions.
Here is Herschel. Among the theories which philosophers have imagined to account for the phenomena of light, two principally have commanded attention: the one conceived by Newton . . . in which light is conceived to consist of excessively minute molecules of matter projected from luminous bodies. . . . The other hypothesis is that of Huygens . . . which supposes light to consist, like sound, of undulations or pulses, propagated through an elastic medium.
Unlike Laudan, these authors had no difficulty reducing the (most plausible) competing theories to two, and then arguing that one is superior to the other. Now for my inductive take on this. I claim that step (i), noted above, can be, and was in fact, inductively supported by wave theorists. (Such theorists did not say simply: let’s just make this assumption and see what follows.) For example, Lloyd supported it by saying that it was empirically established that light travels from one point to another with finite velocity,
Achinstein’s Replies
275
and that in nature one observes motion from one point to another occurring either by the motion of a body or by the motion of vibrations through a set of bodies or medium. (“Nature,” says Lloyd, “affords numerous examples of each of these modes of propagated movement.” If there were others that had been observed presumably Lloyd would have mentioned them.) I take Lloyd to be offering inductive support for his claim. A second part of inductive support appears in the defense of (iii). The reason for the claim of improbability in the type of force particle theorists introduce to explain diffraction, for example, is that the force postulated is unlike any observed forces. It requires that “bodies of different forms and of various refractive powers should possess an equal force of inflection” (Young), which is not observed to be the case with other known forces acting at a distance. (In Particles and Waves I show that if the probability that such a force exists is much, much greater on the assumption of the particle theory than it is without that assumption, then the probability of the particle theory is close to zero.) By contrast, the wave theory introduces no such forces or other improbable causes to explain diffraction It is step (ii), the explanatory one, that Laudan pushes (perhaps to the exclusion of the others). His 19th century wave theorist (a model hypothetico-deductivist) argues for his theory (whether some generalized wave theory or a particular version) simply by showing that it can explain and predict a variety of known phenomena. And, Laudan claims, my characterization of step (ii) downplays its importance. If the wave theorist establishes the high probability of his theory W on the basis of inductive background information b, and inductively supported claims about diffraction d, then I argue that p(W/b & d) is high. However, the wave theorist wants this probability to be high not just on this basis but on the basis of all known and successfully predicted optical phenomena. If his theory can explain and predict these other optical phenomena by deriving them from the theory, then he gets what he wants. Where O1, . . . ,On are these additional phenomena, he obtains the result that p(W/O1, . . . ,On & d &b) is at least as high as p(W/b&d), which follows probabilistically. It is not my claim, of course, that the wave theorist uses the probability calculus, only that this is a way of understanding the wave theorist’s strategy in steps (i)–(iv) so that it gives that theorist what he is seeking. In Particles and Waves I argue that Laudan’s explanatory-predictive account by itself cannot give the wave theorist high probability. Finally, there are several claims that Laudan makes about my view that are too strong. One is that I am a Bayesian. I am not. Another is that, on my view, an eliminative strategy is the only one that can be used to show high probability for any theory. I make no such claim. Indeed in other writings, for example on Newton’s argument for his law of gravity, I analyze
276
Philosophy of Science Matters
the inductive strategy in a non-eliminative way. These claims, and other differences between us, I look forward to discussing with Larry on another occasion.
LONGINO’S CONTEXTUALISM Helen Longino sees the need for what she calls a “contextualist” account of evidence. In The Book of Evidence there is a chapter entitled “Old-Age and New-Age Holism” to which she refers. One version of new-age holism that she mentions is this: where h is some “isolated” hypothesis and e is some “isolated” fact, it is never true that e is evidence that h. Rather e is evidence that h, relative to some set of assumptions b. She calls this a form of contextualism, since on this view, “the relevance of some state of affairs to an hypothesis requires background assumptions.” How and why is this a “contextualist” position? Her answer is not always clear to me. At one point she writes that (1) The truth of an evidential claim of the form “e is evidence that h” depends on “a context of background assumptions that are themselves empirical in nature, not known to be true, and that might be false.” More often her view seems to be that (2) When we assess or determine the truth of an evidential claim of this form, we make background assumptions of the sort described in (1). How is (1) a “contextualist” position? Is Longino saying that an evidential claim of the form “e is evidence that h” is incomplete as it stands, that to have a truth-value it must be relativized to some set of background assumptions b, and that which set this is depends on the context of utterance? But then the “completed” relativized claim “e is evidence that h, given b,” which is the sort of claim I am concerned with, is not “contextual.” It is either true or false independent of any context, indeed, independent of whether anyone knows the truth of b or of the evidential claim itself. Where, then, is the “contextualism” in (1)? To be sure, whether someone is in a position to utter an evidential claim (whether relativized or not) depends on contextual conditions satisfied by the utterer. But that is so for any claim whatever, whether evidential or otherwise. Accordingly, if this all that evidential contextualism amounts to, then the doctrine is, I think, minimal, and not one I would dispute. What about position (2)? There are usually different ways we can assess or determine the truth of a claim that e is evidence that h (different pieces
Achinstein’s Replies
277
of background information we can appeal to). Which we choose depends not only on the evidential claim itself but also on our own knowledge and interests as well as on those of a potential audience, which, of course, can vary from one context to another. But this is true of any claim we make, whether evidential or not. So if (2) is an expression of, or leads to, evidential contextualism, then, again, the doctrine is, I think, minimal, and not one I would dispute. One of Longino’s main concerns (suggested in the latter part of the quote in (1)) may be this. Evidential claims are useful to us only if we can determine whether they are true or false, which involves determining whether some background assumption is true or false. But, given our epistemic situation, we are often not in a position to do the latter. My response is this. Surely sometimes we are in a position to know that some particular background assumption being made is true (I am not a skeptic). But even if we are not in a position with respect to a particular b, we may be in a position to determine whether there is evidence that b is true. Longino is right in worrying about how, given the epistemic situation we happen to be in, we are to determine the truth of an evidential claim. (My concern in the book was not with this question, but with the question of what it means to make evidential claims of the sorts I have in mind.) But I am not as pessimistic as she is about the question of assessing the truth of evidential claims of the form “e is veridical evidence that h, given b.” Contrary to what Longino seems to suggest, such assessments are not restricted to cases involving lotteries where probabilities are usually easy to determine. One does not have to be able to assign a probability to an hypothesis h to determine whether, given e and b, it is more probable than not that there is an explanatory connection between h and e (a crucial necessary condition for veridical evidence, on my view). Nobel Prize committees depend on evidential assessments all the time. Perrin was given the Nobel Prize for his experiments on Brownian motion. The Nobel Committee, or its referees, were able to assess the claim that the results of these experiments provided (veridical) evidence that chemical substances are composed of molecules the number N of which in a gram molecular weight of any substance is approximately 6 x 1023. My theory of evidence explains what such an evidential claim means. It does not say how to go about assessing its truth. The evidential claim itself is empirical, and so, needs to be defended, if it does, empirically – presumably by scientists themselves. In The Book of Evidence I offer no general account of how empirical claims, whether evidential or otherwise, are to be defended. My own view is indeed contextual in a sense indicated earlier: what, if anything, needs
278
Philosophy of Science Matters
to be defended and how depends on the knowledge and questions of the defender and the audience, which vary from one context to another. For this reason I do not think any general philosophical account of how evidence claims are to be defended can be provided. If this is Helen’s position on contextualism, or part thereof, we are in agreement.
MAYO ON SEVERE TESTING As Deborah Mayo notes, she and I are in agreement about various important issues regarding evidence. For example, we both hold that evidence (at least in the most central sense) is an objective, not a subjective, concept, and that a claim that e is evidence that h (in typical cases) is empirical, not a priori. We both agree that, contrary to the standard Bayesian idea, increase in probability is neither necessary nor sufficient for evidence, and that high probability is not sufficient. There is, however, one significant difference between us. Mayo believes that e is evidence that h only if e results from “highly probing” hypothesis h (in a sense that she defines). I believe that e is (potential or veridical) evidence that h only if e provides a good reason for believing that hypothesis h is true. On my view, this occurs if e is true, e does not entail h, and (most importantly) given e, the probability is high that there is an explanatory connection between h and e. (The latter entails that h is highly probable, given e, which I take to be a necessary condition for evidence.) For her, “passing a severe test” is both necessary and sufficient for evidence; for me, it is neither. Mayo and I have been debating this for some time now, as the references in her paper indicate. Here I will simply focus on one example that will bring out this difference between our views, and show why I cannot accept her “severe testing” as a basis for evidence. (I have used this example in “Mill’s Sins or Mayo’s Errors,” in Deborah Mayo and Aris Spanos, eds., Error and Inference. For my purposes in this reply, it is an easier one to construct, and, I hope, a better one in terms of which to appreciate our contrasting views than her “college ready” and my “Stand and Deliver” examples cited in her present reply.) Suppose there is a very rare disease S, which is such that only one person in ten million in the general population has S. In such a case, I will assume, we can assign a prior probability for having S, (1) P(S) = .0000001. (In what follows probability denoted by a capital P represents a relative frequency). Now suppose there is only one test T for this disease, which
Achinstein’s Replies
279
can yield one of two results: the test result turns either red or blue. Of those with S, 80% who take test T test red, so that (2) P(red result with test T/disease S) = .8, and of those who don’t have S, 2 out of 100,000 test red, so that (3) P(red result with test T/-S) = .00002. Now we consider a particular individual, Irving, for whom we can write the following probabilities, based on those above. (4) P(Irving’s getting a red result with test T/Irving has disease S) = .8 (5) P(Irving’s getting a red result with test T/Irving does not have disease S) = .00002. Because the probability in (4) is much greater than that in (5), Irving’s test result of red “fits” the hypothesis that Irving has S in a sense of “fits” that seems to conform to Mayo’s idea. Furthermore, because the probability in (5) is so low, Irving’s getting the red result with test T should count for Mayo as passing a severe test for the hypothesis that he has disease S. On her view (as expressed, e.g., in her “Evidence as Passing Severe Tests,” in Achinstein, ed. Scientific Evidence, 99), if we have a hypothesis h, a test T for that hypothesis, and some data D produced as a result of employing the test, then for h to pass a severe test T yielding D, it is required that the data “fit” the hypothesis—in a fairly weak sense of “fit” such as the one suggested above; and it is required that the probability be very low that T would yield results that fit h as well as data D if h were false. In her contribution for the present volume, Mayo avoids talk of probabilities in characterizing a severe test. Such probabilities presuppose priors, which, being a frequentist about probability, she doesn’t much like, since frequency priors are often not “ascertainable.” However, in the present example, we have such a prior, since I am stipulating a base rate for disease S. To continue the example, using Bayes’ theorem, from the probabilities (1)–(3) above we compute that P(S/red result with test T) = .004; that is, four out of 1000 people who get the red result from test T (less than half of 1%) have disease S. My own take on this is that T is a very poor test for disease S, despite the fact that it seems to satisfy Mayo’s criteria for being a good test. (If it doesn’t, I invite her to say why.) To see why I regard this as a poor test, let us use these frequency probabilities as a basis for what I call objective epistemic ones (represented with a small p). Doing so, we obtain (6) p(Irving has disease S) = .0000001 (7) p(Irving has disease S/Irving got a red result with test T) = .004.
280
Philosophy of Science Matters
That is, (6), the prior epistemic probability (the degree of reasonableness of believing) that Irving has the disease, is miniscule; and (7), the posterior epistemic probability that he has the disease, given that he got the red result with the test, is still very small. Despite the fact that Irving has passed a “severe test” for disease S in what I take to be Mayo’s sense, Irving’s test result gives very little reason to believe that he has disease S. As (7) indicates, the degree of reasonableness of believing that he has S, even given that he tested red, is very low. On my view, passing a severe test for a hypothesis h, and hence having evidence for h, should mean that there is a good reason to believe that h is true. On her view, as characterized above, passing a severe test T for hypothesis h with data D, and hence having evidence for the hypothesis, entails this: given that test T continues to be repeated, the relative frequency of outcomes fitting hypothesis h as well as data D will at some point in the testing sequence be and remain very low, under the assumption that h is false (see, e.g., her major work Error and the Growth of Experimental Knowledge, 121–4). Sometimes she puts this by saying that passing a severe test means that the hypothesis is “reliable,” in the sense that “what h says about certain experimental results will often be close to the results actually produced” (10). But, as my example is intended to show, this does not entail that passing a severe test (in Mayo’s sense) provides a good reason to believe that the hypothesis is true. That, as I see it, is a significant difference between her concept of evidence and mine. I will now comment on three issues Mayo raises in her paper. First, she asks whether I am claiming that in the sort of case above when I assert that Irving’s testing red is not evidence that he has disease S, I am committed to saying that the fact that he tested red is evidence that he does not have disease S. After all, the probability that he does not have S, given that he tested red is very high (.996). To her that would be paradoxical, since testing red is passing a “severe test” for having disease S. The answer is that I am not committed to saying this, because in this case my explanation condition for evidence is violated. Given that Irving tested red, the probability is not high that there is an explanatory connection between his not having the disease S and his testing red. (It is not probable that the reason he does not have the disease is that he tested red, or conversely, or that some hypothesis explains both why he does not have the disease and why he tested red.) Second, my example avoids what Mayo calls the “fallacy of probabilistic instantiation,” in which a relative frequency probability for a single case is inferred from a relative frequency for a type of property or situation. It avoids this because the probability “instantiated” is not a relative frequency, but an objective epistemic probability, which is applicable to
Achinstein’s Replies
281
hypotheses about single cases (e.g., the probability that Irving has disease S). My claim is that, in the case I am imagining, the objective epistemic probability statement (7) above can be defended empirically by appeal to the relative frequency probability statement (5) above. I see no fallacy in such a defense. Third, Mayo raises a general “ascertainability” question for anyone appealing to numerical probabilities in a definition of evidence: how are such probabilities to be determined (whether these are conditional probability statements of the form p(h/e) = r, or priors of the form p(h) = r)? This is a large and controversial topic, but let me respond briefly by saying that I distinguish the question of “ascertainability” from that of the meaning of probability, and evidence, claims. Moreover, if we do consider the question of how to ascertain the truth of evidence claims, my account of evidence does not require ascertaining numerical probability values, whether or not these are conditional probabilities or priors. To ascertain the truth of “e is (potential or veridical) evidence that h,” what is required is to determine whether given e, it is more probable than not that there is an explanatory connection between h and e, which does not require determining how probable it is. To give a simple example, let e = I own more than 90% of the tickets in a fair lottery, one ticket to be drawn at random. Let h = I will win. To determine whether it is true that e is (potential or veridical) evidence that h, we don’t need to compute p(h/e), p(h), or p(there is an explanatory connection between h and e/e). Finally, in this connection, evidential claims of the form “e is evidence that h,” being empirical (in the kinds of cases Mayo and I are concerned with), there are usually many different empirical ways to “ascertain” their truth, or to provide evidence for such claims. I do not subscribe to the idea that the philosopher of science can provide a general account of such ways. (See my replies to Doppelt, Kitcher, and Longino.) Although there are other issues raised in Deborah’s present contribution to which I would like to respond, induction tells me that we will have an opportunity to continue our debate on another occasion, even if that hypothesis has not been severely tested.
MORGAN ON COHERENCE William Whewell claimed that the coherence of a theory is a sign of its truth (indeed a pretty darn good sign). But he never argued for such a claim, much less proved it. Whewell’s concept of coherence is rather vague. In Particles and Waves, in attempting to assess his claim, I offer a precise probabilistic definition of coherence that Gregory Morgan cites in
282
Philosophy of Science Matters
his thought-provoking paper. I argue that at least with this notion of coherence, Whewell cannot establish his claim that coherence is a sign of truth—either in the sense that it makes a theory highly probable or in the sense that it increases its probability. Morgan and I are in general agreement about this conclusion, although unlike me, he does not define coherence. (He regards my definition as a “sanitized” version of Whewellian coherence, since it does not capture certain “subjective” features of that concept.) Morgan considers a version of Whewell’s claim that coherence is a sign of truth to be committed to saying that for any theory T (1) If T is coherent, then there is a reason to believe T. One way Morgan expresses this idea probabilistically is this: (2) p(T/T is coherent) > .5 He calls such coherence a sign1 of truth. Another way he expresses the idea in (1) probabilistically is this: (3) p(T/T is coherent) > p(T), in which case he calls coherence a sign2 of truth. Morgan is assuming that in (1) “a reason to believe T” either makes T more probable than not, or increases its probability. But as I see it, (1) itself is a conditional statement, with the idea of probability, if any, expressed or presupposed in the consequent of the conditional; it does not seem to be expressing or presupposing a conditional probability. Accordingly, if we are to construe it probabilistically, I would understand (1) to be saying not (2) or (3) but (4) If T is coherent, then p(T) > .5, or in the case of the weaker probabilistic version, something like (5) If T is coherent and T’ is not, or if T is more coherent than T’, then, other things being equal, p(T) > p(T’). Even though Morgan does not do so, I propose to assess the weaker claim (5), which I will do in a way that does not make use of my definition of coherence. Such an assessment requires saying something about coherence, even if we do not supply a full-scale definition. Among other things, Whewellian coherence pertains to how well a given theory explains something, which in turn depends on how it is formulated (a point Morgan recognizes in his paper). Indeed, two logically equivalent formulations may not be equally explanatory. For example, let T be a conjunction of Newton’s laws of motion and his law of gravity. Let T’ be a conjunction of T together with
Achinstein’s Replies
283
the tautology “either Socrates was bald or he wasn’t.” Now one thing we demand of a good explanation (even Hempel demands it in his D-N model) is that all parts of the explanation be relevant, that is, be used to generate the explanation. (Whewell himself writes that in a system in which coherence is lacking “the new suppositions are something altogether additional;—not suggested by the original scheme. . . .”) The tautology in T’ is irrelevant for an explanation of why the planetary orbits are elliptical, as well as for explanations of numerous other phenomena explained by Newton. T, let us say, is coherent, T’ is not, or at least T is more coherent than T’ with respect to these phenomena. Since T and T’ are otherwise the same, on the basis of (5) we can assume that p(T) > p(T’) or at least that these probabilities are not the same. But this violates the rules of probability, since T and T’ are logically equivalent. The argument above does not impugn (4). My own preference would be to try to give a complete formal definition of coherence to see if Whewell can or cannot justify the probabilistic claims (4) and (5). But on this occasion I will settle for the brief informal remarks above to show a difficulty Whewell faces in making coherence a sign of truth via probability, even in the weaker sense of “sign.” (A more formal approach with more elaborate arguments appears in my Particles and Waves, which includes a definition of coherence, and in Morgan’s paper, which does not.) Having said this my claim would be that coherence (as well as other virtues, such as simplicity, generality, and precision) are important considerations in defending a theory non-epistemically. However they are defined, and even if they are not but are recognizable without definition, they can be appealed to in defending a given theory without affecting its probability or believability. In my paper, “What to Do If You Want to Defend a Theory You Cannot Prove” (reprinted in my Evidence, Explanation, and Realism), I show how James Clerk Maxwell appeals to the precision and the completeness of his development of the kinetic-molecular theory in its defense, in addition to less than decisive experimental evidence. Maxwell does not claim or assume that the non-epistemic virtues are signs of truth, but only that they are legitimate criteria of goodness in a theory. Criteria of both epistemic and non-epistemic sorts can be invoked in evaluating a theory. However, the non-epistemic ones may carry at least as much weight as (or even more weight than) the epistemic ones in such an evaluation at a given moment of time if the evidence for the truth of the theory at that time is less than decisive. Trial lawyers say: if the facts are not as strong as you would like in favor of your client, pound the law; if the law is not as strong as you would like, pound the facts; if neither is, pound the table. Maxwell emphasized both epistemic and non-epistemic virtues in his presentation of kinetic theory, without pounding anything, let alone the table.
284
Philosophy of Science Matters
NORTON’S “INTRACTABLE PROBLEM OF INDUCTION” In a series of very striking papers, John Norton has defended the thesis that inductive inferences are justified, when they are, not by any universal formal principles of induction (such as those Mill and Newton seem to be supplying), but only by material facts that vary from one induction to another. Universal principles of induction are worthless, and even dangerous for Norton, since they suggest the possibility of justifying particular inductions by reference to formal schemas that cannot possibly do any justificatory work. In the present essay, Norton presents a new sort of case, that, on his view, is inductively “intractable.” It involves observationally indistinguishable spacetimes in which, for example, it is natural to infer inductively to a fully extended Minkowski spacetime rather than to an extendable halfMinkowski spacetime. But, unlike indeterministic cases in quantum mechanics, no material fact seems to ground this inference, nor do empirical laws, or any appeals to universal principles of induction, metaphysical principles of simplicity, or anything else. I am tempted to say that whereas Hume found that although we make inductive inferences none is justified, Norton restricts his “inductive skepticism” to a particular case (or perhaps a few) in theoretical physics. I find relief in that. Let us suppose that Norton is right about his case, and that we find the induction to a fully extended Minkowski spacetime natural and compelling. Let us also suppose that no material warrant exists (or we can’t find any) for such an induction, and that no general principles of induction can justify such an induction. What follows? Norton’s answer: we have an intractable problem, “the ground crumbles around our feet.” But there is another possibility. In such a case, even though we find it natural and compelling to make the inductive inference in question, since there is, or we can find, no inductive warrant, what this shows is that we should not make the induction despite the temptation to do so. Why exactly is this an “intractable problem” for an inductivist, even of the “material” sort that Norton identifies himself as being? After all, Norton, unlike Hume, is not saying that no induction is rationally justified. Why should it be a problem if there is a case in which although we find the induction natural, we also find that it has no justification in fact. That is just an interesting case! It becomes “intractable” only if we further assume that for any induction we find natural and compelling there will always be a justification. If that is what Norton is assuming, he needs more argument. For more general remarks about the differences between Norton’s view of induction and my own, and for a defense of Newton’s methodological rules and Mill’s methods against Norton’s sort of criticism, see my essay
Achinstein’s Replies
285
“The War on Induction,” in my Evidence, Explanation, and Realism, and my reply to Di Fate in the present volume. I find a place for such rules in the justificatory process; Norton does not.
PSILLOS ON PERRIN I welcome Psillos’s stimulating essay, both for its historical background on Perrin and for its attempt to provide a probabilistic account of Perrin’s reasoning that is better than mine. Let me start with Psillos’s claim about my account. As he notes, I formulate Perrin’s major experimental result concerning Brownian motion as follows: C = the calculation of N (Avogadro’s number) by means of Perrin’s experiments using equation (9) is 6x1023 and this number remains constant when values for various empirically determined parameters in equation (9) are varied. Equation (9) relates Avogadro’s number (the number of molecules in a substance whose weight in grams equals its molecular weight) to various other quantities measurable in the experiments Perrin performed. (Strictly speaking, equation (9) represents a number for visible Brownian particles, where Perrin is assuming that any quantity of these particles equal to their molecular weight will contain the same number N of particles—a number that will be the same as Avogadro’s number for molecules.) On my reconstruction, experimental result C is supposed to provide evidence for the following theoretical claim. T = Chemical substances are composed of molecules, the number N of which in a gram molecular weight of any substance is (approximately) 6x1023. Now, according to my definition of potential evidence, C is (potential) evidence that T, given the background information b Perrin cites, if and only if (i) C and b are true; (ii) C does not entail T; (iii) p(E(T,C)/C&b) > 1/2. (The latter is equivalent to the requirement that the product p(T/ C&b)xp(E(T,C)/T&C&b) > ½. (E(T,C) means there is an explanatory connection between T and C—see my reply to Cartwright for a definition of this.) What I argue in my writings is that C is indeed potential evidence that T, given the background information employed by Perrin. Psillos says that on my view all that Perrin achieved is to show that the atomic hypothesis is more likely than not (which would follow from the satisfaction of condition (iii) above). I don’t accept that characterization. On my view, although satisfaction of (iii) is necessary for evidence, it doesn’t follow that Perrin simply established that the probability cited in (iii) is greater
286
Philosophy of Science Matters
than ½. As I argue, what probability Perrin in fact established for an explanatory connection between C and T is an empirical matter. One needs to look at the specific empirical arguments he employed to determine whether this probability is simply greater than ½ or much greater (which Perrin himself claimed). Accordingly, I believe that my position on evidence can readily deal with Psillos’s concern. Let me comment on one important feature in Psillos’s own probabilistic reconstruction of Perrin’s argument. Psillos claims that Perrin argued that the probability of getting the experimental results he did (my C above), given the atomic hypothesis (I will consider simply T above), is very high, while the probability of getting these results, given the falsity of T, is low, that is, (i) p(C/T) is high; (ii) p(C/–T) is low. Using Bayes’s theorem, together with some other plausible assumptions, we may conclude that the probability of T, given Perrin’s experimental results, that is, p(T/C), is high. A crucial passage Psillos cites from Perrin states that before his experimental results no one would have expected that the Brownian particles would not have fallen to the bottom of the vessel (in which case N would be infinite), or that they would not have fallen at all (in which case N would be zero). Perrin concludes: “It cannot be supposed that out of the enormous number of values [of N] a priori possible, values so near to the predicted number have been obtained by chance for every emulsion and under the most varied experimental conditions.” (Atoms, 105). Psillos interprets this as saying that the probability of obtaining result C, given the atomic hypothesis, is high, and the probability of obtaining C, given that the atomic hypothesis is false, is low. This passage, I believe, does not say, or support, the second half of that conclusion. Perrin does not here argue that if the atomic theory were false then it would be very unlikely he would have obtained the value he did for N over his range of experiments, but only that the atomic theory, or in my more specific formulation, T, will make it likely that one will get consistent values for N, that is, p(C/T&b) is high. (I relativize the latter probability to background information b that includes results of other experiments in the nineteenth century that make it probable that Brownian motion is not caused by various external factors.) If, as Perrin argues, p(T/b) is (very) high, then it can be demonstrated that p(T/C&b) is too, which in my writings is how I suggest reconstructing Perrin’s argument to the high probability of T. (For the proof, see “Jean Perrin and Molecular Reality,” reprinted in my Evidence, Explanation, and Realism.) Finally, although I do not have the space to give a decisive argument here, I am skeptical of what Psillos calls second-order evidence from the track record of scientific theories generally and the reliability of scientific methodology. (He claims that philosophers need to balance first- and second-order
Achinstein’s Replies
287
evidence, something that is contextual.) As an example of second-order evidence Psillos mentions the “pessimistic induction” from the failure of past theories to the probable failure of a particular one. I don’t regard the failure of past theories as potential or veridical evidence of the failure of any particular theory we are considering. Let e be the fact that most theories have turned out false, and let h be some particular theory we are now considering, say T above. For e to be (potential or veridical) evidence that –T, it must be the case that p(there is an explanatory connection between –T and e/e) > ½. But I regard this probability as very low if not zero. Even given e (that most theories have turned out false), it is not probable that a correct explanation for (–T), that chemical substances are not composed of molecules and so on, is that most theories have turned out false; or that a correct explanation of why most theories have turned out false is that chemical substances are not composed of molecules, or that some hypothesis correctly explains why –T and e are both true. (Just to take the first part of this disjunction, it is probable that a correct explanation of why chemical substances are not composed of molecules, assuming they are not, would be some particular chemical one, not the fact that most theories have turned out false.) Is “second-order evidence” evidence in some different sense? I invite Stathis to relieve me of my skepticism concerning such evidence.
RICHARDS’S DARWINIAN PUZZLE I am grateful to Richard Richards not only for his present contribution, but for his friendship over the years as well as for his encouragement, advice, and philosophical insights. In his essay Richards concentrates on a very interesting but puzzling claim made by Darwin, namely, that although he (Darwin) at one point came to believe in branching evolution on the basis of the taxonomic facts, and he believed that he was justified in doing so, he also believed that these same taxonomic facts were not sufficient for other scientists to believe in branching evolution. By the latter, according to Richards, Darwin means not that these taxonomic facts would fail to persuade others to believe, but that others would not be justified in so believing. Richards rightfully finds this puzzling and asks how it can be explained. It cannot be explained, he says, using standard a priori theories of evidence. For example, on Carnap’s logical theory of evidence, or on a hypothetico-deductive view, if e is evidence that h, then this is so “for everyone,” no matter who believes what. Richards’s solution to the problem is rather simple. Using my concept of ES (epistemic situation) evidence, he claims that what Darwin was saying is that relative to the sort of epistemic
288
Philosophy of Science Matters
situation he (Darwin) was in, the taxonomic facts did provide ES-evidence; but others were in a different sort of epistemic situation, relative to which the taxonomic facts did not provide such evidence. Richards draws an interesting conclusion from this regarding the history of science. Historians of science should be interested not only in the (epistemic) reasons scientists had for their beliefs—which he associates with my concept of subjective evidence; they should also be concerned with the question of whether, given their epistemic situations, the scientists were justified in those beliefs—which he associates with my concept of ES-evidence. For example, in describing how Hertz came to believe that cathode rays are electrically neutral, the historian of science Jed Buchwald not only gives Hertz’s own epistemic reasons for that belief, but he also evaluates those reasons as good ones, given Hertz’s particular situation. My only comment here is that not very many historians of science do this sort of thing. More frequently they are concerned with simply identifying the reasons scientists in fact had for believing what they did, whether these reasons are epistemic or non-epistemic (the latter including causal factors influencing those beliefs). But evaluating those beliefs, particularly in the way Richards suggests, is a much more difficult task. It involves not only (1) discovering what reason e the scientist had for believing h, but also (2) identifying what particular epistemic situation the scientist was in (which, among other things, involves identifying what beliefs other than e and h he had or was in a position to have), and (3) determining whether anyone in that epistemic situation would have been justified in believing h for the reason e. It is one thing for e to be ES-evidence for some particular ES. It is quite another to satisfy conditions (1)–(3). Is that what historians of science should be doing to do good history of science? Speaking as a philosopher of science, I am quite satisfied if the historians tell me what reasons scientists in fact had for their beliefs (i.e., (1) above). I would also like to learn from them whether other scientists at the time or later regarded e as a good reason for believing h, and why they did so. And, of course, as I have emphasized elsewhere, what scientists want to know is whether e really is a good reason for believing h (or, in terms of evidence, whether e is veridical evidence that h), which is a scientific issue that historians of science do not usually investigate. But none of these tasks requires what Richards
Achinstein’s Replies
289
seems to have in mind (i.e., satisfying all the conditions in (1)–(3)). He sets a very high goal indeed for the historian.
RUSE’S GOOD BREEDING As an eminent philosopher of biology, Michael Ruse’s influence in his own field has been considerable. But he is also one of the very few philosophers who has had a significant impact on applications of philosophy of science to important practical and political issues such as those taken up in the courts regarding the teaching of evolution. I admire him for being able to wear both hats, and I am glad to learn that my first book, Concepts of Science, played at least some role in his formative years, especially in weaning him from logical positivism. In his essay, Ruse discusses the vexing question of the role of artificial selection by breeders in Darwin’s development of the theory of natural selection. According to Ruse, “the world of the breeders provides a crucial piece of Darwin’s thinking. It offers a model or analogy leading to the central evolutionary force of natural (and its sister sexual) selection.” Ruse sees three roles for artificial selection as a model or analogy. One he calls “heuristic”: Darwin claimed that he got the idea of natural selection by considering artificial selection. A second he calls “pedagogical”: the model or analogy of artificial selection is used to introduce us to, and teach us about, the more general idea of natural selection. The third role is that of evidential support for his theory: “Darwin always looked upon the world of the breeders as part of the justification for his theory.” It is this third role that I want to focus on. Philosophically speaking, this I take to be the most interesting and controversial. I begin with a few remarks about models and analogies generally, using some simple examples from physics. One of the most famous models discussed by philosophers of science is what came to be known as the “billiard ball model of a gas,” developed by James Clerk Maxwell in his first paper on kinetic-molecular theory in 1860. Another often cited example is Rutherford’s “solar system” model of the atom developed in the early years of the twentieth century. In considering these we need to make an important distinction between the model itself and an analogy that it may employ. The expression “billiard ball model,” for example, is used to refer to a set of simplifying assumptions that Maxwell (among others) made about the molecules that, he postulated, comprise a gas. These assumptions included that molecules are spherical, that they exert no forces on one another except at impact, that they travel in straight lines except at impact, that they are subject to standard laws of dynamics,
290
Philosophy of Science Matters
and so forth. The term “analogy” here I take to refer to the analogy between molecules so described and billiard balls exerting no forces on one another except at impact, traveling in straight lines otherwise, and being subject to standard laws of dynamics. A similar distinction can be made in the case of Rutherford’s planetary model of the atom. Now consider the notion of “support for the theory.” Maxwell was explicitly agnostic regarding whether the ideas of his 1860 simplified billiard ball model of a gas could be applied to real gases. Writing to Stokes in 1859 just before the publication of his paper he says: “I do not know how far such speculations may be found to agree with facts. . . . I have done so as an exercise in mechanics. Now do you think there is any so complete a refutation of this theory of gases as would make it absurd to investigate it further so as to found arguments upon measurements of strictly “molecular” quantities before we know whether there be any molecules?” By the 1870s, Maxwell had more indirect evidence for the existence of molecules, and so he was able to make claims about them that he thought were justified. But he did so without use of the original model. More generally, whether a model can be extended to the “real world” depends crucially on the assumptions made in the model, and on whether there is any justification for them. Sometimes there is, sometimes not. What about the “analogy” here? (Keep in mind this is our analogy, not Maxwell’s.) The analogy consists in the (partial) similarity we draw between molecules and their motions in a gas and billiard balls and their motions on a billiard table. Here it is questionable what, if anything, we can infer about real world molecules from billiard balls and their motions. Obviously we cannot infer that molecules have numbers imprinted on them, that they all move in the same plane, that some of them fall into pockets, or, most importantly, that there are forces exerted by humans, in addition to intermolecular ones, responsible for their motions—as in the case of billiard balls. Analogies, which are by their very nature partial and frequently loose, are not, in general, particularly reliable support for the theory about the entities for which they are providing an analogy. (“It’s only an analogy,” we say, “don’t expect too much.”) They may help one to think about the entities in the theory— they may be of “heuristic” and “pedagogical” value—but that doesn’t mean they provide evidential support for the theory. If the analogy drawn between two systems is perfect or close to it, and, if we have reason to suppose that both systems are governed by some common or similar mechanism, then we can use our knowledge of one system to make inferences about the other. But the closer to perfection the analogy, the less inclined are we to speak of it as an analogy, but simply as two systems subject to the same principles.
Achinstein’s Replies
291
My own theory of veridical evidence provides a justification for claims in the last paragraph. For e to be veridical evidence that h, the high probability of an explanatory connection is required between e and h. Let h be the claim that the hypothesized entities (e.g., the molecules) have certain properties P, and let e be fact that the “analog” entities (e.g., the billiard balls) have certain analogous properties P’. It is not that the putative fact (h) that the molecules have the properties P probably correctly explains why (e) billiard balls have the analogous properties P’; nor is the reverse true. For the required explanatory connection, the best we can hope for is that some hypothesis h’ correctly explains both h and e. The closer to perfection the analogy is—the more similar the properties in P and P’, the larger the sets in question, and the less dissimilar are other properties (the “negative analogy”)—the more probable it is that we are dealing with two systems subject to the same principles or “mechanism,” which can explain properties of both systems. In such a case the behavior of one system can supply evidence for claims about the behavior of the other. But then, as noted above, we are not likely to speak of this as an analogy. The more dissimilar the two systems, the less likely it is that there is some common explanation for analogous properties of each, and hence the less justified we would be in the claim that the behavior of the analog system provides evidence for corresponding claims about the hypothesized system. Turning finally to artificial selection, when Ruse claims that this “model” or “analogy” provides support for Darwin’s general theory of natural selection—or at least that Darwin thought it did—to what is he (Ruse) referring by “model” and “analogy”? Let us suppose first that it is intended by Darwin simply as an analogy, in roughly the same way as we use the idea of billiard balls as an analog of molecules, or solar systems as an analog of atoms. There is a partial similarity between what happens in breeding and, according to the theory, what happens in nature. Both involve certain changes in the characteristics of animals and plants, and both involve selection broadly speaking. But as with analogies generally, there are dissimilarities as well. For example, there is a human agent selecting in the case of breeding, none postulated in nature; there is no evolution to a new species Darwin observed in breeding, although this is what he postulated for the case of natural selection in the long run; breeders breed for one or a few characteristics, none of which may involve selective advantages for survival and reproduction, and indeed may be harmful for these purposes; nature, according to Darwin, works very differently; and so forth. Such disanalogies, of course, were noted by Darwin’s opponents. So if Ruse is claiming that for Darwin artificial selection served as an analogy that provided evidential support for the general theory of natural selection as the mechanism for evolution, then I want to know a good deal more
292
Philosophy of Science Matters
about how such support is supposed to work. Why aren’t the disanalogies sufficient to make it unlikely that there is a common explanation for the effects of artificial and natural selection? What happens if we treat artificial selection as a model? How would that work? Here we need to be careful if we have in mind something like the billiard ball model of a gas or the planetary model of an atom. A model in this sense is a set of simplifying assumptions about the real world. So in the Darwin case the model would assume that in nature, members of the same species differ at least slightly with respect to many different characteristics; that some of these differences have selective advantages for survival and reproduction; that over long periods of time changes favoring these characteristics will transpire forming new species; and so forth. Perhaps these were regarded by Darwin as simplifying assumptions, and perhaps he was led to think about them by thinking about artificial selection. But the problem here is that “the model” (even if considered simplified and suggested in part by considering artificial selection) is simply Darwin’s theory of natural selection. I don’t see how conceived as a model it provides support for the theory of natural selection—any more than treating Maxwell’s (simplifying) assumptions about gas molecules as a model provides support for the claim that real-world molecules behave (approximately) in these ways. I leave Michael with a challenge, which I feel sure he can meet: tell me more about your notion of “support” and about how the “model/analogy” of artificial selection is supposed to provide support for natural selection—whether in Darwin’s mind or any other.
STALEY ON “SECURING EVIDENCE” Kent Staley is right to point out (as Kitcher does as well) that my account of ES-evidence appeals (among other things) to being “justified in believing” something, which I don’t further explicate. Staley makes a very interesting attempt to bolster my account of evidence by providing such an explication. He arrives at this result, which he calls JE (justified evidence): an assertion of h as a conclusion inferred from observed fact(s) e is fully justified relative to epistemic situation K if (1) e is potential evidence for h (using my concept of potential evidence); and (2) the proposition “e is potential evidence for h” is secure throughout all scenarios that are epistemically possible relative to K. Prior to arriving at this result, he offers both a formal and an informal definition of a proposition being “secure throughout a scenario relative to an epistemic situation K.” The informal definition is this: a proposition is secure for an epistemic agent
Achinstein’s Replies
293
“just in so far as, whatever might be the case for all that the agent knows, that proposition remains true.” He then illustrates what he has in mind with a more detailed account of Hertz’s cathode ray experiments than I offered in my book. I will discuss Staley’s proposal using an example from Galileo’s Dialogues on the Two Chief World Systems. In that work Simplicio, following Aristotle, asserts that the earth is stationary and does not turn on its axis (h). He infers this from the observed fact that a stone dropped from the top of a tower falls at the base of the tower, not a distance away (e). Is Simplicio’s assertion h “fully justified” by e relative to his own epistemic situation (in Staley’s sense)? Not if we understand Staley to be providing necessary as well as sufficient conditions. (He uses “if,” but I want to consider “if and only if”). The reason is that on my view, e is not potential (or veridical) evidence that h in this case, so Staley’s first condition in JE is not satisfied. Therefore, no matter how well Simplicio argues for his conclusion h given his epistemic situation, his assertion of h will not be justified even relative to his own epistemic situation K. Similarly, and perhaps even more strikingly in the Hertz case, on the present conception of justification, Hertz’s claim that cathode rays are not charged is not justified, even relative to his own epistemic situation, since his experimental results were not potential evidence for this conclusion. This is too rich a sense of “justification” for my blood, even if Staley calls it “ideal justification.” It is not what I had in mind for ES-evidence. (On my view, Hertz’s results were ES-evidence for his hypothesis, relative to his ES; hence, he was justified in believing that hypothesis, even though the results were not potential evidence for the hypothesis.) On the other hand, if Staley’s conditions for being “fully justified” are not necessary conditions, but only sufficient ones, what conditions are necessary? Staley needs to tell us. Suppose that we drop Staley’s first condition for justification and keep the second as both necessary and sufficient. Could my concept of ES-evidence be understood using this idea? We could say that e is ES-evidence that h, relative to a given ES, if and only if e is true and the proposition “e is potential evidence that h” is secure throughout all scenarios that are epistemically possible relative to that ES (or just secure relative to that ES). Presumably, the fact that the stone drops to the base of the tower would then be ES-evidence for the stationary earth hypothesis for someone in Simplicio’s epistemic situation. That would be okay with me. Now my worry pertains to the importance Staley wants to place on “justified evidence” (or someone might want to place on my notion of ES-evidence) in this sense. Let me put the point perhaps more boldly than I should. Besides a few historians and philosophers of science (and Simplicio himself), who cares (and who should care) whether, relative to
294
Philosophy of Science Matters
his own epistemic situation, Simplicio was justified in concluding that the earth is stationary? Who cares if this e is ES-evidence relative to his ES? Even if this is a sense of “evidence,” as I think it is, the major concern (it was certainly Galileo’s) was whether the fact that the stone lands at the base of the tower is veridical (and hence potential) evidence that the earth is stationary. Galileo famously argued that it is not. In short, Staley’s concept of “justified evidence,” if understood as requiring the satisfaction of both of his conditions, yields much too strong a sense of justification, at least for my purposes. The weaker concept in the last paragraph leaves me with the question: If this were to be Staley’s idea of “justified evidence,” why should one value such a concept as much as he seems to?
VAN FRAASSEN’S PERRIN I am a scientific realist. With respect to Perrin, van Fraassen places me in the good company of Wesley Salmon, Clark Glymour, and Penelope Maddy in subscribing to the “philosophy lore” that understands Perrin’s work as “epistemically legitimating the conclusion” that atoms and molecules are real. In his challenging response to this “lore,” van Fraassen proposes a very different, and quite interesting, interpretation of Perrin and other scientists as well. Tracing his ideas to a 1927 work of Herman Weyl, van Fraassen claims that when scientists “describe the acceptance of a scientific theory,” even if, as with Perrin, they make explicitly realist assertions, another, and I take it better, way to understand what they are claiming is only (what van Fraassen calls) “empirical grounding” for the theory. Basically (if I understand this correctly) the idea is that each significant “theoretical” parameter in the theory must be such that its value can be determined on the basis of measurement (of some “observational” quantities or other), where this determination is theory-relative, that is, is made with the aid of theoretical assumptions. So, for example, Perrin derives the formula: (1) n¢/n = 1—Nmg(1-d/D)h/RT in which n´ represents the number of Brownian particles per unit volume at the upper level of his experimental tube and n the same at the lower level; D is the density of the material comprising the Brownian particles; d is the density of the liquid in the tube; and N is Avogadro’s number (the number of molecules in a gram molecular weight of a substance—the theoretical parameter of interest). All of the other parameters are measur-
Achinstein’s Replies
295
able experimentally. In different experiments using different fluids, and Brownian particles with different sizes and masses, Perrin calculated N to be approximately the same in the various experiments. In doing so, van Fraassen emphasizes, Perrin made various “theoretical” assumptions, including that molecules exist, that they cause the Brownian motion, and that the visible Brownian particles comprising the liquid will behave like molecules in a gas with respect to their vertical distribution. For van Fraassen, this is what is necessary for “empirical grounding” for N. The satisfaction of this condition for other theoretical parameters as well is all that Perrin needed or should have wanted for the acceptance of molecular theory. Before responding, let me mention another example cited by van Fraassen. It involves Maxwell’s determination of molecular speeds. From a virial equation of Clausius, Maxwell derives a formula (2) Vp = 1/3Mc2 where V = volume of the gas, p = pressure of the gas, M = mass of the whole quantity of the gas, and c2 = mean square velocity of a gas molecule. Using this, and assuming a very rarefied gas and that the velocity of every molecule is the same, from the measurable quantities V, p, and M, Maxwell calculates that the velocity of a molecule of oxygen is 461 meters per second, that of nitrogen is 492, and hydrogen is 1844, at 0 degrees centigrade. (See The Scientific Papers of James Clerk Maxwell, vol. 2, W. D. Niven, ed. (Dover, 1965), 426.) So here again we have “empirical grounding” for the theoretical parameter “molecular velocity.” And, in this case, like that of Perrin, we have such grounding using theoretical assumptions. There is, I would claim, an important difference between these cases. Intuitively, Maxwell had no empirical method for determining whether the molecular velocities he computed were correct or approximately so, even for the idealized case he considers. By contrast, Perrin, I would claim, had, or at least thought he had, a good empirical method for determining a value for Avogadro’s number. What’s the difference? One difference is mentioned by van Fraassen: “concordance.” Perrin got the same number with different experiments, whereas Maxwell had no experiments whatever to determine the correctness of his claims about molecular velocities. But there is another difference as well. The theoretical assumptions Perrin uses to derive his formula are ones for which he offers arguments: for example, arguments from experiments of Gouy and others that Brownian motion is caused internally by bombardment with molecules, and at least some argument for the claim that Brownian particles can be treated like large molecules for purposes of the equation (1) above. Maxwell does
296
Philosophy of Science Matters
offer some general arguments for the existence of molecules, and for the assumption that they satisfy Newtonian dynamical laws, that are independent of the kinetic-molecular theory. (See my “What to Do if You Want to Defend a Theory You Cannot Prove: A Method of Physical Speculation,” The Journal of Philosophy 107 (2010), 35–55, reprinted in my Evidence, Explanation, and Realism.) But these are not of the strength of those supplied by Perrin (or at least that is what Perrin would have claimed). Even if we grant Maxwell some justification for assuming that the Clausius virial equation that holds for macro-particles in an enclosed vessel will also hold for molecules, he offers none for the supposition that the molecules in the enclosure will all have the same velocity—a theoretical assumption he needs to get his conclusion regarding molecular velocities. More generally, contrary to what van Fraassen seems to be suggesting, I would claim that whether “empirical grounding” of theoretical parameters yields enough for “acceptance”—at least for acceptance of the values of those parameters—depends importantly on whether a justification is given for the theoretical assumptions employed, and on how strong that justification is. This can vary significantly from one case to another. A larger question, of course, pertains to the realist-anti-realist debate. Here I will simply say that both Maxwell and Perrin clearly believed that they themselves were realists. They sought to establish that molecules are real and that the claims each was trying to make about them were true. As van Fraassen agrees, there are plenty of quotes from the writings of Perrin, as well as from Maxwell, that attest to this. They believed that “empirical grounding” (together even with “concordance”) is not sufficient for acceptance of a theory. And, since “empirical grounding” (even, I would argue, with “concordance”) can vary considerably in its strength, I think they were right. However, I will not pursue the larger question here. (My own defense of realism, as well as a critique of van Fraassen’s anti-realism, can be found in my “Is there a Valid Experimental Argument for Scientific Realism?” Journal of Philosophy 99 (2002), 470–95, reprinted in my Evidence, Explanation, and Realism.)
WOODWARD’S CAUSES VS. CONDITIONS Among “causally relevant” factors that play roles in producing events, Woodward wants to distinguish causes (in some narrow sense) from enabling conditions. So, for example, in the robbery example he cites from Judith Thomson, the construction of the bridge was an enabling condition of the robbery, but not its cause. Some philosophers claim that this dis-
Achinstein’s Replies
297
tinction is “capricious and arbitrary” with no objective basis whatever. Woodward argues that there are objective factors—what he calls “stability” and “specificity”—that play an important role in selecting causes from mere conditions. I have not discussed this distinction in my own work, but I have talked about “pragmatic” conditions for giving evaluations of explanations, including causal explanations, as being good or appropriate for certain contexts and not others. Woodward connects this with his discussion of the cause-conditions distinction by admitting that both “stability” and “specificity” are subject to “pragmatic” conditions: what counts as stable or specific in one context may not in another. Yet, on the “objective” side, he thinks that it is an empirical fact about ordinary people and scientists alike that for a range of different contexts there will be agreement about what is stable and specific. I agree with Woodward in recognizing the importance of pragmatic factors in determining causes. What caused the soldier’s death? Was it the terrorists’ roadside bomb, or a lack of proper body armor? Which is selected as the cause, as opposed to a mere condition, although not “arbitrary and capricious,” depends in some measure on the context of inquiry. The military officer making the report to his superiors will cite the bomb as the cause; a congressional committee investigating the army might pick out the lack of proper body armor. Whether Woodward’s (pragmatically influenced) factors of “stability” and “specificity” yield more of the objectivity that Woodward seeks in this distinction is an interesting question for which I can here offer no general answer. I will raise just one question. Can’t there be “unstable” causes? Woodward’s “stability” condition is concerned with the extent to which the relationship between cause C and effect E will continue to hold as various background factors change, where a background factor is any factor distinct from C and E. Now think of those scary warnings on prescription blurbs for pills, for example, “this pill can cause stomach upset.” Suppose that the pill caused John’s stomach upset yesterday, but not today when he took another one. (We rule out other causes one by one.) C can cause E (the blurb tells us, although this is rare), and yesterday it did in John’s case, but there isn’t much “stability” here, since if background conditions distinct from C and E change, the causal connection will probably not exist. To be sure, Woodward says that “stability” is not a necessary condition for the selection of a cause, only a “relevant” one. Whether this is so, and also what exactly is supposed to count as “stable,” are questions I will leave for Jim for another occasion. But I do welcome his pragmatism in these matters.
Index
Achinstein, Peter, his influences, 3–6, 259 on analogy, 204–5, 214, 289–92 on evidence, see Evidence on explanation, 15, 72, 74–6, 247, 255, 267–69 on Jean Perrin, 177, 185–87, 231, 241, 241, 285–87, 294–96 on James Clerk Maxwell, 29–34, 40, 260–64 on John Stuart Mill, 96–99, 106, 147–48 on Isaac Newton, 44–55, 96–99, 264–65 on William Whewell, 106, 152–53, 281–83 on the wave-particle debate, 109–20, 273–76 on scientific realism, 188, 231, 241, 294–96 on rules of induction, 44–55, 110, 164–65, 264–65, 284–85 Analogy, 30, 38, 41, 178, 205–12, 214, 260–61, 289–92 Anti-realism, 294–96 Anscombe, G. E. M., 17 Avogadro’s number, 68, 132, 180–88, 238–42, 285, 294–95 Aristotle, 293, 52–3 Artificial selection, 205–10, 214, 289–92 Atwood, Rev. George, 233–234 Austin, John L., 5–10, 13, 259 Ayer, A. J., 5–6 Bacon, Francis, 110, 194–195, 209 Bayesian methods, 59, 105, 112–13, 116, 135–37, 143–45, 148, 157, 176n, 183, 195, 266, 275, 278–79 Bentham, George, 211 Bootstrapping, 236, 244, 266 Braithwaite, Richard, 203
Brownian motion, 10, 132, 177, 179–185, 188, 231, 241–43, 246n, 277, 285–86, 294–95 Buchwald, Jed Z., 89, 121n, 222, 226–228, 230n, 288 Brush, Steven, 231, 246n Cantor, Geoffrey, 110 Carnap, Rudolf, 3–10, 13, 45, 56n, 135, 194, 259, 266, 287 Cartwright, Nancy, 258, 259–260, 285 Cat, Jordi, 258, 260–64 Cathode rays, 11–12, 46–47, 61, 70, 87, 93, 197–200, 218, 221–28, 270–71, 288, 293 Causation, 4, 7–8, 16, 18, 21, 35, 51, 97, 99, 101–2, 187, 211, 247–257, 265, 295 Coherence, 66, 113, 151–62, 234, 281–83 Confirmation, See Evidence Consilience, 34, 36, 102–4, 110, 152, 271–72 Darwin, Charles, 75, 87, 89–90, 93, 193–201, 205–14, 287–88, 289–92 Davy, Humphry, 263 De Sitter spacetime, 170–71 Dewey, John, 108n De Morgan, Augustus, 41 Descartes, Rene, 48–49, 52–54, 194–95, 265 Di Fate, Victor, 258, 264–65 DNA, 94, 249, 253 Doppelt, Gerald, 258, 265–67, 281 Dorling, Jean, 33, 40, 43n Duhem, Pierre, 105, 127, 178, 184, 189, 233, 246n Dutch books, 158 Earman, John, 156 Electron, 128, Also see Cathode rays
298
Index Eliminative causal reasoning, 66–70, 90, 185, 187 Einstein, Albert, 166–67, 232, 241–43, 246n Euclid, 34 Explanation, D-N explanation, 4, 17, 72–3, 283 Pragmatics of, 79, 255–257, 268 Experimental philosophy, 13 Evidence, and explanation, 15, 63–66 epistemic situation (ES) evidence, 11, 60–2, 64–70, 88–90, 125, 127, 132–33, 192, 216–19, 221, 228, 266, 269–70, 288, 292–94 justified evidence, 221 potential evidence 57n, 60–69, 90, 125, 197–98, 201, 217–21, 228, 229n, 265–268, 270, 278, 281, 285, 287, 292–94 subjective evidence, 60–1, 67, 69, 192, 197–9, 202, 288 veridical evidence, 8, 10, 12, 57n, 60–70, 88–9, 197–98, 216–19, 221, 228–29, 230n, 259, 265–67, 268, 270, 277, 278, 281, 287–88, 291, 293, 294 Fodor, Jerry, 205, 213–14 Friedman, Michael, 74 Fresnel, Augustin-Jean, 115, 274 Galileo, 192–4 Geroch, Robert, 173 Gimbel, Steven, 258, 259 Glymour, Clark, 165, 231, 235–237, 246n, 266, 294 Goldstein, Adam M., 258, 267–68 Goldstein, Eugen, 222 Gouy, Léon, 179 Grue, 176n, 195 Hamilton, William, 41 Hampshire, Stuart, 26 Hare, Richard, 5 Hempel, Carl G., 3–5, 45, 72, 73–80, 194, 203, 244, 259, 266, 268, 283 Hanson, Norwood Russell, 204, 246n Herschel, John, 118, 121n, 194, 195, 211, 274 Hertz, Heinrich, 11–12, 46–47, 61, 85–93, 197–298, 200, 221–28, 270–71 Heuristics, viii, 41, 203, 209–10, 213, 289–90
299 Holism, 124, 127–33, 259, 276 Hume, David, 5, 110, 195, 284 Huygens, Christian, 48, 111, 274 Hypothetico-deductive (H-D) method, 31–2, 35–41, 50, 59, 96, 99, 106, 110, 127, 140, 179, 187, 194, 203, 244, 261–63, 265–66, 271–75, 287 Indeterminism, 165, 168–69, 171–72, 176n Induction, 5, 45, 47–48, 54, 59, 96–99, 101–106, 110, 116, 121n, 147–48, 164, 171, 194–5, 225, 259, 263, 264–65, 273, 281 Norton’s material theory of, 47, 164–65, 173, 284–85 pessimistic meta-induction, 68–70, 287 Indeterminism, 38, 165, 168–69, 171–72, 176n Inference to the best explanation (IBE), 58n, 59, 63–4, 67, 189, 195, 265 Inference to the best explanation realism (IBER), 63–69 Intelligence quotient (IQ), 95 Janiak, Andrew, 48 Joule, James Prescott, 263 Kant, Immanuel, 10 Kelvin, Lord, 35, 38–9, 42 Kepler, Johannes, 17, 101, 212 Kitcher, Philip, 74, 258, 266, 269–71 Kuhn, Thomas, 86–7, 103, 106, 108n, 204, 205, 210, 213, 246n Kronz, Frederick M., 258, 271–73 Lalande, André, 177 Laudan, Larry, 44, 53, 68, 111, 122n, 258, 273 Lakatos, Imre, 105, 108n Laws of nature, 7, 73, 76–82, 101, 147, 269 Leibniz, Gottfried, 48 Lewis, David, 248–250, 256–57, Logical positivism, 3–4, 259, 289 Logical empiricism, 3, 10, 14, 45, 203–5, 209–14 Longino, Helen, 258, 276–78, 281 Lottery examples, 90, 127–28, 130, 133, 281 Lloyd, Humphrey, 122n, 274–75 Mach, Ernst, 178, 246n, Maddy, Penelope, 231–33, 246n, 294
300 Malament, David, 165 Maxwell, James Clerk, 29–42, 67, 239–40, 260–64, 283, 289–92, 295–96 on method of physical speculation 31, 260–261 1860 kinetic theory, 30, 261, 289–90 Maynes, Jeffrey, 259, 260 Mayo, Deborah G., 127, 229n, 258, 278–81 McMullan, Ernan, 57n Mill, John Stuart, 41, 47, 164–65, 194–95, 248–9, 256–57, 264–65, 271–73, 284 as a sever tester, 147–148 Mill’s methods of experimental inquiry, 97–102 on the hypothetical method, 96–103 Minkowski space-time, 169–73, 175n, 176n, 284 Morgan, Gregory J., 258, 281–83 Newton, Isaac, rules for natural philosophy, 51–53, 98 Neyman-Pearson statistical methods, 105 Nagel, Ernest, 203 Nagel, Thomas, 5 Natural selection, 79–81, 192–93, 212, 289–92 Darwin on, 205–210 as vera causa, 211 Norton, John D., 194–6, 258, 264–65, 284–85 Nyhof, John, 231, 246n Objective epistemic probability, 135–6, 142 Occam’s razor, 171–72 Ostwald, Wihelm, 185 Paradox of the ravens, 84, 195 Peacock, George, 41 Piattelli-Palmarini, Massimo, 205 Perrin, Jean, 10, 68–9, 85, 90, 132, 177–89 and realism, 231–33, 238–44 Pierce, Charles, 103–6, 107n, 142, 150n, 194–95 Poincare, Henri, 182, 233, 146n Popper, Karl, 92, 103, 105, 106, 108n, 194 Psillos, Stathis, 258, 285–87 Putman, Hilary, 68 Quarks, 220 Quine, W.V., 3, 259 Duhem-Quine thesis, 105, 127
Index Random controlled trial (RCT), 15–26 Railton, Peter, 73–4 Reichenbach, Hans, 246n Realism, 29, 30, 39, 42, 58n, 262, 296 Achinstein, Maxwell and, 29–34, 262 No miracles argument, 68, 189 Reductionism, 7, 261 Richards, Richard A., 258, 287–89 Rumford, Count, 263 Ruse, Michael, 258, 289–292 Rutherford, Ernest, 244 planetary model of the atom, 289–90 Staley, Kent, 258, 292–94 Stokes, George Gabriel, 35, 290 Stokes equation, 181 Strawson, Peter, 5–7, 259 Salmon, Wesley, 68, 73–74, 156, 187, 231, 268, 294 Severe tests, 137–148 Sexual selection, 207, 214, 289 Sign of truth, 153–162, 282–83 Simplicity, 283, 284, See also Occam’s razor Sober, Elliott, 195, 202 Tait, Peter Guthrie, 38–9, 42 Thomson, J. J., 85, 93, 266 cathode ray experiments, 11–2, 45–7, 197–98, 218, 221, 226–28, 270–71 Thomson, Judith, 247, 251, 254, 296 Thomson, William, See Kelvin, Lord Toulmin, Stephen, 204 Under-determination, 167 van Fraassen, Bas, 72, 134n, 189, 258, 294–96 Wallace, Alfred Russel, 209, 213–14, Wave and particle theories of light, 66–70, 109–120, 273–76 Weyl, Hermann, 235, 241–42, 246n, 294 Whewell, William, 36–37, 41, 57n, 96–106, 107n, 110, 113, 118–19, 121n, 151–55, 161–62, 194–95, 265, 271–72, 281–83 Wolfram, Stephen, 90 Woodward, James, 258, 296–97 Wright, Sewall, 79, 81 Young, Thomas, 111, 274–75