Article
Author: Gordon Belot (University of Michigan)
This paper revisits the bearing of underdetermination arguments on scientific realism. First it argues that underdetermination considerations provide good reason to doubt that science is objective in the strong sense that anyone following the its methods will be led closer and closer (without bound) to the truth about any given question within the purview of those methods, as more relevant data are considered. Then it argues that scientific realism is difficult to maintain in the absence of this sort of objectivity.
Keywords:
How to Cite: Belot, G. (2022) “Gravity and GRACE”, Philosophers' Imprint. 22(0). doi: https://doi.org/10.3998/phimp.2120
But what about such a proposition as ‘I know I have a brain’? Can I doubt it? Grounds for doubt are lacking! Everything speaks in its favour, nothing against it. Nevertheless it is imaginable that my skull should turn out empty when it was operated on.
—Wittgenstein, On Certainty §4
i. This is a paper about the bearing of the underdetermination of theory by data on the question of scientific realism. But the approach taken will be oblique. Initially, the focus will be on the thesis that anyone following the methods of science will be led closer and closer (without bound) to the truth about any given question within the purview of those methods, as more relevant data are considered. I will call this the thesis that science is alethotropically objective (i.e., objective in the sense that it has a tendency to turn towards the truth).
The centrality of the thesis of the alethotropic objectivity to philosophy of science has fluctuated over the years. But it seems fair to say that its popularity among philosophers is currently at a low point. My first goal below will be to kick it while it is down. I will provide some reason to think that science cannot be alethotropically objective in the actual world and some reason to think that science couldn’t be objective in that sense in any world as complex as ours. And I will argue that two popular manoeuvres that promise to save at least the spirit of alethotropic objectivity can do no such thing.
Why bother with all this if, as I say, few philosophers today endorse this sort of strong objectivity thesis? Because many philosophers today espouse scientific realism and dismiss worries about underdetermination. My final task below will be to make a case that it is not so easy to dismiss underdetermination arguments once one has faced up to what scientific realism looks like in the absence of alethotropic objectivity.
ii. Some preliminaries before proceeding. First, note that even if science is alethotropically objective, there may be many questions whose answers remain forever unknown to us. There may be questions outside the purview of scientific methods. There may be questions within the purview of scientific methods that are not pursued indefinitely. There may be questions within the purview of scientific methods that are pursued indefinitely but for which new evidence eventually ceases to come to light. And there may be questions whose answers we approach asymptotically but never reach.
iii. Next, since alethotropic objectivity is by no means the most common notion of scientific objectivity a few words may be order about its career (and about its credentials to be considered a species of objectivity).1
According to one important strand of usage, reality is objective to the extent that it is mind-independent and methods of inquiry are objective to the extent that they allow us to discover the nature of this reality.2
When scientists are driven to discuss objectivity, it is often in reaction to some view on which the human or the mental is at least partially constitutive of the apparently non-human and non-mental aspects of the world described by physics. Thus, some early structural realists, writing in opposition to various forms of positivism, argued that scientific progress consisted in: (i) the gradual increase in objectivity via the elimination from the scientific world picture of anthropocentric elements; and (ii) the production of world pictures that more and more accurately represent the objective structural relations that are in principle accessible to all intelligent investigators, human and non-human.3 In more recent times, it has been the spectre of social constructivism, rather than positivism, that has driven Nobel Prize winners to pro nounce their faith that extra-terrestrial scientists should discover the same laws that we do.4 So it is natural to interpret them as committed at least to the view that the methods of science are alethotropic when it comes to laws (and it is tempting to think that the limitation to laws is an inessential aspect of the view).
Peirce was uninterested in the label ‘objective,’ but he took science to be alethotropic:
And on this he based his accounts of the pragmatic content of the notions of truth and reality.all the followers of science are fully persuaded that the processes of investigation, if only pushed far enough, will give one certain solution to every question to which they can be applied.5
Alethotropic objectivity has also figured in recent philosophy of science (even among authors who don’t share Peirce’s views about truth and reality). Earman (1992, 138) takes it to be part of the “popular image of science” and urges Bayesian philosophers to have it on their wish lists. Lloyd (1995, §2.3) considers it one of the principal senses of the polysemous notion of objectivity. Railton (2000, 186) identifies it as a component or consequence of the logical empiricist account of objectivity.The opinion which is fated to be ultimately agreed by all who investigate, is what we mean by the truth, and the object represented in this opinion is the real. That is the way I would explain reality.6
iv. A final preliminary point. For present purposes, I will identify scientific realism with the thesis that: (i) the sentences expressing a scientific theory are true or false; (ii) special cases aside, the relevant truth conditions are mind-independent; and (iii) the empirical success of our theories gives us (defeasible) reason to think them true.7 The early years of the twentieth century saw a flourishing debate over scientific realism, in which many anti-realist participants denied clauses (i) or (ii).8 But, for better or worse, in recent decades almost all scientific anti-realists have endorsed (i) and (ii)—and the scientific realism debate has largely been concerned with the epistemic virtues (or lack thereof) of our most successful scientific theories. That is the point at which we will be joining the dialectic below.
v. It will be helpful for what follows to recall the distinction between direct problems and inverse problems. In fields such as medical imaging, geophysics, and astrophysics, we are often interested in the set of measurements that can be made in the region exterior to an object with a given internal structure.
So long as the laws of physics take their usual form, the direct problem will be solvable in principle—if one specifies the set of possible internal structures the object of interest might have, there will be a mathematically well-defined map K that determines for each possible internal structure X, the set K(X) of outcomes of ideal measurements in the region exterior to the body of interest, should it have internal structure X. But the inverse problem need not be solvable in this sense. Even if K exists as a well-defined map, it need not have a well-defined inverse—if multiple possible internal structures of our body give rise to the same set of exterior measurements, then the inverse mapping is undefined. In this case we say that the inverse problem is underdetermined or unsolvable. Among applied mathematicians, it is generally thought that solvable inverse problems are the exception rather than the rule—and that the more complex and non-linear the direct problem, the more likely the inverse problem is to be unsolvable.9The Direct Problem: Given the laws and the internal structure X of object S, to determine the set O = K(X) of outcomes of measurements that can be made exterior to S.
The Inverse Problem: Given the laws and the outcomes O of the measurements that can be made exterior to a body S, to determine its internal structure X = K−1(O).
vi. Many of the inverse problems that arise in medical imaging are solvable. Suppose that you want to reconstruct the density of matter within my head, having measured the degree to which x-rays are attenuated as they are shot through it from various directions. This inverse problem, known as the problem of x-ray tomography, is solvable: if you know for each of infinitely many directions in space the degree to which x-rays travelling in that direction are attenuated as they travel through my head, then you can uniquely reconstruct the density of matter within it.10 The situation is quite different if you are restricted to work with a finite set of directions: in that case, for any region inside my head, any set of measurements is consistent with that region being void of matter.11
vii. An iconic unsolvable inverse problem is the geophysical problem of gravimetry: to determine the pattern of mass density within the Earth from knowledge of the gravitational field exterior to the planet. Newton (1729, Book 1, Section XII) already observed that this problem is unsolvable in the special case of a spherically symmetric distribution of matter (a uniformly dense sphere determines the same exterior gravitational field as does an equally massive hollow shell of constant thickness and uniform density). Stokes (1867, 482 f.) observed that this result can be generalized—given any distribution of mass within the Earth and any region in the interior of the Earth, there is a second distribution of mass within the Earth that gives rise to the same exterior field and in which the given interior region is void of matter.12
viii. We are interested here in the phenomenon of underdetermination of theory by evidence. It is important to distinguish two species of underdetermination.
ix. In one sense, underdetermination is endemic in science: whenever the space of hypotheses under consideration is sufficiently rich, we do not expect that collecting a finite amount of data will allow us to conclusively determine the true hypothesis. Our primary interest below will be in the distinctive difficulties associated with extra-strength underdetermination. These go beyond those raised by Hume’s problem of induction.
A standard presentation of Hume’s problem runs along the following lines: (i) consider the inference from the past track record of bread (nourishing every day!) to the expectation that bread will be nourishing tomorrow; (ii) no matter how large our data set, we may be disappointed tomorrow—the inference is ampliative; and (iii) the obvious ways of trying to justify this inference are unsatisfactory because they turn out to be rule-circular. At the second stage, we appeal to the fact that knowing that a binary sequence begins with a certain number of consecutive ones (encoding that fact that bread has always nourished thus far) does not guarantee that it will not eventually contain a zero. So we have a form of underdetermination (a data set consistent with multiple answers to a question of interest).
But we do not here have the feature characteristic of extra-strength underdetermination, since the strategy of guessing that bread will always nourish unless and until a day comes on which it fails to nourish is a strategy that is guaranteed to output a sequence of conjectures that eventually settle permanently on the truth concerning whether or not bread always nourishes. Extra-strength underdetermination brings with it further difficulties beyond those attendant upon Hume’s problem.
x. If we were to imagine for a second that we had no way of discovering what was inside of human heads other than shooting x-rays through them, then the problem of x-ray tomography would constitute an example of regular-strength underdetermination.13 And if we were to imagine for a second that we had no way of discovering what was inside the Earth other than measuring the gravitational field exterior to it, then the problem of gravimetry would constitute an example of extra-strength underdetermination.
xi. Here is a simple recipe for constructing examples of extra-strength underdetermination: find a physical system such that (i) our only way finding out about its internal structure is by making measurements exterior to it and (ii) the relevant inverse problem is unsolvable. This recipe is not altogether easy to follow. It is easy enough to find physical systems with unsolvable inverse problems, harder to convince yourself that the evidence in principle available about them is restricted to measurements made exterior to them. There are, after all, ways of finding out what is inside a planet (or a head) other than making measurements exterior to it. We want examples of systems for which it is in principle impossible to directly inspect their interiors.
xii. Stars are natural candidates to be such systems, since it is, presumably, impossible even in principle to examine their internal structure directly—our evidence is of necessity restricted to what we can measure exterior to them. But there is a hitch: not just any star will do. The Sun is hot enough to be opaque to photons—so we cannot investigate its interior using giant versions of our medical imaging machines. But ordinary stars like the Sun are transparent to neutrinos—and so in principle we could investigate them via neutrino tomography.14 The relevant inverse problem is unsolvable in general but solvable in some special cases.15 It is, I believe, an open question whether this map is invertible when restricted to the stellar regime. If it is not, then such stars are good candidates to be obstructions to the alethotropic objectivity of science. But in any case, there exist objects—such as very newly formed neutron stars—so hot that they are opaque even to neutrinos.16 It is not implausible that the inverse problems relevant to investigating such extreme objects are unsolvable.17
xiii. Another sort of candidate is the global structure of our spacetime. Our universe (i) started out very hot and dense, (ii) will undergo permanent exponential expansion towards the future, (iii) looks approximately the same in every spatial direction at large scales, and (iv) appears to be very large.
There are many general relativistic worlds with these features. These differ from one another about various important questions, including the topology of space. But in every general relativistic universe satisfying the conditions above, there is an early time t0 prior to which the universe is so hot as to be opaque to radiation. Let us agree that observers can exist in such universes only to the future of t0 and that ideal observers in such universes exist eternally to the future of t0. Relative to any ideal observer I in such a universe, we can divide the post-t0 universe into two regions: the region observable by I (consisting of events that can send signals to I’s worldline) and the region unobservable to I (consisting of events that cannot send signals to I’s worldline).
Our observer I can make measurements only within —which is to say that I can investigate only by making observations in the region external to . Under reasonable assumptions we find that so long as I is able to make measurements of only finite precision, knowledge of the physical state within fails to determine the topology of the universe (even though in imposing (i)–(iii) above we have imposed very strong global constraints).18
So the inverse problem of determining cosmic topology (which we can think of as being information about ) from measurements made within the observable universe (i.e., exterior to ) is unsolvable. And here it is very natural to maintain that information about the unobservable portion of the universe is in principle inaccessible to direct measurement. So in universes like our own, our inability to determine the topology of space is an obstruction to the alethotropic objectivity of science.19
xiv. It would appear, then, that for some objects in our world, even the in-principle evidence must come from the region exterior to them—and in some cases the relevant inverse problems are unsolvable. So we have extra-strength underdetermination and hence a limit on the alethotropic objectivity of science: some scientific questions are intractable, and convergence to the truth cannot be guaranteed. The alethotropic objectivity of science fails! But for all that has been said so far, it fails only because it frays a little bit around the edges.
xv. It is natural to worry that we have gone astray. After all, people do build reliable medical imagining devices and do make reliable geophysical inferences from gravimetric data. So there must be practical ways around the fact that a finite data set hardly ever logically determines a unique model.20
xvi. Already in the eighteenth century, precision pendulum measurements of gravity were used to constrain models of the shape and internal structure of the Earth—and these methods were elaborated greatly in the nineteenth century as theory and instruments improved and the project of making widely distributed precision measurements was taken up by an international collaboration.21 In the present century, the pair of satellites that made up the Gravity Recovery and Climate Experiment (GRACE) made it possible, for fifteen years, to make continual measurements of the Earth’s gravitational field, accurate enough to allow month-by-month determination of changes in glaciers and much else besides.22
How is all of this possible? Roughly speaking: each month, data from GRACE gave us, more or less directly, values for the strength of the gravitational field at a finite number of locations 500 km above the surface of the Earth. This information, supplemented by auxiliary hypotheses, allowed the fitting of a smoothed-out model of the Earth’s gravitational field for that month.
The default assumption is that month-by-month changes in the Earth’s gravitational field are driven by processes localized near the surface of the Earth. If we idealize such processes as occurring exactly at the Earth’s surface, we can translate month-by-month changes in the models of the gravitational field into models of the month-by-month redistribution of mass at the Earth’s surface.23
Given that we have excellent reason to think that large-scale changes of mass distribution at the surface of the Earth are primarily due to changes in the distribution of frozen and liquid water, this procedure allows us to give reasonable estimates of the rates at which glaciers are melting, sea levels are rising, aquifers are being depleted, and water is being lost to drought. In effect, we impose very strong auxiliary constraints in order to select among the vast family of models consistent with our data.
Sometimes this procedure leads to anomalous results. Over the southern portion of Hudson Bay, for instance, only a small fraction of the rate of increase of the gravitational field can be traced to increased water storage. There is, however, strong independent evidence that in this region the Earth’s crust is still undergoing rebound from the withdrawal of the ice sheet at the end of the most recent glacial period. There are a number of competing models of this process—differing, e.g., as to the viscosity in various parts of the Earth’s mantle. These lead to distinct predictions for the dynamics of the exterior gravitational field, so that GRACE data can be used to put constraints on these models.24
xvii. We see something similar in the case of medical imaging. Notwithstanding the Wittgensteinian result mentioned above (see fn. 11), actual medical imaging machines get by with making only finitely many measurements—and, indeed, reliably identify pneumocephaly. How do they do it? In effect, by imposing assumptions that rule out large variations of mass density on short length-scales within human bodies.25
xviii. In practice, underdetermination of models by data is resolved by appeal to independently well-supported empirical results from other parts of science. In medical imaging, these will include facts about fine structure of the tissues that compose the human body. In geophysics, these will include, e.g., facts about the behaviour of materials and facts concerning the structure and dynamics of the Earth.26
In the best case, these assumptions will rule out all but one of the models consistent with the data. This would appear to leave room for hope that even extra-strength underdetermination can be tamed. Some scientific problems, considered in isolation, may suffer from extra-strength underdetermination. But what matters is how such problems look when we take into account the full range of our empirical knowledge. It seems at least possible that when we do so, all underde-termination is resolved in the infinite-data limit.
Or rather, this seems possible for some putative examples of extra-strength underdetermination—perhaps the problem of determining the structure of a proto-neutron star falls in this category. But for a problem like the determination of the topology of space, it is hard to see how appeal to far-flung empirical results could help at all: we are already modelling observers as having access to all evidence available in the part of their past lightcones lying to the future of the time at which the cosmos becomes transparent to radiation.
xix. Now I would like to raise a more global worry.27 What is science except the attempt to determine (aspects of) the unobserved structure of a single system (the world) from the class of all measurements that will ever be made? Presumably there is a map that takes as input possible global ways that the world could be and gives as output the set of observations that would be made were the world that way. Science is the inverse problem: from the set of all observations that will be made, to determine the relevant unobserved structure of our world. Many ordinary inverse problems are unsolvable: the salient hidden aspects of the target system cannot be reconstructed from the accessible evidence, even in the infinite-data limit. Surely it is natural, then, to suspect that this mother of all inverse problems is likewise unsolvable—and hence that the threat posed by underdetermination to the alethotropic objectivity of science is a deep and global one. At any rate, the burden of proof would appear to rest with advocates of alethotropy: here, as in ordinary cases, it appears reasonable to presume that the more complex the direct problem, the less plausible it is that the inverse problem should be solvable. And it should be clear that there is no prospect in this global case of underdetermination being resolved via appeal to independently well-supported empirical findings—there are no empirical questions beyond the global one we are asked to solve.
xx. Before pressing on to discuss realism, it will be profitable to discuss a couple of possible responses that aim to save the spirit, if not the letter, of the thesis of the alethotropic objectivity of science.
xxi. The first response takes as its point of departure the thesis that we should believe claims produced by methods guaranteed to eventually lead us to the truth and that we should not believe claims generated by methods that lack this guarantee. Without any pretence to complete historical accuracy, I will call this Reichenbach’s thesis.28 Adopting Reich-enbach’s thesis would serve to protect the alethotropic objectivity of science from the threat of extra-strength underdetermination—the idea being that, properly understood, science simply does not pronounce on questions involving extra-strength underdetermination.
xxii. Sadly, this response is untenable. The problem is that Reichenbach’s thesis implies that scientific warrant fails to be closed under known logical implication in an unacceptably strong way. The problem can be seen as follows.29 Suppose that Nature is revealing an infinite binary sequence, one bit at a time. Suppose also that these bits are generated by tossing a coin with some fixed bias in favour of heads, given by a number between zero and one. Before each new bit is revealed, we are required to conjecture answers to two questions:
Suppose that I adopt the straight rule as my means of generating guesses about the bias of the coin: if the coin comes up heads k times in the first n tosses, I conjecture that the coin has bias k/n. Now, the law of large numbers tells us that the straight rule guarantees success for each possible bias that the coin might have. So Reichenbach’s thesis advises me to believe its output. Whatever data I see, I will know that the conjecture I make by following this rule is a rational number. So if scientific warrant is closed under known logical implication, I should believe, no matter what data I see, that the bias of the coin is rational. But the policy of guessing the bias of the coin is rational, no matter what data I see, is not a method that is guaranteed to succeed, no matter what the bias of the coin (since it is guaranteed to fail for each irrational bias that the coin might have). So when combined with the straight rule, Reichenbach’s thesis gives me advice inconsistent with the closure of scientific warrant under known logical implication: upon seeing k heads in n tosses, believe that the bias of the coin is k/n but do not believe that the bias of the coin is a rational number.
Of course, the straight rule is just one way of handling the first question. But any method M for handling that question induces a method M′ for handling the second question: guess that the bias of the coin is rational if your current best estimate of its bias is rational; otherwise guess that it is irrational. Reichenbach’s thesis and the closure of scientific warrant under known logical implication jointly imply that if you are committed to a method M for handling the first problem that guarantees success for each possible bias of the coin, then the method M′ that M induces for handling the second problem must also guarantee success for each possible bias the coin might have. But no method for handling the second problem has this feature: for every method of handling the second problem, there are infinitely many possible values for r for which that method does not guarantee success; furthermore, if a method guarantees success for each rational number (or even for some dense set of rational numbers), then there is a dense and uncountable (indeed, co-meagre) set of irrational numbers for which it guarantees failure.30 So Reichenbach’s thesis is inconsistent with closure in the following strong sense: if you have a method that underwrites belief in your best estimate of the coin’s bias, then you must not believe any conjecture you make concerning whether the coin is rational or irrational.
Closure principles are by no means sacrosanct. Consideration of skeptical scenarios have driven some philosophers to deny closure.31 And certain closure principles are inconsistent with standards of statistical inference that are common in the sciences.32 Indeed, in the case at hand, it is far from unintuitive that there should be some failure of closure: if I am forced to estimate the bias r of the coin and to guess whether that bias is rational or irrational, knowing that there is a sense in which the second task is more difficult than the first, it seems only reasonable that in some situations I will take a more positive attitude towards my estimate of the bias r than I take towards my guess as to whether r is rational or not (even while knowing whether is rational). Reichenbach’s thesis demands something much stronger: that no matter what data you see, you must not believe your conjecture as to whether the bias of the coin is rational. This is a very strange restriction. For many people, at least, there are data sets that they could see that would render them practically certain that the coin being tossed is fair.33 Upon seeing such data, it would be very strange if we were not willing to affirm the weaker proposition that the bias r of the coin was a rational number. It would be analogous to Newton saying: I have discovered that gravity varies inversely as the square of distance—but don’t ask me whether it varies inversely as some polynomial or other.
xxiii. The response we have just considered involved restraining the ambitions of science in order to protect the alethotropic objectivity of science from the threat posed by extra-strength underdetermination. A second response proceeds rather by restraining the ambitions of objectivity, maintaining that, properly understood, scientific objectivity requires not convergence to the truth, but mere convergence of opinion of scientists. As Hempel (1983, 75) characterizes a common view, the practice and products of science are objective to the extent that they are “independent of idiosyncratic beliefs and attitudes on the part of the scientific investigators.” Consider, then, the thesis that either there is just one scientific method of inference for handling any given problem, or there are multiple such methods, but any disagreement between them evaporates in the infinite-data limit. I will call this the thesis that science is symphonotropically objective (i.e., objective in the sense that it has an inherent tendency to turn towards harmony). Extra-strength underdetermination, as such, presents no threat to the symphonotropic objectivity of science.
xxiv. Sadly, this second response doesn’t work either. It requires that for any question, there be some learning method M with which all scientifically permitted methods agree in the infinite-data limit. Here the stumbling block is that contexts exist in which there is no optimal method of learning.34 Suppose that Nature is revealing an infinite binary sequence to an agent one bit at a time and that immediately before each bit is revealed, the agent is required to guess whether it will be a zero or a one. An extrapolator is a map that takes as input a finite binary string (the bits seen so far) and gives as output a single bit (the guess as to the identity of the next bit to be revealed). Let us say that extrapolator M learns binary sequence σ in the infinite long run, when shown σ, M correctly predicts the next bit at least two-thirds of the time. This learning problem is intractable: each extrapolator learns uncountably many binary sequences and fails to learn uncountably many binary sequences—but there is sense in which the sequences learnable by M (forming a meagre subset of the space of binary sequences) are incomparably more rare than those unlearnable by M. Furthermore, there is no best extrapolator: for any extrapolator and any countable set of sequences, there is another extrapolator that learns each of the given sequences while also learning every sequence learned by the first extrapolator. So each extrapolator M is dominated by some extrapolator M*, which is in turn dominated by some extrapolator M**, and so on without end.35
Let us call two extrapolators asymptotically equivalent if for each binary sequence σ, if fed sufficiently long initial segments of σ, the two extrapolators always agree about what they expect the next bit to be. For the problem at hand (learning patterns in infinite binary sequences), symphonotropic objectivity requires that all scientifically permitted extrapolators be asymptotically equivalent. Suppose that M is a scientifically permitted extrapolator. Then we know that there is an extrapolator M* that learns a strictly larger set of sequences than M does. M and M* will not be asymptotically equivalent: if σ is a sequence that M* learns but M does not, then there will be arbitrarily long initial segments of σ that lead M and M* to make different predictions as to the next bit. So, according to the proposal at hand, M* cannot be scientifically permitted. At this point, one can only wonder why we should care more about the proffered variety of scientific permissibility than we care about the ability of our methods to arrive at the truth.
xxv. I think, then, that we should accept at face value examples of extra-strength underdetermination as constituting failures of the alethotropic objectivity of science. Let me turn, finally, to the argument from under-determination against scientific realism.
xxvi. How should realists react to instances of extra-strength under-determination?39 Advocates of (iii) above tell us that we should be untroubled, because these are cases in which evidence gives us more reason to believe one of the hypotheses at hand.
But I think that we should be troubled. Consider, for instance, how things look from a Bayesian perspective. Consider a Bayesian agent with prior probability distribution Pr who faces a case of extra-strength underdetermination. Divide the hypotheses under consideration into classes of empirically equivalent hypotheses E1, E2, ….40 For each k, let Ak be the hypothesis in Ek to which Pr assigns maximum probability, and let Bk be some other hypothesis in Ek to which Pr assigns positive probability. Define A to be the disjunction of the Ak’s and B to be the disjunction of the Bk’s. A and B are contingent propositions, and our agent assigns each of them non-zero prior probability. We have set things up so that Pr(A) > Pr(B) (i.e., our agent assigns greater prior probability to A than to B). But from the definitions of A and B, it follows for any data set D, Pr(A∣D) > Pr(B∣D) > 0: our agent assigns higher posterior probability to A than to B, no matter what evidence turns up. In other words, our agent, while regarding A and B as contingent propositions (either one of which might be true) also thinks that no evidence could show that B was more plausible than A.
The realists’ reply (iii) above to the argument from underdetermination has it that this is a case where evidence D gives us more reason to believe A than to believe B. But is this really a case in which the evidence gives our agent reason to prefer A to B? It seems, rather, like a built-in initial bias in favour of A over B is doing the work (and, of course, the same sort of issue arises on non-Bayesian approaches).
I claim that realists should find something unsettling in cases like this, where an a priori bias cannot be washed away by evidence. In the toy case above, agents with this prior begin life with a bias that favours hypothesis A over hypothesis B. That in itself is not disturbing: it is a truism that successful inductive learning is possible only against a background of biases that favour some hypotheses over others.41
But in the presence of extra-strength underdetermination, there is no optimal method and learning requires luck: our beliefs may depend on the evidence we see, but not in a way that tracks the truth. Where we end up is determined by our starting point rather than by the world. At this point, some may be tempted to retreat to the position that the success of science is a reason for belief only in those parts of science that will ultimately face some sort of empirical test—but as we have seen above, this commits one to accepting that in science, rational belief fails to be closed under known logical implication in a bizarrely strong way. Others may be tempted to gesture towards an epistemology on which there is a unique rationally permitted initial bias (or, perhaps, a unique family of rationally permitted biases, sharing an asymptotic behaviour). But as we saw above, this too would be costly: to adopt this kind of view requires sometimes counting as irrational methods known to be more reliable than one’s own.
xxvii. The foregoing is not an argument against scientific realism. Rather, it is a plea for an end to the complacency with which realists have tended to dismiss underdetermination arguments—a plea for a recognition of the real threat such arguments pose to a comfortable realist view of science. Here is how Peirce (1877, 11) saw our situation:
I think that most of us will admire the impulse expressed in the first quoted paragraph—even if we would not subscribe to a view quite this stern.43 But those of us who have given up on the full alethotropic objectivity of science (and also on the fantasy of an account of rationality on which there is only one rational response to any given body of evidence) have to think that Peirce is asking for too much in the second paragraph.Now, there are some people, among whom I must suppose that my reader is to be found, who, when they see that any belief of theirs is determined by any circumstance extraneous to the facts, will from that moment not merely admit in words that that belief is doubtful, but will experience a real doubt of it, so that it ceases to be a belief.
To satisfy our doubts, therefore, it is necessary that a method should be found by which our beliefs may be caused by nothing human, but by some external permanency—by something upon which our thinking has no effect.42
What stable resting place is there for those of us who wish that we could stay true to the spirit of Peirce’s first paragraph but know that the goal he sets in the second is beyond all hope?44
Anderson, E. (2004) “Uses of Value Judgments in Science: A General Argument, with Lessons from a Case Study of Feminist Research on Divorce.” Hypatia 19 1–24.
Bell, R. (1896) “Proofs of the Rising of the Land around Hudson Bay.” American Journal of Science 1 219–228.
Belot, G. (2015) “Down to Earth Underdetermination.” Philosophy and Phenomenological Research XCI 456–464.
Belot, G. (2017) “Objectivity and Bias.” Mind 126 655–695.
Belot, G. (2020) “Absolutely No Free Lunches!” Theoretical Computer Science 845 159–180.
Belot, G. (forthcoming) Lambda.
Chakravartty, A. (2017) “Scientific Realism.” In E. Zalta (ed.), The Stanford Encyclopedia of Philosophy. https://plato.stanford.edu/entries/scientific-realism/https://plato.stanford.edu/entries/scientific-realism/
Chomsky, N. (1965) Aspects of the Theory of Syntax. Cambridge MA: MIT Press.
Daston, L., and P. Galison (2007) Objectivity. Brooklyn: Zone Books.
Dretske, F. (1970) “Epistemic Operators.” The Journal of Philosophy LXVII 1007–1023.
Earman, J. (1992) Bayes or Bust? A Critical Examination of Bayesian Confirmation Theory. Cambridge MA: MIT Press.
Earman, J. (1993) “Underdetermination, Realism, and Reason.” Midwest Studies in Philosophy XVIII 19–38.
Earman, J. (2020) “Quantum Sidelights on the Material Theory of Induction.” Studies in History and Philosophy of Science 82 9–16.
Easwaran, K. (2014) “Regularity and Hyperreal Credences.” The Philosophical Review 123 1–41.
Friedl, J., and H. Rutte (eds.) (2013) Moritz Schlick: Erkenntnistheoretische Schriften 1926–1936. Wien: Springer.
Gell-Mann, M. (2007) “Truth, Beauty, and … Physics?” TED Talk. https://www.ted.com/talks/murray_gell_mann_beauty_truth_and_physicshttps://www.ted.com/talks/murray_gell_mann_beauty_truth_and_physics
Glashow, S. (1992) “The Death of Science!?” In R. Elvee (ed.), Nobel Conference XXV: The End of Science? Lanham MD: University Press of America 23–32.
Hacking, I. (1990) The Taming of Chance. Cambridge: Cambridge University Press.
Heilbron, J. L. (1986) The Dilemmas of an Upright Man: Max Planck as Spokesman for German Science. Berkeley: University of California Press.
Hempel, C. G. (1966) Philosophy of Natural Science. Englewood Cliffs NJ: Prentice–Hall.
Hempel, C. G. (1983) “Valuation and Objectivity in Science.” In R. S. Cohen and L. Laudan (eds.), Physics, Philosophy, and Psychoanalysis: Essays in Honor of Adolf Grünbaum. Dordrecht: Reidel 73–100.
Hookway, C. (2004) “Truth, Reality, and Convergence.” In C. Misak (ed.), The Cambridge Companion to Peirce. Cambridge: Cambridge University Press 127–149.
Houser, N., and C. Kloesel (eds.), The Essential Peirce: Selected Philosophical Writings, Volume 1. Bloomington: Indiana University Press.
Jeffreys, H. (1933) “Probability, Statistics, and the Theory of Errors.” Proceedings of the Royal Society of London. Series A 140 523–535.
Koplowitz, J., J. E. Steif, and O. Nerman (1995) “On Cover’s Consistent Estimator.” Scandinavian Journal of Statistics 22 395–397.
Kuhn, T. S. (1962) The Structure of Scientific Revolutions. Chicago: University of Chicago Press.
Kukla, A. (1996) “Does Every Theory Have Empirically Equivalent Rivals?” Erkenntnis 44 137–166.
Kurylev, Y., M. Lassas, and G. Uhlmann (2018) “Inverse Problems for Lorentzian Manifolds and Non-Linear Hyperbolic Equations.” Inventiones Mathematicae 212 781–857.
Lattimer, J. M., and M. Prakash (2004) “The Physics of Neutron Stars.” Science 304 536–542.
Le Cam, L., and G. Yang (2000) Asymptotics in Statistics: Some Basic Concepts (second edition). New York: Springer.
Lenzen, V. F., and R. P. Multhauf (1966) “Development of Gravity Pendulums in the 19th Century.” United States National Museum Bulletin 240 301–345.
Le Roy, É. (1901) “Un positivisme nouveau.” Revue de métaphysique et de morale 9 138–153.
Leweke, S., V. Michel, and R. Telschow (2018) “On the Non-Uniqueness of Gravitational and Magnetic Field Data Inversion.” In W. Freeden and M. Zuhair (eds.), Handbook of Mathematical Geodesy: Functional Analytic and Potential Theoretic Methods. Cham: Birkhäuser 883–919.
Lloyd, E. A. (1995) “Objectivity and the Double Standard for Feminist Epistemologies.” Synthese 104 351–381.
Louis, A. K., and F. Natterer (1983) “Mathematical Problems of Computerized Tomography.” Proceedings of the IEEE 71 379–389.
Mach, E. (1897) The Analysis of Sensations. Chicago: Open Court. Originally published as Beiträge zur Analyse der Empfinungen. Jena: G. Fischer (1886).
Manchak, J. B. (2009) “Can We Know the Global Structure of Spacetime?” Studies in History and Philosophy of Modern Physics 40 53–56.
Mayo-Wilson, C. (2018) “Epistemic Closure in Science.” Philosophical Review 127 73–114.
Natterer, F. (2001) The Mathematics of Computerized Tomography. Philadelphia: Society for Industrial and Applied Mathematics.
Newton, I. (1729) The Mathematical Principles of Natural Philosophy. London: Benjamin Motte.
Nozick, R. (1981) Philosophical Explanations. Cambridge MA: Harvard University Press.
Paulson, A., S. Zhong, and J. Wahr (2007) “Inference of Mantle Viscosity from GRACE and Relative Sea Level Data.” Geophysics Journal International 171 497–508.
Peirce, C. S. (1877) “The Fixation of Belief.” Popular Science Monthly XII 1–15.
Peirce, C. S. (1878) “How to Make Our Ideas Clear.” Popular Science Monthly XII 286–302.
Planck, M. (1915) Eight Lectures on Theoretical Physics. New York: Columbia University Press. Originally published as Acht vorlesungen über theoretische physik, gehalten an der Columbia university in the city of New York im frühjahr 1909. Leipzig: S. Hirzel (1910).
Planck, M. (1949) “The Meaning and Limits of Exact Science.” In Scientific Autobiography and Other Papers. New York: Philosophical Library 80–120. Originally published as Sinn und Grenzen der exacten Wissenschaft (second edition). Leipzig: Johann Ambrosius Barth (1947).
Planck, M. (1970) “The Unity of the Physical World-Picture.” In S. Toulmin (ed.), Physical Reality: Philosophical Essays on Twentieth-Century Physics. New York: Harper & Row 1–27. Originally published as “Die Einheit des physikalischen Weltbildes.” Phyikalische Zeitschrift 10 (1909) 62–75.
Poincaré, H. (1907) The Value of Science. New York: Science Press. Originally published as La valeur de la science. Paris: E. Flammarion (1905).
Psillos, S. (1999) Scientific Realism: How Science Tracks Truth. London: Routledge.
Putnam, H. (1963) Probability and Confirmation. Washington: U.S. Information Agency.
Quine, W. V. O. (1957) “The Scope and Language of Science.” The British Journal for the Philosophy of Science 8 1–17.
Railton, P. (2000) “Scientific Objectivity and the Aims of Belief.” In P. Engel (ed.), Believing and Accepting. Dordrecht: Kluwer 179–208. This is a revised version of “Truth, Reason, and the Regulation of Belief.” Philosophical Issues 5 (1994) 71–93.
Reichenbach, H. (1938) Experience and Prediction. Chicago: University of Chicago Press.
Reiss, J., and J. Sprenger (2020) “Scientific Objectivity.” In E. Zalta (ed.), The Stanford Encyclopedia of Philosophy. https://plato.stanford.edu/entries/scientific-objectivity/https://plato.stanford.edu/entries/scientific-objectivity/
Ringström, H. (2013) On the Topology and Future Stability of the Universe. Oxford: Oxford University Press.
Ringström, H. (2014) “On the Future Stability of Cosmological Solutions to Einstein’s Equations with Accelerated Expansion.” In S.Y. Jang, Y. R. Kim, D.-W. Lee, and I. Yie (eds.), Proceedings of the International Congress of Mathematicians: Seoul 2014, Volume II. Seoul: Kyung Moon SA 983–999.
Scales, J. A., and R. Snieder (2000) “The Anatomy of Inverse Problems.” Geophysics 65 1708–1710.
Schervish, M. J., T. Seidenfeld, R. B. Stern, and J. B. Kadane (2020) “What Finite Additivity Can Add to Decision Theory.” Statistical Methods & Applications 29 237–263.
Schlick, M. (1979) “The Universe and the Human Mind.” In H. Mulder and B. van de Velde-Schlick (eds.), Moritz Schlick: Philosophical Papers, Volume 2. Dordrecht: Reidel 499–513. Translation of “Weltall und Menschengeist” (a manuscript of 1936).
Schoenfield, M. (2019) “Permissivism and the Value of Rationality.” Philosophy and Phenomenological Research XCIX 286–297.
Schoenfield, M. (forthcoming) “Meditations on Beliefs Formed Arbitrarily.” To appear in Oxford Studies in Epistemology.
Sellars, W. (1956) “Empiricism and the Philosophy of Mind.” In H. Feigl and M. Scriven (eds.), The Foundations of Science and the Concepts of Psychology and Psychoanalysis. Minneapolis: University of Minnesota Press 253–329.
Smith, K. T., D. C. Solmon, and S. L. Wagner (1977) “Practical and Mathematical Aspects of the Problem of Reconstructing Objects from Radiographs.” Bulletin of the American Mathematical Society 83 1227–1270.
Snieder, R., and J. Trampert (1999) “Inverse Problems in Geophysics.” In A. Wirgin (ed.), Wavefield Inversion. Vienna: Springer-Verlag 119–190.
Stokes, G. G., (1867) “On the Internal Distribution of Matter which Shall Produce a Given Potential at the Surface of a Gravitating Mass.” Proceedings of the Royal Society of London 15 482–486.
Tapley, B. D., M. M. Watkins, F. Flechtner, C. Reigber, S. Bettadpur, M. Rodell, I. Sasgen, J. S. Famiglietti, F. W. Landerer, D. P. Chambers, J. T. Reager, A. S. Gardner, H. Save, E. R. Ivins, S. C. Swenson, C. Boening, C. Dahle, D. N. Wiese, H. Dobslaw, M. E. Tamisiea, and I. Velicogna (2019) “Contributions of GRACE to Understanding Climate Change.” Nature Climate Change 9 358–369.
Uhlmann, G., and H. Zhou (2016) “Journey to the Center of the Earth.” arXiv:1604.00630
van Fraassen, B. (1980) The Scientific Image. Oxford: Oxford University Press.
van Fraassen, B. (2007) “From a View of Science to a New Empiricism.” In B. Monton (ed.), Images of Empiricism: Essays on Science and Stances, with a Reply from Bas C. van Fraassen. Oxford: Oxford University Press 337–384.
Völkel, S. H., and K. D. Kokkotas (2019) “On the Inverse Spectrum Problem of Neutron Stars.” Classical and Quantum Gravity 36 115002.
Wahr, J., M. Molenaar, and F. Bryan (1998) “Time Variability of the Earth’s Gravity Field: Hydrological and Oceanic Effects and Their Possible Detection Using GRACE.” Journal of Geophysical Research 103 30,205–30,229.
Weinberg, S. (1996) “Sokal’s Hoax.” The New York Review of Books August 8 11–15.
Weinberg, S. (1999) “Illustrative Prediction.” Science 285 1013.
Winter, W. (2006) “Neutrino Tomography—Learning about the Earth’s Interior Using the Propagation of Neutrinos.” Earth, Moon, and Planets 99 285–307.
Wittgenstein, L. (1969). On Certainty. Oxford: Basil Blackwell.