1 Scientific Realism

advertisement
12.02.2016
1
The Pessimistic Meta-Induction
and the Exponential Growth of Science
Ludwig Fahrbach
PLEASE DO NOT QUOTE
Abschnitte und Absätze, die fürs Seminar nicht wichtig sind, sind mit zwei Sternen ** gekennzeichnet.
I aim to defend scientific realism against the argument of pessimistic meta-induction. I start
form a preliminary definition of scientific realism according to which empirical success of a scientific
theory licenses an inference to its approximate truth. Pessimistic meta-induction, then, is the argument
that this inference is undermined by numerous counterexamples, i.e., by theories from the history of
science that were successful, but false. To counter pessimistic meta-induction I argue that the realist
should refine his position by using a graded notion of empirical success. I then examine the history of
science and formulate some claims about the pattern of how the degrees of success of the best theories
have developed over the history of science.
I proceed as follows. First, I define scientific realism, and present the No-miracles argument.
Second, I formulate a simple version of pessimistic meta-induction (simple PI, for short), and examine
how it undermines the realist’s position. Third, I sketch a counterstrategy for realists against this version of PI. This counterstrategy relies on the notion of degrees of success of scientific theories. It
states that in the history of science degrees of success have been continuously growing, and that our
current best theories have higher degrees of success then any of their predecessors. In response antirealists may present what I call the sophisticated PI. The sophisticated PI states that the growth of degrees success of theories has been continually accompanied by theory changes, therefore we should
extrapolate the existence of refuted theories to current degrees of success undermining the inference to
truth. In the second half of the paper, I argue that the case for the sophisticated PI has not been made.
At the centre of my counterargument will be the observation that the increase in degrees of success
over the history of science has by no means been uniform, quite the opposite, most of its increase occurred in very recent times.
1 Scientific Realism
1 Definition of scientific realism and definition of success
In this paper, I start from the definition of scientific realism as the position that the following inductive principle which I call the success-to-truth principle is correct: Empirically successful theories
are probably approximately true. Another formulation of this principle that I will use is that the empirical success of a scientific theory is a good indicator of the approximate truth of the theory. Later this
definition will be refined. Examples for successful theories that realists have in mind here are theories
such as the atomic theory of matter, the theory of evolution or claims about the role of viruses and
bacteria in infectious diseases.
The success-to-truth principle is not meant to be a full-blown account of confirmation or induction; it is only meant to capture the common core relevant for the realism debate as discussed in this
paper of all those inductive principles or accounts of theory confirmation, such as inference to the best
explanation, hypothetico-deductivism, etc., that realists typically offer to formulate their respective
variants of scientific realism.
12.02.2016
2
I adhere to the following conventions. Because it plays only a minor role in this paper, I generally omit the term “approximately” in “approximately true” and simply use “true”.1 Furthermore, I use
the term “theory” in a rather generous sense so that it also denotes laws of nature, theoretical statements, sets of theoretical statements and even classification systems such as the Periodic Table of Elements the reason being that realists usually want to endorse the truth of the statements involved in
these things as well. For example, the term “Periodic Table of Elements” is meant to refer to the
statements involved in that classification system.
We need to define the notion of empirical success. I will start here with a simple definition of
success, which will be refined in some ways over the course of the paper. The definition employs entirely standard ideas of how theories are tested by observation. Thus, a theory is empirically successful
(or simply successful) at some point in time, just in case its known observational consequences fit with
all the data gathered until that time, i.e., the theory describes correctly, as far as scientists know at that
time, all observations and experimental results gathered by scientists until that time, and there are sufficiently many such cases of fit. In other words, a theory is successful at some point in time, just in
case scientists have been able to compare sufficiently many consequences of the theory with observations until that time, and all of them have turned out to be correct.
2 Comments on the definitions
I defined scientific realism as a position which endorses a certain inductive principle, so my definition of realism is a purely epistemic one. Many definitions of realism offered in the literature involve, in addition, semantic, pragmatic and other conditions, but in this paper I focus on epistemic
topics. 2 Put in general terms, the question at issue is how far inductive inference can take us beyond
all observations that scientists have gathered so far, or, in other word, which forms of inductive inferences are reliable and which are not. Realists are typically optimistic about the reach of inductive inference, and therefore endorse the success-to-truth principle or similar principles of induction. In contrast, anti-realists typically doubt in different ways and to different degrees that inductive inference can
take us very far beyond observation, and therefore normally reject the success-to-truth principle and all
the realist principles of inference, of which it is the common core. They maintain, for example, that it
is nothing but the fallacy of affirming the consequent (e.g., Laudan 1981, Alan Musgrave 1988, 230).
There are many different forms of anti-realism, of course, but in this paper I will understand the
position of anti-realism entirely negatively as the rejection of the success-to-truth principle, because I
only want to discuss realism, and arguments for and against realism, and none of the different forms of
anti-realism. So, anti-realists only occur as people who oppose scientific realism and offer counterarguments and other challenges to realism.3
3 The No-miracles Argument
The most important argument offered by realists to support the success-to-truth principle and
similar principles is the no-miracles argument (NMA for short) (Putnam 1978, Smart 1960). The simplest version consists in pointing out the inherent plausibility of the success-to-truth principle: “Given
that a theory enjoys empirical success wouldn’t it be a miracle if it nevertheless were false? Wouldn’t
it be miracle, if infectious diseases behaved all the time, as if they are caused by viruses and bacteria,
but there are no viruses and bacteria.” This argument appeals directly to our confirmational intuitions.
For what follows, we need not engage in a detailed examination of the different versions of the NMA,
1
Realists usually admit that a general explication of the notion of approximate truth has not yet been devised and may even
be impossible to devise, but they think that our intuitive grasp of that notion and scientists’ successful application of it in
many specific situations suffice to justify its use for defining realism. An obstacle for devising a general explication of the
notion of approximate truth is identified in Alexander Bird (2007).
2 For recent book length treatments of the scientific realism debate see Jarrett Leplin (1997), André Kukla (1998), Stathis
Psillos (1999), Jaako Niiniluoto (1999), and Kyle Stanford (2006).
3 For a comparison of different forms of anti-realism with regard to the pessimistic meta-induction see Fahrbach (2009a,
2009b).
12.02.2016
3
all we need to note is that, in the end, they are all based on confirmational intuitions, which are by and
large shared by most realists. Let us call those intuitions the “shared realist intuitions”.4
As anti-realists reject the success-to-truth principle and all the confirmational principles of
which it is the common core they also reject whatever arguments realists offer in support of these
principles, that is they reject all versions of the NMA. For example, they reject inference to the best
explanation. Thus, realists and anti-realists are engaged in an ongoing dispute over the success-to-truth
principle, the principles of which it is the common core, and the possibility of their justification, in
which no side can convince the other side, and which goes on without any signs of a resolution. Their
disagreement can be traced back to a clash of intuitions, where the “shared realist intuitions” on which
the different versions of the NMA are ultimately based are rejected by anti-realists, who simply do not
share them.
2 The pessimistic meta-induction
1 The simple PI
But now, the anti-realist thinks she can offer an independent argument – independent of her rejection of the NMA, inference to the best explanation, the realist’s confirmational intuitions, etc. –
which undermines the success-to-truth principle. This argument is the simple PI. The simple PI moves
from the premise that the history of science is full of theories that were once successful and accepted
by scientists as true, but later refuted and abandoned. Let’s assume for the time being that this premise
is correct. Then these successful, but false theories constitute counterinstances to the inference from
success to truth. In other words, the success-to-truth principle has had a really bad track-record, which
counts strongly against its being valid. This is the simple PI.5
The premise of the simple PI about the widespread occurrence of successful but false theories in
the history of science requires evidence. Thus, antirealists present long lists of examples of such theories. Larry Laudan (1981) famously presents the following list of theories, all of which were once successful, and all of which are now considered to have been refuted:
•
•
•
•
•
•
•
•
•
•
•
•
The crystalline spheres of ancient and medieval astronomy
The humoral theory of medicine
The effluvial theory of static electricity
’Catastrophist’ geology (including Noah’s deluge)
The phlogiston theory of chemistry
The caloric theory of heat
The vibratory theory of heat
The vital force theories of physiology
Electromagnetic ether
Optical ether
The theory of circular inertia
Theories of spontaneous generation.
The anti-realist then argues that, even if judged from the perspective of the realist, i.e., starting
from the confirmational views and intuitions of the realist (his optimism about the reach of inductive
inference, the NMA, the shared realist intuitions) and disregarding the confirmational views and intuitions of the anti-realist, the success-to-truth principle has to be given up. From this perspective two
arguments concerning the success-to-truth principle have to be considered, the NMA and the simple
PI. The NMA supports the success-to-truth principle, the simple PI undermines it. The two arguments
have to be balancee against each other. The anti-realist maintains that the result of the balancing is that
the simple PI is much stronger than the NMA. Whereas the NMA is apriori and theoretical, and ulti4
The nature of the intuitions, whether they are a apriori or somehow rooted in experience, need not concern us here.
Compare, among others, Psillos (1999, ch. 5), Psillos (2000), Peter Lewis (2001), Michael Devitt (2005), Kitcher (2001),
??? Lyons and Juha Saatsis (2004).
5
12.02.2016
4
mately based on intuitions, the premise of the simple PI is based on empirical evidence from the history of science, and provides many concrete counterexamples against the inference from success to
truth.6 What better case against an inference than counterexamples can you provide? Hence, the success-to-truth principle’s support by the NMA is trumped by its negative track record, the wide-spread
incidence of successful, but false theories in the past. The anti-realist concludes that even if one endorses the realist’s confirmational views and intuitions, one has to change one’s view about the success-to-truth principle, and admit that is undermined by the past of science.
What follows for our current successful theories which are so dear to the realist’s heart? The anti-realist can offer what he might call the No-indicators argument: Empirical success is the only promising indicator of truth in empirical science. It is neutralized by the history of science. No other indicator of truth is available to justify our belief in our current successful theories. Therefore, belief in the
truth of those theories is not justified.
2 Three Counterstrategies
**The counterstrategies realists have devised to defend their position against the simple PI fall
roughly into three kinds. Consider the success-to-truth principle: If a theory is successful, then it is
true. Counterstrategies of the first kind restrict the consequent, i.e., restrict what can be inferred from
success; counterstrategies of the second kind restrict the antecedent, i.e., restrict from what truth can
be inferred; counterstrategies of the third kind are combinations of the first and second counterstrategy.
**Counterstrategies of the first kind restrict the consequent of the success-to-truth principle:
They weaken it to an inference from success to some diminished form of truth, such as reference, truth
restricted to some sub-domain of the original domain of the theory, or partial truth (truth about structure, about classification, or about other parts that contributed to the success of the theory). To defend
the inference to diminished truth, proponents of this counterstrategy aim to show that the theory
changes of the past were mostly not as deep as might have seemed at first sight, that the successful
refuted theories of the past, although strictly speaking false, were – judged from today – not entirely
false, but had terms that referred, were approximately true in restricted domains, had parts that were
true, etc. In pursuit of such strategies, realists have developed elaborate accounts of reference, partial
truth, and so on.7 If any of these accounts work, the position of the realist (the success-to-truth principle) is weakened somewhat, but we can still infer some diminished form of truth for our current successful theories. I will later (mis-)use this kind of strategy in my defence of the “shared realist intuitions”.
**The counterstrategies of the second kind restrict what truth can be inferred from by working
on the notion of empirical success. The simplest version consists in noting that several theories on
Laudan’s list enjoyed very little success. For example, Kitcher writes “Laudan’s list includes such
things as the humoral theory of medicine, catastrophist geology, and theories of spontaneous generation. In none of these examples does it seem right to hail the theory as successful…”.8 This reduces the
inductive base of the simple PI somewhat, and therefore weakens it somewhat, but it leaves many
counterexamples untouched, so this reply cannot help much. Other versions of the second counterstrategy, also meant to decrease the number of counterexamples, consist in raising the standards for
counting a theory as successful. The most prominent version of this strategy relies on the demand that
a theory only count as successful, if it has made novel predictions. Unfortunately there remain important cases of refuted theories which produced true novel predictions before being refuted, such as
theories of light and the caloric theory of heat (see Lyons). What is more, the value of novel predictions is quite contested (Hempel, Hull). Therefore, this attempt also succeeds only partially, at best.
6
See also Ladyman/Ross 200??, p. 83
There is a lot of literature on these approaches. See, for example, McMullin (1984, p. 18), Kitcher (1993), Leplin (1997),
Psillos (1999), John Worall (19??), Martin Carrier (2004), Gerhard Schurz (2004, 2009), Nola, Theo Kuipers, Ladyman/Ross (2007??), and so on.
8 Kitcher (2001, footnote 31); likewise McMullin (1984??), McAllister (1993), Kukla (1998), Psillos (1999, Ch. 5), Devitt
(2005, 772).
7
12.02.2016
5
**Finally, counterstrategies of the third kind combine the first and the second counterstrategy.
For example, they refine the success-to-truth principle so that it asserts that an increase in success
leads to an increase in truth-likeness. This is the popular idea that increase in success leads to convergence to truth.
In this paper, I want to develop a version of the second counterstrategy. It relies on grading the
notion of success. Let us begin with the following simple, standard account of theory testing. In order
to test a theory, scientists derive observable consequences, i.e., predictions, from the theory and make
observations. A particular test of a theory consists in comparing some prediction of the theory with
some observation. I use the notion of a test in a rather broad sense here to cover all cases in which
scientists are aware of the theory coming into contact with experience so that it is possible for the theory to fail.9 Also I use the term “prediction of a theory” in a broad sense to denote any observable consequence of the theory scientists are aware of.10 The reason to do so will become apparent later. If
prediction and observation agree, the theory enjoys some measure of success. As a theory passes more
and more tests its degree of success increases. Hence, different theories can differ with respect to the
degree of success they enjoy at some point in time, and the same theory can enjoy different levels of
success at different points in time. As we will see later, these differences can be quite large.
If prediction and observation don’t agree, the theory suffers from an anomaly. As long as the
anomaly is not significant or the anomalies do not accumulate, they can be tolerated. If the anomaly is
significant or the anomalies do accumulate, the theory counts as refuted. In that case scientists have to
look for alternative theories, and a theory change may take place. Of course, this account of theory
testing and empirical success is rather minimal in several respects and could be made more precise in
many ways, but it is all we will need at the moment. Later on, I will refine it further in several respects.
3 The modified success-to-truth inference
Given the grading of success, the assertion is very plausible that the degrees of success of the
theories accepted by scientists in the history of science have by and large grown steadily over time.
They have grown both during theory changes and between theory changes. During theory changes,
successor theories have usually incorporated the successes of their predecessors, and were moreover
successful where the predecessors failed, for the simple reason that they have typically been designed
to take over the successes of their predecessors and to avoid the failures of their predecessors. As for
the times in between theory changes, theories have usually become more successful while they were
accepted (before possibly significant anomalies showed up or anomalies began to pile up), because the
amount and diversity of data, the precision of measurements, the precision of the predictions, etc.,
have been growing all the time in the history of science.
Realists can then offer the following idea of a counterargument to the simple PI: Because degrees of success have increased during the history of science our current best, i.e., most successful,
theories enjoy higher degrees of success than past theories; therefore, we are warranted to infer their
probable truth; in contrast theories of the past were not sufficiently successful to warrant an inference
to their probable truth (as the theory changes of the past show). Hence, according to this idea realists
modify the success-to-truth principle so that it allows the inference to truth for current levels of success only. This idea to counter the simple PI belongs to the counterstrategies of the second kind in that
it does something with the notion of success. To support the modified success-to-truth principle realists invoke, as before, the NMA (in whatever version), although modified so that it only supports the
revised success-to-truth principle. Remember that we are only concerned with defending and restoring
Kitcher’s (2001) notion of success includes practical success in our dealings with the world to reach our practical goals. I
don’t include practical success in my definition, because to every case of practical success which somehow depends on a
theory corresponds a true consequence of the theory, so that my notion of success captures all practical successes of theories
as well. The practical success of science and technology, especially over the last 100-200 years, is nevertheless a good indicator of success in my sense, and could be used alongside the other indicators of success I discuss below.
10 One kind of observational consequence has to be excluded, though, namely irrelevant consequences of a theory which, if
true, offer no support for the theory at all. For example, T implies TO, where O is any true observation sentence, but T is
usually not confirmed by TO. For further discussion of this problem and an account of irrelevant consequences see Schurz
(1991).
9
12.02.2016
6
the realist position, hence we assume that realists are allowed to call upon their confirmational views
and intuitions, and disregard the intuitions of anti-realists. The “shared realist intuitions”, then, feed
into the modified version of the NMA so that it supports the modified position that only current levels
of success indicate truth. If this counterargument works, the realist position is not incompatible with
the history of science, and is saved from the attack by the simple PI.11 As we will see this idea is on
the right track. But it is seriously underdeveloped and certainly looks rather ad hoc so far. It is open to
objections from the anti-realist. Let us look at the objections.
4 Three objections**
The anti-realist can raise three objections. The first objection reiterates the argument from earlier
that the realist applied the NMA to past successful theories some of which were refuted showing that
the „shared realist intuitions” on which it is based are not trustworthy. Hence, the realist can no longer
use them to support anything at all, specifically not the inference to truth for current levels of success.
The realist may respond that a realist in 1800 or 1900, for example, should not have made the inference to truth and should not have supported it by the NMA, because, if he had really possessed shared
realist intuitions and had carefully attended to them, he would have recognized that those levels of
success were actually not high enough to indicate truth; if we listen carefully to shared realist intuitions we recognize that they tell us that only current levels of success suffice to indicate truth.
This response of my realist is certainly not satisfactory. We may wonder whether instead of past
realists not carefully attending to shared realist intuitions, it is my realist who is not carefully attending
to them and only hears from them what he wants to hear, namely that they differentially support the
success-to-truth inference for present, but not for past levels of success. In any case the realist intended
the shared realist intuitions to accord with the intuitions of the working scientists. But the scientists
themselves accepted the theories of the past that were later refuted. Therefore, the shared realist intuitions are such that they did back the inference to truth for past levels of success. The realist cannot
simply change his intuitions as it suits him. Thus, he is still faced with a conflict between the conclusion of the NMA, applied to the theories of the past, and the historical track record, which still throws
the trustworthiness of the shared realist intuitions into doubt.
For the second objection the anti-realist grants that the first objection can be met somehow, that
the shared realist intuitions are not discredited by their application via the NMA to past refuted theories. The second objection starts with the observation that in his defence the realist uses a simple distinction between past and present science. This distinction is based on a distorted view of the relative
weights of past, present and future times. Science has been a human endeavor for at least some centuries now, and will probably go on for some more centuries (or so we can assume). The present is only
a small part of this whole time period. The objection then is that the realist’s preference for the present
looks entirely arbitrary and ad hoc. Why should we believe that it is precisely now, instead of at some
future time, some decades or centuries from now, that a level of success sufficient for truth is reached?
The realist invokes the NMA, but that only pushes the problem back one step: If it is granted that the
NMA does not apply to past levels of success, why does it support the inference from success to truth
for current theories, and not just for theories of some future time? What is special about the present
levels of success? The anti-realist submits that my realist’s whole defence looks as if he adjusted his
position a bit too flexibly to the history of science.
5 The sophisticated PI
The third objection of the anti-realist is as follows. The anti-realist objects that a realist in 1900,
for example, could have reasoned in exactly the same way as my realist just did: “Success of scientific
theories has generally been growing until now. All our current best theories enjoy higher degrees of
success today than any of the refuted theories of the past. Consequently we can infer the truth of our
current best theories from their success, (where that inference is supported by the NMA), whereas their
predecessors’ lower degrees of success in the past didn’t suffice for an inference to their truth.” This
reasoning of a realist of 1900 would have been futile, says the anti-realist, because many of the best
11
Leplin (1997, p. 141), Kyle Stanford (BJPS??, Buch??), Psillos and Gerard Doppelt (2007), among others, mention or
discusses variants of this argument.
12.02.2016
7
theories of 1900 were refuted after 1900. In general, at any point in the history of science, whether
1700 or 1800 or 1900, realists could have reasoned in exactly the same way as the realist just did
(namely that the levels of success of their respective successful theories had risen and now sufficed to
indicate truth), but all these realists would have been refuted by later theory changes. The anti-realist
asks: What is different about current levels of success that they – all at once – suffice to indicate truth,
whereas past levels of success did not suffice to indicate truth? This rhetorical question means, of
course, that nothing is different today, and that we should expect that our current best theories will
suffer the same fate that befell the theories of the past.
The realist will immediately reply that there does exist a relevant difference between past and
present, namely precisely the degrees of success of the respective best theories; they have increased
between the past and the present. So, the epistemic situation of a realist in 1900 and the epistemic situation of a contemporary realist are actually not similar. To this the anti-realist replies that the very
same dissimilarity, namely a difference in degrees of success, also existed for a realist in 1900, namely
with respect to a realist in 1800, say, but after 1900 refutations did occur, showing that the difference
did not save the realist of 1900 from being refuted. So, what is going on here? I take it that the third
objection is meant to be the following piece of reasoning: In the history of science, degrees of success
have been growing continuously for several centuries, and all the time right to the present while they
have been growing, theories enjoying those degrees of success kept being refuted. We should extrapolate the incidence of false theories from past levels of success to current levels of success.12 Such an
extrapolation along degrees of success does justice to what is similar and what is not similar between
the epistemic situations of the different times. It supports that many of our current best theories will
also be refuted at some future time. It is another version of the PI. I will call it “the sophisticated PI”.
The sophisticated PI is an argument in support of the claim that there exist many counterexamples to
the inference from success to truth for current levels of success, and thereby undermines that inference. If it is correct the defence of the realist presented above does not work.13
**Like the second objection the third objection is independent from the first objection; hence the
third objection may work, even if the first does not. (I will later provide an independent reply to the
first objection.) Thus, for the third objection it can be granted that the NMA still has force. This means
that we have to balance two arguments once again, in this case the sophisticated PI with the NMA.
The result of the balancing is less clear than in the first case of balancing, the simple PI with the NMA,
but I will simply assume that the sophisticated PI trumps the NMA also in this case. An additional
reason to do so is that the sophisticated PI can be strengthened somewhat. It is not plausible that the
increase in levels of success has been uniform for all successful theories across all scientific fields.
Instead, it is quite clear that among our current best theories those of some fields have reached higher
levels of success earlier than those of other fields, and have differing levels of success today. In general, the distribution of levels of success across scientific fields at all points in time is far from uniform, even if we only consider the best theories of some time. Therefore, the anti-realist could claim
that it is plausible that some of the refuted theories already enjoyed degrees of success that are typical
of our current best theories; these cases can be used to undermine the success-to-truth inference for
current levels of success directly without invoking an extrapolation. This strengthens the third objection. Anyway, I will ignore these complications (balancing and non-uniformity) from now on.
It is clear that the arguments and objections presented so far assume quite some degree of simplification of the history of science. However this makes the arguments and objections clearer, and it
can, as we will now see, be tolerated. I will now tackle the third objection. This will be the main part
of the paper. It will also provide a response to the second objection. Afterwards I will tackle the first
objection.
12
See also Stanford (2006, Ch. 1.2), Gerald Doppelt (2007), Bird (2007, Section 4.1).
A similar extrapolation threatens to refute a counterargument by Devitt against the PI. Devitt wants to argue against the PI
by invoking the improvement of scientific methods (see Devitt 1991, p. 163, and Psillos 1999, p. 104). His suggestion can be
challenged by an argument similar to the sophisticated PI: In the last few centuries, scientific methods have improved all the
time, and while they have improved, the theories accepted by scientists kept being refuted, therefore, we should extrapolate
this pattern of failures to current levels of quality of scientific methods.
13
12.02.2016
8
3 The Growth of Success
1 Problems for the sophisticated PI
How good an argument against the success-to-truth inference is the sophisticated PI? It has the
premise that “all the time right to the present theories have been refuted”. So far this premise is just a
claim. What is its support from the history of science? The anti-realist will probably assert that it is
supported by the same data from the history of science that supported the simple PI, namely the refuted theories on Laudan’s list and other such examples offered in the philosophical literature. But is that
really so? Let has have another look at those examples. If we do so we notice that all the theories on
Laudan’s list are actually rather old, namely more than 100 years old. The same holds for practically
all examples of theory changes offered in the philosophical literature.14 Kyle Stanford (2006) extensively discusses three further examples, but they are older than 100 years as well. So far anti-realists
have not shown that theory changes occurred “right to the present”, i.e., they have not offered sufficient support for the sophisticated PI so far.
Anti-realists could react by trying to show that despite the difference in time between the refuted
theories and our current best theories, the difference in degrees of success is not that big, so it is still
true that we should extrapolate the occurrence of false theories to current levels of success. At this
point I could stop. The burden of proof is clearly on the side of anti-realists to show that this reply can
be made to work. As long as he has not shown this there is no attack on realism.15 But I won’t stop
here. Instead I want to show that the prospects for anti-realists to succeed with this task are not good at
all. To do so I will now aim to compare the degrees of success of the refuted theories with those of our
current best theories.
In preparation of the comparison, let me present some further examples of what I consider to be
our current best theories. Here is a list meant to be representative of our current best theories (remember that the realist endorses the approximate truth of those theories):










the Periodic Table of Elements16
the theory of evolution17
“Stars are like our sun.”
the conservation of mass-energy
the germ theory of disease
the kinetic gas theory
“All organisms on Earth consist of cells.“
E = mc2
“The oceans of the Earth have large-scale systems of rotating ocean currents.”
And so on18
The list does not include any theories which are simple generalizations from just one kind of observation such as “all pieces of copper expand when heated”. The abandonment of such low-level empirical generalizations is sometimes called “Kuhn-loss”. There seems to be wide agreement that Kuhnlosses are rare in the history of science, i.e., such low-level empirical generalizations have rarely been
abandoned. Therefore, we need not take them into account. Consequently, all theories on my list are
more than just low-level empirical generalizations: They have a certain measure of unifying power
14
As is often the case in philosophy, the same examples are discussed again and again. Here the most discussed examples by
far are the cases of phlogiston, caloric, and ether, all of which are older than 100 years. Examples that are also discussed a lot
are the cases of Newtonian mechanics, QM and GTR, but I want to exclude them from my discussion, because theories in
fundamental physics seem to me to represent a special case with respect to the success-to-truth inference, and should therefore be examined separately.
15 The anti-realist may also try to search for examples of refuted theories which enjoyed degrees of success typical of our
current best theories from the recent history of science. I will later examine this second possible reply by the anti-realist.
16 As remarked at the beginning of the paper, I use the term “theory” in a rather broad sense.
17 The theory of evolution may be taken to consist of two claims, that all organisms on earth are related by common ancestors, and that natural selection is an important force for change.
18 The list does not contains any theories from fundamental physics such as Quantum Theory and the Special Theory of Relativity, because, as just remarked, I think that theories from fundamental physics represent a special case.
12.02.2016
9
(generalize over several distinct domains), or are about unobservables, or are about facts that are otherwise not easily accessible, etc.
In order to compare the degrees of success of current best theories with the successful, but refuted theories of the past I want to employ five “indicators of success”. I call them “indicators of success”, because they are positively correlated with the degrees of success of theories. The first indicator
is the amount of scientific word done by scientists until some time. The second, third, and fourth indicator are the amount, diversity and precision of scientific data and observations gathered by scientists
until some time. The fifth indicator is the amount of computing power available to scientists at some
time. I will examine the growth of these indicators over the history of science. From their growth I will
draw an inference to the degrees of success of both the refuted theories of the past and our current best
theories. This inference will constitute the main argument of this paper. Before turning to discuss the
indicators I have to warn the reader. He will now be showered by a great deal of figures.
2 First indicator: the exponential growth of scientific work
The first indicator of success I want to examine is the amount of scientific work done by scientists in some period of time. Here, “scientific work” means such things as making observations, performing experiments, constructing and testing theories, etc. The amount of scientific work done by
scientists in some period of time can be measured with the help of various quantities. Two such quantities are especially pertinent: the number of journal articles published in the respective period of time
and the number of scientists working in the respective period of time. Because we are only interested
in very rough estimates of the amount of scientific work done in different periods of time, both quantities are plausible ways to measure over-all scientific work during those times. 19 Consider the number
of journal articles published by scientists every year. Over the last few centuries, this number has
grown in an exponential manner. The doubling rate of the number of journal articles published every
year has been 15 – 20 years over the last 300 years. Before 1700, there were, of course, even less publications per year than anytime after 1700. Roughly the same growth rates hold for the number of scientists over the last 300 years (although we can assert this with less confidence for times before the
20th century, because for those times not much good data is available). 20
1900
1950
1970
1990
2010
-∞
Figure 1. The time-line weighted in such a way that the length of any interval is proportional to
the amount of scientific work done in that interval.
Growing with a doubling rate of 15-20 years is a very strong sort of growth. It means that half
of all scientific work ever done was done in the last 15-20 years, while the other half was done in all
the time before; and three quarters of all scientific work ever done was done in the last 30-40 years,
while one quarter was done in all the time before. Figure 1 provides an idea of this sort of growth.
Scientific work is connected with success of theories in the following way. Our definition of degrees of success implies that a theory gains in degree of success, if it passes more tests,
where a test of a theory occurs whenever scientists compare a consequence of the theory with some
piece of data. Hence in the testing of theories two kinds of scientific activities are involved: gathering
data and deriving consequences from theories. Scientists need to gather data, i.e., make observations
and perform experiments. This is certainly an important kind of scientific work for which scientists
often have to invest considerable effort and time. But in addition scientists need to make calculations
19
There are also other ways of measuring scientific work such as government and industry expenditures on research, the
number of scientific journals (rather than scientific journal articles), the number of universities and the number of doctorates.
Where data is available it shows that all these ways of measuring the amount of scientific work yield essentially the same
results.
20 For more details and references see Fahrbach (2009a).
12.02.2016
10
and derivations to arrive at consequences from the theories they want to test. The latter is often not
easy to do either. In most physical sciences and increasingly in many other natural sciences, theories
are formulated using mathematical equations, for example differential equations, therefore scientists
can only arrive at predictions from the theories by solving those equations. But developing and applying such methods is often hard work and may require a great effort on the part of the scientists.
The increase in amount of scientific work done over some time is then linked with the increase
in degrees of success of theories over that time via three steps: (1) the increase in amount of scientific
work leads to an increase in both the amount and the quality of both observations and calculations. (2)
An increase in the amount and quality of both observations and calculations leads to an increase in
both the number and quality of tests of theories. (3) If the number and quality of tests of theories increases, and the tests are passed, then the degrees of success of the tested theories increases, see Figure
2.
More scientific work → more and better observations/calculations
→ more and better tests
→ more success
Figure 2.
3 First part of main argument
Using the link between scientific work and degrees of success, we can now formulate the first
part of the main argument (where the main argument is the inference from the indicators of success to
the degrees of success of the two classes of theories we want to compare). It’s only the first part, because it relies only on the first indicator of success. The rest of the main argument will then be developed over the next few sections.
The first part of the main argument proceeds as follows. As we just saw, by far most of the
growth of scientific work has occurred in the recent past, i.e., in the last 50 to 80 years. Our current
best theories have profited from this increase via the three steps: The increase in scientific work has
meant a huge increase in relevant observations and computations; this has resulted in a huge increase
in tests of those theories; that all these theories have been approximately stable in the last 50 years
(and often much longer) shows that practically all these tests were passed; finally, the high amount of
passed tests has resulted in a huge increase in degrees of success for these theories. So, our current
best theories have received a big boost in degrees of success in the recent past, an increase that is far
greater than any increase in success for any theories of earlier times. In contrast, the refuted theories
discussed by philosophers were all held prior to the big boost of success, therefore they could not partake in it. Therefore, their degrees of success were quite modest. Figure 1 can be used to represent the
big boost of success by reinterpreting the x-axis as depicting the degrees of success of the respective
best theories at different points in time. The conclusion of the main argument is the main thesis of this
paper: Our current best theories enjoy far higher degrees of success than any of the successful, but
refuted theories of the past, which enjoyed only quite modest degrees of success. The main thesis provides us with the sought comparison between current best and refuted past theories.
An objection to the first part of the main argument may be that in many scientific fields more
scientific work has obviously not lead to more successful theories. There are many domains of reality
such as the economy and human society where a lot of efforts by scientists investigating those domains has lead to quite moderately successful theories at best. The same holds for many theories in the
natural sciences such as theories about the origin of life on earth, the future weather or the fundamental constitution of everything. Generally speaking there are, quite obviously, still a lot of large gaps in
our knowledge of the world, despite the exponential growth of scientific work. Indeed, most hypotheses considered at the “research frontier” of science are further examples. So, the two quantities, scientific work and degree of success, seem to be fairly independent from each other.
In reply it has to be conceded that the relationship between scientific work and success is not as
straightforward as the three steps may suggest. Instead, it is a complex and contingent one and varies
with the part of reality scientists are dealing with. Certainly an increase in amount of scientific activity
12.02.2016
11
in a scientific field does not automatically lead to an increase in empirical success of the theories in
that field. However, it is quite implausible that the enormous increase in scientific work had no or little
effect on the degrees of success of all theories of all scientific fields. It would be very surprising, if no
theory profited substantially from all the efforts of scientists in the last few decades and centuries. And
it is, of course, the theories that did profit that realists want to focus on. And these theories are precisely our current best theories.
Still, the first part of the main argument relies solely on the connection between amount of scientific work and degrees of success, and this connection is of a rather indirect sort, as just conceded.
We need to develop the main argument further. I will do so by first looking at what I call core theories.
This will serve to further develop the first part of the main argument. Afterwards I look at the growth
of the other four indicators of success. This will serve to further develop the rest of the main argument.
4 Core theories, weak confirmation, and confirmation en passant
In order to further develop the first part of the main argument we need a couple of further notions. For one kind of theory it is especially plausible that they profited from the huge increase in scientific work, namely those of our current best theories that are highly unifying in the sense that they
are involved in much of what is going on in their respective scientific disciplines. For example, the
periodic Table of Elements plays a role in practically everything chemists do and think. Also, the theory of evolution is involved in much of what is going on in biology.21 A third example: the conversation
of mass-energy plays a role in very many situations in physics, chemistry, engineering, etc. Finally,
plate tectonics is connected with or explains much of the topography of the Earth (location of mountains, shape of seafloor, etc.), locations of earthquakes and volcanoes, magnetic stripes on the seafloor, etc. I call theories that are highly unifying in this sense “core theories” of their respective scientific discipline, and every situation in which a core theory plays a role an “application” of the core
theory.
Even though many of the applications of a core theory don’t constitute tests of the core theory, a
large number of them do. However, almost all of the latter applications are not “severe” tests of the
core theory in any way, but only tests of a weak or moderate kind. Based on the core theory scientists
have expectations about the application (about observations and outcomes of experiments), but these
expectations are typically not of a strong sort. For example, on the basis of the periodic Table of Elements chemists expect that every new substance has a chemical structure in terms of the 92 elements,
and every chemical reactions is describable in terms of the 92 elements. Likewise, based on the theory
of evolution, biologists form expectations about the features of organisms, fossils, genomes, species,
etc. Lets call such tests of a theory “gentle tests” of the theory. If the theory fulfils the respective expectations, i.e., passes the gentle test, it only receives a small or moderate increase in its success. If it
fails a gentle test, it only suffers from a non-significant anomaly, i.e., an anomaly which only means
some kind of trouble for the theory, which, in Bayesian terms, may lower its probability somewhat,
but does usually not lead to its refutation, or only leads to its refutation, if the theories already suffers
from a lot of other anomalies. Thus, as long as theory and observation come into contact in the application somehow, so that it is possible for the application to become an anomaly for the theory that
scientists can notice22, the application counts as a gentle test of the theory.
The term “prediction of a theory” will also be used in a rather broad sense. It will denote any observable statement which scientists arrive at by deriving it from the theory. This includes any expectations scientists form on the basis of the theory. In particular, this includes two things. First, predictions
need not be “novel” in the sense of Worall (1989), i.e., need not be about new kinds of phenomena,
but rather include all consequences derived from the theory even if they are about the same kind of
observations as the observable statements used in the theory‘s construction. Second, if the derivation
As Theodosius Dobzhansky famously said (exaggerating somewhat) “Nothing in biology makes sense except in the light
of evolution”.
22 Kuhn thought that scientists engaged in normal science don’t test their paradigm, and when anomalies arise usually don’t
blame the paradigm. Hence, he depicts scientists engaged in normal science as uncritical and close-minded. I think this is a
misrepresentation. Anyway according to Hoyningen-Huene, Kuhn also thought that scientists “trained in normal science …
are … extraordinarily suited to diagnosing” anomalies of theories (1993, p. 227).
21
12.02.2016
12
from the theory is only approximate, then the prediction is the result of this approximate derivation. In
that case the prediction is not strictly a logical consequence of the theory.23
When scientist consider a core theory to be well-confirmed and accept it, they usually don’t devote their time and energy to further testing of the theory. Therefore, when an application of a core
theory is a test of it, testing it is usually not the scientist’s main aim in the scientific project of which
the application is a part, and mostly not even any one of her aims for that project at all. Instead she
may pursue other aims in the project, such as explaining some phenomenon with the help of the theory, filling in details in the theory, or testing some other entirely different theories. For example, when
scientists study malaria bacteria they don’t intend to test once again that malaria is caused by bacteria,
nevertheless, a lot of studies about the (rather complicated) behaviour of the malaria bacterium are
also (usually rather weak) tests of the germ theory of disease for the case of malaria. Likewise, when
palaeontologists look for fossils nowadays, they generally don’t have the aim of testing the theory of
evolution once more, nevertheless, any fossil they find is also a (usually very weak) test of the theory
of evolution, because some of its features may turn out to contradict or undermine the theory of evolution, e.g., may suggest really big jumps in the fossil record, which would not accord well with the byand-large gradual nature of evolution. When an application of a theory is a test of the theory, but testing the theory is not among the main purposes of the scientists in applying the theory, and the test is
passed, I will call the resulting kind of confirmation of the theory “confirmation en passant”.
5 The main argument again
We can now develop the first part of the main argument further, as follows. Consider those theories among our current best theories that are core theories in their respective scientific disciplines.
Most of their applications have only been gentle tests. So, in most of them they have received only a
small or moderate increase in degree of success. However as the very strong growth of scientific work
shows, the number of applications has increased enormously. In almost all of them, the expectations
were met, and the gentle tests were passed. Some applications may not have been successful, but in
those cases there was almost never a reason to blame the core theory, otherwise it would not have been
as stable as we observe it to have been. Hence, the small or moderate successes have accumulated to a
very high overall amount of increase in degrees of success for the core theory. In this way, weaker
kinds of confirmation of core theories have contributed strongly to the increase in their degrees of
success. I will now present an example for this reasoning, and then strengthen the main argument by
examining the other four indicators of success.
The core theory of chemistry is the periodic Table of Elements. Like in the rest of science, manpower and publications in chemistry have risen exponentially. Jochen Schummer observes that “[o]nly
during the past 15 years [i.e., until 1999] we saw more chemistry publications than had been written
ever before… [T]his year chemists will publish a hundred times as many papers as in 1901, when
van’t Hoff received the first chemistry Nobel prize.“ (1999) Schummer shows that during the past 200
years the growth of scientific work in chemistry meant that the number of newly discovered or produced chemical substances has risen exponentially with a growth rate of around 13 years. “During the
whole period the total curve corresponds quite well to a stable exponential growth … with an annual
rate of 5.5% and doubling time of 12.9 years.” (1997, p. 111)24
As remarked above the Periodic Table of Elements implies constraints about, for example, the
features of any chemical substance and what can happen in any chemical reaction. Hence, every new
substance provides an occasion for a gentle test of the Periodic Table of Elements. That the Periodic
Table of Elements has been entirely stable for many decades shows that the gentle tests have always
been passed. Mostly they have only provided a weak increase of degree of success. But because the
number of such gentle tests has been huge – as witnessed by the number of new substances that have
been found or produced over the last 150 years, all of them different from each other – the over-all
23
Of course, a theory rarely implies an observational statement all by its own, but needs boundary conditions, auxiliary theories, etc., to do so. I ignore such complications here.
24 A growth rate of 12.9 years seems to be higher than that of scientific manpower or journal articles. Schummer provides a
plausible explanation for this discrepancy. He observes that the number of reported new substances per article, and also the
number of reported new substances per chemist has been growing over time, i.e., the productivity of chemists has increased
over time. (Schummer 1997b, p. 118)
12.02.2016
13
increase in degrees of success of the Periodic Table of Elements has been huge. Note that nowadays
and in the more recent past testing the periodic Table of Elements has clearly not been a purpose of
chemists on any of these applications, as chemists have accepted it for a long time now, so these applications are all cases of confirmation en passant.
4 The other four indicators of success
1 Amount of data
The second indicator of success is the amount of data gathered by scientists until some time.
Here, we observe that in many scientific disciplines the amount of data has grown at a very strong rate.
First, in some disciplines such as palaeontology or chemistry, it is often still the scientists themselves
who gather or produce the data, e.g., searching for fossils or synthesizing chemical substances. For
such kinds of data, it is often plausible that the amount of data gathered or produced by scientists has
grown very roughly proportional to the number of scientists in the field. As we saw earlier manpower
has increased with a doubling rate of 15-20 years over the last few centuries. A doubling time of 20
years means that today there are around 30 times as many scientists as in 1900, around 1.000 times as
many scientists as in 1800, and around 30.000 times as many scientists as in 1700.25 Therefore, in such
disciplines the amount of data has often risen in a similar fashion. For example, figure 3 depicts the
growth of the number of a certain type of fossil over the last two centuries, where this kind of growth
is entirely typical for the growth of the fossil record in general.26
Secondly and more importantly, in many scientific disciplines the growth of amount of data has
been far stronger than the growth of scientific man power due to better instruments and computer
technology.27 In many disciplines, data are nowadays gathered automatically.28 During the last six
years the Sloan Digital Sky Survey, the “most ambitious astronomical survey project ever undertaken
… measured precise brightnesses and positions for hundreds of millions of galaxies, stars and quasars…. it mapped in detail one-quarter of the entire sky. … The total quantity of information produced, about 15 terabytes (trillion bytes), rivals the information content of the Library of Congress.”29
By comparison, the most ambitious such project at the beginning of the 20th century, a survey of the
sky conducted at Harvard and completed in 1908, measured and cataloged the brightnesses and positions of 45 000 stars.30 The future has even more in stock: The “Large Synoptic Survey Telescope,
scheduled for completion atop Chile’s Cerro Pachon in 2015, will gather that much data [as did the
Sloan Digital Sky Survey over the last six years] in one night”.31
25
A doubling rate of 15 years means that today there are around 100 times as many scientists as in 1900, around 10.000 as
many scientists as in 1800, and around 1.000.000 as many scientists as in 1700. Here are the calculations: For a doubling rate
of 20 years we get 25  30, 210  1000, and 215  30.000; for a doubling rate of 15 years we get 13 doublings in 195 years,
hence a factor of 213  8.000, or around 10.000 in 200 years.
26 Timothy Rowe (2005), Book Review of Mammals From the Age of Dinosaurs: Origins, Evolution, and Structure by Zofia
Kielan-Jaworowska, Richard L. Cifelli and Zhe-Xi Luo. Nature 438, 426 (24 November 2005). For similar numbers about
the growth of dinosaur fossils, see New Scientist, (21 May 2005, pp. 36, 38).
27 See also the fourth indicator which concerns computing power.
28 See also Humphreys (2004, pp. 6-8)
29 http://www.sdss.org/. See also Robert Kennicutt (2007).
30 Jones and Boyd, (1971, p. 202). Cited from Johnson (2005)
31 Nature (2009), vol. 460, no. 7255, p. 551. This represents an increase of time efficiency by a factor of around 2000, namely
5 times 365. The Large Synoptic Survey Telescope is just one of several large astronomical projects planned for the next 10
years.
12.02.2016
14
Figure 3 The growth of the fossil record of mammals from the age of dinosaurs (245 – 66 million years ago). From Kielan-Jaworowska, Cifelli, Luo, 2005, S.7
Another example is provided by the sequencing of DNA. Here, over the last 20 years the overall number of decoded DNA sequences has grown with a fairly stable doubling rate of 18 months.32
This means a growth by a factor of 100 every 10 years from 1984 (around 4000 sequences) to 2004
(around 40 000 000 sequences). Again, this is not possible without automation. To illustrate, “geneticists spent more than a decade getting their first complete reading of the 3 billion base pairs of the
human genome, which they finally published in 2003. But today’s rapid sequencing machines can run
through that much DNA in a week.”33 The costs for sequencing the first human genome was probably
at least $500 million, whereas the costs for sequencing the eighth human genome in 2009 was around
$50,000. In the next few years, the costs are expected to decrease by a factor of two each year (Nicholas Wade 2009).
As a final example consider the Argo system which was installed in the years 2005 to 2009 and
which consists of a network of 3000 robotic probes floating in the Earth’s oceans. The probes continuously measure and record salinity and temperature of the upper 2000m of the ocean, surfacing once
every 10 days to transmit the collected data via satellite to some stations on land. In these and many
other fields, the automatic gathering of data has lead to truly gigantic amounts of data. 34 Relating my
examples from palaeontology, astronomy, genetics, and oceanography of increase of data with the
respective entries on my list of our current best theories supports the main thesis that those theories
enjoy far higher degrees of success than any theories some decades ago.
2 The worry of diminishing returns and the diversity of data
At this point the following worry might arise. Even if in the examples above the amount of data
has increased enormously, still every new pieces of data may be rather similar to data gathered previously, and, as is well known, if new data is of the same kind as old data its confirmational value converges to zero very quickly. For an extreme example, the results of the same kind of experiment repeated again and again only varying times or locations or other irrelevant characteristics quickly loose
32
See http://www.ncbi.nlm.nih.gov/Genbank/genbankstats.html.
Nature (2009), vol. 460, no. 7255, p. 551
34 Sazlay and Gray (2006, pp. 413-4), in a report about a Microsoft workshop on the future of science, claim that the amount
of scientific data is doubling every year.
33
12.02.2016
15
any confirmational value. This is the worry of diminishing returns.35 If the worry were correct, the
increase in amount of data would not mean that there was a big boost in degrees of success in the recent past. In reply note the following: It is surely the case for the above sets of data (fossils, stars, etc.)
that for every piece of data gathered nowadays there are almost always many pieces of data gathered
earlier that look rather similar to it. But this way of viewing the confirmational situation is misleading.
We need to adopt a global perspective, and determine for each whole set of data whether it exhibits
high diversity or not.
So, how much diversity do the data sets presented above (data from fossils, chemical elements,
oceans, astronomical objects) possess? Surely they are not such that all the pieces of data are of the
same narrow kind. It is not the case that palaeontologists examine the same features of the same fossil
again and again, or that all fossils they find are of the same species and the same age. Such deeds
would correspond to repeating the same sort of experiment again and again varying only insignificant
properties such as time or location. Instead palaeontologists look for fossils from different locations
and strata, and what they find are mostly different species. Thus, in Figure 3 the y-axis depicts the
number of genera of mammalian fossils. (Genera are a more general classification grouping than species). Although some important parts of the fossil record are missing, especially from earlier times,
e.g., from the beginning of life on earth, we possess fossils for many important parts of the tree of life,
enough to have a rough outline of it.36 Also it is not the case that chemists examine the same features
of the same substance again and again; instead they create, as we saw, new substances incessantly. The
millions of chemical substances that have been synthesized so far clearly represent an extremely high
variety of evidence. Likewise it is not the case that all probes of ARGO were released at the very same
location in the same ocean, but instead were, of course, distributed widely over many different locations in the oceans of the earth so as to maximize the worth of the data produced. Finally, astronomical
projects such as the Sloan Digital Sky Survey don’t record the same star again and again, but record
features of hundreds of millions of different stars and galaxies.
The rise in the diversity of data manifests itself also in the very strong rise in the number of
kinds of instruments and kinds of measurement techniques in many scientific disciplines, especially in
the last few decades. While 100 years ago, we just had light microscopes, after World War II many
new kinds of microscopes have been developed: many types of light microscopes (polarization, fluorescence, phase contrast, etc., etc.), many types of electron microscopes, scanning probe microscopes
and acoustic microscopes.37 Similarly in astronomy: Until some decades ago, we had only light telescopes, while today, astronomers have many different kinds of instruments which not only cover most
of the electromagnetic spectrum, from radio-telescopes to gamma-ray telescopes, but also detect neutrinos, muons, Oh-My-God particles, and so on. To get an impression of the diversity of microscopes,
telescopes and other measurement techniques, have a look at the respective entries in Wikipedia, e.g.
at the entries of “measuring instruments” and “non-destructive testing”.
Finally, we can point to a further feature of science that supports that the evidence gathered by
scientists show a huge and ever growing variety and breadth. This feature is the ever increasing specialization in science. In practically all scientific disciplines we see an ever increasing number of approaches, techniques, instruments, problems, etc. All the time we see fields splitting into sub-fields. It
is very plausible that if there is a core theory in the respective scientific disciplines specialization often
results in more diversity of evidence for the core theory.
A particular salient indication of the specialisation is the problem of communication between
scientists. Scientists of different scientific disciplines, and even scientists of the same discipline, but
different sub-disciplines or sub-sub-discipline, have an ever harder time to communicate the particulars of their work with each other. This can easily be witnessed at scientific conferences. (Just go to
any talk of any conference and ask any member of the audience how much he really understood of the
talk.) In general, communication is possible for the basics of the respective field, the general aims,
35
This worry was first suggested to me by Rachel Cooper. See Hempel XXX
The fossil record as evidence for the theory of evolution has only a limited diversity. However, fossils are just one kind of
evidence among very many very different kinds of evidence from very different sources such as biogeography, genetics,
embryology, molecular biology, etc. To get an impression of the diversity of evidence for the theory of evolution compare
such sites as www.talkorigins.org, or any textbook on evolution such as Douglas Futuyama (2009).
37 See Ian Hacking’s “Do We See Through a Microscope”, in particular the section titled “A Plethora of Microscopes”. Microscopy is nowadays being developed in many different sometimes surprising directions, see Alison Abbott (2009).
36
12.02.2016
16
theories and problems of the respective scientific field. In contrast, communication about the specifics
of the daily work in the field, the particular techniques, data, instruments, particular problems and
special theories is mostly not possible without being provided a lot of special background information.
Compare this situation with the situation 100 years (or even 200 years ago), when it was far easier for
scientists to keep up with developments in their own discipline and even in other scientific disciplines.
The omnipresent difficulties of communication between scientists of different fields show just
how specialized science has become. And the ever higher specialisation is a clear indication of the
ever growing diversity of approaches, techniques, instruments, etc, and hence of the ever growing
diversity of evidence for the core theories. As with growing scientific work in general the growing
specialization does not automatically lead to more diversity of evidence. It is only an indicator of diversity. Surely, there are many scientific fields in which the large increase in specialization has not
lead to much more diversity of evidence for the theories in that field that result in a very degree of
high success for any of the theories. But, as with the increase of scientific work in general, it is also
implausible that no scientific theories profited in this way. Especially for the core theories of the respective disciplines such as the theory of evolution or the periodic Table of Elements it is highly plausible that they profited in this way.
3 The final definition of success
Before turning to the fourth and fifth indicator of success, I want to improve the notion of success one last time. To do so I will once more use confirmational ideas that are entirely standard (compare, e.g., Hempel 1966). So, here is the final characterization of success. It is still a partial characterization in that it only specifies some factors on which the degree of success of a scientific theory depends. First, as was noted earlier the degree of success of a theory at a given time depends on the total
number and diversity of all the tests that the theory has passed until that time. In every test a prediction38 of the theory and a piece of data are compared, hence the degree of success of a theory at a given
time depends on both the total number and variety of the predictions and the total number and variety
of the data involved in all the passed tests.
Second, individual tests differ with respect to their quality. I will here consider only one aspect
of the quality of tests, namely the precision (or specificity) of the data and predictions involved in
tests. The notion of precision applies to data of both a qualitative and a quantitative sort. If the data is
of a quantitative sort, then obviously its level of precision may vary, but the same is true for data of a
qualitative sort: It can vary in its precision from rather unspecific to very specific. For example, the
precise description of all features of a fossil can constitute a quite precise piece of data. Likewise for
the predictions derived by scientists from the given theory: they may be of a qualitative or a quantitative sort, and in both cases their precision may vary. For example, when scientists have to solve equations to derive predictions from a theory, they often have to use approximations and simplifications
which usually reduce the precision of the predictions.
In general, if the data is of higher precision than the prediction, then the quality of the test is determined by the precision of the prediction, and if the data is of lower precision than the prediction,
then the quality of the test is determined by the data precision of the data. If in a test either data or
predictions are of low or modest precision, we have a gentle test of the theory. Such a test, if passed,
can only result in a weak or modest increase in degrees of success for the theory. Confirmation en
passant is typically of this sort.
My partial characterization of the notion of success is of a rather basic sort. It could surely be refined in many ways, e.g., with the help of probabilistic means. Some ways of refining it are not necessary for our purposes here, quite the opposite some measure of generality is welcome having the advantage of making the arguments of this paper compatible with a large range of realist theories of confirmation. However other ways of refining the notion would be important to consider. It should be
taken into account that confirmation is holistic (that theories are not tested singly, but in bundles). The
same holds for the distinction between data and phenomena (Bogen/Woodward 198XX). Also relationships such as mutual support of theories at different levels of generality and between neighbouring
38
As noted earlier the notion of a prediction of a theory is used in a rather broad sense: Any observation statement which
scientists have derived from the theory counts as a prediction.
12.02.2016
17
scientific areas have to be examined. And so on. Taking all these complications into account has to be
left for several other occasions.
We can discern two extreme ways of how a theory can acquire very high degrees of success.
The two extreme ways form the two end points of a continuum. At one end we find theories that have
passed only gentle tests – all the data or predictions of the tests have had a low or moderate precision –
, but there are a very large number and variety of such tests. At the other end are theories that have not
passed very many tests, but passed tests of a very precise sort. In between the two ends of the continuum are a lot of possible mixtures of the two extreme ways. The support for our current best theories
are typically mixtures. A plausible conjuncture would be that for the best theories of the physical sciences we find both high precision and strong diversity of evidence. By contrast, the current best theories from other sciences, such as the theory of evolution or the periodic table of elements, profited less
from tests with very high precision, and more from a very high diversity of evidence.
4 Precision
The fourth indicator of success is the precision of data. Data becomes more precise, when scientists improve already existing kinds of instruments and measurement techniques, or develop new kinds
of instruments and measurement techniques. This happens all the time, of course, and has lead to constant improvement in the precision of data over the last few centuries, and especially the last few decades. Often the improvements were by great leaps. Examples abound. Let me just mention three especially interesting ones.
The first example concerns the measurement of distances between places on the surface of the
earth for the purpose of determining details about the movement of tectonic plates. The kinds of measurement available until the 1980s required years to produce meaningful data. In the 1980s this
changed dramatically through the advent of GPS. The precision increased 100fold. In consequence
determining the movements of tectonic plates became rather easy and very reliable.39 In this case we
have a substantial increase in the precision of the data which exhibits at the same time a large variety,
namely from thousands of GPS stations from many different places on the earth. This example is interesting, because it concerns the only interesting theory change discussed in the philosophical literature of recent times: In the 1960s geologists changed their view from the belief that the earth crust
does not move to the theory of moving tectonic plates. The example also illustrates the notion of confirmation en passant. Clearly, the purpose of the measurements is not to confirm once again that plate
tectonics is true, but instead to determine details of the tectonic plates such as their precise borders,
direction and velocity of their movements, etc., nevertheless plate tectonics is reconfirmed all the time
by these measurements.
The second example is the increase in precision of time measurement. Figure 4 offers a rough
idea of the immense increase in precision of time measurement over the last 400 years. Note that the yaxis has a logarithmic scale so that the increase is actually hyper-exponential. Furthermore, since the
1950s the precision of the best clocks has increased by at least one digit per decade (Sullivan, 2001, p.
6). Today the best clocks, so-called optical clocks, reach a precision of 1 in 1017. They are, of course,
expected to become still better soon (T. Rosenband, et al 2008). Needless to say precise measurements
of time are vital in very many different scientific fields (for example for GPS), and have led to a strong
increase in success for many of our current best theories over last few decades.
The third example is a very specific one, namely a test of Einstein’s equation E = mc2 in the year
2005. Using “improved crystal Bragg spectrometers” and a “Penning trap” (whatever those are), E =
mc2 was reconfirmed with an accuracy of at least 0.00004%.40 This accuracy is 55 times higher than
the accuracy of the previous best test from 1991, which, by the way, used an entirely different method.
39
GPS became available much earlier for scientists than for the general public. Incidentally, GPS works only, because several relativistic effects are taken into account (Neil Ashby 2003). So, every time you use GPS successfully, you reconfirm en
passant the approximate truth of general relativity, although only a tiny little bit.
40 NATURE, Vol. 438, (22/29 December 2005, pp. 1096-7). See New Scientist (4 March 2006, pp. 42-43) for the story
behind this recent validation of E = mc2 in which two teams, one in the US and the other in France made their respective
measurements entirely independently from each other until they deemed them stable, before simultaneously faxing the results
to each other.
12.02.2016
18
The last example shows that even for some of our best theories scientists don’t stop intentionally
testing them. As remarked earlier, reconfirmation of our best theories mainly occurs as a side-effect of
scientific projects, as “confirmation en passant”. But the last example shows that scientific projects in
which scientists make a deliberate effort to recheck a well-established theory, so that rechecking of the
theory is the main purpose or one of the main purposes of the project, do exist. While such intentional
tests are not considered to provide great advances, and are rarely in the limelight of the scientific
community (precisely because the theories are considered to be already sufficiently confirmed), still,
those tests are performed when the occasion to do so arises (which they do time and again, because of
new instruments and techniques), and their results are noticed and acknowledged. 41 It is simply interesting (and also satisfying) to see that a theory you already accept passes further more stringent tests.
And it is always at least possible that the theory does not pass a further test, in which case, if you can
really establish that this is so, you will become a candidate for the Nobel prize. Nevertheless, this does
not happen, and every such passed test constitutes a further boost for the success of the respective theory.
Figure 4: Improvement in the quality of artificial clocks and comparison with the clock provided
by terrestrial rotation (from XX)
5 Computing power
The fifth indicator of success is computing power. As I use the term here, the computing power
of the scientists at some time has two components, the software known to the scientists at that time
and the hardware available to the scientists at that time. The software comprises the methods and algorithms known to scientists to solve equations (and, more generally, methods to derive predictions from
theories). The hardware comprises the devices on which software are implemented such as abacuses,
logarithmic tables, computers, and the human brain. The computing power – software and hardware –
available to scientists at some time determines three things: the kinds of equations scientists of that
Another example is the recent test of Newton’s law of gravitation gravitational forces between masses separated by 55m
(“Gravity passes a little test“, NATURE Vol 446, 1 March 2007, pp. 31-32). This test also served to rule out some versions of
string theory, so testing Newton’s law of gravitation was not the only purpose of the project.
41
12.02.2016
19
time can solve, the precision of the solutions, and the amount of time needed to obtain the solutions
(efficiency).
So, let us look how computing power has developed over the last few centuries. First, software.
Mathematics and the other disciplines responsible for finding methods and algorithms to solve equations have, like the rest of science, gone through exponential growth during the history of science.
Hence, the number and efficiency of methods and algorithms to solve mathematical equations has
certainly been growing very strongly. However, it is difficult to quantify this growth in any meaningful way on a global level. Describing the growth of hardware over the history of science is far easier.
Until 50 years ago computations were done by humans, therefore overall human computing power
rose at least at the rate of the number of scientists, and actually considerably more due to instruments
like abacuses, logarithmic tables, slide rules, etc. Furthermore, for the last 50 years hardware power of
computers certainly rose exponentially: As is well know, hardware power of computers doubled
roughly every two years over the last 50 years. This growth is, of course, much stronger than the
growth of the number of scientists and journal articles.
6 The benefits of the increase in computing power
The increase in computing power is then connected with increase in success of theories in
straightforward ways. New and better software, i.e., new and better methods for solving equations, can
provide solutions for kinds of equations which were not solvable before, and it can lead to solutions of
already solvable equations, but with higher efficiency (i.e., consuming less time and other resources);
both kinds of improvement can then result in a higher number of predictions, which can thereby exhibit more diversity. Similarly, new and better software can lead to more precise solutions of already
solvable equations, hence to more precise predictions. The same holds for hardware: more and better
hardware leads to a higher number and diversity of predictions, and to more precise predictions.
Note an important difference between computing power and data. Every piece of data is taken to
have a specific content, i.e., is taken to carry information about a specific object or phenomenon such
as some star, some stretch of DNA, or some fossil. Therefore every piece of data is confirmationally
relevant for a very narrow set of theories only, mostly in just one scientific area. By contrast, neither
software nor hardware are taken to have a specific content.42 A method to solve some kind of equation
can, of course, be used in every scientific field in which such equations of that kind arise. And very
many methods for solving equations are in fact used in many different scientific areas.43 Even more so
for hardware. The same piece of hardware can usually implement a large number of different kinds of
software. This is certainly true for the human brain and for electronic computers.
All of this makes it highly plausible that the growth of software and hardware has contributed
strongly to the increase in degrees of success of our current best theories. Paul Humphreys remarks
that “much of the success of the modern physical sciences is due to calculation” (2004, especially Ch.
3). He observes that our knowledge of the numerical methods needed to solve equations numerically
have made big advances.44 Thus, he writes:
… [B]ehind a great deal of physical science lies this principle: It is the invention and deployment of tractable mathematics that drives much progress
in the physical sciences. Whenever you have a sudden increase in usable
42
Here are the correspondences between the elements on the prediction side of testing and the elements on the observation
side of testing: Software corresponds to observational and measurement techniques, hardware (including the human brain)
corresponds to measuring instruments (including the human perceptual apparatus). On both sides we can draw the distinction
between process and product. Making a calculation means running some piece of software on some piece of hardware. This is
a process the product of which is some computational result. Likewise, making an observation or measurement means employing a measuring device. This is a process the product of which is some piece of data. Finally, comparing the results of
calculations with the data constitutes a test of the respective theory. – Needless to say, I chose to examine those elements on
the two sides of testing for which some kind of quantitative statement is meaningful and for which I could find numerical
information.
43 See Paul Humphrey (2004, Ch ??)
44 The vast majority of equations are nowadays solved numerically. For example, the equations of Quantum Mechanics are
almost exclusively solved numerically today (Humphreys 2004, p. 60).
12.02.2016
20
mathematics, there will be a sudden, concomitant increase in scientific progress in the areas affected. And what has always been true of the physical
sciences is increasingly true of the other sciences. Biology, sociology, anthropology, economics, and other areas now have available, or are developing for themselves, computational mathematics that provide a previously unavailable dimension to their work. (p. 55, emphasis in original)
Humphreys provides ample evidence for the huge difference increase in computing power has
made for the empirical success of theories.
An important use of computing power in both physical and non-physical sciences is the construction of computer models and computer simulations as an intermediary between theory and observation. When constructing computer models scientists often have to make strong approximations and
idealizations. If so, their confidence that the predictions of the models are actually the predictions of
the theory cannot be that high. In case data is available for comparison with the predictions of the
model, i.e., in case we have a test of the respective theory, this test is a gentle test at best. Therefore, if
the predictions of the model accord with the respective observations or experimental results, the
amount of success the respective theory enjoys is limited. Still, if the theory is a core theory that is
profitably used in very many such models, and if the cases where the predictions are wrong are rare or
the fault can be traced to other culprits, such models of the theory can strongly contribute to the rise in
its degree of success. What is more, because computing power has increased so strongly, especially
over the last 50 years, the approximations have become ever better and the necessary idealizations
ever weaker, leading in many scientific fields to ever more precise models, which therefore constitute
increasingly stringent tests of the respective theories and have contributed ever more strongly to an
increase in their degrees of success.
5 Saving Realism
1 Completing the main argument
Let us complete the main argument. The main argument is an inference from statements about
the five indicators of success to statements about the degrees of success of the two classes of theories
we want to compare. As we saw, the indicators of success have enjoyed an enormous increase over the
last few decades. Before that almost all of them were quite low, today they are very high. From this we
can infer the degrees of success of both our current best theories and the refuted theories of the past.
On the one hand, we can infer that our current best theories profited from the enormous increase in the
indicators in the recent past and therefore received a big boost in their degrees of success in the recent
past, an increase that is far greater than any increase in success of any theories of earlier times. So they
enjoy very high degrees of success today. This inference proceeds on a general level, but we could
also show directly for a number of specific theories, e.g., some core theories, that their degrees of success profited from the increase of the indicators. On the other hand, the refuted theories of the past
discussed by philosophers were all refuted before the big boost of success took place. At those times
practically all indicators were quite low. Therefore, the degrees of success of the refuted theories were
quite modest. This completes the main argument. The conclusion of the main argument is the main
thesis of the paper: All our current best theories enjoy far higher degrees of success than any of the
successful, but refuted theories of the past, which enjoyed only quite modest degrees of success.
2 Reply to sophisticated PI
We can now reply to the sophisticated PI. The sophisticated PI is the argument that right to the
present theories kept being refuted, therefore we should extrapolate the incidence of false theories
from past degrees of success to current degrees of success. The target of the sophisticated PI is the
modified success-to-truth principle which restricts the inference from success to truth to current degrees of success, and which threatens to be undermined by the extrapolation of false theories. But the
sophisticated PI is challenged by the main thesis. The main thesis states that there is a large difference
12.02.2016
21
in degrees of success between the refuted theories of the past and our current best theories. It implies
that the claim of the sophisticated PI that theories kept being refuted as success kept growing is simply
not true. Quite the opposite is true: theory change among theories stopped at rather low levels of success. Therefore, the extrapolation of the existence of false theories from past degrees of success to
current degrees of success is not plausible. This rebuts the sophisticated PI, and saves the modified
success-to-truth principle.
3 How Realism is saved
Because the modified success-to-truth principle is not threatened by the existence of counterexamples, the realist can modify his position so that it consists in the endorsement of the modified success-to-truth principle. This modified form of realism is save from the sophisticated PI. It is a version
of realism which is compatible with the history of science.
The realist wants to support his position, i.e., the modified success-to-truth principle, with the
NMA, but the NMA and the intuitions behind it were attacked in the first objection. In a moment I will
offer a reply to the first objection. I want to show that the NMA and the intuitions behind it appear
slightly scathed, yet basically intact from the confrontation with the history of science. Assuming I can
show this, the realist can, after all, use the NMA to support the modified success-to-truth principle.
Finally, he can apply the modified success-to-truth principle to our current best theories, and infer that
they are approximately true.
6 Restoring the Intuitions**
1 The weak success-to-truth principle**
Let us now deal with the first objection. The first objection aims to undermine what I called the
shared realist intuitions by pointing out that they support via the NMA an inference to truth for the
levels of success enjoyed by the refuted theories of the past, so that the shared realist intuitions are at
odds with the historic track record. In this conflict, the intuitions lose, rendering them untrustworthy
and no longer able to support anything at all, in particular not the inference from success to truth for
current levels of success.
To restore the integrity of the shared realist intuitions, I want to consider what I will call the
weak success-to-truth principle. This principle states what we can infer for theories with moderate
degrees of success, as typically possessed by the successful, but refuted theories of the past. The principle is a conjunctions of two statements. First, we can infer that such a theory is partly true45 in the
sense that important components of such a theory are probably correct. The correct components may
be part of its ontology (some of its central terms refer), or part of its classification system, or they may
be structural claims such as its equations. Scientist accepting such a theory may not be able to determine which of its components are the correct ones, but rather know only in hindsight. 46 Also the theory may be true or partially true not in the whole domain of application as it was initially intended, but
only in a sizeable part of it, where this part may also be known in hindsight only. Second, the weak
success-to-truth principle states that there is a substantial probability (to fix ideas arbitrarily, between
20% and 80% say) that the theory is not just partially true, but actually fully true. A person is warranted to have some, though not that high level of confidence in its full truth, e.g., an attitude of tentative
belief may not be altogether unreasonable.
To show that realists can hold on to the weak success-to-truth principle I have to show two
things: that it is by and large in line with shared realist intuitions, and that it is not undermined by the
historical track record. First, we need to show that it is by and large in line with shared realist intuitions. When Smart and Putnam presented the NMA the levels of success they had in mind were, of
course, the present ones. I showed above that by far most of the growth in success of the respective
As before, “true” always means “approximately true”.
Kitcher compares this situation of not knowing which components are the correct ones with the preface paradox (2001,
170-71).
45
46
12.02.2016
22
best theories occurred in the recent past, that levels of success of the refuted theories of the more distant past were much lower than that of our current best theories. Also it is clearly the case that the lower the degree of success of some theory, the weaker the shared realist intuitions’ support for its truth
via the NMA. For these two reasons, the claim is plausible that the shared realist’s intuitions are such
that they don’t offer that much support for a NMA in support of the success-to-truth inference for the
refuted theories of the past. A realist 100 or 200 years ago, for example, if he had inferred the truth of
some successful theory of those times, should have been aware that he could have done so only rather
tentatively, and that anti-realism was a real option for many of the best theories at that time. When
some theories with so much lower amounts of success turned out to be wrong, it should not have been
such a big surprise for him, far less a miracle. Thus, for past refuted theories the common realist intuitions seem to support at most a watered-down NMA (which may therefore not really deserve its name
in this case).
Having said this it must be admitted that it is difficult to determine the shared realist intuitions
with any precision. It must also be admitted that current realists often seem to identify with past realists and realist-minded scientists, and may think that, at least in some cases, success of past refuted
theories already sufficed to have greater confidence in their truth than the weak success-to-truth principle recommends. Therefore, this principle may not be entirely in line with common realist intuitions.
Still, I think we can say that it is by and large in line with them.
2 The weak success-to-truth principle and the history of science**
We further have to show that the weak success-to-truth principle is not undermined by the history of science, but on the whole accords with it. So, let us examine the history of science and therein
the theories with moderate levels of success. What we then can do is use the first counterstrategy
against the PI mentioned earlier. That counterstrategy consists in showing that judged from today the
successful refuted theories of the past, although strictly speaking false, had parts that were true, e.g.,
had terms that referred, got the structure right, etc., or were approximately true in restricted domains.
Many philosophers have develop versions of this counterstrategy and have provided convincing historical evidence for its viability. Therefore I can be brief here. Thus, for example, important claims of the
phlogiston theory of combustion are still accepted today. Likewise many central claims of the caloric
theory of heat are true (in many situations heat behaves like a substance).47 Another example is the
sequence of theories of light since Maxwell: All its members agree on the structure of light, i.e., on
Maxwell’s equations.48 Furthermore, the ontologies of many of the successful abandoned theories of
the past overlap to a considerable extent with the ontologies of the theories we accept today, e.g. theories about the electron (Nola 19XX, Norton 19XX in Nola eds). Finally, some theories can only be
considered approximately true, if their domain of application is significantly restricted. Newtonian
mechanics is approximately true in a very large domain including the domains of most engineering
sciences. Hence, the first part of the weak success-to-truth principle, according to which theories with
moderate success usually have important components that are correct, is in accord with the history of
science.
In order to assess the second part of the weak success-to-truth principle, we not only have to
look at the incidence of false theories among theories with moderate degrees of success, but also at the
theories that were not refuted; we have to determine the ratio of the number of non-refuted theories to
the number of refuted theories. The weak principle is only compatible with the history of science if
that ratio has been not much lower than one, i.e., if among theories with moderate success the number
of refuted theories has at most the same order of magnitude as the number of non-refuted theories. Of
course, to show this is once again an ambitious thing to do, with numerous problems, such as how
theories are individuated, what levels of success precisely count as moderate, how degrees of success
are to be determined in the first place, etc. Hence, the following (all to brief) observations from the
history of science can only provide a crude estimation of that ratio. Still, I think they are already suffi47
Martin Carrier (2004) defends the claim that phlogiston theory and the caloric theory of heat already got significant statements about classification, i.e., natural kinds, right. This idea is further developed into a formal proof showing that under
certain conditions the successor theory in a theory has to retain some elements of the abandoned theory, see Schurz (2004,
2009). See also, among others, Ladyman (2009, §5).
48 Worall (1989), French and Ladyman (19XX). Many more examples are provided by Ladyman 1998XX.
12.02.2016
23
ciently good to show that the realist’s case is promising. As my aim is only to defend realism, the burden of proof that the ratio is considerably lower than one is once again on the side of the antirealist, as
he wants to attack the intuitions, therefore it suffices for the realist to throw sufficient doubt on the
anti-realist’s attack on the intuitions.
So, let us examine the moderately successful theories of the history of science. Here are four observations. First, inspecting the set of abandoned theories, for instance those on Laudan’s list or the
other cases offered in the philosophical literature, we observe that most of the successor theories of
the abandoned theories are still held today. Judged from today, many of the abandoned theories, especially the most interesting and most frequently discussed ones, were already “finalists” or, at the very
least, “semi-finalists” in the “theory contests” of the respective scientific fields. For example, the theory of phlogiston, the caloric theory of heat and the geocentric system were finalists (or semi-finalists,
depending inter alia on how theories are individuated and which theories count as versions of each
other) in their respective fields.
Second observation. As a more recent set of examples consider the field of clinical studies. In a
recent meta-study, 49 highly cited (i.e., cited more than 1000 times) original clinical studies claiming
that a drug or other treatment worked were examined. We can classify the results of these studies as
cases of moderate success, because they were usually supported by just one study and did not enjoy
strong support from independent evidence. It then turned out that subsequent studies of comparable or
larger sample size and with similarly or better-controlled designs contradicted the results of 16% of
the earlier studies and reported weaker results for another 16%. (For example, the refuted studies had
seemingly shown that hormone pills protect menopausal women from heart disease and that vitamin A
supplements reduce the risk of breast cancer.) This means that nearly two-thirds of the original results
held up.49
Third, there are only very few scientific fields which experienced more than one or two theory
changes among successful theories. The three to five changes of theories of light are quite an exception in this regard. Very few scientific fields exhibit such a large number of theory changes. Because
of the low number of such fields, their contribution to the ratio of non-refuted to refuted theories is
rather small, and barely lowers it.
Fourth, there are some fields in which the first really successful theories scientists hit on were
already the theories we still accept today. Examples of such theories are provided by discoveries of
entities and phenomena to which scientists had no access before, e.g., the discovery of X-rays, the
double helix, or double stars. If such discoveries were based on at least moderate evidence, they mostly have survived until today, hence in these cases there were sometimes no theory changes at all.
Hence, these cases either increase or, at least, do not lower the ratio of non-refuted to refuted theories.
Putting these four observations together (which, of course, have to be checked more thoroughly
than I can do here), we arrive at the (admittedly preliminary) estimation of the ratio of non-refuted to
refuted theories among moderately successful theories of very roughly one. It follows that scientists
could be reasonably confident that the moderate success of those theories sufficed to eliminate all but
two or three theories, that the two or three known and seriously entertained theories of their times
probably included the true theory, or, in other words, that the respective small disjunction of those
theories was probably true. For example, Priestly and Lavoisier could both be reasonably confident
that one of them was right. So, although Priestly, for example, was not right to have a strong conviction in phlogiston theory, he was not only justified to believe that important parts of his theory were
true, but also a tentative acceptance of his theory as a whole would not have been entirely unreasonable for him. Hence, the first part of the weak success-to-truth principle which allows an attitude of
tentative belief towards theories with a moderate amount of success is compatible with the history of
science.
From this and my conclusion above about the first part of the weak success-to-truth principle, it
follows that, although the common realist’s intuitions may not escape entirely unscathed from the confrontation with the past of science, they are not invalidated by the confrontation either, and can still be
used by the realist as reasons for supporting the full (as well as the weak) success-to-truth principle.
49
Paraphrased from John Ioannidis (2005) and Lindey Tanner (2005). Ironically whereas I use the results of this meta-study
to support realism, its real import is that it shows that clinical studies are not as reliable as commonly thought.
12.02.2016
24
7 Five objections
(…)
(…)
8 Conclusion
(…)
12.02.2016
25
9 References
Abbott, Alison (2009). „Magnifying Power“, Nature Vol. 459, no. 7247, (4 June 2009), pp. 629-39.
Ashby, Neil (2003). “Relativity in the Global Positioning System”, Living Reviews in Relativity 6, 1.
Bird, A., (2007). "What is Scientific Progress?" Noûs
Bolton, A., Scott Burles, Leon V.E. Koopmans, Tomaso Treu, and Leonidas A. Moustakas, (2006). “The Sloan Lens ACS
Survey. I. A Large Spectroscopically Selected Sample of Massive Early-Type Lens Galaxies”, The Astrophysical Journal, Volume 638, Issue 2, pp. 703-724.
Bovens, L. und Hartman, S. (2004), Bayesian Epistemology. Oxford University Press.
Boyd, R., 1983. "On the Current Status of the Issue of Scientific Realism." Erkenntnis 19: 45-90.
Carrier, M. (2004). “Experimental Success and the Revelation of Reality”, in: M. Carrier, J. Roggenhofer, G. Küppers, P.
Blanchard, Challenges beyond the Science Wars. Berlin: Springer-Verlag.
Dobzhansky Theodosius (1973). “Nothing in biology makes sense except in the light of evolution”. American Tiology Teacher 35, pp. 125-129.
Robert C. Kennicutt Jr (2007). “Sloan at five”, Nature, vol 450, Nov 22th 2007, p. 488-9.
Devitt, M. (1991). Realism and Truth. 2nd edn Oxford: Basil Blackwell.
Devitt, M. (2005), “Scientific Realism”, In The Oxford Handbook of Contemporary Analytic Philosophy, Frank Jackson and
Michael Smith, eds. Oxford: Oxford University Press pp. 767-91.
de Solla Price, D.J. (1963). Little Science, Big Science. New York: Columbia University Press.
Dunbar, R. (2004). The Human Story, Faber & Faber.
Egghe, L. and I.K. Raviachandra Rao, (1992). “Classification of Growth Models Based on Growth Rates and its Applications”, Scientometics, Vol. 25. No. 1, 5-46.
Forster, M. R. (2000. “Hard Problems in the Philosophy of Science: Idealization and Commensurability”. In R. Nola and H.
Sankey (eds.) (2000) After Popper, Kuhn & Feyerabend: Issues in Theories of Scientific Method, Australasian Studies
in History and Philosophy of Science, Kluwer.
Frigg, R. and Stephan Hartmann, (2005), “Scientific Models“, in: S. Sarkar et al. (eds.), The Philosophy of Science: An Encyclopedia, Vol. 2. New York: Routledge 2005, 740-749.
Furner, J. (2003). “Little Book, Big Book: Before and After Little Science, Big Science: A Review Article, Part II“, Journal
of Librarianship and Information Science, Vol. 35, No. 3, 189-201.
Futuyma, D.J. (2009). Evolution. 2nd edition, Sinauer Associates, Sunderland, Massachusetts.
Godfrey-Smith, P. (2003). Theory and Reality: An Introduction to the Philosophy of Science, The University of Chicago
Press, Chicago.
Humphreys,
P.
(2004).
Extending
Ourselves:
Computational
Science,
Empiricism,
and
Scientific Method, Oxford University Press.
Hardin, C., and Rosenberg, A., 1982. “In Defense of Convergent Realism” Philosophy of Science 49: 604-615.
Hempel, Carl Gustav (1966). Philosophy of Natural Science. Prentice Hall.
Hoyningen-Huene, P. (1993). Reconstructing Scientific Revolutions: Thomas S. Kuhn's Philosophy of Science, Chicago:
University of Chicago Press.
Ioannidis, John P. A. 2005. “Contradicted and Initially Stronger Effects in Highly Cited Clinical Research” Journal of the
American Medical Association, July 13, 2005; 294: 218 - 228.
Kitcher, P. (1993). The Advancement of Science New York: Oxford University Press
Kitcher, P. (2001). ‘Real realism: The Galilean Strategy”. The Philosophical Review, 110(2):151–197.
Klee, R. (1997). Introduction to the Philosophy of Science: Cutting Nature at Its Seams, Oxford University Press.
Kukla, A. (1998). Studies in Scientific Realism. Oxford: Oxford University Press.
Ladyman, J., (2002). Understanding Philosophy of Science. London: Routledge
Ladyman, James (2009). “Structural realism versus standard scientific realism: the case of phlogiston and dephlogisticated
air”, Synthese, XXX.
Lange, M. (2002). “Baseball, Pessimistic Inductions, and the Turnover Fallacy“ Analysis 62, 281-5.
Laudan, L. (1981) “A Refutation of Convergent Realism”, Philosophy of Science, 48 March, 19-49.
Laudan, L and J. Leplin (1991), “Empirical Equivalence and Underdetermination” in Journal
of Philosophy 1991 vol. 88,.no. 9 pp. 449-473
Leplin, J., (1997). A Novel Defence of Scientific Realism, Oxford: Oxford University Press.
Lewis, P. (2001). “Why The Pessimistic Induction Is A Fallacy”, Synthese 129: 371-380.
Mabe, M. and M. Amin, (2001). “Growth Dynamics of Scholarly and Scientific Journals”, Scientometrics. Vol. 51, No. 1,
147-162.
McAllister, J. W. (1993) ‘Scientific Realism and the Criteria for Theory Choice’, Erkenntnis 38, 2, pp.203-222.
McClellan III, James E. and Harold Dorn (1998). Science and Technology in World History, The Johns Hopkins University
Press, Baltimore and London.
McMullin, E. (1984). ‘A Case for Scientific Realism’ in Leplin, J., ed. Scientific Realism (Berkeley: University of California
Press), pp. 8-40.
McMullin,
E.
(1993).
‘Rationality
and
paradigm
change
in
science’.
In
World
Changes:
Thomas Kuhn and the Nature of Science, ed. P. Horwich, 55-78.
Meadows, J. (1974). Communication in Science. Butterworths, London.
12.02.2016
26
Meadows, J. (2000). “The Growth of Journal Literature: A Historical Perspective” in: The Web of Knowledge: A Festschrift
in honor of Eugene Garfield Edited by Blaise Cronin and Helen Barsky Atkins ASIS&T Monograph Series, Information Today, Inc., Medford, New Jersey.
Musgrave, A. (1988). “The Ultimate Argument for Scientific Realism,” in Robert Nola (ed.), Relativism and Realism in
Science Kluwer Academic Publishers.
Musgrave, A. (1999). Essays on realism and rationalism. Amsterdam & Atlanta: Rodopi.
Newton-Smith, W. H. (1987). “Realism and Inference to the Best Explanation”, Fundamenta Scientiae, Vol. 7, No. 3/4, 305316.
Niiniluoto, Ilkka (1999). Critical Scientific Realism, Oxford University Press
Norton, J. (2003). ‘Must Evidence Underdetermine Theory?’. Preprint on http://philsci-archive.pitt.edu.
Okasha, S. (2002). “Underdetermination, Holism and the Theory/Data Distinction”, Philosophical Quarterly, 52, 208, 303319.
Papineau, D. (1993), Philosophical Naturalism. Oxford: Blackwell.
Psillos, P. (1999). Scientific Realism: How Science Tracks Truth. New York and London: Routledge.
Psillos, P. (2000). “The Present State of the Scientific Realism Debate”, British Journal for the Philosophy of Science, 51
(Special Supplement), pp.705-728.
T. Rosenband, D. B. Hume, P. O. Schmidt, C. W. Chou, A. Brusch, L. Lorini, W. H. Oskay, R. E. Drullinger, T. M. Fortier,
J. E. Stalnaker,|| S. A. Diddams, W. C. Swann, N. R. Newbury, W. M. Itano, D. J. Wineland, J. C. Bergquist (2008).
“Frequency Ratio of Al+ and Hg+ Single-Ion Optical Clocks; Metrology at the 17th Decimal Place”. Science 28 March
2008: Vol. 319. no. 5871, pp. 1808 – 1812.
Rowe, T. (2005), Book Review of Mammals From the Age of Dinosaurs: Origins, Evolution, and Structure by Zofia KielanJaworowska, Richard L. Cifelli and Zhe-Xi Luo. Nature 438, 426 (24 November 2005)
Saatsi, J. (2004). “On the Pessimistic Induction and Two Fallacies”. In Proceedings Philosophy of Science Assoc. 19th Biennial Meeting - PSA2004: PSA 2004 Contributed Papers.
Sanderson, Allen R. and Bernard L. Dugoni, Thomas B. Hoffer Sharon L. Myers, (1999). Doctorate Recipients from United
States Universities: Summary Report 1999. http://www.norc.uchicago.edu/studies/sed/sed1999.htm.
Sazlay, A. and J. Gray (2006), Nature, vol. 440, 23 March 2006
Schummer, Jochen (1999). “Coping with the Growth of Chemical Knowledge”
Schurz, Gerhard (1991). “Relevant Deduction”, Erkenntnis 35, 391-437.
Schurz, G. (2004). „Theoretical Commensurability By Correspondence Relations: When Empirical Success Implies Theoretical Reference”, in: Kolak, D. and Symons, J (Eds.), Quantifiers, Questions and Quantum Physics Essays on the Philosophy of Jaakko Hintikka. Kluwer Academic Publishers
Schurz, Gerhard (2008). „Patterns of Abduction“, Synthese, 164: 201-234.
Schurz, Gerhard (2009). „When Empirical Success Implies Theoretical Reference: A Structural Correspondence Theorem“,British Journal for the Philosophy of Science 60/1, 2009, 101-133
Stanford, P. Kyle (2001). “Refusing the Devil's Bargain: What Kind of Underdetermination Should We Take Seriously?”,
Philosophy of Science 68 (Proceedings): S1-S12..
Stanford, P. Kyle (2003), „Pyrrhic Victories for Scientific Realism” Journal-of-Philosophy. N 03; 100(11): 553-572
Stanford, P. Kyle (2006). Exceeding Our Grasp: Science, History, and the Problem of Unconceived Alternatives. Oxford
University Press.
Sullivan, D.B., (2001). “Time and Frequency Measurement at NIST: The first 100 Years”.
Tannehill, J., Anderson. D. and R. Pletcher. (1997). Computational Fluid Mechanics and Heat Transfer. Taylor francis.
Tanner, Lindey (2005). „Review of Medical Research Turns Up Contradictory Results”, The Associate Press, July 13, 2005.
Tenopir, C. and D. W. King, (2004). “Communication Patterns of Engineers“, IEEE Press, Wiley-Interscience.
Trout, J. D. (2002). „Scientific Explanation and the Sense of Understanding“,Philosophy of Science 69: 212-233.
van Fraassen, B., 1980. The Scientific Image. Oxford: Oxford University Press.
van Fraassen, B., 1998. “The agnostic subtly probabilified”, Analysis 58, 212-220.
van Fraassen. 2003. “On McMullin's Appreciation of Realism.” Philosophy of Science 70: 479-492
Vickery, B.C. (1990). “The Growth of Scientific Literature, 1660-1970”, in: The Infomration Environemnt: A World View,
Studies in Honour of Professor A. I. Mikhailov, Elsevier Science Publishers B.V. North-Holland.
Vickery, B.C. (2000). Scientific Communication in History, The Scarecrow Press, Lanham, Maryland.
Wade, Nicholas (2009). “Cost of Decoding a Genome Is Lowered”, New York Times, August 11.
Williamson, T. (2006). “Must Do Better.” In: Greenough, P. and M. Lynch Truth and Realism, Oxford University Press.
Worrall, J. (1989). “Structural realism: The best of two worlds?” Dialectica, 43, 99-124.
Download