Philosophy of science, the study, from a philosophical perspective, of the elements of scientific inquiry. This article discusses metaphysical, epistemological, and ethical issues related to the practice and goals of modern science. For treatment of philosophical issues raised by the problems and concepts of specific sciences, see biology, philosophy of; and physics, philosophy of.
From Natural Philosophy To Theories Of Method
Philosophy and natural science

René Descartes (1596–1650)

Gottfried Wilhelm Leibniz (1646–1716)
The history of philosophy is intertwined with the history of the natural sciences. Long before the 19th century, when the term science began to be used with its modern meaning, those who are now counted among the major figures in the history of Western philosophy were often equally famous for their contributions to “natural philosophy,” the bundle of inquiries now designated as sciences. Aristotle (384–322 BCE) was the first great biologist; René Descartes (1596–1650) formulated analytic geometry (“Cartesian geometry”) and discovered the laws of the reflection and refraction of light; Gottfried Wilhelm Leibniz (1646–1716) laid claim to priority in the invention of the calculus; and Immanuel Kant (1724–1804) offered the basis of a still-current hypothesis regarding the formation of the solar system (the Kant-Laplace nebular hypothesis).

Immanuel Kant (1724–1804)

Francis Bacon (1561–1626)
In reflecting on human knowledge, the great philosophers also offered accounts of the aims and methods of the sciences, ranging from Aristotle’s studies in logic through the proposals of Francis Bacon (1561–1626) and Descartes, which were instrumental in shaping 17th-century science. They were joined in these reflections by the most eminent natural scientists. Galileo (1564–1642) supplemented his arguments about the motions of earthly and heavenly bodies with claims about the roles of mathematics and experiment in discovering facts about nature. Similarly, the account given by Isaac Newton (1642–1727) of his system of the natural world is punctuated by a defense of his methods and an outline of a positive program for scientific inquiry. Antoine-Laurent Lavoisier (1743–94), James Clerk Maxwell (1831–79), Charles Darwin (1809–82), and Albert Einstein (1879–1955) all continued this tradition, offering their own insights into the character of the scientific enterprise.

Galileo (1564–1642)

Isaac Newton (1642–1727)
Although it may sometimes be difficult to decide whether to classify an older figure as a philosopher or a scientist—and, indeed, the archaic “natural philosopher” may sometimes seem to provide a good compromise—since the early 20th century, philosophy of science has been more self-conscious about its proper role. Some philosophers continue to work on problems that are continuous with the natural sciences, exploring, for example, the character of space and time or the fundamental features of life. They contribute to the philosophy of the special sciences, a field with a long tradition of distinguished work in the philosophy of physics and with more-recent contributions in the philosophy of biology and the philosophy of psychology and neuroscience. General philosophy of science, by contrast, seeks to illuminate broad features of the sciences, continuing the inquiries begun in Aristotle’s discussions of logic and method. This is the topic of the present article.
Logical positivism and logical empiricism

Gottlob Frege (1848–1925)

Bertrand Russell (1872–1970)
A series of developments in early 20th-century philosophy made the general philosophy of science central to philosophy in the English-speaking world. Inspired by the articulation of mathematical logic, or formal logic, in the work of the philosophers Gottlob Frege (1848–1925) and Bertrand Russell (1872–1970) and the mathematician David Hilbert (1862–1943), a group of European philosophers known as the Vienna Circle attempted to diagnose the difference between the inconclusive debates that mark the history of philosophy and the firm accomplishments of the sciences they admired. They offered criteria of meaningfulness, or “cognitive significance,” aiming to demonstrate that traditional philosophical questions (and their proposed answers) are meaningless. The correct task of philosophy, they suggested, is to formulate a “logic of the sciences” that would be analogous to the logic of pure mathematics formulated by Frege, Russell, and Hilbert. In the light of logic, they thought, genuinely fruitful inquiries could be freed from the encumbrances of traditional philosophy.

David Hilbert (1862–1943)
To carry through this bold program, a sharp criterion of meaningfulness was required. Unfortunately, as they tried to use the tools of mathematical logic to specify the criterion, the logical positivists (as they came to be known) encountered unexpected difficulties. Again and again, promising proposals were either so lax that they allowed the cloudiest pronouncements of traditional metaphysics to count as meaningful, or so restrictive that they excluded the most cherished hypotheses of the sciences (see verifiability principle). Faced with these discouraging results, logical positivism evolved into a more moderate movement, logical empiricism. (Many historians of philosophy treat this movement as a late version of logical positivism and accordingly do not refer to it by any distinct name.) Logical empiricists took as central the task of understanding the distinctive virtues of the natural sciences. In effect, they proposed that the search for a theory of scientific method— undertaken by Aristotle, Bacon, Descartes, and others—could be carried out more thoroughly with the tools of mathematical logic. Not only did they see a theory of scientific method as central to philosophy, but they also viewed that theory as valuable for aspiring areas of inquiry in which an explicit understanding of method might resolve debates and clear away confusions. Their agenda was deeply influential in subsequent philosophy of science.
Discovery, Justification, And Falsification
Logics of discovery and justification

Hans Reichenbach (1891–1953)

August Kekulé (1829–96)
An ideal theory of scientific method would consist of instructions that could lead an investigator from ignorance to knowledge. Descartes and Bacon sometimes wrote as if they could offer so ideal a theory, but after the mid-20th century the orthodox view was that this is too much to ask for. Following Hans Reichenbach (1891–1953), philosophers often distinguished between the “context of discovery” and the “context of justification.” Once a hypothesis has been proposed, there are canons of logic that determine whether or not it should be accepted—that is, there are rules of method that hold in the context of justification. There are, however, no such rules that will guide someone to formulate the right hypothesis, or even hypotheses that are plausible or fruitful. The logical empiricists were led to this conclusion by reflecting on cases in which scientific discoveries were made either by imaginative leaps or by lucky accidents; a favourite example was the hypothesis by August Kekulé (1829–96) that benzene molecules have a hexagonal structure, allegedly formed as he was dozing in front of a fire in which the live coals seemed to resemble a snake devouring its own tail.

Thomas Kuhn (1922–96)
Although the idea that there cannot be a logic of scientific discovery often assumed the status of orthodoxy, it was not unquestioned. As will become clear below, one of the implications of the influential work of Thomas Kuhn (1922–96) in the philosophy of science was that considerations of the likelihood of future discoveries of particular kinds are sometimes entangled with judgments of evidence, so discovery can be dismissed as an irrational process only if one is prepared to concede that the irrationality also infects the context of justification itself.
Sometimes in response to Kuhn and sometimes for independent reasons, philosophers tried to analyze particular instances of complex scientific discoveries, showing how the scientists involved appear to have followed identifiable methods and strategies. The most ambitious response to the empiricist orthodoxy tried to do exactly what was abandoned as hopeless—to wit, specify formal procedures for producing hypotheses in response to an available body of evidence. So, for example, the American philosopher Clark Glymour and his associates wrote computer programs to generate hypotheses in response to statistical evidence, hypotheses that often introduced new variables that did not themselves figure in the data. These programs were applied in various traditionally difficult areas of natural and social scientific research. Perhaps, then, logical empiricism was premature in writing off the context of discovery as beyond the range of philosophical analysis.
In contrast, logical empiricists worked vigorously on the problem of understanding scientific justification. Inspired by the thought that Frege, Russell, and Hilbert had given a completely precise specification of the conditions under which premises deductively imply a conclusion, philosophers of science hoped to offer a “logic of confirmation” that would identify, with equal precision, the conditions under which a body of evidence supported a scientific hypothesis. They recognized, of course, that a series of experimental reports on the expansion of metals under heat would not deductively imply the general conclusion that all metals expand when heated—for even if all the reports were correct, it would still be possible that the very next metal to be examined failed to expand under heat. Nonetheless, it seemed that a sufficiently large and sufficiently varied collection of reports would provide some support, even strong support, for the generalization. The philosophical task was to make precise this intuitive judgment about support.

Rudolf Carnap (1891–1970)

Carl Hempel (1905–97)
During the 1940s, two prominent logical empiricists, Rudolf Carnap (1891–1970) and Carl Hempel (1905–97), made influential attempts to solve this problem. Carnap offered a valuable distinction between various versions of the question. The “qualitative” problem of confirmation seeks to specify the conditions under which a body of evidence E supports, to some degree, a hypothesis H. The “comparative” problem seeks to determine when one body of evidence E supports a hypothesis H more than a body of evidence E* supports a hypothesis H* (here E and E* might be the same, or H and H* might be the same). Finally, the “quantitative” problem seeks a function that assigns a numerical measure of the degree to which E supports H. The comparative problem attracted little attention, but Hempel attacked the qualitative problem while Carnap concentrated on the quantitative problem.
It would be natural to assume that the qualitative problem is the easier of the two, and even that it is quite straightforward. Many scientists (and philosophers) were attracted to the idea of hypothetico-deductivism, or the hypothetico-deductive method: scientific hypotheses are confirmed by deducing from them predictions about empirically determinable phenomena, and, when the predictions hold good, support accrues to the hypotheses from which those predictions derive. Hempel’s explorations revealed why so simple a view could not be maintained. An apparently innocuous point about support seems to be that, if E confirms H, then E confirms any statement that can be deduced from H. Suppose, then, that H deductively implies E, and E has been ascertained by observation or experiment. If H is now conjoined with any arbitrary statement, the resulting conjunction will also deductively imply E. Hypothetico-deductivism says that this conjunction is confirmed by the evidence. By the innocuous point, E confirms any deductive consequence of the conjunction. One such deductive consequence is the arbitrary statement. So one reaches the conclusion that E, which might be anything whatsoever, confirms any arbitrary statement.
To see how bad this is, consider one of the great predictive theories—for example, Newton’s account of the motions of the heavenly bodies. Hypothetico-deductivism looks promising in cases like this, precisely because Newton’s theory seems to yield many predictions that can be checked and found to be correct. But if one tacks on to Newtonian theory any doctrine one pleases—perhaps the claim that global warming is the result of the activities of elves at the North Pole—then the expanded theory will equally yield the old predictions. On the account of confirmation just offered, the predictions confirm the expanded theory and any statement that follows deductively from it, including the elfin warming theory.
Hempel’s work showed that this was only the start of the complexities of the problem of qualitative confirmation, and, although he and later philosophers made headway in addressing the difficulties, it seemed to many confirmation theorists that the quantitative problem was more tractable. Carnap’s own attempts to tackle that problem, carried out in the 1940s and ’50s, aimed to emulate the achievements of deductive logic. Carnap considered artificial systems whose expressive power falls dramatically short of the languages actually used in the practice of the sciences, and he hoped to define for any pair of statements in his restricted languages a function that would measure the degree to which the second supports the first. His painstaking research made it apparent that there were infinitely many functions (indeed, continuum many—a “larger” infinity corresponding to the size of the set of real numbers) satisfying the criteria he considered admissible. Despite the failure of the official project, however, he argued in detail for a connection between confirmation and probability, showing that, given certain apparently reasonable assumptions, the degree-of-confirmation function must satisfy the axioms of the probability calculus.
Bayesian confirmation
That conclusion was extended in the most prominent contemporary approach to issues of confirmation, so-called Bayesianism, named for the English clergyman and mathematician Thomas Bayes (1702–61). The guiding thought of Bayesianism is that acquiring evidence modifies the probability rationally assigned to a hypothesis.
For a simple version of the thought, a hackneyed example will suffice. If one is asked what probability should be assigned to drawing the king of hearts from a standard deck of 52 cards, one would almost certainly answer 1/52. Suppose now that one obtains information to the effect that a face card (ace, king, queen, or jack) will be drawn; now the probability shifts from 1/52 to 1/16. If one learns that the card will be red, the probability increases to 1/8. Adding the information that the card is neither an ace nor a queen makes the probability 1/4. As the evidence comes in, one forms a probability that is conditional on the information one now has, and in this case the evidence drives the probability upward. (This need not have been the case: if one had learned that the card drawn was a jack, the probability of drawing the king of hearts would have plummeted to 0.)
Bayes is renowned for a theorem that explains an important relationship between conditional probabilities. If, at a particular stage in an inquiry, a scientist assigns a probability to the hypothesis H, Pr(H)—call this the prior probability of H—and assigns probabilities to the evidential reports conditionally on the truth of H, PrH(E), and conditionally on the falsehood of H, Pr−H(E), Bayes’s theorem gives a value for the probability of the hypothesis H conditionally on the evidence E by the formula:
One of the attractive features of this approach to confirmation is that when the evidence would be highly improbable if the hypothesis were false—that is, when Pr−H(E) is extremely small—it is easy to see how a hypothesis with a quite low prior probability can acquire a probability close to 1 when the evidence comes in. (This holds even when Pr(H) is quite small and Pr(−H), the probability that H is false, correspondingly large; if E follows deductively from H, PrH(E) will be 1; hence, if Pr−H(E) is tiny, the numerator of the right side of the formula will be very close to the denominator, and the value of the right side thus approaches 1.)
Any use of Bayes’s theorem to reconstruct scientific reasoning plainly depends on the idea that scientists can assign the pertinent probabilities, both the prior probabilities and the probabilities of the evidence conditional on various hypotheses. But how should scientists conclude that the probability of an interesting hypothesis takes on a particular value or that a certain evidential finding would be extremely improbable if the interesting hypothesis were false? The simple example about drawing from a deck of cards is potentially misleading in this respect, because in this case there seems to be available a straightforward means of calculating the probability that a specific card, such as the king of hearts, will be drawn. There is no obvious analogue with respect to scientific hypotheses. It would seem foolish, for example, to suppose that there is some list of potential scientific hypotheses, each of which is equally likely to hold true of the universe.
If one supposes that the evidence obtained is like that acquired in the decades after the publication of Newton’s hypothesis in his Principia (Philosophiae naturalis principia mathematica, 1687), it may seem possible to resolve the issue as follows: even though both investigators were initially skeptical (both assigned small prior probabilities to Newton’s hypothesis), one gave the hypothesis a serious chance and the other did not; the inquirer who started with the truly minute probability made an irrational judgment that infects the conclusion. No subjective Bayesian can tolerate this diagnosis, however. The Newtonian hypothesis is not a logical or mathematical truth (or a logical or mathematical falsehood), and both scientists give it a probability different from 0 and 1. By subjective Bayesian standards, that is all rational inquirers are asked to do.
The orthodox response to worries of this type is to offer mathematical theorems that demonstrate how individuals starting with different prior probabilities will eventually converge on a common value. Indeed, were the imaginary investigators to keep going long enough, their eventual assignments of probability would differ by an amount as tiny as one cared to make it. In the long run, scientists who lived by Bayesian standards would agree. But, as the English economist (and contributor to the theory of probability and confirmation) John Maynard Keynes (1883–1946) once observed, “in the long run we are all dead.” Scientific decisions are inevitably made in a finite period of time, and the same mathematical explorations that yield convergence theorems will also show that, given a fixed period for decision making, however long it may be, there can be people who satisfy the subjective Bayesian requirements and yet remain about as far apart as possible, even at the end of the evidence-gathering period.
Eliminativism and falsification
Subjective Bayesianism is currently the most popular view of the confirmation of scientific hypotheses, partly because it seems to accord with important features of confirmation and partly because it is both systematic and precise. But the worry just outlined is not the only concern that critics press and defenders endeavour to meet. Among others is the objection that explicit assignments of probabilities seem to figure in scientific reasoning only when the focus is on statistical hypotheses. A more homely view of testing and the appraisal of hypotheses suggests that scientists proceed by the method of Sherlock Holmes: they formulate rival hypotheses and apply tests designed to eliminate some until the hypothesis that remains, however antecedently implausible, is judged correct. Unlike Bayesianism, this approach to scientific reasoning is explicitly concerned with the acceptance and rejection of hypotheses and thus seems far closer to the everyday practice of scientists than the revision of probabilities. But eliminativism, as this view is sometimes called, also faces serious challenges.
The first main worry centres on the choice of alternatives. In the setting of the country-house murder, Sherlock Holmes (or his counterpart) has a clear list of suspects. In scientific inquiries, however, no such complete roster of potential hypotheses is available. For all anyone knows, the correct hypothesis might not figure among the rivals under consideration. How then can the eliminative procedure provide any confidence in the hypothesis left standing at the end? Eliminativists are forced to concede that this is a genuine difficulty and that there can be many situations in which it is appropriate to wonder whether the initial construction of possibilities was unimaginative. If they believe that inquirers are sometimes justified in accepting the hypothesis that survives an eliminative process, then they must formulate criteria for distinguishing such situations. By the early 21st century, no one had yet offered any such precise criteria.

Karl Popper (1902–92)
An apparent method of avoiding the difficulty just raised would be to emphasize the tentative character of scientific judgment. This tactic was pursued with considerable thoroughness by the Austrian-born British philosopher Karl Popper (1902–92), whose views about scientific reasoning probably had more influence on practicing scientists than those of any other philosopher. Although not himself a logical positivist, Popper shared many of the aspirations of those who wished to promote “scientific philosophy.” Instead of supposing that traditional philosophical discussions failed because they lapsed into meaninglessness, he offered a criterion of demarcation in terms of the falsifiability of genuine scientific hypotheses. That criterion was linked to his reconstruction of scientific reasoning: science, he claimed, consists of bold conjectures that scientists endeavour to refute, and the conjectures that survive are given tentative acceptance. Popper thus envisaged an eliminative process that begins with the rival hypotheses that a particular group of scientists happen to have thought of, and he responded to the worry that the successful survival of a series of tests might not be any indicator of truth by emphasizing that scientific acceptance is always tentative and provisional.
Popper’s influence on scientists reflected his ability to capture features that investigators recognized in their own reasoning. Philosophers, however, were less convinced. For however much he emphasized the tentative character of acceptance, Popper—like the scientists who read him—plainly thought that surviving the eliminative process makes a hypothesis more worthy of being pursued or applied in a practical context. The “conjectures” are written into textbooks, taught to aspiring scientists, relied on in further research, and used as the basis for interventions in nature that sometimes affect the well-being of large numbers of people. If they attain some privileged status by enduring the fire of eliminative testing, then Popper’s view covertly presupposes a solution to the worry that elimination has merely isolated the best of a bad lot. If, on the other hand, the talk about “tentative acceptance” is taken seriously, and survival confers no special privilege, then it is quite mysterious why anybody should be entitled to use the science “in the books” in the highly consequential ways it is in fact used. Popper’s program was attractive because it embraced the virtues of eliminativism, but the rhetoric of “bold conjectures” and “tentative acceptance” should be viewed as a way of ducking a fundamental problem that eliminativists face.

Pierre Duhem (1861–1916)
A second major worry about eliminativism charged that the notion of falsification is more complex than eliminativists (including Popper) allowed. As the philosopher-physicist Pierre Duhem (1861–1916) pointed out, experiments and observations typically test a bundle of different hypotheses. When a complicated experiment reveals results that are dramatically at odds with predictions, a scientist’s first thought is not to abandon a cherished hypothesis but to check whether the apparatus is working properly, whether the samples used are pure, and so forth. A particularly striking example of this situation comes from the early responses to the Copernican system. Astronomers of the late 16th century, virtually all of whom believed in the traditional view that the heavenly bodies revolved around the Earth, pointed out that if, as Copernicus claimed, the Earth is in motion, then the stars should be seen at different angles at different times of the year; but no differences were observed, and thus Copernicanism, they concluded, is false. Galileo, a champion of the Copernican view, replied that the argument is fallacious. The apparent constancy of the angles at which the stars are seen is in conflict not with Copernicanism alone but with the joint hypothesis that the Earth moves and that the stars are relatively close. Galileo proposed to “save” Copernicanism from falsification by abandoning the latter part of the hypothesis, claiming instead that the universe is much larger than had been suspected and that the nearest stars are so distant that the differences in their angular positions cannot be detected with the naked eye. (He was vindicated in the 19th century, when improved telescopes revealed the stellar parallax.)

W.V.O. Quine (1908–2000)
The complexities of the notion of falsification, originally diagnosed by Duhem, had considerable impact on contemporary philosophy of science through the work of the American philosopher W.V.O. Quine (1908–2000). Quine proposed a general thesis of the underdetermination of theory by evidence, arguing that it is always possible to preserve any hypothesis in the face of any evidence. This thesis can be understood as a bare logical point, to the effect that an investigator can always find some consistent way of dealing with observations or experiments so as to continue to maintain a chosen hypothesis (perhaps by claiming that the apparent observations are the result of hallucination). So conceived, it appears trivial. Alternatively, one can interpret it as proposing that all the criteria of rationality and scientific method permit some means of protecting the favoured hypothesis from the apparently refuting results. On the latter reading, Quine went considerably beyond Duhem, who held that the “good sense” of scientists enables them to distinguish legitimate from illegitimate ways of responding to recalcitrant findings.

Samuel Clarke (1675–1729)
The stronger interpretation of the thesis is sometimes inspired by a small number of famous examples from the history of physics. In the early 18th century, there was a celebrated debate between Leibniz and Samuel Clarke (1675–1729), an acolyte of Newton, over the “true motions” of the heavenly bodies. Clarke, following Newton, defined true motion as motion with respect to absolute space and claimed that the centre of mass of the solar system was at rest with respect to absolute space. Leibniz countered by suggesting that, if the centre of mass of the solar system were moving with uniform velocity with respect to absolute space, all the observations one could ever make would be the same as they would be if the universe were displaced in absolute space. In effect, he offered infinitely many alternatives to the Newtonian theory, each of which seemed equally well supported by any data that could be collected. Recent discussions in the foundations of physics sometimes suggested a similar moral. Perhaps there are rival versions of string theory, each of which is equally well supported by all the evidence that could become available.
Such examples, which illustrate the complexities inherent in the notion of falsification, raise two important questions: first, when cases of underdetermination arise, what is it reasonable to believe? And second, how frequently do such cases arise? One very natural response to the motivating examples from physics is to suggest that, when one recognizes that genuinely rival hypotheses could each be embedded in a body of theory that would be equally well supported by any available evidence, one should look for a more minimal hypothesis that will somehow “capture what is common” to the apparent alternatives. If that natural response is right, then the examples do not really support Quine’s sweeping thesis, for they do not permit the rationality of believing either (or any) of a pair (or collection) of alternatives but rather insist on articulating a different, more minimal, view.
A second objection to the strong thesis of underdetermination is that the historical examples are exceptional. Certain kinds of mathematical theories, together with plausible assumptions about the evidence that can be collected, allow for the formulation of serious alternatives. In most areas of science, however, there is no obvious way to invoke genuine rivals. Since the 1950s, for example, scientists have held that DNA molecules have the structure of a double helix, in which the bases jut inward, like the rungs of a ladder, and that there are simple rules of base pairing. If Quine’s global thesis were correct, there should be some scientific rival that would account equally well for the vast range of data that supports this hypothesis. Not only has no such rival been proposed, but there are simply no good reasons for thinking that any exists.

DNA molecule
Many contemporary discussions in the philosophy of science take up the issues of this section, seeking algorithms for scientific discovery, attempting to respond to the worries about Bayesian confirmation theory or to develop a rival, and exploring the notions of falsification and underdetermination. These discussions often continue the inquiries begun by the principal logical empiricists—Carnap, Hempel, Reichenbach, and Popper—adhering to the conceptions of science and philosophy that were central to their enterprise. For a significant number of philosophers, however, the questions posed in this section were transformed by reactions to logical empiricism, by the historicist turn in the philosophy of science, and by the increasing interest in the social dimensions of scientific research. As will be discussed in later sections, some of the issues already raised arise in different forms and with more disturbing implications.
Explanations, Laws, And Theories
The logical-empiricist project of contrasting the virtues of science with the defects of other human ventures was only partly carried out by attempting to understand the logic of scientific justification. In addition, empiricists hoped to analyze the forms of scientific knowledge. They saw the sciences as arriving at laws of nature that were systematically assembled into theories. Laws and theories were valuable not only for providing bases for prediction and intervention but also for yielding explanation of natural phenomena. In some discussions, philosophers also envisaged an ultimate aim for the systematic and explanatory work of the sciences: the construction of a unified science in which nature was understood in maximum depth.
The idea that the aims of the natural sciences are explanation, prediction, and control dates back at least to the 19th century. Early in the 20th century, however, some prominent scholars of science were inclined to dismiss the ideal of explanation, contending that explanation is inevitably a subjective matter. Explanation, it was suggested, is a matter of feeling “at home” with the phenomena, and good science need provide nothing of the sort. It is enough if it achieves accurate predictions and an ability to control.
Explanation as deduction
The work of Carl Hempel

Ernest Nagel (1901–85)
During the 1930s and ’40s, philosophers fought back against this dismissal of explanation. Popper, Hempel, and Ernest Nagel (1901–85) all proposed an ideal of objective explanation and argued that explanation should be restored as one of the aims of the sciences. Their writings recapitulated in more precise form a view that had surfaced in earlier reflections on science from Aristotle onward. Hempel’s formulations were the most detailed and systematic and the most influential.
Hempel explicitly conceded that many scientific advances fail to make one feel at home with the phenomena—and, indeed, that they sometimes replace a familiar world with something much stranger. He denied, however, that providing an explanation should yield any sense of “at homeness.” First, explanations should give grounds for expecting the phenomenon to be explained, so that one no longer wonders why it came about but sees that it should have been anticipated; second, explanations should do this by making apparent how the phenomenon exemplifies the laws of nature. So, according to Hempel, explanations are arguments. The conclusion of the argument is a statement describing the phenomenon to be explained. The premises must include at least one law of nature and must provide support for the conclusion.
The simplest type of explanation is that in which the conclusion describes a fact or event and the premises provide deductive grounds for it. Hempel’s celebrated example involved the cracking of a car radiator on a cold night. Here the conclusion to be explained might be formulated as the statement, “The radiator cracked on the night of January 10th.” Among the premises would be statements describing the conditions (“The temperature on the night of January 10th fell to −10 °C,” etc.), as well as laws about the freezing of water, the pressure exerted by ice, and so forth. The premises would consitute an explanation because the conclusion follows from them deductively.
Difficulties
One obvious line of objection is that explanations, in ordinary life as well as in the sciences, rarely take the form of complete arguments. A clumsy person, for example, may explain why there is a stain on the carpet by confessing that he spilled the coffee, and a geneticist may account for an unusual fruit fly by claiming that there was a recombination of the parental genotypes. Hempel responded to this criticism by distinguishing between what is actually presented to someone who requests an explanation (the “explanation sketch”) and the full objective explanation. A reply to an explanation seeker works because the explanation sketch can be combined with information that the person already possesses to enable him to arrive at the full explanation. The explanation sketch gains its explanatory force from the full explanation and contains the part of the full explanation that the questioner needs to know.
A second difficulty for Hempel’s account resulted from his candid admission that he was unable to offer a full analysis of the notion of a scientific law. Laws are generalizations about a range of natural phenomena, sometimes universal (“Any two bodies attract one another with a force that is proportional to the product of their masses and inversely as the square of the distance between them”) and sometimes statistical (“The chance that any particular allele will be transmitted to a gamete in meiosis is 50 percent”). Not every generalization, however, counts as a scientific law. There are streets on which every house is made of brick, but no judgment of the form “All houses on X street are made of brick” qualifies as a scientific law. As Reichenbach pointed out, there are accidental generalizations that seem to have very broad scope. Whereas the statement “All uranium spheres have a radius of less than one kilometre” is a matter of natural law (large uranium spheres would be unstable because of fundamental physical properties), the statement “All gold spheres have a radius of less than one kilometre” merely expresses a cosmic accident.

David Hume (1711–76)
Intuitively, laws of nature seem to embody a kind of necessity: they do not simply describe the way that things happen to be, but, in some sense, they describe how things have to be. If one attempted to build a very large uranium sphere, one would be bound to fail. The prevalent attitude of logical empiricism, following the celebrated discussion of “necessary connections” in nature by the Scottish philosopher David Hume (1711–76), was to be wary of invoking notions of necessity. To be sure, logical empiricists recognized the necessity of logic and mathematics, but the laws of nature could hardly be conceived as necessary in this sense, for it is logically (and mathematically) possible that the universe had different laws. Indeed, one main hope of Hempel and his colleagues was to avoid difficulties with necessity by relying on the concepts of law and explanation. To say that there is a necessary connection between two types of events is, they proposed, simply to assert a lawlike succession—events of the first type are regularly succeeded by events of the second, and the succession is a matter of natural law. For this program to succeed, however, logical empiricism required an analysis of the notion of a law of nature that did not rely on the concept of necessity. Logical empiricists were admirably clear about what they wanted and about what had to be done to achieve it, but the project of providing the pertinent analysis of laws of nature remained an open problem for them.
Nor is it obvious that the fundamental idea of explaining through making the phenomena expectable can be sustained. To cite a famous example, one can explain the fact that the mayor contracted paresis by pointing out that he had previously had untreated syphilis, even though only 8 to 10 percent of people with untreated syphilis go on to develop paresis. In this instance, there is no statistical argument that confers high probability on the conclusion that the mayor contracted paresis—that conclusion remains improbable in light of the information advanced (85 percent of those with untreated syphilis do not get paresis). What seems crucial is the increase in probability, the fact that the probability of the conclusion rose from truly minute (paresis is extremely rare in the general population) to significant.
Other approaches to explanation

Wesley Salmon (1925–2001)
By the early 1970s, Hempel’s approach to explanation (known as the covering-law model) seemed to be in trouble on a number of fronts, leading philosophers to canvass alternative treatments. An influential early proposal elaborated on the diagnosis of the last paragraph. Wesley Salmon (1925–2001) argued that probabilistic explanation should be taken as primary and that probabilistic explanations proceed by advancing information that raises the probability of the event (or fact) to be explained. Building on insights of Reichenbach, Salmon noted that there are cases in which giving information that raises probability is not explanatory: the probability that there is a storm goes up when one is told that the barometer is falling, but the fall of the barometer does not explain the occurrence of the storm. Reichenbach had analyzed such examples by seeing both the barometer’s fall and the storm as effects of a common cause and offering a statistical condition to encompass situations in which common causes are present. Salmon extended Reichenbach’s approach, effectively thinking of explanation as identifying the causes of phenomena and, consonant with empiricist scruples, attempting to provide an analysis of causation in terms of statistical relations. Unfortunately, it proved very difficult to reconstruct causal notions in statistical terms, and by the 1980s most philosophers had abandoned the attempt as hopeless.
Many, however—including Salmon—remained convinced that the notion of causation is central to the understanding of explanation and that scientific explanation is a matter of tracing causes. They were divided (and continue to be divided) into two groups: those who believed that Humean worries about causation are important and that, in consequence, a prior analysis of causation is needed, and those who think that Hume and his successors adopted a faulty picture of human knowledge, failing to recognize that people are capable of detecting causal relations perceptually. Salmon was the most prominent member of the first group, offering an intricate account of causal processes, causal propagation, and causal interaction by appealing (in later work) to the conservation of physical quantities. He also argued, against his earlier view, that causal explanation can sometimes proceed by making the event explained appear less probable than it formerly seemed. (Imagine a golfer whose ball strikes a tree and is deflected into the hole; a description of the initial trajectory of the ball would decrease the probability that the result will be a hole in one.)
Although regarding explanation as a matter of tracing causes responds in a very direct way to several of the problems encountered by Hempel’s approach, it was not the only program in the recent theory of explanation. Some philosophers attempted to remain closer to Hempel’s project by thinking of explanation in terms of unification. Especially concerned with examples of theoretical explanation in the sciences, they proposed that the hallmark of explanation is the ability to treat from a single perspective phenomena previously seen as highly disparate. They elaborate on the remark of the English biologist T.H. Huxley (1825–95) that “in the end, all phenomena are incomprehensible and that the task of science is to reduce the fundamental incomprehensibilities to the smallest possible number.” This view, however, faced considerable technical difficulties in addressing some of the problems that arose for Hempel’s approach. Its principal merits lay in the avoidance of any reliance on causal concepts and in the ability to give an account of explanation in areas of theoretical science in which talk of causation seems strained.
A different strategy began by questioning the Hempelian proposal that ordinary explanations consist in explanation sketches whose force derives from an unarticulated ideal explanation. Philosophers such as Peter Achinstein and Bas van Fraassen offered pragmatic theories, according to which what counts as an explanation is contextually determined. Their accounts remained close to the everyday practice of explaining, but, to the extent that they eschewed context-independent conditions on explanation, they encouraged a return to the idea that explanation is a purely subjective business, a matter of what an audience will be satisfied with. Indeed, van Fraassen welcomed a conclusion of this type, holding that explanatory power is not an objective virtue of scientific theories.
The current state of scientific explanation is thus highly fragmentary. Although many philosophers hold that explanations trace causes, there is still considerable disagreement about whether or not the notion of causation should be analyzed and, if so, how. The question of whether theoretical explanation can always be construed in causal terms remains open. It is unclear whether unifying the phenomena is an explanatory virtue and how a satisfactory notion of unification should be understood. Perhaps most fundamentally, there are controversies about whether there is a single notion of explanation that applies to all sciences, all contexts, and all periods and about whether explanatory power counts as an objective quality of theories.
Scientific Laws
Similar uncertainties affect recent discussions of scientific laws. As already noted, logical empiricism faced a difficult problem in distinguishing between genuine laws and accidental generalizations. Just as theorists of explanation sometimes liberated themselves from hard problems by invoking a concept hitherto held as taboo—the notion of causation—so too some philosophers championed an idea of natural necessity and tried to characterize it as precisely as possible. Others, more sympathetic to Hume’s suspicions, continued the logical-empiricist project of analyzing the notion independently of the concept of natural necessity. The most important approach along these lines identifies the laws of nature as the generalizations that would figure in the best systematization of all natural phenomena. This suggestion fits naturally with the unificationist approach to explanation but encounters similar difficulties in articulating the idea of a “best systematization.” Perhaps more fundamentally, it is not obvious that the concept of “all natural phenomena” is coherent (or, even if it is, whether this is something in which science should be interested).
There is an even more basic issue. Why is the notion of a scientific law of any philosophical interest? Within the framework of logical empiricism, and specifically within Hempel’s approach to explanation, there was a clear answer. Explanations depend on laws, and the notion of law is to be explicated without appeal to suspect notions such as natural necessity. But Hempel’s approach is now defunct, and many contemporary philosophers are suspicious of the old suspicions, prepared to be more tolerant of appeals to causation and natural necessity. What function, then, would an account of laws now serve?

Francis Crick (1916–2004)
Perhaps the thought is that the search for the laws of nature is central to the scientific enterprise. But, to begin with, the scientific habit of labeling certain statements as “laws” seems extremely haphazard. There are areas, moreover, in which it is hard to find any laws—large tracts of the life and earth sciences, for example—and yet scientists in these areas are credited with the most important discoveries. James Watson and Francis Crick (1916–2004) won a Nobel Prize for one of the greatest scientific achievements of the 20th century (indeed, arguably the most fruitful), but it would be hard to state the law that they discovered. Accordingly, philosophers of science are beginning to abandon the notion that laws are central to science, focusing instead on the search for symmetries in physics, on the differing uses of approximate generalizations in biology, and on the deployment of models in numerous areas of the sciences.
Scientific theories
The axiomatic conception

Georg Wilhelm Friedrich Hegel (1770–1831)
In similar fashion, contemporary philosophy of science is moving beyond the question of the structure of scientific theories. For a variety of reasons, that question was of enormous importance to the logical positivists and to the logical empiricists. Mathematical logic supplied a clear conception: a theory is a collection of statements (the axioms of the theory) and their deductive consequences. The logical positivists showed how this conception could be applied in scientific cases—one could axiomatize the theory of relativity, for example. Nor was the work of axiomatization an idle exercise, for the difficulties of formulating a precise criterion of cognitive significance (intended to separate good science from meaningless philosophical discussion) raised questions about the legitimacy of the special vocabulary that figures in scientific theories. Convinced that the sound and fury of German metaphysics—references to “Absolute Spirit” by Georg Wilhelm Friedrich Hegel (1770–1831) and talk of “the Nothing” by Martin Heidegger (1889–1976)—signified, indeed, nothing, logical positivists (and logical empiricists) recognized that they needed to show how terms such as electron and covalent bond were different.

Martin Heidegger (1889–1976)
They began from a distinction between two types of language. Observational language comprises all the terms that can be acquired by presentation of observable samples. Although they were skeptical about mixing psychology and philosophy, logical empiricists tacitly adopted a simple theory of learning: children can learn terms such as red by being shown appropriate swatches, hot by holding their hands under the right taps, and so forth. Logical empiricists denied that this observational vocabulary would suffice to define the special terms of theoretical science, the theoretical language that seemed to pick out unobservable entities and properties. Conceiving of theories as axiomatic systems, however, they drew a distinction between two types of axioms. Some axioms contain only theoretical vocabulary, while others contain both theoretical and observational terms. The latter, variously characterized as “correspondence rules” or “coordinating definitions,” relate the theoretical and observational vocabularies, and it is through them that theoretical terms acquire what meaning they have.
The last formulation blurs an important difference between two schools within logical empiricism. According to one school, the theoretical terms are “partially interpreted” by the correspondence rules, so, for example, if one such rule is that an electron produces a particular kind of track in a cloud chamber, then many possibilities for the meaning of the previously unfamiliar term electron are ruled out. A more radical school, instrumentalism, held that, strictly speaking, the theoretical vocabulary remains meaningless. Instrumentalists took scientific theories to be axiomatic systems only part of whose vocabulary—the observational language—is interpreted; the rest is a formal calculus whose purpose is to yield predictions couched in the observational vocabulary. Even instrumentalists, however, were able to maintain a distinction between serious theoretical science and the much-derided metaphysics, for their reconstructions of scientific theories would reveal the uninterpreted vocabulary as playing an important functional role (a result not to be expected in the metaphysical case).
Logical empiricists debated the merits of the two stances, exploring the difficulties of making precise the notion of partial interpretation and the possibility of finding axiomatic systems that would generate all the observational consequences without employing any theoretical vocabulary. Their exchanges were effectively undercut by the American philosopher Hilary Putnam, who recognized that the initial motivation for the approach to theories was deeply problematic. In their brief sketches of the differences between the two languages, logical empiricists had conflated two distinctions. On the one hand there is a contrast between things that can be observed and things that cannot—the observable-unobservable distinction; on the other hand, there is the difference between terms whose meanings can be acquired through demonstration and those whose meanings cannot be acquired in this way—the observational-theoretical distinction. It is a mistake to believe that the distinctions are congruent, that observational terms apply to observable things and theoretical terms to unobservable things. In the first place, many theoretical terms apply to observables (spectroscope is an example). More important, many terms learnable through demonstration apply to unobservables—in Putnam’s telling example, even small children learn to talk of “people too little to see.”

John Dalton (1766–1844)
Once the second point was appreciated, the way was open for introducing theoretical vocabulary that logical empiricism had never taken seriously (even though many eminent scientists and gifted science teachers had often developed such modes of conveying meaning). One can see that the term part might be learned in connection with pieces of observable objects and that its use might cover unobservable things as well, so the specification of atoms as “parts of all matter that themselves have no parts” (whatever its merits today) might have served the contemporaries of John Dalton (1766–1844), an early developer of atomic theory, as a means of appreciating what he was claiming.
The semantic conception
Starting in the 1960s, philosophers of science explored alternative approaches to scientific theories. Prominent among them was the so-called semantic conception, originally formulated by Patrick Suppes, according to which theories are viewed as collections of models together with hypotheses about how these models relate to parts of nature. Versions of the semantic conception differ in their views about the character of models, sometimes taking models to be abstract mathematical structures, susceptible to precise formal specifications, and sometimes taking them to be more concrete (as chemists do, for example, when they build models of particular molecules).
The semantic conception of theories has several attractive features. First, unlike the older approach, it provides a way of discussing aspects of science that are independent of the choice of a particular language. Second, it appears to do far more justice to areas of science in which theoretical achievements resist axiomatization. Darwinian evolutionary theory is a case in point. During the heyday of the axiomatic approach, a few philosophers attempted to show how the theory of evolution could be brought within the orthodox conception of theories, but their efforts tended to produce formal theories that bordered on triviality. The consequent debates about whether the theory of evolution was more than a tautology should have generated serious philosophical embarrassment. Philosophers deploying the semantic conception, by contrast, shed light on theoretical issues that arise in contemporary evolutionary biology.
Finally, the semantic conception is far better suited to an aspect of the sciences that was frequently neglected, the practice of idealization. Instead of thinking of scientists as aspiring to offer literally correct descriptions of general features of the world, the semantic conception supposes that they propose models accompanied by claims that particular parts of nature correspond to these models in specific respects and to specific degrees.
When the ways in which meaning accrued to theoretical vocabulary constituted a burning question for the philosophy of science, it was natural to adopt an axiomatic approach to scientific theories and to focus on the connections between theoretical terms and language that are more readily understood (and, to the extent that questions remain in the wake of Putnam’s insights about the theoretical-observational and observable-unobservable distinctions, the axiomatic approach can still be of value in this area). Similarly, when a philosopher (or scientist) wonders whether a specific assumption or a particular choice of a parameter value is necessary, the device of axiomatization helps to resolve the question; given an axiomatic presentation, one can explore whether every derivation using the assumption can be transformed into one without. However, when the topic under study is a science in which there are few generalizations, or when one is concerned to elucidate issues about idealization in science, the semantic conception seems much more illuminating. Finally, in probing the dynamics of large-scale change in science—reconstructing the ways in which Darwin won acceptance for his evolutionary theory, for example—the concepts introduced by Kuhn and those who reacted to his work seem more readily applicable. The insistence that there must be a unique answer to what scientific theories really are seems like misplaced dogmatism that obstructs philosophical inquiry.
Unification and reduction
One large question about scientific theories that excites philosophical and scientific attention concerns the possibility of producing a single theory that will encompass the domains of all the sciences. Many thinkers are attracted by the idea of a unified science, or by the view that the sciences form a hierarchy. There is a powerful intuitive argument for this attitude. If one considers the subject matter of the social sciences, for example, it seems that social phenomena are the product of people standing in complicated relations to each other and acting in complicated ways. These people, of course, are complex biological and psychological systems. Their psychological activity is grounded in the neural firings in their brains. Hence, people are intricate biological systems. The intricacies of biology are based on the choreography of molecular reactions within and between individual cells. Biology, then, is very complicated chemistry. Chemical reactions themselves involve the forming and breaking of bonds, and these are matters of microphysics. At the end of the day, therefore, all natural phenomena, even those involving interactions between people, are no more than an exceptionally complicated series of transactions between the ultimate physical constituents of matter. A complete account of those ultimate constituents and their interactions would thus amount to a “theory of everything.”
This argument builds on some important scientific discoveries. Whereas earlier generations thought that living things must contain something more than complex molecules (some “vital substance,” say), or that there must be something more to thinking beings than intricate brains (an “immaterial mind,” for example), contemporary biology and contemporary neuroscience showed that there is no need for such hypotheses. Given the firm consensus of contemporary science, there is a constitutive hierarchy: all molecules are made out of fundamental particles; all organic systems are made out of molecules; people are organic systems; and societies are composed of people. Yet there is a difference between a constitutive hierarchy of the things studied by various sciences and a reductive hierarchy of those sciences. Biology studies organisms, entities composed of molecules (and nothing more); it does not follow that biology can be reduced to the science that studies molecules (chemistry).
To understand this distinction it is necessary to have a clear concept of reduction. The most influential such proposal, by Ernest Nagel, was made within the framework of the axiomatic conception of scientific theories. Nagel suggested that one theory is reduced to another when the axioms of the reduced theory can be derived from the axioms of the reducing theory, supplemented with principles (“bridge principles”) that connect the language of the reduced theory with that of the reducing theory. So, for example, to reduce genetics to biochemistry, one would show how the principles of genetics follow from premises that include the principles of biochemistry together with specifications in biochemical language of the distinctive vocabulary of genetics (terms such as gene, allele, and so forth).
Many philosophers criticized the idea of unified science by arguing that, when reduction is understood in Nagel’s sense, the constitutive hierarchy does not correspond to a reductive hierarchy. They focused specifically on the possibility of reducing biology to physics and chemistry and of reducing psychology to neuroscience. Attempts at reduction face two major obstacles. First, despite serious efforts to formulate them, there are as yet no bridge principles that link the vocabulary of biology to that of chemistry or the vocabulary of psychology to that of neuroscience. It is evidently hard to think of chemical specifications of the property of being a predator, or neurological specifications of the generic state of desiring to eat ice cream, but the problem arises even in more tractable cases, such as that of providing chemical conditions for being a gene. Every gene is a segment of nucleic acid (DNA in most organisms, RNA in retroviruses); the challenge is to find a chemical condition that distinguishes just those segments of nucleic acid that count as genes. Interestingly, this is a serious research question, for, if it were answered, molecular biologists engaged in genomic sequencing would be able to discover the genes in their sequence data far more rapidly than they are now able to do. The fact that the question is still unanswered is due to the fact that genes are functional units that lack any common chemical structure (beyond being nucleic acids, of course). The language of genetics and the language of chemistry classify the molecules in different ways, and, because of this cross-classification, there is no possibility of reduction.
The second difficulty turns on points about explanation. Imagine a small child who is tired and hot. He is dragged by his harried parent past an ice-cream stand. The child starts to scream. One might explain this behaviour by saying that the child saw the ice-cream stand and expressed a desire for ice cream, and the parent refused. Suppose further that a friendly neuroscientist is able to trace the causal history of neural firings in the child’s brain. Would this replace the everyday explanation? Would it deepen it? Would it even constitute an intelligible account of what had happened? A natural inclination is to suspect that the answer to all these questions is no.
A friend of the unity of science, on the other hand, might respond by claiming that this natural inclination arises only because one is ignorant of the neuroscientific details. If one were able actually to formulate the account of the neural causes and to follow the details of the story, one would obtain greater insight into the child’s behaviour and perhaps even be inclined to abandon the explanation based in everyday psychological concepts (“folk psychology”).

Gregor Mendel (1822–84)
Once again, the objection to unified science can be posed in a case in which it is possible to give at least some of the biochemical details. One of the best candidates for a regularity in genetics is a revised version of the rule of independent assortment devised by Gregor Mendel (1822–84): genes on different chromosomes are distributed independently when the gametes are formed (at meiosis). Classical (premolecular) genetics provides a satisfying account of why this is so. In sexually reproducing organisms, the gametes (sperm and ova) are formed in a process in which the chromosomes line up in pairs; after some recombination between members of each pair, one chromosome from each pair is transmitted to the gamete. This kind of pairing and separation will produce independent assortments of chromosomal segments (including genes), no matter what the chromosomes are made of and no matter what the underlying molecular mechanisms. If one were now told a complicated story about the sequence of chemical reactions that go on in all instances of meiosis—it would have to be very complicated indeed, since the cases are amazingly diverse—it would add nothing to the original explanation, for it would fail to address the question “Why do genes on different chromosomes assort independently?” The question is completely resolved once one understands that meiosis involves a specific type of pairing and separation.
The points just made do not imply that ventures in molecular biology are unfruitful or that future research in neuroscience will be irrelevant to psychology. To say that not all explanations in genetics can be replaced by molecular accounts is quite compatible with supposing that molecular biology often deepens the perspective offered by classical genetics (as in the cases of mutation, gene replication, gene transcription and translation, and a host of other processes). Moreover, to deny the possibility of reduction in Nagel’s sense is not to exclude the possibility that some other notion might allow reducibility on a broader scale. It is important, however, to understand this particular failure of the idea of unified science, because when scientists (and others) often think about a “theory of everything,” they are envisaging a set of principles from which explanations of all natural phenomena may be derived. That kind of “final theory” is a pipe dream.

Nancy Cartwright
Proponents of the semantic conception of theories explored alternative notions of reduction. For some philosophers, however, conceiving of theories as families of models provided a useful way of capturing what they saw as the piecemeal character of contemporary scientific work. Instead of viewing the sciences as directed at large generalizations, they suggested that researchers offer a patchwork of models, successful in different respects and to different degrees at characterizing the behaviour of bits and pieces of the natural world. This theme was thoroughly pursued by the American philosopher Nancy Cartwright, who emerged in the late 20th century as the most vigorous critic of unified science.
Cartwright opposed the kind of reduction considered above (“vertical reduction”), but she believed that the standard critiques did not go far enough. She argued that philosophers should also be skeptical of “horizontal reduction,” the idea that models and generalizations have broad scope. Traditional philosophy of science took for granted the possibility of extrapolating regularities beyond the limited contexts in which they can be successfully applied. As a powerful illustration, Cartwright invited readers to consider their confidence in Newton’s second law, which states that force is equal to the product of mass and acceleration (see Newton’s laws of motion). The law can be used to account for the motions of particular kinds of bodies; more exactly, the solar system, the pendulum, and so forth can be modeled as Newtonian systems. There are many natural settings, however, in which it is hard to create Newtonian order. Imagine, for example, someone dropping a piece of paper money from a high window overlooking a public square. Does Newton’s second law determine the trajectory? A standard response would be that it does in principle, though in practice the forces operating would be exceedingly hard to specify. Cartwright questioned whether this reponse is correct. She suggested instead that modern science should be thought of in terms of a history of successful building of Newtonian models for a limited range of situations and that it is only a “fundamentalist faith” that such models can be applied everywhere and always. It is consistent with current scientific knowledge, she argued, that the world is thoroughly “dappled,” containing some pockets of order in which modeling works well and pockets of disorder that cannot be captured by the kinds of models that human beings can formulate.
Scientific Change
Although some of the proposals discussed in the previous sections were influenced by the critical reaction to logical empiricism, the topics are those that figured on the logical-empiricist agenda. In many philosophical circles, that agenda continues to be central to the philosophy of science, sometimes accompanied by the dismissal of critiques of logical empiricism and sometimes by an attempt to integrate critical insights into the discussion of traditional questions. For some philosophers, however, the philosophy of science was profoundly transformed by a succession of criticisms that began in the 1950s as some historically minded scholars pondered issues about scientific change.
The historicist critique was initiated by the philosophers N.R. Hanson (1924–67), Stephen Toulmin, Paul Feyerabend (1924–94), and Thomas Kuhn. Although these authors differed on many points, they shared the view that standard logical-empiricist accounts of confirmation, theory, and other topics were quite inadequate to explain the major transitions that have occurred in the history of the sciences. Feyerabend, the most radical and flamboyant of the group, put the fundamental challenge with characteristic brio: if one seeks a methodological rule that will account for all of the historical episodes that philosophers of science are inclined to celebrate—the triumph of the Copernican system, the birth of modern chemistry, the Darwinian revolution, the transition to the theories of relativity, and so forth—then the best candidate is “anything goes.” Even in less-provocative forms, however, philosophical reconstructions of parts of the history of science had the effect of calling into question the very concepts of scientific progress and rationality.
A natural conception of scientific progress is that it consists in the accumulation of truth. In the heyday of logical empiricism, a more qualified version might have seemed preferable: scientific progress consists in accumulating truths in the “observation language.” Philosophers of science in this period also thought that they had a clear view of scientific rationality: to be rational is to accept and reject hypotheses according to the rules of method, or perhaps to distribute degrees of confirmation in accordance with Bayesian standards. The historicist challenge consisted in arguing, with respect to detailed historical examples, that the very transitions in which great scientific advances seem to be made cannot be seen as the result of the simple accumulation of truth. Further, the participants in the major scientific controversies of the past did not divide neatly into irrational losers and rational winners; all too frequently, it was suggested, the heroes flouted the canons of rationality, while the reasoning of the supposed reactionaries was exemplary.
The work of Thomas Kuhn
In the 1960s it was unclear which version of the historicist critique would have the most impact, but during subsequent decades Kuhn’s monograph emerged as the seminal text. The Structure of Scientific Revolutions offered a general pattern of scientific change. Inquiries in a given field start with a clash of different perspectives. Eventually one approach manages to resolve some concrete issue, and investigators concur in pursuing it—they follow the “paradigm.” Commitment to the approach begins a tradition of normal science in which there are well-defined problems, or “puzzles,” for researchers to solve. In the practice of normal science, the failure to solve a puzzle does not reflect badly on the paradigm but rather does so on the skill of the researcher. Only when puzzles repeatedly prove recalcitrant does the community begin to develop a sense that something may be amiss; the unsolved puzzles acquire a new status, being seen as anomalies. Even so, the normal scientific tradition will continue so long as there are no available alternatives. If a rival does emerge, and if it succeeds in attracting a new consensus, then a revolution occurs: the old paradigm is replaced by a new one, and investigators pursue a new normal scientific tradition. Puzzle solving is now directed by the victorious paradigm, and the old pattern may be repeated, with some puzzles deepening into anomalies and generating a sense of crisis, which ultimately gives way to a new revolution, a new normal scientific tradition, and so on indefinitely.
Kuhn’s proposals can be read in a number of ways. Many scientists have found that his account of normal science offers insights into their own experiences and that the idea of puzzle solving is particularly apt. In addition, from a strictly historical perspective, Kuhn offered a novel historiography of the sciences. However, although a few scholars attempted to apply his approach, most historians of science were skeptical of Kuhnian categories. Philosophers of science, on the other hand, focused neither on his suggestions about normal science nor on his general historiography, concentrating instead on Kuhn’s treatment of the episodes he termed “revolutions.” For it is in discussing scientific revolutions that he challenged traditional ideas about progress and rationality.
At the basis of the challenge is Kuhn’s claim that paradigms are incommensurable with each other. His complicated notion of incommensurability begins from a mathematical metaphor, alluding to the Pythagorean discovery of numbers (such as Square root of√2) that could not be expressed as rationals; irrational and rational lengths share no common measure. He considered three aspects of the incommensurability of paradigms (which he did not always clearly separate). First, paradigms are conceptually incommensurable in that the languages in which they describe nature cannot readily be translated into one another; communication in revolutionary debates, he suggested, is inevitably partial. Second, paradigms are observationally incommensurable in that workers in different paradigms will respond in different ways to the same stimuli—or, as he sometimes put it, they will see different things when looking in the same places. Third, paradigms are methodologically incommensurable in that they have different criteria for success, attributing different values to questions and to proposed ways of answering them. In combination, Kuhn argued, these forms of incommensurability are so deep that, after a scientific revolution, there will be a sense in which scientists work in a different world.
These striking claims are defended by considering a small number of historical examples of revolutionary change. Kuhn focused most on the Copernican revolution, on the replacement of the phlogiston theory with Lavoisier’s new chemistry, and on the transition from Newton’s physics to the special and general theories of relativity. So, for example, he supported the doctrine of conceptual incommensurability by arguing that pre-Copernican astronomy could make no sense of the Copernican notion of planet (within the earlier astronomy, the Earth itself could not be a planet), that phlogiston chemistry could make no sense of Lavoisier’s notion of oxygen (for phlogistonians, combustion is a process in which phlogiston is emitted, and talk of oxygen as a substance that is absorbed is quite wrongheaded), and that theories of relativity distinguish two notions of mass (rest mass and proper mass), neither of which makes sense in Newtonian terms.
All of these arguments received detailed philosophical attention, and it became apparent that the conclusions can be met by adopting a more sophisticated approach to language than that presupposed by Kuhn. The crucial issue is whether the languages of rival paradigms suffice to identify the objects and properties referred to in the terms of the other. Although Kuhn was right to see difficulties here, it is an exaggeration to suppose that the identification is impossible. From Lavoisier’s perspective, for example, the antiquated term dephlogisticated air sometimes means “what remains when phlogiston is removed from the air” (in which case, because there is no such substance as phlogiston, the term fails to pick out anything in the world). But at other times it is used to designate a specific gas (oxygen) that both groups of chemists have isolated. As far as conceptual incommensurability is concerned, it is possible to see Kuhn’s examples as cases in which communication is tricky but not impossible and in which the parties respond to and talk about a common world.

Johannes Kepler (1571–1630)

Tycho Brahe (1546–1601)
The thesis of observational incommensurability is best illustrated via Kuhn’s example of the Copernican revolution. In the late 16th century, Johannes Kepler (1571–1630), a committed follower of Copernicus, assisted the great astronomer Tycho Brahe (1546–1601), who believed that the Earth is at rest. Kuhn imagined Tycho and Kepler watching the sunrise together, and, like Hanson before him, suggested that Tycho would see a moving Sun coming into view, while Kepler would see a static Sun becoming visible as the Earth rotates.
Evidently Tycho and Kepler might report their visual experiences in different ways. Nor should it be supposed that there is some privileged “primitive” language—a language that picks out shapes and colours, perhaps—in which all observers can describe what they see and reach agreement with those who are similarly situated. But these points, while they may have been neglected in earlier philosophy of science, do not yield the radical Kuhnian conclusions. In the first place, the difference in the experiential reports is quite compatible with the perception of a common object, possibly described correctly by one of the participants, possibly accurately reported by neither; both Tycho and Kepler see the Sun, and both perceive the relative motion of Sun and Earth. Furthermore, although there may be no bedrock language of uncontaminated observation to which they can retreat, they have available to them forms of description that presuppose only shared commonsense ideas about objects in the vicinity. If they become tired of exchanging their preferred reports—“I see a moving Sun,” “I see a stationary Sun becoming visible through the Earth’s rotation”—they can both agree that the orange blob above the hillside is the Sun and that more of it can be seen now than could be seen two minutes ago. There is no reason, then, to deny that Tycho and Kepler experience the same world or to suppose that there are no observable aspects of it about which they can reach agreement.
The thesis of methodological incommensurability can also be illustrated through the Copernican example. After the publication of Copernicus’s system in 1543, professional astronomers quickly realized that, for any Sun-centred system like Copernicus’s, it would be possible to produce an equally accurate Earth-centred system, and conversely. How could the debate be resolved? One difference between the systems lay in the number of technical devices required to generate accurate predictions of planetary motions. Copernicus did better on this score, using fewer of the approved repertoire of geometrical tricks than his opponents did. But there was also a tradition of arguments against the possibility of a moving Earth. Scholars had long maintained, for example, that, if the Earth moved, objects released from high places would fall backwards, birds and clouds would be left behind, loose materials on the Earth’s surface would be flung off, and so forth. Given the then-current state of theories of motion, there were no obvious errors in these lines of reasoning. Hence, it might have seemed that a decision about the Earth’s motion must involve a judgment of values (perhaps to the effect that it is more important not to introduce dynamical absurdities than to reduce the number of technical astronomical devices). Or perhaps the decision could be made only on faith—faith that answers to questions about the behaviour of birds and clouds would eventually be found. (This illustrates a point raised in an earlier section: namely, that attempts to justify the choice of a hypothesis rest on expectations about future discoveries.
Methodological incommensurability presents the most severe challenge to views about progress and rationality in the sciences. In effect, Kuhn offered a different version of the underdetermination thesis, one more firmly grounded in the actual practice of the sciences. Instead of supposing that any theory has rivals that make exactly the same predictions and accord equally well with all canons of scientific method, Kuhn suggested that certain kinds of large controversies in the history of science pit against each other approaches with different virtues and defects and that there is no privileged way to balance virtues and defects. The only way to address this challenge is to probe the examples, trying to understand the ways in which various kinds of trade-offs might be defended or criticized.
Responses
One way to think about the Copernican example (and other Kuhnian revolutions) is to recognize the evolution of the debates. In 1543 the controversy might have seemed quite unsettled; the simplification of technical machinery might have inspired some people to work further on the Copernican program, while the dynamical problems posed by the moving Earth might have prompted others to articulate the more traditional view. If neither choice can be seen as uniquely rational, neither can be dismissed as unreasonable.
Later, after Kepler’s proposals of elliptical orbits, Galileo’s telescopic observations, and Galileo’s consideration of the dynamical arguments, the balance shifted. Copernicanism had shed a number of its defects, while the traditional view had acquired some new ones. Since both approaches still faced residual problems—sciences rarely solve all the problems that lie within their domain, and there are always unanswered questions—it would still have been possible in principle to give greater weight to the virtues of traditional astronomy or to the defects of Copernicanism. By the mid-17th century, however, it would have been unreasonable to adopt any value judgment that saw the achievements of the tradition as so glorious, or the deficiencies of the rival as so severe, that Copernicanism should still be rejected. That type of valuation would be akin to preferring a decrepit jalopy, with a nonfunctioning engine and a rusting chassis, to a serviceable new car solely on the grounds that the old wreck had a more appealing hood ornament.
Although a few philosophers of science tried to make this line of response to Kuhn’s challenge more general and more precise, many contemporary discussions seem to embody one of two premature reactions. Some hold that the worries about revolutionary change have been adequately addressed and that the philosophy of science can return to business as usual. Others conclude that Kuhn’s arguments are definitive and that there is no hope of salvaging the progressiveness and rationality of science (some more-radical versions of this position will be considered in the next two sections).
Kuhn’s discussions of incommensurability challenge claims about the rationality of science by asking whether it is possible to show how the accepted views of method and justification would allow the resolution of scientific revolutions. The philosophical task here is to adapt one of the existing approaches to confirmation (Bayesianism or eliminativism, for example) to the complex contexts Kuhn presents or, if that cannot be done, to formulate new methodological rules, rules that can be defended as conditions of rationality that will apply to these contexts.
Equally, the points about incommensurability challenge the thesis that the sciences are progressive by denying the possibility of understanding the history of science as a process of accumulating truth. Here the philosopher of science needs to provide an account of progress in terms of convergence on the truth or to show how progress can be understood in other terms.
Scientific Realism
Issues about scientific realism had already emerged within the logical-empiricist discussions of scientific theories. Philosophers who held that theoretical language was strictly meaningless, taking theories to be instruments for the prediction of statements formulated in an observational vocabulary, concluded that the theoretical claims of the sciences lack truth value (i.e., are neither true nor false) and that use of the formalism of theoretical science does not commit one to the existence of unobservable entities. Instrumentalists suggested that terms such as electron should not be taken to refer to minute parts of matter; they simply function in a formal calculus that enables one to make true predictions about observables. By contrast, philosophers who emphasized the explanatory power of scientific theories argued that one cannot make sense of theoretical explanation unless one recognizes the reality of unobservable entities; one can understand the character of chemical bonds and see why elements combine in the ways they do if one takes the proposals about electrons filling shells around nuclei seriously but not if one supposes that electron, shell, and nucleus are mere façons de parler.
An initial dispute about scientific realism thus focused on the status of unobservables. In an obvious sense this was a debate about democracy with respect to scientific language: realists and instrumentalists alike believed that the concept of truth made good sense for a portion of scientific language—the observation language—though they differed as to whether this privileged status should be extended to scientific language as a whole.
The antirealism of van Fraassen, Laudan, and Fine
In the 1990s, however, the controversy about the reality of unobservables was revived through the development of sophisticated antirealist arguments. Van Fraassen advocated a position that he called “constructive empiricism,” a view intended to capture the insights of logical empiricism while avoiding its defects. A champion of the semantic conception of theories, he proposed that scientists build models that are designed to “save the phenomena” by yielding correct predictions about observables. To adopt the models is simply to suppose that observable events and states of affairs are as if the models were true, but there is no need to commit oneself to the existence of the unobservable entities and processes that figure in the models. Rather, one should remain agnostic. Because the aim of science is to achieve correct predictions about observables, there is no need to assume the extra risks involved in commitment to the existence of unobservables.
A different antirealist argument, presented by Laudan, attacks directly the “ultimate argument” for realism. Laudan reflected on the history of science and considered all the past theories that were once counted as outstandingly successful. He offered a list of outmoded theories, claiming that all enjoyed successes and noting that not only is each now viewed as false, but each also contains theoretical vocabulary that is now recognized as picking out nothing at all in nature. If so many scientists of past generations judged their theories to be successful and, on that basis, concluded that they were true, and if, by current lights, they were all wrong, how can it be supposed that the contemporary situation is different—that, when contemporary scientists gesture at apparent successes and infer to the approximate truth of their theories, they are correct? Laudan formulated a “pessimistic induction on the history of science,” generalizing from the fact that large numbers of past successful theories have proved false to the conclusion that successful contemporary theories are also incorrect.
A third antirealist objection, formulated by both Laudan and Arthur Fine, charges that the popular defenses of realism beg the question. Realists try to convince their opponents by suggesting that only a realist view of unobservables will explain the success of science. In doing so, however, they presuppose that the fact that a certain doctrine has explanatory power provides a reason to accept it. But the point of many antirealist arguments is that allegations about explanatory power have no bearing on questions of truth. Antirealists are unpersuaded when it is suggested, for example, that a hypothesis about atoms should be accepted because it explains observable chemical phenomena. They will be equally unmoved when they are told that a philosophical hypothesis (the hypothesis of scientific realism) should be accepted because it explains the success of science. In both instances, they want to know why the features of the hypotheses to which realists draw attention—the ability of those hypotheses to generate correct conclusions about observable matters—should be taken as indicators of the truth of the hypotheses.
“Piecemeal” realism
Realists tried to respond to these powerful points. One popular rejoinder is that antirealists cannot account for important facets of scientific practice. Thus, it is sometimes suggested that the routine method of conjoining theoretical claims from different scientific theories (as, for example, when earth scientists draw on parts of physics and chemistry) would not make sense unless there was a serious commitment to the approximate truth of the theoretical principles. Alternatively, one may take the practice of choosing certain kinds of experiments (experiments taken to be particularly revealing) to reflect a belief in the reality of underlying entities; thus, a medical researcher might choose a particular class of animals to inject with an antibiotic on the grounds that the concentration of bacteria in those animals is likely to be especially high.
Or the realist can attempt to argue that the kinds of inferences that the antirealist will acknowledge as unproblematic—for example, the generalization from observed samples to conclusions about a broader population of observable things—can be made only in light of an understanding of unobservable entities and mechanisms. One cannot tell what makes a sample suitable for generalization unless one has views about the ways in which that sample might be biased, and that will typically entail beliefs about relevant unobservable causes. Antirealists must either show that they have the resources to make sense of these and other features of scientific practice or offer reasons for thinking that the procedures in question should be revised.
Laudan’s pessimistic induction on the history of science attracted considerable scrutiny. Realists pointed out, correctly, that his list of successful past theories contains a number of dubious entries. Thus, it would be hard to defend the medieval theory of disease as caused by an imbalance of humours as particularly successful, and similar judgments apply to the geological catastrophism of the 18th century and the phlogiston theory of chemical combination.
Yet it is impossible to dismiss all of Laudan’s examples. One of his most telling points is that the account of the wave propagation of light of Augustin-Jean Fresnel (1788–1827) was spectacularly successful in explaining and predicting facts about diffraction and interference; one of its most dramatic successes, for example, was the prediction of the Poisson bright spot, a point of light at the centre of the shadow of a small rotating disk. (Ironically, the French mathematician for whom the spot is named, Siméon-Denis Poisson [1781–1840], believed that Fresnel was wrong and that the prediction of the spot was an absurd consequence of a false theory.) Fresnel, however, based his theory on the hypothesis that light waves are propagated in an all-pervading ether. Since contemporary science rejects the ether, it must also reject Fresnel’s theory as false.
This example is especially instructive, because it points to a refinement of realism. Contemporary optics takes over Fresnel’s mathematical treatment of wave propagation but denies the need for any medium in which the propagation takes place. So part of his theory is honoured as approximately correct, while the rest is seen as going astray because of Fresnel’s belief that any wave motion needs a medium in which the waves are propagated. Faced with a choice between saying that Fresnel’s theory is correct and saying that it is wrong, contemporary scientists would opt for the negative verdict. One would do greater justice to the situation, however, not by treating the theory as a whole but by judging some parts to be true and others false. Furthermore, when Fresnel’s work is analyzed in this way, it can be seen that the correct parts are responsible for its predictive successes. Appeals to the ether play no role when Fresnel is accounting for experimental data about interference bands and diffraction patterns. Hence, this example supports the realist linkage of success and truth by revealing that the parts of theory actually put to work in generating successful predictions continue to be counted as correct.

These sophisticated proposals and the intricate arguments urged in favour of them contrast with a more widely accessible critique of the idea of “scientific truth” that also starts from Kuhn’s suspicion that the idea of truth as correspondence to mind-independent reality makes no sense. Inspired by Kuhn’s recognition of the social character of scientific knowledge (a paradigm is, after all, something that is shared by a community), a number of scholars proposed a more thoroughly sociological approach to science. Urging that beliefs acclaimed as “true” or “false” be explained in the same ways, they concluded that truth must be relativized to communities: a statement counts as true for a community just in case members of that community accept it. (For an account of this view in the context of ethics.
The proposal for a serious sociology of scientific knowledge should be welcomed. As the sociologists David Bloor and Barry Barnes argued in the early 1970s, it is unsatisfactory to suppose that only beliefs counted as incorrect need social and psychological explanation. For it would be foolish to suggest that human minds have some attraction to the truth and that cases in which people go astray must be accounted for in terms of the operation of social or psychological biases that interfere with this natural aptitude. All human beliefs have psychological causes, and those causes typically involve facts about the societies in which the people in question live. A comprehensive account of how an individual scientist came to some novel conclusion would refer not only to the observations and inferences that he made but to the ways in which he was trained, the range of options available for pursuing inquiries, and the values that guided various choices—all of which would lead, relatively quickly, to aspects of the social practice of the surrounding community. Barnes and Bloor were right to advocate symmetry, to see all beliefs as subject to psychological and sociological explanation.
Yet another attempt to argue that the only serviceable notion of truth reduces to social consensus begins from the strong Quinean thesis of the underdetermination of theories by experience. Some historians and sociologists of science maintained that choices of doctrine and method are always open in the course of scientific practice. Those choices are made not by appealing to evidence but by drawing on antecedently accepted social values or, in some instances, by simultaneously “constructing” both the natural and the social order. The best versions of these arguments attempt to specify in some detail what the relevant alternatives are; in such cases, as with Kuhn’s arguments about the irresolvability of scientific revolutions, philosophical responses must attend to the details.
Unfortunately, such detailed specifications are relatively rare, and the usual strategy is for the sociological critique to proceed by invoking the general thesis of underdetermination and to declare that there are always rival ways of going on. As noted earlier, however, a blanket claim about inevitable underdetermination is highly suspect, and without it sociological confidence in “truth by consensus” is quite unwarranted.
Issues about scientific realism and the proper understanding of truth remain unsettled. It is important, however, to appreciate what the genuine philosophical options are. Despite its popularity in the history and sociology of science, the crude sociological reduction of truth is not among those options. Yet, like history, the sociological study of science can offer valuable insights for philosophers to ponder.
Science, Society, And Values
Many philosophers believe that it can. It is worth recalling, however, that one of the principal influences on the development of modern science, Francis Bacon, was explicitly concerned with science as a social endeavour and that the founders of the Royal Society attempted to create an institution that would follow Bacon’s direction. Furthermore, as the discussion of the Copernican revolution above seems to show, the notion of social (or collective) rationality is philosophically important. As of 1543, the choice between Copernicanism and the traditional Earth-centred astronomy was unclear; the discussion evolved because some scientists were willing to commit themselves to exploring each of the two views. That was a good thing—but the good was a feature of the community and not of the individuals. Had one of the rival positions languished and all of the community members dedicated themselves to a single point of view, it would have been hard to accuse any single individual of a failure of rationality. It would not, however, have been a rational community.
This is an elementary example of a social feature of science that calls for a broader approach to rationality than what is standard in philosophical discussions. One way of understanding why some methods or principles deserve the label “rational” is to suggest that the ultimate standard for appraising them is in terms of their capacity to yield true beliefs. By the same token, one could suppose that institutions or methods of organizing inquiry count as rational if they are likely to enhance the chances of a future state in which members of the community believe the truth. (There are lurking complications here, which will emerge shortly, but they can be ignored for the moment.) It is not hard to think of ways of promoting diversity in a scientific community. Perhaps the educational system could encourage some people to take large risks and others to pursue relatively safe strategies. Perhaps the system of rewards for scientific achievement could be set up in such a way that individuals would gravitate to lines of research that looked neglected. Standard techniques of mathematical modeling reveal that institutional structures like these produce collectively rational outcomes in situations that seem to recur in the history of the sciences. One thus discovers that factors one might have thought of as antithetical to the rational pursuit of truth—individual biases or interest in social rewards—actually play a positive role in the collective venture.
Detailed sociological investigation is required to discover the ways in which scientists interact with each other and with parts of the broader society; detailed psychological investigations are needed to understand the ways in which they make choices. A satisfactory philosophical account of the sciences should be just as interested in whether the sociopsychological matrix is conducive to the attainment of truth by the community as it is in whether particular lines or styles of reasoning lead individuals to correct beliefs. At present, however, the sociology and psychology of science are in their infancy, and philosophy has little by way of data on which to build. It is already possible, however, to envisage a future philosophical account that avoids the limitations of the individualistic perspective now current.
Such an account might find that the social structures inherited from the early-modern period are quite satisfactory as a means of pursuing the aims of the sciences (although that would be surprising). Some contemporary philosophers believe that good reasons for thinking this will not be so are already apparent. Pointing to the exclusion, or marginalization, of some groups of people, they suggest that the current collective practice of science is biased toward the realization of a partial set of values. The most vigorous articulation of this perspective is offered in recent feminist philosophy of science.
Feminist themes
There are various ways of pursuing feminist themes in connection with the sciences. An important project often dismissed as too limited, is to document the ways in which women have been excluded from participation in research projects. More philosophically ambitious is the attempt to show how women’s exclusion led to a bias in the conclusions that scientists accept. Here there is a classic and compelling example: during the 1950s and ’60s, (male) primatologists arrived at hypotheses about territoriality and aggression in the troops of primates they studied; as an increasing number of women entered the field in the 1970s, aspects of primate social life that had been invisible came to be noted, and the old hypotheses were forced to undergo radical revision. The specific moral of this case is that pooling the observations of both men and women may enlarge the range of evidence available to the scientific community; the more general point is that a diversity of social backgrounds and social roles can sometimes provide the most inclusive body of data.
Feminists sometimes wanted to argue for a bolder thesis. Appealing to the general thesis of the underdetermination of theories by evidence, they claimed that choices between equally good rivals are made by introducing considerations of value that reflect the masculine bias of the scientific community. Yet this style of argument works no better in this context than it did in the blanket sociological invocation of underdetermination considered in the last section. Where feminists can make a detailed case for the existence of equivalent rivals, it is important to probe their decision making to see whether an arbitrary choice is being grounded in a problematic way. There is no general reason for believing that evidential considerations always fall short, creating a vacuum that can be filled only by the irruption of masculine values.
The feminist argument does, however, point toward a deeper issue. Once it is understood that science is a social enterprise, it may be supposed that the institutions that guide the development of the sciences absorb major features of the background society, including the privileged position of men, and that this affects the goals set for the sciences and the values placed on certain types of scientific achievements. This form of the feminist critique is extremely important in bringing into the open issues that were skirted in previous discussions and that have been neglected in the traditional philosophy of science. They can best be approached by returning to the unfinished question of the nature of scientific progress.
Progress and values
Suppose that scientific realism succeeds in fighting off challenges to the view that the sciences attain (or accumulate, or converge on) truth. Does this mean that there is now a satisfactory understanding of scientific progress as increasing grasp of truth? Not necessarily. For the truths about nature are too many, and most of them are not worth knowing. Even if one focuses on a small region of the universe—a particular room, say, during the period of an hour—there are infinitely many languages for describing that room and, for each such language, infinitely many true statements about the room during that time. Simply accumulating truth about the world is far too easy. Scientific progress would not be made by dispatching armies of investigators to count leaves or grains of sand. If the sciences make progress, it is because they offer an increasing number of significant truths about the world.
The question of scientific progress is unfinished because this notion of significance was not sufficiently analyzed. Many philosophers wrote either as if the aim of the sciences is to deliver the complete truth about the world (a goal that is not obviously coherent and is surely unattainable) or as if there is some objective notion of significance, given by nature. What might this notion of significance be? Perhaps that the truths desired are the laws of nature or the fundamental principles that govern natural phenomena. But proposals like this are vulnerable to the worries about the role of laws and about the possibility of unified science discussed above. Moreover, many thriving sciences do not seem to be in the business of enunciating laws; there appear to be large obstacles to finding some “theory of everything” that will integrate and subsume all the sciences that have been pursued (let alone those that might be pursued in the future). A sober look at the variety of scientific research undertaken today suggests that the sciences seek true answers to questions that are taken to be significant, either because they arouse people’s curiosity or because they lend themselves to the pursuit of practical goals that people want to achieve. The agenda for research is set not by nature but by society.
At this point, the feminist critique obtains a purchase, for the picture just outlined identifies judgments of value as central to the direction of scientific inquiry—we pursue the truths that matter to us. But who are the “we” whose values enter into the identification of the goals of the sciences? To what extent do the value judgments actually made leave out important constituencies within the human population? These are serious questions, and one of the main contributions of feminist philosophy of science is to bring them to philosophical attention.
The main point, however, is general. An account of the goals of science cannot rest with the bare assertion that the sciences seek truth. Philosophers should offer an analysis of which kinds of truths are important, and, unless they can revive the idea of an “objective agenda set by nature,” they will have to conclude that judgments about human interests and values are part of a philosophical account of science. This means that philosophy of science can no longer confine itself to treating issues that relate to logic, epistemology, and metaphysics (questions about the reconstruction of scientific theories, the nature of natural necessity, and the conditions under which hypotheses are confirmed). Moral and political philosophy will also enter the philosophy of science.
Insofar as philosophers have reflected on the ethics of science, they have often regarded the questions as relatively straightforward. Application of virtually any major moral theory will support restrictions on the kinds of things that can be done to people in scientific experimentation; everyday maxims about honesty will generate the conclusions about fraud and misrepresentation that are routinely made when cases of scientific misconduct surface. These issues about the ways in which scientists are expected to behave in their daily work are superficial; the deeper moral and political questions concern the ways in which the goals of inquiry are set (and, correspondingly, in which progress is understood). One might say, vaguely, that the sciences should pursue those truths whose attainment would best promote the collective good; but this, of course, leaves the hard philosophical task of understanding “the collective good.” How should the divergent interests of different groups of people be weighed? How should the balance between satisfying human curiosity and solving practical problems be struck? How should future gains be judged in relation to short-term demands? Philosophy of science has so far said too little in response to these questions.
Many of the philosophical topics so clearly formulated by the logical positivists and logical empiricists are, rightly, still the focus of 21st-century concern. Increased understanding of the history of the sciences and of the social character of scientific practice has set broader tasks for the philosophy of science. In a world in which the power of scientific research, for good and for ill, is becoming increasingly obvious, it is to be hoped that issues about the values adopted in the pursuit of science will become more central to philosophical discussion.
Source: britannica.com