Ockham’s Razor

Ockham’s Razor, otherwise called the principle of the economy of thought, is often invoked in debate or arguments, usually to discard or count against one or more theories on the basis that another exists that is simpler or more parsimonious. This entry considers the principle, its domain of application and some associated philosophical concerns, using examples from the history of science to illustrate some of the points at issue. It is argued that Ockham’s Razor is neither as useful nor as important as is sometimes claimed.

The Simplest Explanation

The principle of parsimony is typically stated in the form Entia non sunt multiplicanda praeter necessitatem (“Entities are not to be multiplied beyond necessity”). Although referred to as Ockham’s Razor after William of Ockham, a Franciscan living at the turn of the fourteenth century (c.1285–c.1349), this version has not been found in any of his extant works. The closest match (Frustra fit per plura quod potest fieri per pauciora or “It is pointless to do with more what can be done with fewer”) may have been written in quoting others and indeed the general principle was common among Aristotelians. In brief, when explaining a phenomenon or developing a theory, we should refrain from invoking entities that are not necessary in doing so.

As a contemporary example, some people suspect that crop circles are due to extraterrestrial influence, whether directly or otherwise; others suggest that the patterns are the work of dedicated artists or hoaxers and very much an earthly occurrence. Given that the latter group have been able to demonstrate the construction of a crop circle, there is no need to posit aliens to account for why farmer’s fields are routinely invaded in this fashion and hence this explanation is discounted via Ockham’s Razor. If we wish to hold to the economy of thought, we should pick the simpler explanation.

Ockham’s Razor is only a principle; that is, it does not tell us that the simplest explanation is true but instead that we ought to prefer it on methodological grounds. We are counselled to adopt theories which are more efficient, insofar as they can do the same explanatory work with less. Note that this does not imply that we must do so: a direct route to a destination is neither better nor worse than a diversion unless we include the criterion that we wish to get there by the most direct route. Nevertheless, it seems plain enough that we are inclined to favour the simpler explanation, other things being equal. It is this latter ceteris paribus assumption that we now examine.

Applying Ockham’s Razor, part 1: Astronomical theories

Perhaps the best-known example of two competing theories between which a decision had to be made was the seventeenth century controversy over astronomical systems. The long-standing Ptolemaic/Aristotelian model of the heavens was challenged by the Copernicans, who insisted that heliocentrism was simpler than geocentrism. (Note that the question of geostaticism – or the fixed (or otherwise) nature of the Earth itself – was a separate issue.) Since that time, much effort has gone into demonstrating (or refuting) that either system was more parsimonious than the other.

Although Copernicus believed that a sun-centred universe consisting in circular orbits was the most beautiful that could be created, he did so on the basis of thematic assumptions derived from neo-platonic influences and not as a result of any new observations, of which there were none until some years later. (Max Jammer has shown – see this entry for more detail – that Copernicus’s reasoning resulted in his being faced with having to reject either geocentrism or the Aristotelian conception of space. Having no metaphysical substitute for the latter, he was forced to dispense with the former. Ptolemy had actually considered the possibility of circular motion but dismissed it precisely because it did not agree with what was seen in the night sky.) On making the change to heliocentrism, Copernicus found that he still required the assistance of devices like epicycles to save the phenomenon; that is, to make the predictions of his theory agree with what was actually discerned by astronomers. The issue of relative simplicity has subsequently been reduced by some commentators to comparing the number of epicycles but this is beside the point: neither the Ptolemaic nor Copernican system was empirically adequate, leading Kepler to produce another.

The basic error inherent in this counting approach is that of considering theories in isolation, since they always include a host of ancillary presuppositions and exist within metaphysical systems. A comparison with an alternative implicitly or otherwise assumes that all other things are equal when they are not (or, at the very least, we need to show that this requirement is satisfied). Copernicus himself was wary of asserting the truth of his system and only received a copy of his De revolutionibus orbium celestium on his deathbed. When the issue was forced during the so-called “Galileo Affair“, a judgment was sought between two systems whose empirical base was the same and whose practical utility was identical at that time. Galileo sought to delay any choice by invoking the Augustinian principle that it would be folly to ground theological certainties on physical propositions that might subsequently be shown to be false, but his pleas were not heard.

There are several lessons to take from this historical episode. In the first place, we have two competing theories with the same content, and thus a prime candidate for the application of Ockham’s razor. Upon consideration, however, we immediately note that the ceteris paribus clause was not satisfied, for many reasons: the theological consequences were (ostensibly) very different; the political outcome moreso, particularly against the backdrop of the Reformation; the implications for morality were easy to predict but harder to judge; and the metaphysical fallout was just beginning to be investigated. The decision made on this basis did not count the number of postulated entities (which were effectively the same) and did not include an assessment of the relative economies of each theory, since they were also equivalent. In any event, Copernicanism was rejected with scarcely a mention of William of Ockham.

We know now, of course, that a variant of heliocentrism eventually won the day. Galileo’s warning to the Church was not heeded and its decision to assert the reality of geocentrism had catastrophic results for its authority and – later – its credibility. Nevertheless, the history of this change is also illustrative: at no time was there an invocation of the “decisive experiment”, dreamt of by many a philosopher of science. By the time Foucault’s investigations with his pendulum showed the movement of the Earth, confidence in geocentrism had already been slowly eroded over the years. At the only stage in this entire episode where a comparison between rival theories had been insisted upon, the question was decided by “non-scientific” means (notwithstanding the anachronism implying the inverted commas) with Ockham’s Razor playing no part.

The general point raised by this brief study is that Copernicanism required time to develop. Attempting to make a straightforward comparison was disastrous for the Church and for astronomy (and subsequently science) in Italy. Kepler was able to refine the basic Copernican insight because the theory was not limited to the narrow domain in which it was judged. No one needed Ockham’s Razor.

Applying Ockham’s Razor, part 2: Theories of gases

To take a second example from the history of science, consider now a theory, which we call T1, that applies within a domain D. T1 predicts P but the actual state of affairs is in fact P’, which is close to P but such that their difference is beyond experimental possibilities; that is, there is a difference but it is so slight that we could never notice it by investigation. In such circumstances it would be of little use to hope (or even expect) that an increase in experimental capabilities will lead to the discovery that P’ actually obtains because there is no apparent need to refine T1.

Now suppose instead that we propose additional theories T2, T3…, etc, each of which differs from T1 within D and which predicts P’. Ockham’s Razor cannot help us decide whether or not to pursue these new theories. However, when we investigate them further we may find that T2, say, is confirmed where T1 was but also makes novel predictions not given by T1, or else suggests answers to extant problems for T1. In that case, then, we may chose to reject T1 and adopt T2, even though no refuting case has been made against T1.

Although this hypothetical example may be considered fanciful, it is illustrative of what occurred when the kinetic theory of gases was proposed in opposition to the prevailing phenomenological theory. For the phenomenological theory of gases (i.e. based on describing the behaviour of gases via the laws of thermodynamics), Brownian motion was an instance of a perpetuum mobile that refuted the second law of thermodynamics, which expressly disallows perpetual motion. (In brief, the apparently random movement of the Brownian particle seems to go on indefinitely, suggesting that somehow the particle does not run out of energy. In kinetic terms, however, we now say that it is being “bumped” by other molecules, explaining both its behaviour and where its energy comes from). Following his studies of Brownian motion, Einstein was able to entirely recast the phenomenological theory in kinetic terms (see the 1956 edition of his Investigations on the Theory of the Brownian Motion), in spite of having no experimental motivation to do so beyond the known difficulties; after all, the differences in temperature expected, if the kinetic theory was correct, were below the range of detection of thermometers (see Fürth, 1933). Nevertheless, the new theory prevailed when Einstein used it to derive statistical predictions for the behaviour of the Brownian particle by assuming that molecules existed and a mechanical account of the motion could be given. (Feyerabend (1963 (1999, pp.92-94)) made this argument for a different reason, which Laymon (1977) disputed.) This decision could later be justified by the eventual successes of the kinetic programme, but this is only to say that parsimony was discussed after the event, if at all. The possibility of applying Ockham’s Razor was again not considered, nor could it be of any use.

Applying Ockham’s Razor, part 3: The special theory of relativity

By way of a third example, consider that when Einstein published his 1905 paper on special relativity the first response remarked on how his ideas had been decisively refuted by Kaufman’s papers of that year and the next in the Annalen der Physik (in issues 19 and 20, especially his Über die Konstitution des Electrons (1906, p.487)). Kaufman began, in italics, by saying that the “measurement results are not compatible with the Lorentz-Einstein fundamental assumptions”. To see how convincing Kaufman’s work was considered at the time, note that Lorentz wrote to Poincaré in March of 1906, saying that his theory was “in contradiction with Kaufman’s results, and I must abandon it.” The latter agreed and could offer no advice. A glance through the journal and the absence of significant (indeed, for quite some time, any) response shows how seriously Kaufman’s objections were taken. (See Feyerabend, 1999, pp.146-148 for more detail on this and the below.)

Planck, however, was committed to Einstein’s ideas because he thought their “simplicity and generality” meant that they should be preferred, even in the face of experimental refutation. He attempted to re-examine Kaufman’s data and demonstrate that there were flaws, but instead he found that they were far closer to Abraham’s rival theory. Thereafter he presented his findings at the Deutsche Naturvorscherversammlung in Stuttgart in September 1906, which proved to be an amusing affair in which Abraham drew much applause by observing that since the Lorentz-Einstein theory was twice as far from Kaufman’s data as his own, it followed that his theory was twice as good (Physikalische Zeitschrift 7, 1906, pp.759-761). Planck tried but ultimately failed to convince Sommerfeld, Abraham or Bucherer that Einstein’s ideas should be given time to develop. Ultimately, of course, they were accepted because of their “inner consistency” (Wien, 1909) or because Kaufman’s experiments lacked “the great simple universal principle” of relativity theory (von Laue – see below), so that the matter was decided well before Kaufman’s results were finally shown to have been flawed (Guye and Lavanchy, 1916).

Thus we find that Einstein’s ideas succeeded because of a large measure of rhetoric from him, Bohr, Planck and others, and because of a commitment to the presuppositions of relativity theory, long after there had been very little doubt (on the parts of very many great and distinguished physicists) that experimental considerations had killed it. Indeed, by 1911 von Laue was writing that “a really experimental decision between the theory of Lorentz and the Relativity Theory is indeed not to be gained; and that the first of these nevertheless had receded into the background is chiefly due to the fact that, close as it comes to the Relativity Theory, yet it lacks the great simple universal principle, the possession of which lends the Relativity Theory from the start an imposing appearance” (see Das Relativitätsprinzip, 1911). Physicists were more interested in how they could use Einstein’s ideas to explain the result of the Michelson-Morley experiment, even though they were still confusing Lorentz’s and Einstein’s theories in 1921 significantly enough for von Laue to address it (see the fourth edition of his text, then entitled Das Relativitätstheorie as acceptance of the theory had grown and hence changed its status from a mere “principle”). As a result of these theoretical and thematic factors, D.C. Miller’s later (apparent) falsification of Einstein was given very little attention at all, even though it again took a long time (almost thirty years) for Shankland to find the mistake (1955, pp.167ff). (See Holton, 1988, for more discussion of these episodes in the history of physics.)

We see, then, that even in this instance in which the notion of simplicity was relied upon throughout, no actual comparison of the number of entities or parsimony took place. The special theory was held to possess greater inherent simplicity both before and after any experiment and in spite of the negative results of Kaufman’s work. Simplicity was invoked as a metaphysical principle rather than a methodological one.

The general case

From these examples we can say that there are two key difficulties with Ockham’s Razor. The first is that, as we have seen, other things are rarely (if ever) equal. The second – and perhaps still more important – objection is that the unknown (or additional) entities parsed away may have explanatory power outside the domain of consideration, or else they may offer further methodological suggestions which subsequently show that the utility (or even truth) granted to the former explanation was too narrow. The extra terms, which rigorous application of Ockham’s Razor would have us discard, may be methodologically interesting and stimulating even if they turn out to be completely in error. As Niels Bohr was fond of saying, parsimony is something we judge after the event. It makes little methodological sense to disallow additional entities before their consequences have been investigated; indeed, the application of parsimony in the examples we have considered above would likely have proved disastrous, at least with the benefit of hindsight.

The lack of evidence for a posited entity is hardly a problem for scientists who are both willing and able to continue their efforts regardless. Moreover, this risks putting the cart before the horse: a theory may predict the existence of an entity for which there is no evidence but – as a result – is subsequently discovered. While there may be a limitless supply of alternative hypotheses (as asserted by the strong underdetermination of theories), or at least enough to require a decision between them (even if only on practical or financial grounds), not all of them will (or may be suspected to) have interesting enough consequences to pursue. The methodological point, once again, is to ask how we can know the utility (or truth) of apparently un-evidenced or unwarranted theories/entities before the fact? Given that so many have turned out to be of benefit in the past (or so goes the historical argument), why assume to the contrary now?

Theories are surrounded by anomalies and additional entities are postulated to explain them; sometimes these are ad hoc, thus maintaining the theory, while sometimes they necessitate a replacement. Since the resulting alternative theories are empirically equivalent and adequate within the domain satisfied by the current theory, disallowing hypotheses that fail the requirement for parsimony presupposes that they will also fail to address the anomalies or predict novel facts. This is the restriction that the Church imposed upon Galileo and hence following Ockham’s Razor leaves us with a dilemma: should we reject theories that appear to violate parsimony and risk stifling (or ending) their development, which may subsequently show otherwise; or should we instead reject the requirement for parsimony and accept that matters are more complex than methodologically preferring what appears to be the simplest theory?

If we return to the theories T1 and T2 in the second example of applying Ockham’s Razor, it could be that T1 employs different assumptions to T2 such that a straightforward comparison is not possible. Moreover, two hypotheses may be successful in different domains but mutually exclusive within their intersection, if there is one (consider complementarity, for instance). The believer in God or in aliens declaring that an agency other than man was responsible for a phenomenon does not make a straightforward choice restricted to explaining a single instance but involves their additional entity in an entire worldview (incorporating the existence of God or extraterrestrials respectively), which also explains or makes sense of a whole range of phenomena. The ceteris paribus clause here might also turn out to have failed: perhaps the confirming instances of T1 are apparently refuted (as with special relativity) but the inclusion of further assumptions can explain these anomalies, or else we could say that neither theory is satisfactory and the proper response might be to withhold judgment. In addition, T2 might have greater predictive and/or explanatory power outside the domain of comparison, making the evaluation within D an interesting but not particularly devastating factor. Rather than straightforwardly dismissing T2 because of its auxiliary (and apparently unnecessary) assumptions, it may instead make methodological sense to investigate what consequences these have.

Moving beyond any actual results or possibility thereof, the additional entities rejected by parsimony may not explain any other data in further domains but could nevertheless still provide a stimulus to work that subsequently uncovers further domains in which they are needed, or that shows the previous theory to have been but an approximation. While methodological concerns are important, scientists press on with developing theories, perhaps unaware of or unconcerned with the notion that they should first have applied Ockham’s Razor. This is merely to say that science involves elements of guesswork and certainty of resolve (tenacity, as Feyerabend called it), which prompt scientists to continue working on ideas rejected by many of their contemporaries (plate tectonics, say, or Pauli’s positing of the neutrino), possibly reminded of the changing fortunes of atomism over millennia. In short, epistemological considerations are not sufficient to choose between theories and cannot be expected to account for scientific practice.

Still another way to introduce or justify Ockham’s Razor is to assert that parsimonious theories are more likely to be correct. This is a problematic claim. Suppose we take the case of a theory which is regarded by all as highly successful, but which relies upon unobservable entities (such as sub-atomic particles, say). Is the theory true or just a useful instrument? Is it more parsimonious to suppose that these entities do or do not exist? In the absence of an ability to divine the fortunes of a theory in the years to come (or, in the case of atomism, the thousands of years), how are we to decide? To assume, as many apparently do, that parsimony is important because the universe is fundamentally simple, rather than complex (hence the search for grand theories, underlying all others), merely begs the question.

To summarise, the important point which renders parsimony methodologically unhelpful, if not explicitly detrimental, is that the consequences of additional entities or assumptions are impossible to state a priori. Since science is never complete, we are always trying to assess matters beforehand and never get to a comfortable position in which we can count entities in competing theories, which is the only place parsimony can be introduced, much less judged.

References:

Einstein, A., Investigations on the Theory of the Brownian Motion (New York: Dover, 1956).
Einstein, A., Über das Relativitätprinzip und die aus demselben gezogene Folgerungen in Jahrbuch der Radioaktivität, vol. 4, 1907.
Feyerabend, P.K., Knowledge, Science and Relativism (Cambridge: Cambridge University Press, 1999).
Fürth, R., Über einige Beziehungen zwischen klassischer Statistik und Quantenmechanik in Zeitschrift für Physik, vol. 81, 1993.
Guye, C.-E. and Lavanchy, C., Verification experimentale de la
formule de Lorentz-Einstein par les rayons cathodiques de grande vitesse
in Archives des sciences physiques et naturelles, 42: 286–299, 353–373, 441–448, 1916.
Holton, G., Thematic Origins of Scientific Thought (Cambridge: Harvard University Press, 1988).
Jammer, M., Concepts of space: the history of theories of space in physics (Mineola, N.Y.: Dover, 1993).
Kaufman, W., Über die Konstitution des Electrons in Annalen der Physik, vol. 19, 1906.
Kuhn, T.S., The Copernican Revolution: Planetary Astronomy in the Development of Western Thought (Cambridge: Harvard University Press, 1957).
Laue, M. von, Das Relativitätsprinzip (Braunschweig: Friedrich Vieweg & Son, 1911).
Laymon, R., Feyerabend, Brownian Motion and the Hiddenness of Refuting Facts in Philosophy of Science, 44, 225-247, 1977.
Physikalische Zeitschrift, 7, pp.759-761, 1906.
Shankland, R.S., A New Analysis of the Interferometer Observations of Dayton C. Miller in Reviews of Modern Physics, vol. 31, 1963.
Wien, W., Über Elektronen (Leipzig: B.G. Teubner, 1909).

This entry was posted in History and Philosophy of Science, History of Science, Philosophy, Philosophy of Science and tagged , , , , , , , , . Bookmark the permalink.

5 Responses to Ockham’s Razor

  1. Paul Newall says:

    Note that an earlier version of this entry was published at The Galilean Library. Several of the history and philosophy of science essays will be revised and updated here over the coming months.

  2. Pingback: Giant’s Shoulders #34: The Existentialist Edition « From the Hands of Quacks

  3. Pingback: The politics of demarcation | The Kindly Ones

  4. Pingback: On Lutz on Laudan and demarcation | The Kindly Ones

  5. Pingback: Anything goes? Feyerabend and method | The Kindly Ones

Leave a Reply

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s