Crises and Attempted Resolutions of Paradoxes in Foundations of Mathematics By Matt Insall I was recently asked on an email discussion group (http://groups.yahoo.com/group/4DWorldx/) to elaborate on how mathematics was changed between the 19th and 20th century to reestablish consistency. Below is my response. I will give an abbreviated version, for lay readers, of my take on the status of the problem of consistency in mathematics. For many centuries, mathematics developed in fits and starts. The applications of mathematics drove early discovery, and the Greek philosophers then began to make mathematics more an art, or a philosophical study than an applied science. They studied infinity. They studied what it means to be a number. They studied ideal forms, such as the Platonic solids. They studied geometry as a pure science, but in a more philosophical light than in the applied versions that had previously existed. This did not end the applications, or replace them. It only enhanced the ability of mathematicians, both pure and applied, to explain to students, who would then go on to apply mathematics or to teach others themselves. A standard tool of these mathematicians developed (not just in the hands of the Greeks) in the form of argumentation via reductio ad absurdum. The mathematics was more successful with this tool, both as a philosophical pursuit and as an applied science. Other scientists relied on the results of mathematics to support their arguments about the workings of the world, in both a pure and an applied sense. The idea of a consistent system of axioms for the workings of geometry in particular and mathematics in general were emerging. This comfort in the workings of the systems being used and studied for use in mathematics was to last several centuries. It was bolstered by the success of its applications and its popularity as a pure science (or art or philosophical pursuit) among the learned of various societies. The highest form of mathematics was seen to be geometry, which was what we now refer to as euclidean geometry. Then a crisis occurred in geometry. Until Bolyai, Lobachevski and Gauss came on the scene, it was believed that geometry had to be of this type. In particular, there was one postulate of euclidean geometry whose self-evidence had not been assumed, namely the fifth postulate of euclid, sometimes called the parallel postulate. It states that for any line L and any point p not on the line L, there is a unique line K through p that is parallel to L. This seemed to be true in the physical world, but it had never been proved. Attempts had been made to prove it, but all were unsuccessful. Gauss, Lobachevsky and Bolyai showed that the fifth postulate was not provable from the other postulates (or axioms) of euclidean geometry. In doing so, they demonstrated a way to do proofs that we now call ``independence proofs''. In modern parlance, they showed that the parallel postulate is independent of the other axioms of euclidean geometry. (Similar methods may have been employed previously. I will comment on this below.) Their approach was to ``construct'' a model of the other axioms that failed to satisfy the parallel postulate. In a sense, there was not a need to ``construct'' such a model, for one such model turns out to have been under their noses. The surface of a ball is a two-dimensional geometric structure that satisfies all of euclids axioms except the parallel postulate. It had been believed that any geometric structure that satisfied the other axioms of euclidean geometry must also satisfy the parallel postulate, but this was seen now to be not the case. But at least, the axioms of euclid, including the parallel postulate, were consistent, as was evident from the existence of at least one model: the universe. In the early twentieth century, Einstein was to rely on the work of Bolyai, Lobachevski and Gauss, among others, to deny us even this model of euclidean geometry, in the development of his general theory of relativity, designed to account for a new ida of the workings of gravitational force. (Now the universe is thought of as locally euclidean, but not euclidean in the global sense.) In the late nineteenth century, Cantor was studying convergent and divergent trigonometric series, which are a mainstay of a branch of mathematics known as harmonic analysis. Harmonic analysis studies, among other things, the way to design initial-boundary conditions for partial differential equations so that one can have a procedure for solving the given equations. This was an area of what is known now as ``hard analysis'', in contrast to something known as ``soft analysis''. (I will not go into the distinction in detail, in part because I am not an analyst of any sort, though I have done some analysis in the ``hard'' arena.) A special type of set of real numbers was of interest. Now these sets are referred to as ``thin sets'', and in particular, Cantor was studying sets of uniqueness for trigonometric series. He was trying to describe their structure, in terms that other mathematicians would understand and in terms that could be used to explain why they are sets of uniqueness. (A set of uniqueness E is a set of real numbers such that if a trigonometric series converges to zero at all points outside E, then the coefficients of the series must all be zero.) Some of the easiest sets to recognize as sets of uniqueness are the countable ones, and Cantor had shown that. He began to realize that a well-developed theory of sets needed to be developed in order to explain the structure of these sets of uniqueness. Thus he devised what is now referred to as Cantorian Set Theory or Naive Set Theory. In it he found the need to describe a new kind of number, which he called transfinite numbers, for they were larger than any given finite number. This began to be seen as a formulation of a foundation for mathematics in general, and not just for the theory of sets of uniqueness in the real line. In particular, Frege began a program to axiomatize the new ``theory of sets'', and things were going well, including a move by Dedekind to found the theory of real numbers in terms of sets of rational numbers, which had been founded on sets of integers. (Dedekind described the reals as ``cuts'' [now called ``Dedekind cuts''] of rational numbers.) Thus, all of mathematics outside geometry was being reduced to a study of the theory of sets. The beautiful landscape was shattered when Bertrand Russell found a contradiction in the axiom system of Frege. The theory of sets was inconsistent. This was a new crisis. It had already been realized that when a theory was inconsistent, the problem could be remedied by removal of one or more of the foundational assumptions of the theory. The way in which Russell obtained his contradiction, now called ``Russell's Paradox'', led mathematicians to believe that the problem lay in the Fregean axiom of unrestricted comprehension. The comprehension axiom at that time said roughly that for any property P, there is a set whose members are all and nothing but the sets which satisfy property P. Russel obtained his contradiction by considering the property ``is not a member of itself''. In fact, he showed that the set of all sets which are not members of themselves cannot exist. Revisions of Fregean set theory then found ways to restrict the axiom of comprehension so that sets such as this could be proven not to exist and were not in any obvious way postulated by the given axioms. One such axiom system that has survived and is now considered by many to be the most useful foundationally for mathematics is Zermelo-Fraenkel Set Theory, or ZF for short. At the opening of the twentieth century, there was a huge congress of mathematicians in europe (I forget where. I think it was held in Paris.) At the international congress, a well-known mathematician of the time, David Hilbert, proposed that a certain twenty-four problems should be considered top priority for mathematicians in the coming decades. One of these was to show by finitary means that ZF is consistent. By ``finitary means'', it was meant that the proof should be accomplished in a way that can be checked by a (hypothetical) machine, or, what was believed equivalent, by humans working in an entirely mechanical manner. Thus, an area of mathematics grew up, called mathematical logic. In various incarnations, it takes other names, such as ``computer science'' or ``recursive function theory''. But also, these may be taken as parts of mathematical logic, along with the area now known as ``model theory''. In the 1930s, Goedel answered this particular problem of Hilbert negatively, in the following sense. It was believed that the idea of ``provable by finitary means'' should be modeled mathematically by recursive functions. Goedel showed the following things about the first-order theory PA of Peano Arithmetic: 1. in PA, one can interpret the concepts of provability (by finitary means), consistency, etc., and in particular, one may express in PA the statement Con(PA) that PA is consitent. 2. the statement Con(PA) is not provable from the axioms of PA. Then he proceeded to show that the above results could be generalized to many other first-order axiom systems, including ZF. Thus, roughly, ZF can express its own consistency problem, but ZF cannot prove its own consistency. Goedel's theorems made essentail use of an assumption I mentioned above that is known as Church's Thesis: ``Every mechanical procedure corresponds to a recursive function.'' There are ways to state Goedel's results so that they do not use Church's Thesis, and are merely results that show the ``recursive unsolvability'' of the consistency problem in mathematics. Now, many people in mathematics take ZF to be consistent, but also, many point out - rightly - that their work does not use all the axioms of ZF, and they contend that the axioms needed for their work are consistent, and are provably consistent. This is being checked, albeit slowly. Many do not see the need to check it, but there is an area of foundational studies called reverse mathematics, developed by Harvey Friedman and Stephen Simpson, among others, which studies for given mathematical results in the literature exactly what set theoretic axioms are required to prove the given result. I do not know a lot about this area of mathematical logic, but I would enjoy learning more. There is another direction being considered that is closer to the original view. It is a study of a particular set theory that is obtained by restricting the axioms of ZF. This set theory is called Kripke-Platek Set Theory, or KP for short. I do not know a lot about it, but it appears to me to be an attempt to provide a very broad foundation for all of mathematics which is provably consistent. (I am not entirely certain of this, as I am only recently becoming acquainted with notions in the related area of admissible sets.) The culprits are seen by proponents of KP set theory to be not only comprehension, but the power set axiom and the axiom of choice. (The axiom of choice is not a part of ZF. It is part of the modern set theory known as ZFC, Zermelo-Fraenkel Set Theory with the Axiom of Choice.) In any case, theorems of the following form are provable: 1. 2. 3. If PA is consistent, then ZF is consistent, and vice versa. If ZF is consistent, then ZFC is consistent, and vice versa. If PA is consistent, then KP is consistent. (And I think 3 can be strengthened.) Why is it important to try to get a consistent foundation for mathematics? Well, we want to still use the old tool: reductio ad absurdum. If the foundation of mathematics is inconsistent, then this means that there is a reductio ad absurdum argument that proves a statement that is false. (More to the point, if the foundations of mathematics are inconsistent, then every statement, whether true or false, is provable using a reductio ad absurdum argument. This would make such foundations less useful than guessing in practical mathematical applications!) Let me now weigh in, philosophically. I consider the determination of whether our axiom system is consistent to be a valuable question, but after looking at it for quite a while, I am convinced (though I have not proved) that ZFC is consistent. It produces some strange results that make some people uncomfortable, but it is a theory that I believe to be consistent. Thus I expect that no one will ever prove that ZFC is inconsistent, but I admit that I have no (finitary) proof of its consistency. (For PA, there are infinitary proofs of consistency, but I am convinced that on the meta-level, these require an ontological commitment to the consistency of ZF or ZFC.) Moreover, the attempts to provide a (finitarily provably) consistent foundation for mathematics tend to lead either to oversimplified set theories not capable of an appropriate expressiveness or to extremely complicated set theories that will not gain wide acceptance primarily due to the difficulty inherent in just intuitively understanding their axioms. (As it is, some people find the axioms of ZF to be more complicated than they want to deal with. Personally, I find them quite simple to understand, for they were designed to model a consistent portion of the inconsistent form of Naive set theory that was developed by the genius of George Cantor.)