The Church-Turing Thesis (Stanford Encyclopedia of Philosophy)
Excerpt
The Church-Turing thesis (or Turing-Church thesis) is a fundamental claim in the theory of computability. It was advanced independently by Church and Turing in the mid 1930s. There are various equivalent formulations of the thesis. A common one is that every effective computation can be carried out by a Turing machine (i.e., by Turingâs abstract computing machine, which in its universal form encapsulates the fundamental logical principles of the stored-program all-purpose digital computer). Modern reimaginings of the Church-Turing thesis transform it, extending it to fundamental physics, complexity theory, exotic algorithms, and cognitive science. It is important to be aware though that some of the theses nowadays referred to as the Church-Turing thesis are at best very distant relatives of the thesis advanced by Church and Turing.
The Church-Turing thesis (or Turing-Church thesis) is a fundamental claim in the theory of computability. It was advanced independently by Church and Turing in the mid 1930s. There are various equivalent formulations of the thesis. A common one is that every effective computation can be carried out by a Turing machine (i.e., by Turingâs abstract computing machine, which in its universal form encapsulates the fundamental logical principles of the stored-program all-purpose digital computer). Modern reimaginings of the Church-Turing thesis transform it, extending it to fundamental physics, complexity theory, exotic algorithms, and cognitive science. It is important to be aware though that some of the theses nowadays referred to as the Church-Turing thesis are at best very distant relatives of the thesis advanced by Church and Turing.
1. The 1936 Thesis and its Context
The Church-Turing thesis concerns the concept of an effective or systematic or mechanical method, as used in logic, mathematics and computer science. âEffectiveâ and its synonyms âsystematicâ and âmechanicalâ are terms of art in these disciplines: they do not carry their everyday meaning. A method, or procedure, M, for achieving some desired result is called âeffectiveâ (or âsystematicâ or âmechanicalâ) just in case:
- M is set out in terms of a finite number of exact instructions (each instruction being expressed by means of a finite number of symbols);
- M will, if carried out without error, produce the desired result in a finite number of steps;
- M can (in practice or in principle) be carried out by a human being unaided by any machinery except paper and pencil;
- M demands no insight, intuition, or ingenuity, on the part of the human being carrying out the method.
A well-known example of an effective method is the truth-table test for tautologousness. In principle, a human being who works by rote could apply this test successfully to any formula of the propositional calculusâgiven sufficient time, tenacity, paper, and pencils (although in practice the test is unworkable for any formula containing more than a few propositional variables).
1.1 Note on terminology
Statements that there is an effective method for achieving such-and-such a result are commonly expressed by saying that there is an effective method for obtaining the values of such-and-such a mathematical function.
For example, that there is an effective method for determining whether or not any given formula of the propositional calculus is a tautology (such as the truth-table method) is expressed in function-speak by saying there is an effective method for obtaining the values of a function, call it T, whose domain is the set of formulae of the propositional calculus and whose value for any given formula x, written T(x), is 1 or 0 according to whether x is, or is not, a tautology.
1.2 Making the informal concept of an effective method precise
The notion of an effective method or procedure is an informal one, and attempts to characterize effectiveness, such as the above, lack rigor, for the key requirement that the method must demand no insight, intuition or ingenuity is left unexplicated.
One of Alan Turingâs achievements, in his famous paper of 1936, was to present a formally exact predicate with which the informal predicate âcan be done by means of an effective methodâ may be replaced (Turing 1936). Alonzo Church, working independently, did the same (Church 1936a).
The replacement predicates that Church and Turing proposed were, on the face of it, very different from one another. However, these predicates turned out to be equivalent, in the sense that each picks out the same set (call it S) of mathematical functions. The Church-Turing thesis is the assertion that this set S contains every function whose values can be obtained by a method or procedure satisfying the above conditions for effectiveness.
Since it can also be shown that there are no functions in S other than ones whose values can be obtained by a method satisfying the above conditions for effectiveness, the Church-Turing thesis licenses replacing the informal claim âThere is an effective method for obtaining the values of function fâ by the formal claim âf is a member of Sââor by any other formal claim equivalent to this one.
When the Church-Turing thesis is expressed in terms of the replacement concept proposed by Turing, it is appropriate to refer to the thesis also as âTuringâs thesisâ; and as âChurchâs thesisâ when expressed in terms of one or another of the formal replacements proposed by Church.
The formal concept proposed by Turing was that of computability by Turing machine. He argued for the claimâTuringâs thesisâthat whenever there is an effective method for obtaining the values of a mathematical function, the function can be computed by a Turing machine.
The converse claimâamounting to the claim mentioned above, that there are no functions in S other than ones whose values can be obtained by an effective methodâis easily established, since a Turing machine program is itself a specification of an effective method. Without exercising any insight, intuition, or ingenuity, a human being can work through the instructions in the program and carry out the required operations.
If Turingâs thesis is correct, then talk about the existence and non-existence of effective methods and procedures can be replaced throughout mathematics, logic and computer science by talk about the existence or non-existence of Turing machine programs.
Turing stated his thesis in numerous places, with varying degrees of rigor. The following formulation is one of the most accessible:
Turingâs thesis:
L.C.M.s [logical computing machines: Turingâs expression for Turing machines] can do anything that could be described as ârule of thumbâ or âpurely mechanicalâ. (Turing 1948 [2004: 414])
He adds:
This is sufficiently well established that it is now agreed amongst logicians that âcalculable by means of an L.C.M.â is the correct accurate rendering of such phrases. (Ibid.)
1.3 Formulations of Turingâs thesis in terms of numbers
In his 1936 paper, which he titled âOn Computable Numbers, with an Application to the Entscheidungsproblemâ, Turing wrote:
Although the subject of this paper is ostensibly the computable numbers, it is almost equally easy to define and investigate computable functions ⊠I have chosen the computable numbers for explicit treatment as involving the least cumbrous technique. (1936 [2004: 58])
Computable numbers are (real) numbers whose decimal representation can be generated progressively, digit by digit, by a Turing machine. Examples are:
- any number whose decimal representation consists of a finite number of digits (e.g., 109, 1.142)
- all rational numbers, such as one-third, two-sevenths, etc.
- some irrational real numbers, such as Ï and e.
Some real numbers, though, are _un_computable, as Turing proved. Turingâs proof pointed to specific examples of uncomputable real numbers, but it is easy to see in a general way that there must be real numbers that cannot be computed by any Turing machine, since there are more real numbers than there are Turing-machine programs. There can be no more Turing-machine programs than there are whole numbers, since the programs can be counted: 1st program, 2nd program, and so on; but, as Cantor proved in 1874, there are vastly more real numbers than whole numbers (Cantor 1874).
As Turing said, âit is almost equally easy to define and investigate computable functionsâ: There is, in a certain sense, little difference between a computable number and a computable function. For example, the computable number .14159⊠(formed of the digits following the decimal point in Ï, 3.14159âŠ) corresponds to the computable function: f(1)=1, f(2)=4, f(3)=1, f(4)=5, f(5)=9,⊠.
As well as formulations of Turingâs thesis like the one given above, Turing also formulated his thesis in terms of numbers:
[T]he âcomputable numbersâ include all numbers which would naturally be regarded as computable. (Turing 1936 [2004: 58])
and
It is my contention that these operations [the operations of an L.C.M.] include all those which are used in the computation of a number. (Turing 1936 [2004: 60])
In the first of these two formulations, Turing is stating that every number which is able to be calculated by an effective method (that is, âall numbers which would naturally be regarded as computableâ) is included among the numbers whose decimal representations can be written out progressively by one or another Turing machine. In the second, Turing is saying that the operations of a Turing machine include all those that a human mathematician needs to use when calculating a number by means of an effective method.
1.4 The meaning of âcomputableâ and âcomputationâ in Turingâs thesis
Turing introduced his machines with the intention of providing an idealized description of a certain human activity, the tedious one of numerical computation. Until the advent of automatic computing machines, this was the occupation of many thousands of people in business, government, and research establishments. These human rote-workers were in fact called âcomputersâ. Human computers used effective methods to carry out some aspects of the work nowadays done by electronic computers. The Church-Turing thesis is about computation as this term was used in 1936, viz. human computation (to read more on this, turn to Section 7).
For instance, when Turing says that the operations of an L.C.M. include all those needed âin the computation of a numberâ, he means âin the computation of a number by a human beingâ, since that is what computation was in those days. Similarly, ânumbers which would naturally be regarded as computableâ are numbers that would be regarded as being computable by a human computer, a human being who is working solely in accordance with an effective method.
1.5 Churchâs thesis
Where Turing used the term âpurely mechanicalâ, Church used âeffectively calculableâ to indicate that there is an effective method for obtaining the values of the function; and where Turing offered an analysis in terms of computability by an L.C.M., Church gave two alternative analyses, one in terms of the concept of recursion and the other in terms of lambda-definability (λ-definability). He proposed that we
define the notion ⊠of an effectively calculable function of positive integers by identifying it with the notion of a recursive function of positive integers (or of a λ-definable function of positive integers). (Church 1936a: 356)
The concept of a λ-definable function was due to Church and Kleene, with contributions also by Rosser (Church 1932, 1933, 1935c, 1936a; Church & Rosser 1936; Kleene 1934, 1935a,b, 1936a,b; Kleene & Rosser 1935; Rosser 1935a,b). A function is said to be λ-definable if the values of the function can be obtained by a certain process of repeated substitution. The concept of a recursive function had emerged over time through the work of, among others, Grassmann, Peirce, Dedekind, Peano, Skolem, Hilbertâand his âassistantsâ Ackermann and BernaysâSudan, PĂ©ter (nĂ©e Politzer), Herbrand, Kleene, and pre-eminently Gödel (Gödel 1931, 1934). The class of λ-definable functions (of positive integers) and the class of recursive functions (of positive integers) are identical; this was proved by Church and Kleene (Church 1936a; Kleene 1936a,b).
When Turing learned of Churchâs 1936 proposal to identify effectiveness with λ-definability (while preparing his own paper for publication), he quickly established that the concept of λ-definability and his concept of computability are equivalent (by proving the âtheorem that all ⊠λ-definable sequences ⊠are computableâ and its converse; Turing 1936 [2004: 88ff]). Thus, in Churchâs proposal, the words âλ-definable function of positive integersâ (and equally the words ârecursive function of positive integersâ) can be replaced by âfunction of positive integers that is computable by Turing machineâ. What Turing proved is called an equivalence result. There is further discussion of equivalence results in Section 4.1.
Post referred to Churchâs identification of effective calculability with recursiveness and λ-definability as a âworking hypothesisâ, and he quite properly criticized Church for masking this hypothesis as a definition:
[T]o mask this identification under a definition ⊠blinds us to the need of its continual verification. (Post 1936: 105)
This, then, is the âworking hypothesisâ that, in effect, Church proposed:
Churchâs thesis:
A function of positive integers is effectively calculable only if λ-definable (or, equivalently, recursive).
The reverse implication, that every λ-definable function of positive integers is effectively calculable, is commonly referred to as the converse of Churchâs thesis, although Church himself did not so distinguish (bundling both theses together in his âdefinitionâ).
If attention is restricted to functions of positive integers, Churchâs thesis and Turingâs thesis are extensionally equivalent. âExtensionally equivalentâ means that the two theses are about one and the same class of functions: In view of the previously mentioned results by Church, Kleene and Turing, the class of λ-definable functions (of positive integers) is identical to the class of recursive functions (of positive integers) and to the class of computable functions (of positive integers). Notice, though, that while the two theses are equivalent in this sense, they nevertheless have distinct meanings and so are two different theses. One important difference between the two is that Turingâs thesis concerns computing machines, whereas Churchâs does not.
Concerning the origin of the terms âChurchâs thesisâ and âTuringâs thesisâ, Kleene seems to have been the first to use the word âthesisâ in this connection: In 1952, he introduced the name âChurchâs thesisâ for the proposition that every effectively calculable function (on the natural numbers) is recursive (Kleene 1952: 300, 301, 317). The term âChurch-Turing thesisâ also seems to have originated with Kleeneâwith a flourish of bias in favor of his mentor Church:
So Turingâs and Churchâs theses are equivalent. We shall usually refer to them both as Churchâs thesis, or in connection with that one of its ⊠versions which deals with âTuring machinesâ as the Church-Turing thesis. (Kleene 1967: 232)
Some prefer the name Turing-Church thesis.
1.6 Comparing the Turing and Church approaches
One way in which Turingâs and Churchâs approaches differed was that Turingâs concerns were rather more general than Churchâs, in that (as just mentioned) Church considered only functions of positive integers, whereas Turing described his work as encompassing âcomputable functions of an integral variable or a real or computable variable, computable predicates, and so forthâ (1936 [2004: 58]). Turing intended to pursue the theory of computable functions of a real variable in a subsequent paper, but in fact did not do so.
A greater difference lay in the profound significance of Turingâs approach for the emerging science of automatic computation. Churchâs approach did not mention computing machinery, whereas Turingâs introduced the âTuring machineâ (as Church dubbed it in his 1937a review of Turingâs 1936 paper). Turingâs paper also introduced what he called the âuniversal computing machineâ. Now known as the universal Turing machine, this is Turingâs all-purpose computing machine. The universal machine is able to emulate the behavior of any single-purpose Turing machine, i.e., any Turing machine set up to solve one particular problem. The universal machine does this by means of storing a description of the other machine on its tape, in the form of a finite list of instructions (a computer program, in modern terms). By following suitable instructions, the universal machine can carry out any and every effective procedure, assuming Turingâs thesis is true. The functional parts of the abstract universal machine are:
- the memory in which instructions and data are stored, and
- the instruction-reading-and-obeying control mechanism.
In that respect, the universal Turing machine is a bare-bones logical model of almost every modern electronic digital computer.
In his review of Turingâs work, Church noted an advantage of Turingâs analysis of effectiveness over his own:
computability by a Turing machine ⊠has the advantage of making the identification with effectiveness in the ordinary (not explicitly defined) sense evident immediately. (Church 1937a: 43)
He also said that Turingâs analysis has âa more immediate intuitive appealâ than his own (Church 1941: 41).
Gödel found Turingâs analysis superior to Churchâs. Kleene related that Gödel was unpersuaded by Churchâs thesis until he saw Turingâs formulation:
According to a November 29, 1935, letter from Church to me, Gödel âregarded as thoroughly unsatisfactoryâ Churchâs proposal to use λ-definability as a definition of effective calculability. ⊠It seems that only after Turingâs formulation appeared did Gödel accept Churchâs thesis. (Kleene 1981: 59, 61)
Gödel described Turingâs analysis of computability as âmost satisfactoryâ and âcorrect ⊠beyond any doubtâ (Gödel 1951: 304 and 193?: 168). He remarked:
We had not perceived the sharp concept of mechanical procedures sharply before Turing, who brought us to the right perspective. (Quoted in Wang 1974: 85)
Gödel also said:
The resulting definition of the concept of mechanical by the sharp concept of âperformable by a Turing machineâ is both correct and unique. (Quoted in Wang 1996: 203)
And:
Moreover it is absolutely impossible that anybody who understands the question and knows Turingâs definition should decide for a different concept. (Ibid.)
Even the modest young Turing agreed that his analysis was âpossibly more convincingâ than Churchâs (Turing 1937: 153).
1.7 The Entscheidungsproblem
Both Turing and Church introduced their respective versions of the Church-Turing thesis in the course of attacking the so-called Entscheidungsproblem. As already mentioned, the title of Turingâs 1936 paper included âwith an Application to the Entscheidungsproblemâ, and Church went with simply âA Note on the Entscheidungsproblemâ for the title of his 1936 paper. Soâwhat is the Entscheidungsproblem?
The German word âEntscheidungsproblemâ means decision problem. The Entscheidungsproblem for a logical calculus is the problem of devising an effective method for deciding whether or not a given formulaâany formulaâis provable in the calculus. (Here âprovableâ means that the formula can be derived, step by logical step, from the axioms and definitions of the calculus, using only the rules of the calculus.) For example, if such a method for the classical propositional calculus is used to test the formula AâA (A implies A), the output will be âYes, provableâ, and if the contradiction A&ÂŹA is tested, the output will be âNot provableâ. Such a method is called a decision method or decision procedure.
Church and Turing took on the Entscheidungsproblem for a fundamentally important logical system called the (first-order) functional calculus. The functional calculus consists of standard propositional logic plus standard quantifier logic. The functional calculus is also known as the classical predicate calculus and as quantification theory (and Church sometimes used the German term engere FunktionenkalkĂŒl). They each arrived at the same negative result, arguing on the basis of the Church-Turing thesis that, in the case of the functional calculus, the Entscheidungsproblem is unsolvableâthere can be no decision method for the calculus. The two discovered this result independently of one another, both publishing it in 1936 (Church a few months earlier than Turing). Churchâs proof, which made no reference to computing machines, is for that reason sometimes considered to be of less interest than Turingâs.
The Entscheidungsproblem had attracted some of the finest minds of early twentieth-century mathematical logic, including Gödel, Herbrand, Post, Ramsey, and Hilbert and his assistants Ackermann, Behmann, Bernays, and Schönfinkel. Herbrand described the Entscheidungsproblem as âthe most general problem of mathematicsâ (Herbrand 1931b: 187). But it was Hilbert who had brought the Entscheidungsproblem for the functional calculus into the limelight. In 1928, he and Ackermann called it âdas Hauptproblem der mathematischen Logikâââthe main problem of mathematical logicâ (Hilbert & Ackermann 1928: 77).
Hilbert knew that the propositional calculus (which is a fragment of the functional calculus) is decidable, having found with Bernays a decision procedure based on what are called ânormal formsâ (Bernays 1918; Behmann 1922; Hilbert & Ackermann 1928: 9â12; Zach 1999), and he also knew from the work of Löwenheim that the monadic functional calculus is decidable (Löwenheim 1915). (The monadic functional calculus is the fragment involving only one-place predicatesâi.e., no relations, such as â=â and â<â, and no higher-place predicates, such as ââ is between â and ââ.) He thought there must be a decision procedure for the entire functional calculus. The challenge, the main problem of mathematical logic, was to find it. As he and Ackermann wrote in 1928, in their famous book GrundzĂŒge der Theoretischen Logik (Principles of Mathematical Logic):
[I]t is to be expected that a systematic, so to speak computational treatment of the logical formulae is possible âŠ. (Hilbert & Ackermann 1928: 72)
However, their expectations were frustrated by the Church-Turing result of 1936. Hilbert and Ackermann excised the quoted statement from a revised edition of their book. Published in 1938, the new edition was considerably watered down to take account of Turingâs and Churchâs monumental result.
Hilbert knew, of course, that some mathematical problems have no solution, for example the problem of finding a finite binary numeral n (or unary numeral, in Hilbertâs version of the problem) such that n2=2 (Hilbert 1926: 179). He was nevertheless very fond of saying that every mathematical problem can be solved, and by this he meant that
every definite mathematical problem must necessarily be susceptible of an exact settlement, either in the form of an actual answer to the question asked, or by the proof of the impossibility of its solution and therewith the necessary failure of all attempts. (Hilbert 1900: 261 [trans. 1902: 444])
It seems never to have crossed his mind that his âHauptproblem der mathematischen Logikâ falls into the second of these two categoriesâuntil, that is, Church and Turing unexpectedly proved âthe impossibility of its solutionâ.
For more detail on the Entscheidungsproblem, and an outline of the stunning result that Church and Turing independently established in 1936, see the supplement on The Rise and Fall of the Entscheidungsproblem.
2. Backstory: Emergence of the concepts of effective method and decision method
Effective methods are the subject matter of the Church-Turing thesis. How did this subject matter evolve and how was it elaborated prior to Church and Turing? This section looks back to an earlier era, after which Section 3 turns to modern developments.
2.1 From simple rules-of-thumb to Siri and beyond
Effective methods are extremely helpful in carrying out many practical tasks, and their use stretches back into the mists of antiquity, although it was not until the twentieth century that interest began to build in analysing their nature. Perhaps the earliest effective methods to be utilized were rules-of-thumb (as Turing called them) for arithmetical calculations of various sorts, but whatever their humble beginnings, the scope of effective methods has widened dramatically over the centuries. In the Middle Ages, the Catalan philosopher Llull envisaged an effective method for posing and answering questions about the attributes of God, the nature of the soul, the nature of goodness, and other fundamental issues. Three hundred years later, in the seventeenth century, Hobbes was asserting that human reasoning processes amount to nothing more than (essentially arithmetical) effective procedures:
By reasoning I understand computation. (Hobbes 1655 [1839]: ch. 1 sect. 2)
Nowadays, effective methodsâalgorithmsâare the basis for every job that electronic computers do. According to some computer scientists, advances in the design of effective methods will soon usher in human-level artificial intelligence, followed by superhuman intelligence. Already, virtual assistants such as Siri, Cortana and ChatGPT implement effective methods that produce useful answers to a wide range of questions.
In its most sublimely general form, the Entscheidungsproblem is the problem of designing an effective general-purpose question-answerer, an effective method that is capable of giving the correct answer, yes or no, to any meaningful scientific question, and perhaps even ethical and metaphysical questions too. The idea of such a method is almost jaw-dropping. Llull seems to have glimpsed the concept of a general question-answering method, writing in approximately 1300 of a general art (âarsâ), or technique, âby means of which one may know in regard to all natural thingsâ (Lo Desconhort, line 8, in Llull 1986: 99). He dreamed of an ars generalis (general technique) that could mechanize the âone general science, with its own general principles in which the principles of other sciences would be implicitâ (Preface to Ars Generalis Ultima, in Llull 1645 [1970: 1]). Llull used circumscribed fields of knowledge to illustrate his idea of a mechanical question-answerer, designing small domain-specific machines consisting of superimposed discs; possibly his machines took the form of a parchment volvelle, a relative of the metal astrolabe.
In early modern times, Llullâs idea of the ars generalis received a sympathetic discussion in Leibnizâs writings.
2.2 Leibniz
Leibniz designed a calculating machine that he said would add, subtract, multiply and divide, and in 1673 he demonstrated a version of his machine in London and Paris (Leibniz 1710). His interest in mechanical methods led him to an even grander conception, inspired in part by Llullâs unclear but provocative speculations about a general-purpose question-answering mechanism. Leibniz said that Llull âhad scraped the skin offâ this idea, but âdid not see its inmost partsâ (Leibniz 1671 [1926: 160]). Leibniz envisaged a method, just as mechanical as multiplication or division, whereby
when there are disputes among persons, we can simply say: Let us calculate, without further ado, in order to see who is right. (Leibniz 1685 [1951: 51])
The basis of the method, Leibniz explained, was that âwe can represent all sorts of truths and consequences by Numbersâ and âthen all the results of reasoning can be determined in numerical fashionâ (Leibniz 1685 [1951: 50â51]). He hoped the method would apply to âMetaphysics, Physics, and Ethicsâ just as well as it did to mathematics (1685 [1951: 50]). This conjectured method could, he thought, be implemented by what he called a machina combinatoria, a combinatorial machine (Leibniz _n.d._1; Leibniz 1666). However, there was never much progress towards his dreamed-of method, and in a letter two years before his death he wrote:
[I]f I were younger or had talented young men to help me, I should still hope to create a kind of universal symbolistic [spécieuse générale] in which all truths of reason would be reduced to a kind of calculus. (Leibniz 1714 [1969: 654])
In his theorizing Leibniz described what he called an ars inveniendi, a discovering or devising method. The function of an ars inveniendi is to produce hitherto unknown truths of science (Leibniz 1679 [1903: 37]; Leibniz _n.d._2 [1890: 180]; Hermes 1969). A mechanical ars inveniendi would generate true statements, and with time the awaited answer to a scientific question would arrive (Leibniz 1671 [1926: 160]). Blessed with a universal (i.e., complete, and consistent) ars inveniendi, the user could input any meaningful and unambiguous (scientific or mathematical) statement S, and the machine would eventually respond (correctly) with either âS is trueâ or âS is falseâ. As the groundbreaking developments in 1936 by Church and Turing made clear, if the ars inveniendi is supposed to work by means of an effective method, then there can be no universal ars inveniendiâand not even an ars inveniendi that is restricted to all mathematical statements, since these include statements of the form âp is provableâ, or even to all purely logical statements.
2.3 Logic machines
The modern concept of a decision method for a logical calculus did not develop until the twentieth century. But earlier logicians, including Leibniz, certainly had the concept of a method that is mechanical in the literal sense that it could be carried out by a machine constructed from mechanical components of the sort familiar to themâdiscs, pins, rods, springs, levers, pulleys, rotating shafts, gear wheels, weights, dials, mechanical switches, slotted plates, and so forth.
In 1869, Jevons designed a pioneering machine known as the âlogic pianoâ (Jevons 1870; Barrett & Connell 2005). The name arose because of the machineâs piano-like keyboard for inputting logical formulae. The formulae were drawn from a syllogistic calculus involving four positive terms, such as âironâ and âmetalâ (Jevons 1870). Turingâs colleague Mays, who himself designed an influential electrical logic machine (Mays & Prinz 1950), described the logic piano as âthe first working machine to perform logical inference without the intervention of human agencyâ (Mays & Henry 1951: 4).
The logic piano implemented a method for determining which combinations drawn from eight termsâthe four positive terms and the corresponding four negated terms (ânon-metalâ, etc.)âwere consistent with the inputted formulae and which not (although in fact not all consistent combinations were taken into account). The machine displayed the consistent formulae by means of lettered strips of wood, with upper-case letters representing positive terms and lower-case negative. Jevons exhibited the logic piano in Manchester at Owens College, now Manchester University, where he was professor of logic (Mays & Henry 1953: 503). His piano, Jevons claimed with considerable exaggeration, made it âevident that mechanism is capable of replacing for the most part the action of thought required in the performance of logical deductionâ (Jevons 1870: 517).
A decade later, Venn published the technique we now call Venn diagrams (Venn 1880). This technique satisfies the four criteria set out for an effective method in Section 1. The user first diagrams the premisses of a syllogism and then, as Quine put it, âwe inspect the diagram to see whether the content of the conclusion has automatically appeared in the diagram as a resultâ (Quine 1950: 74). Not all formulae of the functional calculus are Venn-diagrammable, and Vennâs original method is limited to testing syllogisms. In the twentieth century, Massey showed that Vennâs method can be stretched to give a decision procedure for the monadic functional calculus (Massey 1966).
Venn, like Jevons, was well aware of the concept of a literally mechanical method. He pointed out that diagrammatic methods such as his âreadily lend themselves to mechanical performanceâ (Venn 1880: 15). Venn went on to describe what he called a âlogical-diagram machineâ. This simple machine displayed wooden segments corresponding to the component areas of a Venn diagram; each wooden segment represented one of four terms. A finger-operated release mechanism allowed any segment selected by the user to drop downwards. This represented âthe destruction of any classâ (1880: 18). Venn reported that he constructed this machine, which measured âbetween five and six inches square and three inches deepâ (1880: 17). When Venn published his description of it, Jevons quickly wrote to him saying that the logical-diagram machine âis exceedingly ingenious & seems to represent the relations of four terms very wellâ (Jevons 1880). Venn himself however was less enthusiastic, saying in his article âI have no high estimate myself of the interest or importance of what are sometimes called logical machinesâ (1880: 15). Baldwin, commenting on Vennâs machine in 1902, complained that it was âmerely a more cumbersome diagramâ (1902: 29). This is quite trueâit would be at least as easy to draw the Venn diagram on paper as to set it up on the machine. But Vennâs article made it very plain that the logical-diagram machine was intended to be a hilarious send-up of Jevonsâ complicated logic piano.
At around the same time, Marquandâa student of Peirceâsâdesigned a logic machine which a Princeton colleague then built (out of wood salvaged from âPrincetonâs oldest homesteadâ, Marquand related in his 1885). Marquand knew of Jevonsâ and Vennâs designs, and said he had âfollowed Jevonsâ in certain respects, and that his own machine was âsomewhat similarâ to Jevonsâ (Marquand 1881, 1883: 16, 1885: 303). Peirce, with customary bluntness, called Marquandâs machine âa vastly more clear-headed contrivance than that of Jevonsâ (Peirce 1887: 166). Again limited to a syllogistic calculus involving only four positive terms, Marquandâs device, like Jevonsâ, displayed term-combinations consistent with the inputted formulae. A lettered plate with sixteen mechanical dials was used to display the combinations.
2.4 Peirce
In 1886, in a letter to Marquand, Peirce famously suggested that Marquand consider an electrical version of his machine, and he sketched simple switching circuits implementing (what we would now call) an AND-gate and an OR-gateâpossibly the earliest proposal for electrical computation (Peirce 1886). Far-sightedly, Peirce wrote in the letter that, with the use of electricity, âit is by no means hopeless to expect to make a machine for really very difficult mathematical problemsâ. Much later, Church discovered a detailed diagram of an electrical relay-based form of Marquandâs machine among Marquandâs papers at Princeton (reproduced in Ketner & Stewart 1984: 200). Whoever worked out the design in this diagramâMarquand, Peirce, or an unknown third personâhas a claim to be an important early pioneer of electromechanical computing.
Peirce, with his interest in logic machines, seems to have been the first to consider the decision problem in roughly the form in which Turing and Church tackled it. From about 1896, he developed the diagrammatic proof procedures he called âexistential graphsâ. These were much more advanced than Vennâs diagrams. Peirceâs system of alpha-graphs is a diagrammatic formulation of the propositional calculus, and his system of beta-graphs is a version of the first-order functional calculus (Peirce 1903a; Roberts 1973). Roberts (1973) proved that the beta-graphs system contains the axioms and rules of Quineâs 1951 formulation of the first-order functional calculus, in which only closed formulae are asserted (Quine 1951: 88).
Peirce anticipated the concept of a decision method in his extensive notes for a series of lectures he delivered in Boston in 1903. There he developed a method (Peirce 1903b,c) that, if applied to any given formula of the propositional calculus, would, he said, âdetermineâ (or âpositively ascertainâ) whether the alpha-graphs system demonstrates that the formula is satisfiable (is âalpha-possibleâ, in Peirceâs terminology), or whether, on the other hand, the system demonstrates that it is unsatisfiable (âalpha-impossibleâ). (See the supplement on The Rise and Fall of the Entschedungsproblem for an explanation of âsatisfiableâ.) Peirce said his method âis such a comprehensive routine that it would be easy to define a machine that would perform itââalthough the âcomplexity of the caseâ, he continued, ârenders any such procedure quite impracticableâ (Peirce 1903c). Perhaps he would not have been completely surprised to learn that within five or six decades, and with the use of electricity, it became far from impractical to run a decision method for the propositional calculus on a machine.
Peirce also searchedâin vain, of courseâfor a corresponding method for his beta-graphs system (Peirce 1903b,c,d; Roberts 1997). Like Hilbert after him, he seems to have entertained no doubt that full first-order predicate logic is amenable to a machine-like method.
Peirce had prescient ideas about the use of machines in mathematics more generally. Around the turn of the century, he wrote:
[T]he whole science of higher arithmetic, with its hundreds of marvellous theorems, has in fact been deduced from six primary assumptions about number. The logical machines hitherto constructed are inadequate to the performance of mathematical deductions. There is, however, a modern Exact Logic which, although yet in its infancy, is already far enough advanced to render it a mere question of expense to construct a machine that would grind out all the known theorems of arithmetic and advance that science still more rapidly than it is now progressing. (Peirce n.d., quoted in Stjernfelt 2022)
Here Peirce seems to be assertingâquite correctlyâthat a machine can be devised to grind out all the theorems of Dedekindâs (1888) axiomatisation of arithmetic (which consisted of six âprimary assumptionsâ in the form of of four axioms and two definitions). This statement of Peirceâs, made almost four decades before Turing introduced Turing machines into mathematics, was well ahead of its time.
As to whether all mathematical reasoning can be performed by a machine, as Leibniz seems to have thought, Peirce was fiercely skeptical. He formulated the hypothesis that, in the future, mathematical reasoning
might conceivably be left to a machineâsome Babbageâs analytical engine or some logical machine. (Peirce 1908: 434)
However, he placed this hypothesis alongside others he deemed âlogical heresiesâ, calling it âmalignantâ (ibid.). His skeptical attitude, if perhaps not his reasons for it, was arguably vindicated by Turingâs subsequent results (Turing 1936, 1939). But before that, a quite different view of matters took root among mathematicians, under the influence of Hilbert and his group at Göttingen.
2.5 Hilbert and the Göttingen group
It was largely Hilbert who first drew attention to the need for a precise analysis of the idea of an effective decision method. In a lecture he gave in Zurich in 1917, to the Swiss Mathematical Society, he emphasized the need to study the concept of âdecidability by a finite number of operationsâ, sayingâaccuratelyâthat this would be âan important new field of research to developâ (Hilbert 1917: 415). The lecture considered a number of what he called âmost challenging epistemological problems of a specifically mathematical characterâ (1917: 412). Pre-eminent among these was the âproblem of the decidability [Entscheidbarkeit] of a mathematical questionâ because the problem âtouches profoundly upon the nature of mathematical thinkingâ (1917: 413).
Hilbert and his Göttingen group looked back on Leibniz as the originator of their approach to logic and the foundations of mathematics. Behmann, a prominent member of the group, said that Leibniz had anticipated modern symbolic logic (Behmann 1921: 4â5). Leibnizâs hypothesized âuniversal characteristicâ or universal symbolistic was a universal symbolic language, in conception akin to languages used in mathematical logic and computer science today. Hilbert and Ackermann acknowledged Leibnizâs influence on the first page of their GrundzĂŒge der Theoretischen Logik, saying âThe idea of a mathematical logic was first put into a clear form by Leibnizâ (Hilbert & Ackermann 1928: 1). Cassirer said that in Hilbertâs work âthe fundamental idea of Leibnizâs âuniversal characteristicâ is taken up anew and attains a succinct and precise expressionâ (Cassirer 1929: 440). It was in the writings of the Göttingen group that Leibnizâs idea of an effective method for answering any specified mathematical or scientific question found its fullest development (see further the supplement on The Rise and Fall of the Entscheidungsproblem).
Hilbertâs earliest publication to mention what we would now call a decision problem is his 1899 book Grundlagen der Geometrie [Foundations of Geometry]. He said that in the course of his investigations of Euclidean geometry he was
guided by the principle of discussing each given question in such a way that we examined both whether it can or cannot be answered by means of prescribed steps using certain limited resources. (Hilbert 1899: 89)
Concerning a specific example, he wrote:
Suppose a geometrical construction problem that can be carried out with a compass is presented; we will attempt to lay down a criterion that enables us to determine [beurteilen] immediately, from the analytical nature of the problem and its solutions, whether the construction can also be carried out using only a ruler and a segment-transferrer. (Hilbert 1899: 85â86)
He described what would now be called an effective method for determining this, and his term âbeurteilenâ could, with a trace of anachronism, be translated as âdecideâ.
Hilbert expressed the concept of a decision method more clearly the following year, in his famous turn-of-the-century speech in Paris, to the International Congress of Mathematicians. He presented twenty-three unsolved problems, âfrom the discussion of which an advancement of science may be expectedâ. The tenth on his list (now known universally as Hilbertâs Tenth Problem) was:
Given a diophantine equation with any number of unknown quantities and with rational integral numerical coefficients: To devise a process according to which it can be determined by a finite number of operations whether the equation is solvable in rational integers. (Hilbert 1900: 276 [trans. 1902: 458])
The Entscheidungsproblem was coming into even clearer focus by the time Hilbertâs student Behmann published a landmark article in 1922, âContributions to the Algebra of Logic, in particular to the Entscheidungsproblemâ. It was probably Behmann who coined the term âEntscheidungsproblemâ (Mancosu & Zach 2015: 166â167). In a 1921 lecture to the Göttingen group, Behmann said:
If a logical or mathematical statement is given, the required procedure should give complete instructions for determining whether the statement is correct or false by a deterministic calculation after finitely many steps. The problem thus formulated I want to call the allgemeine Entscheidungsproblem [general decision problem]. (Behmann 1921: 6 [trans. 2015: 176])
Peirce, as we saw, spoke of a procedureâs forming âsuch a comprehensive routine that it would be easy to define a machine that would perform itâ. His work was well-known in Göttingen: Hilbert and Ackermann said that Peirce âespeciallyâ, and also Jevons, had âenriched the young scienceâ of mathematical logic (1928: 1). Like Peirce, Behmann used the concept of a machine to clarify the nature of the Entscheidungsproblem. âIt is essential to the characterâ of the problem, Behmann said, that âonly entirely mechanical calculation according to given instructionsâ is involved. The decision whether the statement is true or false becomes âa mere exercise in computationâ; there is âan elimination of thinking in favor of mechanical calculationâ. Behmann continued:
One might, if one wanted to, speak of mechanical or machine-like thinking. (Perhaps one can one day even let it be carried out by a machine.) (Behmann 1921: 6â7 [trans. 2015: 176])
Leibnizâs Llullian idea of a machine that could calculate the truth was suddenly at the forefront of early twentieth century mathematics.
2.6 Newman and the Cambridge mathematicians
The connection Behmann emphasized between the decision problem and a machine that carries out an âexercise in computationâ would soon prove crucial in Turingâs hands. What seems to have been Turingâs first significant brush with the Entscheidungsproblem was in 1935, in a Cambridge lecture given by Newman. Newman, a mathematical logician and topologist, was very familiar with the ideas emanating from Göttingen. As early as 1923 he gave a left-field discussion of some of Hilbertâs ideas, himself proposing an approach to the foundations of mathematics that, while radical and new, nevertheless had a strongly Hilbertian flavor (Newman 1923). In 1928, Newman attended an international congress of mathematicians in the Italian city of Bologna, where Hilbert talked about the Entscheidungsproblem while lecturing on his proof theory (Hilbert 1930a; Zanichelli 1929). Hilbertâs leading works in mathematical logicâHilbert and Ackermann (1928) and Hilbert and Bernays (1934)âwere both recommended reading for Newmanâs own lectures on the Foundations of Mathematics (Smithies 1934; Copeland and Fan 2022).
Speaking in a tape-recorded interview about Turingâs engagement with the Entscheidungsproblem, Newman said âI believe it all started because he attended a lecture of mine on foundations of mathematics and logicâ:
I think I said in the course of this lecture that what is meant by saying that [a] process is constructive is that itâs a purely mechanical machineâand I may even have said, a machine can do it.
And this of course led [Turing] to the next challenge, what sort of machine, and this inspired him to try and say what one would mean by a perfectly general computing machine. (Newman _c_1977)
Sadly, there seems to be no record of what else Newman said at that crucial juncture in his lecture. However, his 1923 paper âThe Foundations of Mathematics from the Standpoint of Physicsâ does record some of his related thinking (Copeland & Fan 2023). There he introduced the term âprocessâ (which he also used in the above quotation), saying âAll logic and mathematics consist of certain processesâ (1923: 12). He emphasized the requirement that a process should terminate with the required result (such as a theorem or number); and he gave a formal treatment of processes:
The properties of processes are formally developed from a set of axioms, and a general method reached for attacking the problem of whether a given process terminates or not. (Newman 1923: 12)
Newman did not mention the Entscheidungsproblem in his 1923 paperâlet alone moot its unsolvability (there is no evidence that, pre-Turing, he thought the problem unsolvable)âyet, with hindsight, he certainly laid some suggestive groundwork for an attack on the problem. He wrote:
The information of the first importance to be obtained about a process or segment of a process is whether it is possible to perform itâŠ. The statement that [process] Ί||Î±Ï is possible means that this process terminates: comes to a halt ⊠(Newman 1923: 39)
Newman even proposed an âapparatusâ, a âsymbolic machineâ, for producing numbers by means of carrying out processes of the sort he analysed, and he gave a profound discussion of real numbers from the standpoint of this proposal (1923: 130ff).
Nor was Newman the only person at Cambridge with an interest in the Entscheidungsproblem. The Entscheidungsproblem was âin the airâ there during the decade leading up to Turingâs assault on it. The Sadleirian Professor of Mathematics at Cambridge, Hardy, took an interest in the problem, inspired by von Neumannâs magisterial exposition and critique of Hilbertâs ideas (von Neumann 1927). Ackermann himself had visited Cambridge from Göttingen for the first half of 1925 (Zach 2003: 226). Another visitor, Langfordâwho worked in Cambridge on a fellowship from Harvard for the academic year 1924â25 (Frankena & Burks 1964)âpresented a series of results to the American Mathematical Society not long after his return to Harvard, in effect solving a number of special cases of the Entscheidungsproblem (Langford 1926a, 1927).
The Cambridge logician Ramsey, like Turing a Fellow of Kingâs College, also worked on the Entscheidungsproblem in the latter part of the 1920s. He died in 1930 (the year before Turing arrived in Cambridge as an undergraduate), but not before completing a key paper solving the Entscheidungsproblem in special cases (Ramsey 1930). His work, too, was prominent in the recommended reading for Newmanâs lecture course. Braithwaite, another Fellow of Kingâs College (who had a hand in Turingâs election to a junior research fellowship at Kingâs in 1935), was keenly interested in Ramseyâs work on the Entscheidungsproblem (Copeland & Fan 2022). Also in 1935, von Neumann visited Cambridge from Princeton, for the term following Newmanâs lecture course (Copeland & Fan 2023). Von Neumann, a member of the Göttingen group during the mid-1920s, had called the Entscheidungsproblem âprofound and complexâ, and he voiced doubts that it was solvable (von Neumann 1927: 11; 1931: 120).
He was not alone. Hardy gave this statement of the Entscheidungsproblem, in the course of a famous discussion of Hilbertâs ideas:
Suppose, for example, that we could find a finite system of rules which enabled us to say whether any given formula was demonstrable or not. (Hardy 1929: 16)
Hardy foresaw what Turing, and Church, would soon prove, telling his audience that such a system of rules âis not to be expectedâ.
What Turing showed is that there will never be, and can never be, a computing machine satisfying the following specification: When the user types in a formulaâany formulaâof the functional calculus, the machine carries out a finite number of steps and then outputs the correct answer, either âThis formula is provable in the functional calculusâ or âThis formula is not provable in the functional calculusâ, as the case may be. Given, therefore, Turingâs thesis that if an effective method exists then it can be carried out by one of his machines, it follows that there is no effective method for deciding the full first-order functional calculus.
3. Other Approaches to Computability
Turing and Church were certainly not the only people to tackle the problem of analyzing the concept of effectiveness. This section surveys some other important proposals made during the twentieth and twenty-first centuries.
3.1 Gödel
Gödel was led to the problem of analyzing effectiveness by his search for a means to generalize his 1931 incompleteness results (which in their original form applied specifically to the formal system set out by Whitehead and Russell in their Principia Mathematica). In 1934, he considered an analysis in terms of his generalized concept of recursionâabout a year before Church first publicly announced his thesis that âthe notion of an effectively calculable function of positive integers should be identified with that of a recursive functionâ (Church 1935a; Gödel 1934, fn. 3; Davis 1982).
But Gödel was doubtful: âI was, at the time ⊠not at all convinced that my concept of recursion comprises all possible recursionsâ (Gödel 1965b). It was Turingâs analysis, Gödel emphasized, that finally enabled him to generalize his incompleteness theorems:
due to A. M. Turingâs work, a precise and unquestionably adequate definition of the general concept of formal system can now be given. (Gödel 1965a: 71)
He explained:
Turingâs work gives an analysis of the concept of âmechanical procedureâ (alias âalgorithmâ or âcomputation procedureâ or âfinite combinatorial procedureâ). This concept is shown to be equivalent with that of a âTuring machineâ. A formal system can simply be defined to be any mechanical procedure for producing formulas, called provable formulas. (Gödel 1965a: 71â72)
Armed with this definition, incompleteness can, Gödel said, âbe proved rigorously for every consistent formal system containing a certain amount of finitary number theoryâ (1965a: 71).
3.2 Post
By 1936, Post had arrived independently at an analysis of effectiveness that was substantially the same as Turingâs (Post 1936; Davis & Sieg 2015). Postâs idealized human âworkerââor âproblem solverââoperated in a âsymbol spaceâ consisting of âa two way infinite sequence of spaces or boxesâ. A box admitted
of but two possible conditions, i.e., being empty or unmarked, and having a single mark in it, say a vertical stroke. (Post 1936: 103)
The problem solver worked in accordance with âa fixed unalterable set of directionsâ and could perform a small number of âprimitive actsâ (Post 1936: 103), namely:
- determine whether the box that is presently occupied is marked or not;
- erase any mark in the box that is presently occupied;
- mark the box that is presently occupied if it is unmarked;
- move to the box to the right of the present position; and
- move to the box to the left of the present position.
Postâs paper was submitted for publication in October 1936, some five months after Turingâs. It contained no analog of Turingâs universal computing machine, and nor did it anticipate Churchâs and Turingâs result that the Entscheidungsproblem is unsolvable. Curiously, though, Post had achieved far more than he let on in his 1936 paper. In an article subtitled âAccount of an Anticipationâ, published in 1965 but written in about 1941, he explained that during the early 1920s he had devised a systemâhe called it the âcomplete normal systemâ, because âin a way, it contains all normal systemsââand this, he said, âcorrespond[ed]â to Turingâs universal machine (Post 1965: 412). Furthermore, he argued during the same period that the decision problem is unsolvable in the case of his ânormal systemsâ (1965: 405ff). But it seems he did not extend this argument to anticipate the Church-Turing result that the decision problem for the predicate calculus is unsolvable (1965: 407).
Turing later generously acknowledged Postâs 1936 paper, describing Turing machines as âthe logical computing machines introduced by Post and the authorâ (Turing 1950b: 491).
3.3 Hilbert and Bernays
In 1939, in Volume II of their titanic work Grundlagen der Mathematik (Foundations of Mathematics), Hilbert and Bernays proposed a logic-based analysis of effectiveness. According to this analysis, effectively calculable numerical functions are numerical functions that can be evaluated in what they called a âregelrechtâ manner (Hilbert & Bernays 1939: 392â421). In this context, the German word âregelrechtâ can be translated ârule-governedâ. Hilbert and Bernays offered the concept of the rule-governed evaluation of a numerical function as a âsharpening of the concept of computableâ (1939: 417).
The basic idea is this: To evaluate a numerical function (such as addition or multiplication) in a rule-governed way is to calculate the values of the function, step by logical step, in a suitable deductive logical system. On this approach, effective calculability is analysed as calculability in a logic. (Both Church and Turing had previously discussed the approachâsee Section 4.4.)
The logical system Hilbert and Bernays used to flesh out this idea was an equational calculus, reminiscent of a calculus that Gödel had detailed in lectures he gave in Princeton in 1934 (Gödel 1934). The theorems of an equational calculus are (as the name says) equationsâfor example 23=8 and x2+1=x(x+1)â(xâ1), or in general f(m)=n. The Hilbert-Bernays equational calculus contains no logical symbols (such as negation, conjunction, implication, or quantifiers), and every formula is simply an equation between terms. Three types of equation are permitted as the initial formulae (or premisses) of deductions in the system; and the system is required to satisfy three general conditions that Hilbert and Bernays called ârecursivity conditionsâ. The rules of the calculus concern substitutions within equations and are very simple, allowing steps such as:
a=b,f(a)âąf(b)
On the basis of this calculus (which they called Z0) Hilbert and Bernays established an equivalence result: The numerical functions that are capable of rule-governed evaluation coincide with the (primitive) recursive functions (1939: 403 and 393_n_).
It is perhaps unsurprising that Hilbert, the founder of proof theory, ultimately selected an analysis of effective calculability as calculability within a logic, even though Church and Turing had already presented their analyses in terms of recursive functions and Turing machines respectively. Hilbert and Bernays went on to use their analysis to give a new proof of the unsolvability of the Entscheidungsproblem (Hilbert & Bernays 1939: 416â421). This proof quietly marks what must have been an unsettling, even painful, shift of perspective for them. The idea of a decision procedure for mathematics had until the Church-Turing result been central to their thinking, and in Volume 1 of the Grundlagen, published in 1934, they had assumed the Entscheidungsproblem to be solvable.
3.4 Modern axiomatic analyses
Church reported a discussion he had had with Gödel at the time when it was still wide open how the intuitive concept of effective calculability should be formalized (probably during 1934). Gödel suggested that
it might be possible, in terms of effective calculability as an undefined notion, to state a set of axioms which would embody the generally accepted properties of this notion, and to do something on that basis. (Church 1935b)
Logicians frequently analyse a concept of interest, e.g., universal quantification, not by defining it in terms of other concepts, but by stating a set of axioms that collectively embody the generally accepted properties of (say) universal quantification. To follow this approach in the case of effectiveness, we would âwrite down some axioms about computable functions which most people would agree are evidently trueâ (Shoenfield 1993: 26). Shoenfield continued, âIt might be possible to prove Churchâs Thesis from such axiomsâ, but added: âHowever, despite strenuous efforts, no one has succeeded in doing thisâ.
Moving on a few years, a meeting on The Prospects for Mathematical Logic in the Twenty-First Century, held at the turn of the millennium, included the following among leading open problems:
âProveâ the Church-Turing thesis by finding intuitively obvious or at least clearly acceptable properties of computation that suffice to guarantee that any function so computed is recursive [and therefore can be computed by a Turing machine]. (Shore in Buss et al. 2001: 174â175)
The axiomatic type of approach sketched by Gödel has by now been developed in a number of quite different ways. These axiomatic frameworks go a long way toward countering Montagueâs complaint of over 60 years ago that âDiscussion of Churchâs thesis has suffered for lack of a precise general framework within which it could be conductedâ (Montague 1960: 432). Some examples of the axiomatic approach are as follows (in chronological order):
-
Gandy (Turingâs only PhD student) pointed out that Turingâs analysis of human computation does not immediately apply to computing machines strongly dissimilar from Turing machines. (See Section 4.3 below for details of Turingâs analysis.) For example, Turingâs analysis does not obviously apply to parallel machines which, unlike a Turing machine, are able to process an arbitrary number of symbols simultaneously. Seeking a generalized form of Turingâs analysis that applies equally well to Turing machines and massively parallel machines, Gandy (1980) stated four axioms governing the behaviour of what he called discrete deterministic mechanical devices. (He formulated the axioms in terms of hereditarily finite sets.) Gandy was then able to prove that every device satisfying these axioms can be simulated by a Turing machine: Discrete deterministic mechanical devices, even massively parallel ones, are no more powerful than Turing machines, in the sense that whatever computations such a device is able to perform can also be done by the universal Turing machine. (For more on Gandyâs analysis, see Section 6.4.2.)
-
Engeler axiomatized the concept of an algorithmic function by using combinators (Engeler 1983: ch. III). Combinators were originally introduced by Schönfinkel in 1924, in a paper that a recent book on combinators described as âpresenting a formalism for universal computation for the very first timeâ (Schönfinkel 1924; Wolfram 2021: 214). Schönfinkelâs combinators were extensively developed by Curry (Curry 1929, 1930a,b, 1932; Curry & Feys 1958). Examples of combinators are the âpermutatorâ C and the âcancellatorâ K. These produce the following effects: Cxyz=xzy and Kxy=x.
-
Sieg formalized Turingâs analysis of human computation by means of four axioms (Sieg 2008). The result, Sieg said, is an axiomatic characterization of âthe concept âmechanical procedureââ, and he observed that his system âsubstantiates Gödelâs remarksâ (above) that one should try to find a set of axioms embodying the generally accepted properties of the concept of effectiveness (Sieg 2008: 150).
-
Dershowitz and Gurevich (2008) stated three very general axioms, treating computations as discrete, deterministic, sequentially-evolving structures of states. They called these structures âstate-transition systemsâ, and called the three axioms the âSequential Postulatesâ. They also used a fourth axiom, stipulating that âOnly undeniably computable operations are available in initial statesâ (2008: 306). From their four axioms, they proved a proposition they called Churchâs thesis: âEvery numeric function computed by a state-transition system satisfying the Sequential Postulates, and provided initially with only basic arithmetic, is partial recursiveâ (2008: 327).
Returning to the very idea of proving the Church-Turing thesis, it is important to note that the proposition Dershowitz and Gurevich call Churchâs thesis is in fact not the thesis stated by Church, viz. âA function of positive integers is effectively calculable only if recursiveâ. Crucially, their version of Churchâs thesis does not even mention the key concept of effective calculability. The entire project of trying to prove Churchâs (or Turingâs) actual thesis has its share of philosophical difficulties. For example, suppose someone were to lay down some axioms expressing claims about effective calculability (as Sieg for instance has done), and suppose it is possible to prove from these axioms that a function of positive integers is effectively calculable only if recursive. Churchâs thesis would have been proved from the axioms, but whether the axioms form a satisfactory account of effective calculability is a further question. If they do not, then this âproofâ of Churchâs thesis could carry no conviction. Which is to say, a proof of this sort will be convincing only to one who accepts another thesis, namely that the axioms are indeed a satisfactory account of effective calculability. This is a Churchian meta-thesis. Churchâs thesis would have been proved, but only at the expense of throwing up another, unproved, thesis seemingly of the same nature.
There is further discussion of difficulties associated with the idea of proving the Church-Turing thesis in Section 4.3.5, Section 4.5, and Section 4.6.
4. The Case for the Church-Turing Thesis
4.1 The inductive and equivalence arguments
Although there have from time to time been attempts to call the Church-Turing thesis into question (for example by KalmĂĄr in his 1959; Mendelson replied in his 1963), the summary of the situation that Turing gave in 1948 is no less true today: âit is now agreed amongst logicians that âcalculable by L.C.M.â is the correct accurate renderingâ of the informal concept of effectiveness.
In 1936, both Church and Turing gave various grounds for accepting their respective theses. Church argued:
The fact ⊠that two such widely different and (in the opinion of the author) equally natural definitions of effective calculability [i.e., in terms of λ-definability and recursion] turn out to be equivalent adds to the strength of the reasons adduced below for believing that they constitute as general a characterization of this notion as is consistent with the usual intuitive understanding of it. (Church 1936a: 346, emphasis added)
Churchâs âreasons adduced belowâ comprised two not wholly convincing arguments. Both suffered from the same weakness, discussed in Section 4.4.4.
Turing, on the other hand, marshalled a formidable case for the thesis. Unlike Church, he offered inductive evidence for it, showing that large classes of numbers âwhich would naturally be regarded as computableâ are computable in his sense (1936: 74â75). Turing proved, for example, that the limit of a computably convergent sequence is computable; that all real algebraic numbers are computable; that the real zeroes of the Bessel functions are computable; and that (as previously noted) Ï and e are computable (1936: 79â83). But most importantly of all, Turing gave profound logico-philosophical arguments for the thesis. He referred to these arguments simply as âIâ, âIIâ and âIIIâ. They are described in Section 4.3 and Section 4.4.
By about 1950, considerable evidence had amassed for the thesis. One of the fullest surveys of this evidence is to be found in chapters 12 and 13 of Kleeneâs 1952. As well as discussing Turingâs argument I, and Churchâs two arguments mentioned above, Kleene bolstered Churchâs just quoted equivalence argument, pointing out that âSeveral other characterizations ⊠have turned out to be equivalentâ (1952: 320). As well as the characterizations mentioned by Church, Kleene included computability by Turing machine, Postâs canonical and normal systems (Post 1943, 1946), and Gödelâs notion of reckonability (Gödel 1936). (Turingâs student and lifelong friend Robin Gandy picturesquely called Churchâs equivalence argument the âargument by confluenceâ [Gandy 1988: 78].)
In modern times, the equivalence argument can be presented even more forcefully: All attempts to give an exact characterization of the intuitive notion of an effectively calculable function have turned out to be equivalent, in the sense that each characterization offered has been proved to pick out the same class of functions, namely those that are computable by Turing machine. The equivalence argument is often considered to be very strong evidence for the thesis, because of the diversity of the various formal characterizations involved. Apart from the many different characterizations already mentioned in Section 1 and Section 3, there are also analyses in terms of register machines (Shepherdson & Sturgis 1963), Markov algorithms (Markov 1951), and other formalisms.
The equivalence argument may be summed up by saying that the concept of effective calculabilityâor the concept of computability simpliciterâhas turned out to be formalism-transcendent, or even âformalism-freeâ (Kennedy 2013: 362), in that all these different formal approaches pick out exactly the same class of functions.
Indeed, there is not even a need to distinguish, within any given formal approach, systems of different orders or types. Gödel noted in an abstract published in 1936 that the concept âcomputableâ is absolute, in the sense that all the computable functions are specifiable in one and the same system, there being no need to introduce a hierarchy of systems of different ordersâas is done, for example, in Tarskian analyses of the concept âtrueâ, and standardly in the case of the concept âprovableâ (Gödel 1936: 24). Ten years later, commenting on Turingâs work, Gödel emphasized that âthe great importance ⊠[of] Turingâs computabilityâ is
largely due to the fact that with this concept one has for the first time succeeded in giving an absolute definition of an interesting epistemological notion, i.e., one not depending on the formalism chosen. In all other cases treated previously, such as demonstrability or definability, one has been able to define them only relative to a given languageâŠ. (Gödel 1946: 150)
In his 1952 survey, Kleene also developed Turingâs inductive argument (1952: 319â320). To summarize:
- Every effectively calculable function that has been investigated in this respect has turned out to be computable by Turing machine.
- All known methods or operations for obtaining new effectively calculable functions from given effectively calculable functions are paralleled by methods for constructing new Turing machines from given Turing machines.
Inductive evidence for the thesis has continued to accumulate. For example, Gurevich points out that
As far as the input-output relation is concerned, synchronous parallel algorithms and interactive sequential algorithms can be simulated by Turing machines. This gives additional confirmation of the Church-Turing thesis. (Gurevich 2012: 33)
4.2 Skepticism about the inductive and equivalence arguments
It is a general feature of inductive arguments that, while they may supply strong evidence, they nevertheless do not establish their conclusions with certainty. A stronger argument for the Church-Turing thesis is to be desired. Gandy said that the inductive argument
cannot settle the philosophical (or foundational) question. It might happen that one day some genius established an entirely new sort of calculation. (Gandy 1988: 79)
Dershowitz and Gurevich highlighted the difficulty:
History is full of examples of delayed discoveries. Aristotelian and Newtonian mechanics lasted much longer than the seventy years that have elapsed since Church proposed identifying effectiveness with recursiveness, but still those physical theories were eventually found lacking. (Dershowitz & Gurevich 2008: 304)
Dershowitz and Gurevich presented a highly relevant example of delayed discovery (following Barendregt 1997: 187): Any hope that the effectively calculable functions could be identified with the primitive recursive functionsâintroduced in 1923 (Skolem 1923; PĂ©ter 1935)âevaporated a few years later, when Ackermann described an effectively calculable function that is not primitive recursive (Ackermann 1928).
The equivalence argument has also been deemed unsatisfactory. Dershowitz and Gurevich call it âweakâ (2008: 304). After all, the fact that a number of statements are equivalent does not show the statements are true, only that if one is true, all areâand if one is false, all are. Kreisel wrote:
The support for Churchâs thesis ⊠certainly does not consist in ⊠the equivalence of different characterizations: what excludes the case of a systematic error? (Kreisel 1965: 144)
Mendelson put the point more mildly, saying that the equivalence argument is ânot conclusiveâ:
It is conceivable that all the equivalent notions define a concept that is related to, but not identical with, effective computability. (Mendelson 1990: 228)
Clearly, what is required is a direct argument for the thesis from first principles. Turingâs argument I fills this role.
4.3 Turingâs argument I
The logico-philosophical arguments that Turing gave in Section 9 of âOn Computable Numbersâ are outstanding among the reasons for accepting the thesis.
He introduced argument I as âonly an elaborationâ of remarks at the beginning of his 1936 paperâsuch as:
We may compare a man in the process of computing a real number to a machine which is only capable of a finite number of conditions q1, q2,âŠ, qR which will be called âm-configurationsâ. (Turing 1936 [2004: 59, 75])
He also described argument I as a âdirect appeal to intuitionâ (Turing 1936 [2004: 75]). The appeal he is talking about concerns our understanding of which features of human computation are the essential features (some examples of _in_essential features are that human computers eat, breathe, and sleep).
In outline, argument I runs as follows: Given that human computation has these (and only these) essential featuresâand here Turing supplied a list of featuresâthen, whichever human computation is specified, a Turing machine can be designed to carry out the computation. Therefore, the Turing-machine computable numbers include all numbers that would naturally be regarded as computable (Turingâs thesis).
4.3.1 Turingâs analysis
Turingâs list of the essential features of human computation is as follows (Turing 1936 [2004: 75â76]):
- Computers write symbols on two-dimensional sheets of paper, which may be considered to be (or may actually be) divided up into squares, each square containing no more than a single individual symbol.
- The computer is not able to recognize, or print, more than a finite number of different types of individual symbol.
- The computer is not able to observe an unlimited number of squares all at onceâif he or she wishes to observe more squares than can be taken in at one time, then successive observations must be made. (Say the maximum number of squares the computer can observe at any one moment is B, where B is some positive integer.)
- When the computer makes a fresh observation in order to view more squares, none of the newly observed squares will be more than a certain fixed distance away from the nearest previously observed square. (Say this fixed distance consists of L squares, where L is some positive integer.)
- In order to alter a symbol (e.g., to replace it by a different symbol), the computer needs to be actually observing the square containing the symbol.
- The computerâs behavior at any moment is determined by the symbols that he or she is observing and his or her âstate of mindâ at that moment. Moreover, the computerâs state of mind at any given moment, together with the symbols he or she is observing at that moment, determine the computerâs state of mind at the next moment.
- The number of states of mind that need to be taken into account when describing the computerâs behavior is finite.
- The operations the computer performs can be split up into elementary operations. These are so simple that no more than one symbol is altered in a single elementary operation.
- All elementary operations are of one or other of the following forms:
- A change of state of mind.
- A change of observed squares, together with a possible change of state of mind.
- A change of symbol, together with a possible change of state of mind.
4.3.2 Next step: B-L-type Turing machines
The next step of argument I is to establish that if human computation has those and only those essential features, then, whatever human computation is specified, a Turing machine can be designed to perform the computation. In order to show this, Turing introduced a modified form of Turing machine, which can be called a âB-L-typeâ Turing machine. A B-L-type Turing machine has much in common with an ordinary Turing machine:
- AÂ B-L-type Turing machine consists of a scanner and a one-dimensional paper tape; the tape is divided into squares.
- The scanner contains mechanisms that enable it to move the tape to the left or right.
- The scannerâs mechanisms also enable it recognize, delete, and print symbols.
- The scanner is able to recognize and print only a finite number of different types of individual symbol.
- At any moment, the control mechanism of the scanner will be in any one of a finite number of internal states. Turing terms these âm-configurationsâ. He included an explanatory remark about m-configurations in a summary in French of the central ideas of âOn Computable Numbersâ: Inside the machine, âlevers, wheels, et cetera can be arranged in several ways, called âm-configurationsââ. (The complete summary is translated in Copeland & Fan 2022.)
- The machineâs behavior at any moment is determined by its m-configuration and the symbols it is observing (i.e., scanning).
- The machineâs possible behaviors are limited to moving the tape, deleting the symbol on an observed square, and printing a symbol on an observed square. Each of these behaviors may be accompanied by a change in m-configuration.
Moving on now to the differences between ordinary Turing machines and B-L-type machines:
- The scanner of a B-L-type machine can observe up to B squares at once; whereas the scanner of an ordinary Turing machine can observe only a single square of the tape at any one moment. A Turing machine that is able to survey a sequence of squares all at once like this is sometimes known by the (perhaps inelegant) term âstring machineâ.
- The scanner of a B-L-type machine can, in a single operation, move the tape up to L squares at once (to the left or right of any one of the immediately previously observed squares); whereas the scanner of an ordinary machine can move the tape by only one square in a single elementary operation.
Returning to the argument, Turing asserted that, given his account 1â9 of the essential features of human computation, a B-L-type machine can âdo the workâ of any human computer (1936: 77). This is because the B-L-type machine either duplicates or can simulate each of features 1â9. Let us take these features in turn.
Feature 1 is simulated by the machine: The B-L-type machine uses its one-dimensional tape to mimic the computerâs two-dimensional sheets of paper. Turing said:
I think it will be agreed that the two-dimensional character of paper is no essential of computation. (Turing 1936 [2004: 75])
However, some commentators note that there is room for doubt about this matter. Gandy complained that Turing here argued âmuch too brieflyâ, saying:
It is not totally obvious that calculations carried out in two (or three) dimensions can be put on a one-dimensional tape and yet preserve the âlocalâ properties. (Gandy 1988: 81, 82â83)
Dershowitz and Gurevich ask:
[H]ow certain is it that each and every elaborate data structure used during a computation can be encoded as a string, and its operations simulated by effective string manipulations? (Dershowitz & Gurevich 2008: 305)
Progressing to the other features on Turingâs list: 2, 3, 4 and 5 are straightforwardly duplicated in the machine. Features 6 and 7 are simulated, by letting the machineâs m-configurations do duty for the computerâs states of mind (more on that below). Feature 8 is duplicated in the machine: the machineâs complex operations (such as long multiplication and division) are built up out of elementary operations. Feature 9 is simulated, again by letting the m-configurations to do duty for human states of mind.
4.3.3 Final step
The next and final step of argument I involves the statement that any computation done by a B-L-type machine can also be done by an ordinary Turing machine. This is straightforward, since by means of a sequence of single-square moves, the ordinary machine can simulate a B-L-type machineâs tape-moves of up to L squares at once; and the ordinary machine can also simulate the B-L-type machineâs scanning of up to B squares at once, by means of a sequence of single-square scannings (interspersed where necessary with changes of m-configuration). Thus, if a B-L-type machine can âdo the workâ of a human computer, so can an ordinary Turing machine.
In summary, Turing has shown the followingâprovided his claim is accepted that âTo each state of mind of the computer corresponds an âm-configurationâ of the machineâ: Given the above account of the essential features of human computation, an ordinary Turing machine is able to do the work of any human computer. In other words: Subject to that proviso and that given, he has established his thesis that the numbers computable by an ordinary Turing machine include all numbers which would naturally be regarded as computable.
4.3.4 States of mind, and argument III
But should Turingâs claim about the correspondence of states of mind and m-configurations be accepted? Might not human states of mind greatly surpass arrangements of levers and wheels? Might not the computerâs states of mind sometimes determine him or her to change the symbols in a way that a B-L-type machine cannot?
Turing addressed worries about the correspondence between states of mind and m-configurations in his supplementary argument III, which he said âmay be regarded as a modification of Iâ (1936: 79). Here he argued that reference to the computerâs states of mind can be avoided altogether, by talking instead about what he called a ânote of instructionsâ. A note of instructions, he said, is âa more definite and physical counterpartâ of a state of mind. Each step of the human computation can be regarded as being governed by a note of instructionsâby means of following the instructions in the note, the computer will know what operation to perform at that step (erase, print, or move). Turing envisaged the computer preparing new notes on the fly, as the computation progresses: âThe note of instructions must enable him [the computer] to carry out one step and write the next noteâ. Each note is in effect a tiny computer program, which both carries out a single step of the computation and also writes the program that is to be used at the next step.
Once instruction notes are in the picture, there is no need to refer to the human computerâs states of mind:
the state of progress of the computation at any stage is completely determined by the note of instructions and the symbols on the tape. (Turing 1936 [2004: 79])
Anotherârelatedâway of answering the worry that human states of mind might surpass the machineâs m-configurations is to point out that, even if this were true, it would make no essential difference to argument I. This is because of feature 3 and feature 7 (Section 4.3.1): The number of states of mind that need to be taken into account is finite, and the maximum number of squares that the computer can observe at any one moment is B (a finite number).
Given feature 7, it follows that no matter how fancy a state of mind might be, the computerâs relevant behaviors when in that state of mind can be encapsulated by means of finite table. Each row of the table will be of the following form: If the observed symbols are such-and-such, then perform elementary operation so-and-so (where the elementary operations are as specified in feature 9). Since only a finite number of states of mind are in consideration (feature 3)âsay nâall necessary information about the computerâs states of mind can be encapsulated in a list of n such tables. This list consists of finitely many symbols, and therefore it can be placed on the tape of a B-L-type machine in advance of the machine beginning its emulation of the human computer. (This is akin to writing a program on the tape of a universal Turing machine.) The B-L-type machine consults the list at each step of the computation, and the machineâs behavior at every step is completely determined by the list together with the currently observed symbols.
To conclude: no matter what powers might be accorded to the human computerâs states of mind, a B-L-type machine can nevertheless âdo the workâ of the computer, so long as only finitely many states of mind need be taken into consideration (given, of course, the remainder of Turingâs account of the essential features of computation).
4.3.5 Turingâs theorem
Now that the proviso mentioned above about states of mind has been cleared out of the way, Turingâs achievement in argument I can be summed up like this: He has, in Gandyâs phrase, âoutlined a proofâ of a theorem (Gandy 1980: 124).
Turingâs computation theorem:
This account of the essential features of human computation implies Turingâs thesis.
It should by now be completely clear why Turing called argument I a âdirect appeal to intuitionâ. If oneâs intuition tells one that Turingâs account of the essential features of human computation is correct, then the theorem can be applied and Turingâs thesis is secured.
However, Turingâs account is not immune from skepticism. Some skeptical questions are: Might there be aspects of human computation that Turing has overlooked? Might a computer who is limited by 1â9 be unable to perform some calculations that can be done by a human computer not so restricted? Also, must the number of states of mind that need to be taken into account when describing the computerâs behavior always be finite? Gödel thought the number of Turingâs âdistinguishable states of mindâ may âconverge toward infinityâ, saying
What Turing disregards completely is the fact that mind, in its use, is not static, but constantly developing. (Gödel 1972: 306)
Indeed, what are the grounds supposed to be for thinking that 1â9 are true? Are these claims supposed to be grounded in the nature and limitations of the human sense organs and the human mind? Or are they supposed to be grounded in some other way, e.g., in the fundamental nature of effective methods?
Turingâs argument I is a towering landmark and there is now a sizable literature on these and other questions concerning it. For more about this important argument see, for starters, Sieg 1994, 2008; Shagrir 2006; and Copeland & Shagrir 2013.
4.4 Turingâs argument II
4.4.1 Calculating in a logic
Kleene, in his survey of evidence for the Church-Turing thesis, listed a type of argument based on symbolic logic (Kleene 1952: 322â3). (He called these category âDâ arguments.) Arguments of this type commence by introducing a plausible alternative method of characterizing effectively calculable functions (or, in Turingâs case, computable functions or numbers). The alternative method involves derivability in one or another symbolic logic: The concept of effective calculability (or of computability) is characterized in terms of calculability within the logic (see Section 3.3). Schematically, the characterization is of the form: A function is effectively calculable (or computable) if each successive value of the function is derivable within the logic. The next step of the argument is then to establish that the new characterization (whatever it is) is equivalent to the old. In Churchâs case, this amounts to arguing that the new characterization is equivalent to his characterization in terms of either recursiveness or λ-definability. Finally, the conclusion that the new and previous characterizations are equivalent is claimed as further evidence in favor of the Church-Turing thesis.
In his survey, Kleene illustrated this approach by describing an argument of Churchâs (Church 1936a: 357â358). Turingâs argument II is also of this type, but, curiously, Kleene did not mention it (despite assigning five pages of his 1952 survey to Turingâs argument I).
4.4.2 Churchâs âstep-by-stepâ argument
It is instructive to examine Churchâs argumentâwhich Gandy dubbed the âstep-by-stepâ argument (Gandy 1988: 77)âbefore considering Turingâs II. Church introduced the following alternative method, describing it as among the âmethods which naturally suggest themselvesâ in connection with defining effective calculability:
a function F (of one positive integer) [is defined] to be effectively calculable if, for every positive integer m, there exists a positive integer n such that F(m)=n is a provable theorem. (Church 1936a: 358)
Church did not specify any particular symbolic logic. He merely stipulated a number of general conditions that the logic must satisfy (1936a: 357). These included the stipulations that the list of axioms of the logic must be either finite or enumerably infinite, and that each rule of the logic must specify an âeffectively calculable operationâ (the latter is necessary, he said, if the logic âis to serve at all the purposes for which a system of symbolic logic is usually intendedâ). Having introduced this alternative method of characterizing effective calculability, Church then went on to argue that every function (of one positive integer) that is âcalculable within the logicâ in this way is also recursive. He concluded, in support of Churchâs thesis, that the new method produces âno more general definition of effective calculability than that proposedâ, i.e., in terms of recursiveness (1936a: 358).
4.4.3 Turingâs variant
Turingâs prefatory remarks to argument II bring out its broad similarity to Churchâs argument. Turing described II as being a âproof of the equivalence of two definitionsâ, addingââin case the new definition has a greater intuitive appealâ (1936 [2004: 75]).
Turingâs argument, unlike Churchâs, does involve a specific symbolic logic, namely Hilbertâs first-order predicate calculus. Argument II hinges on a proposition that can be called
Turingâs provability theorem:
Every formula provable in Hilbertâs first-order predicate calculus can be proved by the universal Turing machine. (See Turing 1936 [2004: 77].)
The alternative method considered by Turing (which is similar to Churchâs) characterizes a computable number (or sequence) in terms of statements each of which supplies the next digit of the number (or sequence). The number (sequence) is said to be computable if each such statement is provable in Hilbertâs calculus (the idea being that, if this is so, then Hilbertâs calculus may be used to calculateâor computeâthe digits of the number one by one). Employing the provability theorem, Turing then showed the following: Every number that is computable according to this alternative definition is also computable according to the Turing-machine definition (i.e., the digits of the number can be written out progressively by a Turing machine), and vice versa (Turing 1936 [2004: 78]). In other words, he proved the equivalence of the two definitions, as promised.
4.4.4 Comparing the Church and Turing arguments
Returning to Churchâs step-by-step argument, there is an air of jiggery-pokery about it. Church wished to conclude that functions âcalculable within the logicâ are recursive, and, in the course of arguing for this, he found it necessary to assert that each rule of the logic is a recursive operation, on the basis that each rule is required to be an effectively calculable operation. In a different context, he might have supported this assertion by appealing to Churchâs thesis (which says, after all, that what is effectively calculable is recursive). But in the present context, such an appeal would naturally be question-begging.
Nor did Church make any such appeal. (Sieg described Churchâs reasoning as âsemi-circularâ, but this seems too harshâthere is nothing circular about it; Sieg 1994: 87, 2002: 394.) But nor did Church offer any compelling reasons in support of his assertion. He merely gave examples of systems whose rules are recursive operations; and also saidâhaving stipulated that each rule of procedure must be an effectively calculable operationâthat he will âinterpret this to mean that ⊠each rule of procedure must be a recursive operationâ (1936: 357, italics added.) In short, a crucial step of Churchâs argument for Churchâs thesis receives inadequate support. Sieg famously dubbed this step the âstumbling blockâ in Churchâs argument (Sieg 1994: 87).
There is no such difficulty in Turingâs argument. Having selected a specific logic (Hilbertâs calculus), Turing was able specify a Turing machine that would âfind all the provable formulae of the calculusâ, so making it indubitable that functions calculable in the logic are Turing-machine computable (Turing 1936 [2004: 77]). For this reason, Turingâs argument II is to be preferred to Churchâs step-by-step argument.
4.5 Kripkeâs version of argument II
A significant recent contribution to the area has been made by Kripke (2013). A conventional view of the status of the Church-Turing thesis is that, while âvery considerable intuitive evidenceâ has amassed for the thesis, the thesis is ânot a precise enough issue to be itself susceptible to mathematical treatmentâ (Kripke 2013: 77). Kleene gave an early expression of this now conventional view:
Since our original notion of effective calculability of a function ⊠is a somewhat vague intuitive one, the thesis cannot be proved. ⊠While we cannot prove Churchâs thesis, since its role is to delimit precisely an hitherto vaguely conceived totality, we require evidence âŠ. (Kleene 1952: 318)
Rejecting the conventional view, Kripke suggests that, on the contrary, the Church-Turing thesis is susceptible to mathematical proof. Furthermore, he canvasses the idea that Turing himself sketched an argument that serves to prove the thesis.
Kripke attempts to build a mathematical demonstration of the Church-Turing thesis around Turingâs argument II. He claims that his demonstration is âvery closeâ to Turingâs (Kripke 2013: 80). However, this is debatable, since, in its detail, the Kripke argument differs considerably from argument II. But one can at least say that Kripkeâs argument was inspired by Turingâs argument II, and belongs in Kleeneâs category âDâ (along with II and Churchâs step-by-step argument).
Kripke argues that the Church-Turing thesis is a corollary of Gödelâs completeness theorem for first-order predicate calculus with identity. Put somewhat crudely, the latter theorem states that every valid deduction (couched in the language of first-order predicate calculus with identity) is provable in the calculus. In other words, the deduction of B from premises A1, A2, ⊠An (where statements A1, A2, ⊠An, B are all in the language of first-order predicate calculus with identity) is logically valid if and only if B can be proved from A1, A2, ⊠An in the calculus.
The first step of the Kripke argument is his claim that (error-free, human) computation is itself a form of deduction:
[A] computation is a special form of mathematical argument. One is given a set of instructions, and the steps in the computation are supposed to followâfollow deductivelyâfrom the instructions as given. So a computation is just another mathematical deduction, albeit one of a very specialized form. (Kripke 2013: 80)
The following two-line program in pseudo-code illustrates Kripkeâs claim. The symbol âââ is read âbecomesâ, and â=â as usual means identity. The first instruction in the program is râ2. This tells the computer to place the value 2 in storage location r (assumed to be initially empty). The second instruction râr+3 tells the computer to add 3 to the content of r and store the result in r (over-writing the previous content of r). The execution of this two-line program can be represented as a deduction:
{Execution of râ2, followed immediately by execution of râr+3} logically entails that r=5 in the immediately resulting state.
In the case of Turing-machine programs, Turing developed a detailed logical notation for expressing all such deductions (Turing 1936).
(In fact, the successful execution of any string of instructions can be represented deductively in this fashionâKripke has not drawn attention to a feature special to computation. The instructions do not need to be ones that a computer can carry out.)
The second step of Kripkeâs argument is to appeal to what he refers to as Hilbertâs thesis: this is the thesis that the steps of any mathematical argument can be expressed âin a language based on first-order logic (with identity)â (Kripke 2013: 81). The practice of calling this claim âHilbertâs thesisâ originated in Barwise (1977: 41), but it should be noted that since Hilbert regarded second-order logic as indispensable (see, e.g., Hilbert & Ackermann 1928: 86), the name âHilbertâs thesisâ is potentially misleading.
Applying âHilbertâs thesisâ to Kripkeâs above quoted claim that âa computation is just another mathematical deductionâ (2013: 80) yields:
every (human) computation can be formalized as a valid deduction couched in the language of first-order predicate calculus with identity.
Now, applying Gödelâs completeness theorem to this yields in turn:
every (human) computation is provable in first-order predicate calculus with identity, in the sense that, given an appropriate formalization, each step of the computation can be derived from the instructions (possibly in conjunction with ancillary premises, e.g., well-known mathematical premises, or premises concerning numbers that are supplied to the computer at the start of the computation).
Finally, applying Turingâs provability theorem to this intermediate conclusion yields the Church-Turing thesis:
every (human) computation can be done by Turing machine.
4.6 Turing on the status of the thesis
As Section 3.4 mentioned, Dershowitz and Gurevich have also argued that the Church-Turing thesis is susceptible to mathematical proof (Dershowitz & Gurevich 2008). They offer âa proof of Churchâs Thesis, as Gödel and others suggested may be possibleâ (2008: 299), and they add:
In a similar way, but with a different set of basic operations, one can prove Turingâs Thesis, ⊠. (Dershowitz & Gurevich 2008: 299)
Yet Turingâs own view of the status of his thesis is very different from that expressed by Kripke, Dershowitz and Gurevich. According to Turing, his thesis is not susceptible to mathematical proof. He did not consider either argument I or argument II to be a mathematical demonstration of his thesis: he asserted that I and II, and indeed â[a]ll arguments which can be givenâ for the thesis, are
fundamentally, appeals to intuition, and for this reason rather unsatisfactory mathematically. (Turing 1936 [2004: 74])
Indeed, Turing might have regarded âHilbertâs thesisâ as itself an example of a proposition that can be justified only by appeals to intuition.
Turing discussed a thesis closely related to Turingâs thesis, namely for every systematic method there is a corresponding substitution-puzzle (where âsubstitution-puzzleâ, like âcomputable by Turing machineâ, is a rigorously defined concept). He said:
The statement is ⊠one which one does not attempt to prove. Propaganda is more appropriate to it than proof, for its status is something between a theorem and a definition. (Turing 1954 [2004: 588])
Probably Turing would have taken this remark to apply equally to the thesis (Turingâs thesis) that for every systematic method there is a corresponding Turing machine.
Turing also said (in handwritten material published in 2004) that the phrase âsystematic methodâ
is a phrase which, like many others e.g., âvegetableâ one understands well enough in the ordinary way. But one can have difficulties when speaking to greengrocers or microbiologists or when playing âtwenty questionsâ. Are rhubarb and tomatoes vegetables or fruits? Is coal vegetable or mineral? What about coal gas, marrow, fossilised trees, streptococci, viruses? Has the lettuce I ate at lunch yet become animal? ⊠The same sort of difficulty arises about question c) above [Is there a systematic method by which I can answer such-and-such questions?]. An ordinary sort of acquaintance with the meaning of the phrase âsystematic methodâ wonât do, because one has got to be able to say quite clearly about any kind of method that might be proposed whether it is allowable or not. (Turing in Copeland 2004: 590)
Here Turing is emphasizing that the term âsystematic methodâ is not exact, and so in that respect is like the term âvegetableâ, and unlike mathematically precise terms such as âλ-definableâ, âTuring-machine computableâ, and âsubstitution-puzzleâ. Kleene claimed that, since terms like âsystematic methodâ and âeffectively calculableâ are not exact, theses involving them cannot be proved (op. cit.). Turing however did not voice a similar argument (perhaps because he saw a difficulty). The fact that the term âsystematic methodâ is inexact is not enough to show that there could be no mathematically acceptable proof of a thesis involving the term. Mendelson gave a graphic statement of this point, writing about what is called above âthe converse of Churchâs thesisâ (Section 1.5):
The assumption that a proof connecting intuitive and precise mathematical notions is impossible is patently false. In fact, half of CT (the âeasierâ half), the assertion that all partial-recursive functions are effectively computable, is acknowledged to be obvious in all textbooks in recursion theory. A straightforward argument can be given for itâŠ. This simple argument is as clear a proof as I have seen in mathematics, and it is a proof in spite of the fact that it involves the intuitive notion of effective computability. (Mendelson 1990: 232â233)
Yet the point that the âintuitiveâ nature of some of its terms does not rule out the thesisâs being provable is not to say that the thesis is provable. If the status of the Church-Turing thesis is âsomething between a theorem and a definitionâ, then the definition is presumably Churchâs proposal to âdefine the notion ⊠of an effectively calculable functionâ (Section 1.5) and the theorem is Turingâs computation theorem (Section 4.3.5), i.e., that given Turingâs account of the essential features of human computation, Turingâs thesis is true. This theorem is demonstrable, but to prove the thesis itself from the theorem, it would be necessary to show, with mathematical certainty, that Turingâs account of the essential features of human computation is correct. So far, no one has done this. Propaganda does seem more appropriate than proof.
5. The Church-Turing Thesis and the Limits of Machines
5.1 Two distinct theses
Can the universal Turing machine perfectly simulate the behavior of each and any machine? The Church-Turing thesis is sometimes regarded as providing a statement of the logical limits of machinery, to the effect that the universal Turing machine is the most general machine possible (and so the answer to the question just posed is yes.) For example:
That there exists a most general formulation of machine and that it leads to a unique set of input-output functions has come to be called Churchâs thesis. (Newell 1980: 150)
Yet the Church-Turing thesis is a thesis about the extent of effective methods (therein lies its mathematical importance). Putting this another way, the thesis concerns what a human being can achieve when calculating by rote, using paper and pencil (absent contingencies such as boredom, death, or insufficiency of paper). What a human rote-worker can achieve, and what a machine can achieve, may be different.
Gandy was perhaps the first to distinguish explicitly between Turingâs thesis and the very different proposition that whatever can be calculated by a machine can be calculated by a Turing machine (Gandy 1980). Gandy called this proposition âThesis Mâ. He pointed out that Thesis M is in fact false in the case of some âmachines obeying Newtonian mechanicsâ, where âthere may be rigid rods of arbitrary lengths and messengers travelling with arbitrary large velocitiesâ (1980: 145). He also pointed out that Thesis M fails to apply to what he calls âessentially analogue machinesâ (1980: 125). A most interesting question is whether Thesis M is true of all discrete (i.e., non-analogue) machines that are consistent with the actual laws of physics. This question is discussed in Section 6.4.
Thesis M is imprecise, since Gandy never explicitly specified quite what he meant by âcalculated by a machineâ. It is useful to state a more definite proposition that captures the spirit of Thesis M. This might be called the strong Church-Turing thesis, but on balance it seems preferable to avoid that name, since the proposition in question is very different from the Church-Turing thesis of 1936. The proposition will be called the âmaximality thesisâ.
Some more terminology: A machine m will be said to generate (borrowing this word from Turing 1937: 153) a certain function (e.g., x squared) if m can be set up so that, if m is presented with any of the functionâs arguments (e.g., 4), m will carry out some sequence of processing steps, at the end of which m produces the corresponding value of the function (16 in the example). Mutatis mutandis for functions that, like addition, demand more than one argument.
Maximality thesis:
All functions that can be generated by machine are effectively computable.
âEffectively computableâ is a commonly used term: A function is said to be effectively computable if (and only if) there is an effective method for obtaining its values. When phrased in terms of effective computability, the Church-Turing thesis says: All effectively computable functions are Turing-machine computable.
Clearly the Church-Turing thesis and the maximality thesis are different theses.
5.2 The âequivalence fallacyâ
A common argument for the maximality thesis, or an equivalent, cites the fact, noted above, that many different attempts to analyse the informal notion of computability in precise termsâattempts by Turing, Church, Post, Markov, and othersâturned out to be equivalent to one another, in the sense that each analysis provably picks out the same class of functions, namely those functions computable by Turing machines.
As previously mentioned, this convergence of analyses is often considered strong evidence for the Church-Turing thesis (this is the equivalence argument for the thesisâSection 4.1). Some go further and take this convergence to be evidence also for the maximality thesis. Newell, for example, presented the convergence of the analyses given by Turing, Church, Post, Markov, et al., as showing that
all attempts to ⊠formulate ⊠general notions of mechanism ⊠lead to classes of machines that are equivalent in that they encompass in toto exactly the same set of input-output functions. (Newell 1980: 150)
The various equivalent analyses, said Newell, constitute a
large zoo of different formulations of maximal classes of machines. (ibid.)
Arguably there is a fallacy here. The analyses Newell is discussing are of the concept of an effective method: The equivalence of the analyses bears only on the question of the extent of what is humanly computable, not on the further question whether functions generatable by machines could extend beyond what is in principle humanly computable.
5.3 Watching our words
It may be helpful at this point to survey some standard technical terminology that could set traps for the unwary.
5.3.1 The word âcomputableâ
As already emphasized, when Turing talks about computable numbers, he is talking about humanly computable numbers. He says: âComputing is normally done by writing certain symbols on paperâ (1936 [2004: 75])âand normally done âby human clerical labour, working to fixed rulesâ (1945 [2005: 386]). âThe computerâ, he says, might proceed âin such a desultory manner that he never does more than one step at a sittingâ (1936 [2004: 79]). The work of the human computer is mechanizable: âWe may now construct a machineââthe Turing machineââto do the work of this computerâ (1936 [2004: 77]). See also Section 7 for more quotations relating to this crucial point.
Thus, the various results in âOn Computable Numbersâ to the effect that such-and-such functions are uncomputable are accordingly about human computers. Turing should not be construed as intending to state results about the limitations of machinery. Gandy wrote:
it is by no means obvious that the limitations described in [Section 4.3 above] apply to mechanical devices; Turing does not claim this. (Gandy 1988: 84)
5.3.2 Two instructive quotations
[C]ertain functions are uncomputable in an absolute sense: uncomputable even by [Turing machine], and, therefore, uncomputable by any past, present, or future real machine. (Boolos & Jeffrey 1974: 55)
In the technical logical literature, the term âcomputableâ is usually used to mean âeffectively computableâ (although not alwaysâsee Section 5.3.3). (âEffectively computableâ was defined in Section 5.1.) Since Boolos and Jeffrey are using âcomputableâ to mean âeffectively computableâ, what they are saying in this quotation comes down to the statement that the functions in question are not effectively computable by any past, present, or future real machineâwhich is true, since the functions are, ex hypothesi, not effectively computable. However, to a casual reader of the literature, this statement (and others like it) might appear to say more than it in fact does. That a function is uncomputable (i.e., is effectively uncomputable), by any past, present, or future real machine, does not entail per se that the function in question cannot be generated by some real machine.
The second quotation:
FORMAL LIMITS OF MACHINE BEHAVIORS ⊠There are certain behaviors that are âuncomputableââbehaviors for which no formal specification can be given for a machine that will exhibit that behavior. The classic example of this sort of limitation is Turingâs famous Halting Problem: can we give a formal specification for a machine which, when provided with the description of any other machine together with its initial state, will ⊠determine whether or not that machine will reach its halt state? Turing proved that no such machine can be specified. (Langton 1989: 12)
What is proved is that no Turing machine can always determine, when provided with the description of any Turing machine together with its initial state, whether or not that machine will reach its halt state. Turing certainly proved nothing entailing that it is impossible to specify a machine of some sort or other that can do what Langton describes. Thus, the considerations Langton presents do not impose any general formal limits on machine behaviorsâonly on the behaviors of Turing machines. Yet the quotation gives a different impression. (In passing, it is worth pointing out that although the Halting Problem is very commonly attributed to Turing, as Langton does here, Turing did not in fact formulate it. The Halting Problem originated with Davis in the early 1950s (Davis 1958: 70).)
5.3.3 Beyond effective
Some authors use phrases such as âcomputation in a broad senseâ, or simply âcomputationâ, to refer to computation of a type that potentially transcends effective computation (e.g., Doyle 2002; MacLennan 2003; Shagrir & Pitowsky 2003; Siegelmann 2003; AndrĂ©ka, NĂ©meti, & NĂ©meti 2009; Copeland & Shagrir 2019).
Doyle, for instance, suggested that equilibrating systems with discrete spectra (e.g., molecules or other quantum many-body systems) may illustrate a concept of computation that is wider than effective computation. Since âequilibrating can be so easily, reproducibly, and mindlessly accomplishedâ, Doyle said, we may âtake the operation of equilibratingâ to be a computational operation, even if the functions computable in principle using Turing-machine operations plus equilibrating include functions that are not computable by an unaided Turing machine (Doyle 2002: 519).
5.3.4 The word âmechanicalâ
There is a world of difference between the technical and everyday meanings of âmechanicalâ. In the technical literature, âmechanicalâ and âeffectiveâ are usually used interchangeably: A âmechanicalâ procedure is simply an effective procedure. Gandy 1988 outlines the history of this use of the word âmechanicalâ.
Statements like the following occur in the literature:
Turing proposed that a certain class of abstract machines [Turing machines] could perform any âmechanicalâ computing procedure. (Mendelson 1964: 229)
This could be mistaken for Thesis M. However, âmechanicalâ is here being used in its technical sense, and the statement is nothing more than the Church-Turing thesis:
Turing proposed that a certain class of abstract machines could perform any effective computing procedure.
The technical usage of âmechanicalâ has a tendency to obscure the conceptual possibility that not all machine-generatable functions are Turing-machine computable. The question âCan a machine implement a procedure that is not mechanical?â may appear self-answeringâyet this is what is being asked if Thesis M and the maximality thesis are questioned.
5.4 The strong maximality thesis
The maximality thesis has two interpretations, depending whether the phrase âcan be generated by machineâ is taken in the sense of âcan be generated by a machine conforming to the physical laws of the actual worldâ (the weak form of the thesis), or in a sense that quantifies over all machines, irrespective of conformity to the actual laws of physics (the strong form). (The strong-weak terminology reflects the fact that the strong form entails the weak, but not vice versa.)
The weak form will be discussed in Section 6.4. The strong form is known to be false. This can be shown by giving an example of a notional machine that is capable of generating a function that is not effectively computable. A single example will be provided here; further examples may be found in Andréka et al. 2009, Davies 2001, Hogarth 1994, Pitowsky 1990, Siegelmann 2003, and other papers mentioned below.
5.4.1 Accelerating Turing machines
Accelerating Turing machines (ATMs) are exactly like standard Turing machines except that their speed of operation accelerates as the computation proceeds (Stewart 1991; Copeland 1998a,b, 2002a; Copeland & Shagrir 2011). An ATM performs the second operation called for by its program in half the time taken to perform the first, the third in half the time taken to perform the second, and so on.
If the time taken to perform the first operation is called one âmomentâ, then the second operation is performed in half a moment, the third operation in quarter of a moment, and so on. Since
12+14+18+âŠ+12n+12n+1+âŠâ€1,
an ATM is able to perform infinitely many operations in two moments of operating time. This enables ATMs to generate functions that cannot be computed by any standard Turing machine (and so, by the Church-Turing thesis, these functions are not effectively computable).
One example of such a function is the halting function h. h(n)=1 if the nth Turing machine halts, and h(n)=0 if the nth Turing machine runs on endlessly. It is well known that no standard Turing machine can compute this function (Davis 1958); but an ATM can produce any of the functionâs values in a finite period of time.
When computing h(n), the ATMâs first step is write â0â on a square of the tape called the answer square (A). The ATM then proceeds to simulate the actions of the nth Turing machine. If the ATM finds that the nth machine halts, then the ATM goes on to erase the â0â it previously wrote on A, replacing this by â1â. If, on the other hand, the nth machine does not halt, the ATM never returns to square A to erase the â0â originally written there. Either way, once two moments of operating time have elapsed, A contains the value h(n) (Copeland 1998a).
This notional machine is a counterexample to the strong maximality thesis.
6. Modern Versions of the Church-Turing Thesis
6.1 The algorithmic version
In modern computer science, algorithms and effective procedures are associated not primarily with humans but with machines. Accordingly, many computer science textbooks formulate the Church-Turing thesis without mentioning human computers (e.g., Hopcroft & Ullman 1979; Lewis & Papadimitriou 1981). This is despite the fact that the concept of human computation lay at the heart of Turingâs and Churchâs analyses.
The variety of algorithms studied by modern computer science eclipses the field as it was in Turingâs day. There are now parallel algorithms, distributed algorithms, interactive algorithms, analog algorithms, hybrid algorithms, quantum algorithms, enzymatic algorithms, bacterial foraging algorithms, slime-mold algorithms and more (see e.g., Gurevich 2012; Copeland & Shagrir 2019). The universal Turing machine cannot even perform the atomic steps of algorithms carried out by, e.g., a parallel system where every cell updates simultaneously (in contrast to the serial Turing machine), or an enzymatic system (where the atomic steps involve operations such as selective enzyme binding).
Nevertheless, the universal Turing machine is still able to calculate the behavior of parallel systems and enzymematic systems. The algorithmic version of the Church-Turing thesis states that this is true of every algorithmic system. Thus Lewis and Papadimitriou said: âwe take the Turing machine to be a precise formal equivalent of the intuitive notion of âalgorithmââ (1981: 223). David Harel gave the following (famous) formulation of the algorithmic version of the thesis:
any algorithmic problem for which we can find an algorithm that can be programmed in some programming language, any language, ⊠is also solvable by a Turing machine. This statement is one version of the so-called Church/Turing thesis. (Harel 1992: 233)
Given the extent to which the concept of an algorithm has evolved since the 1930sâfrom the step-by-step labors of human computers to the growth of slime moldâinteresting questions arise. Will the concept continue to evolve? What are the limits, if any, on this evolution? Could the concept evolve in such that a way that the algorithmic version of the Church-Turing thesis is no longer universally true? Returning to Doyleâs suggestions about equilibrating systems (in Section 5.3.3), Doyleâs claim is essentially that the operation of equilibrating could reasonably be regarded as a basic step of some effective procedures or algorithmsâwhether or not the resulting algorithms satisfy the algorithmic version of the Church-Turing thesis. (See Copeland & Shagrir 2019 for further discussion.)
In summary, the algorithmic version of the Church-Turing thesis is broader than the original thesis, in that Church and Turing considered essentially only a single type of algorithm, effective step-by-step calculations on paper. The algorithmic version is also perhaps less secure than the original thesis.
6.2 Computational complexity: the Extended Church-Turing thesis
The Turing machine now holds a central place not only in computability theory but also in complexity theory. Quantum computation researchers Bernstein and Vazirani say:
Just as the theory of computability has its foundations in the Church-Turing thesis, computational complexity theory rests upon a modern strengthening of this thesis. (Bernstein & Vazirani 1997: 1411)
There are in fact two different complexity-theoretic versions of the Church-Turing thesis in the modern computer science literature. Both are referred to as the âExtended Church-Turing thesisâ. The first was presented by Yao in 2003:
The Extended Church-Turing Thesis (ECT) makes the ⊠assertion that the Turing machine model is also as efficient as any computing device can be. That is, if a function is computable by some hardware device in time T(n) for input of size n, then it is computable by a Turing machine in time (T(n))k for some fixed k (dependent on the problem). (Yao 2003: 100â101)
Yao points out that ECT has a powerful implication:
at least in principle, to make future computers more efficient, one only needs to focus on improving the implementation technology of present-day computer designs. (2003: 101)
Unlike the original Church-Turing thesis (whose status is âsomething betweenâ a theorem and a definition) ECT is neither a logico-mathematical theorem nor a definition. If it is true, then its truth is a consequence of the laws of physicsâand it might not be true. (Although it is trivial if, contrary to a standard but unproved assumption in computer science, P = NP.)
The second complexity-theoretic version of the thesis involves the concept of a probabilistic Turing machine (due to Rabin & Scott 1959). Vazirani and Aharonov state the thesis:
[T]he extended Church-Turing thesis ⊠asserts that any reasonable computational model can be simulated efficiently by the standard model of classical computation, namely, a probabilistic Turing machine. (Aharonov & Vazirani 2013: 329)
These two related theses differ considerably from the original Church-Turing thesis, not least in that both extended theses are empirical hypotheses. Moreover, there is ongoing debate as to whether quantum computers in fact falsify these theses. (For an introduction to this debate see Copeland & Shagrir 2019, and for a more detailed treatment see Aharonov & Vazirani 2013.)
6.3 Brain simulation and the Church-Turing thesis
It is sometimes said that the Church-Turing thesis has implications concerning the scope of computational simulation. For example, Searle writes:
Can the operations of the brain be simulated on a digital computer? ⊠The answer seems to me ⊠demonstrably âYesâ ⊠That is, naturally interpreted, the question means: Is there some description of the brain such that under that description you could do a computational simulation of the operations of the brain. But given Churchâs thesis that anything that can be given a precise enough characterization as a set of steps can be simulated on a digital computer, it follows trivially that the question has an affirmative answer. (Searle 1992: 200)
Another example:
we can depend on there being a Turing machine that captures the functional relations of the brain,
for so long as
these relations between input and output are functionally well-behaved enough to be describable by ⊠mathematical relationships ⊠we know that some specific version of a Turing machine will be able to mimic them. (Guttenplan 1994: 595)
Andréka, Németi and Németi state a more general thesis about the power of Turing machines to simulate other systems:
[T]he Physical Church-Turing Thesis ⊠is the conjecture that whatever physical computing device (in the broader sense) or physical thought-experiment will be designed by any future civilization, it will always be simulateable by a Turing machine. (Andréka, Németi, & Németi 2009: 500)
AndrĂ©ka, NĂ©meti, and NĂ©meti even say that the thesis they state here âwas formulated and generally accepted in the 1930sâ (ibid.).
Yet it was not a thesis about the simulation of physical systems that Church and Turing formulated in the 1930s, but rather a completely different thesis concerning human computationâand it was the latter thesis that became generally accepted during the 1930s and 1940s.
It certainly muddies the waters to call a thesis about simulation âChurchâs thesisâ or the âChurch-Turing thesisâ, because the arguments that Church and Turing used to support their actual theses go no way at all towards supporting the theses set out in the several quotations above. Nevertheless, what can be termed the âSimulation thesisâ has its place in the present catalogue of modern forms of the Church-Turing thesis:
Simulation thesis:
Any system whose operations can be characterized as a set of steps (Searle) or whose input-output relations are describable by mathematical relationships (Guttenplan) can be simulated by a Turing machine.
If the Simulation thesis is intended to cover all possible systems then it is surely false, since Doyleâs envisaged equilibrating systems falsify it (Section 5.3.3). If, on the other hand, the thesis is intended to cover only actual physical systems, including brains, then the Simulation thesis is, like the Extended Church-Turing thesis, an empirical thesisâand so is very different from Turingâs thesis and Churchâs thesis. The truth of the âactual physical systemsâ version of the Simulation thesis depends on the laws of physics.
One potential objection that any upholder of the Simulation thesis will need to confront parallels a difficulty that Gandy raised for Thesis M (Section 5.1). Physical systems that are not discreteâsuch as Gandyâs âessentially analogue machinesââappear to falsify the Simulation thesis, since the variables of a system with continuous dynamics take arbitrary real numbers as their values, whereas a Turing machine is restricted to computable real numbers, and so cannot fully simulate the continuous system.
This brings the discussion squarely to one of the most interesting topics in the area, so-called âphysical versionsâ of the Church-Turing thesis.
6.4 The Church-Turing thesis and physics
6.4.1 The Deutsch-Wolfram thesis
In 1985, Wolfram formulated a thesis that he described as âa physical form of the Church-Turing hypothesisâ:
[U]niversal computers are as powerful in their computational capacities as any physically realizable system can be, so that they can simulate any physical system. (Wolfram 1985: 735)
Deutsch (who laid the foundations of quantum computation) independently stated a similar thesis, again in 1985, and also described it as a âphysical versionâ of the Church-Turing thesis:
I can now state the physical version of the Church-Turing principle: âEvery finitely realizable physical system can be perfectly simulated by a universal model computing machine operating by finite meansâ. This formulation is both better defined and more physical than Turingâs own way of expressing it. (Deutsch 1985: 99)
This thesis is certainly âmore physicalâ than Turingâs thesis. It is, however, a completely different claim from Turingâs own, so it is potentially confusing to present it as a âbetter definedâ version of what Turing said. As already emphasized, Turing was talking about effective methods, whereas the theses presented by Deutsch and Wolfram concern all (finitely realizable) physical systemsâno matter whether or not the systemâs activity is effective.
In the wake of this early work by Deutsch and Wolfram, the phrases âphysical form of the Church-Turing thesisâ, âphysical version of the Church-Turing thesisââand even âthe physical Church-Turing thesisââare now quite common in the current literature. However, such terms are probably better avoided, since these physical theses are very distant from Turingâs thesis and Churchâs thesis.
In his 1985 paper, Deutsch went on to point out that if the description âa universal model computing machine operating by finite meansâ is replaced in his physical thesis by âa universal Turing machineâ, then the result:
Every finitely realizable physical system can be perfectly simulated by a universal Turing machine
is not true. His reason for saying so is the point discussed at the end of Section 6.3, concerning non-discrete physical systems. Deutsch argued that a universal Turing machine âcannot perfectly simulate any classical dynamical systemâ, since â[o]wing to the continuity of classical dynamics, the possible states of a classical system necessarily form a continuumâ, whereas the universal Turing machine is a discrete system (Deutsch 1985: 100). Deutsch then went on to introduce the important concept of a universal quantum computer, saying (but without proof) that this is âcapable of perfectly simulating every finite, realizable physical systemâ (1985: 102).
The following formulation differs in its details from both Wolframâs and Deutschâs theses, but arguably captures the spirit of both. In view of the Deutsch-Gandy point about continuous systems, the idea of perfect simulation is replaced by the concept of simulation to any desired degree of accuracy:
Deutsch-Wolfram Thesis:
Every finite physical system can be simulated to any specified degree of accuracy by a universal Turing machine. (Copeland & Shagrir 2019)
Related physical theses were advanced by Earman 1986, Pour-El and Richards 1989, Pitowsky 1990, Blum et al. 1998, and others. The Deutsch-Wolfram thesis is closely related to Gandyâs Thesis M, and to the weak maximality thesis (Section 5.4). In fact the Deutsch-Wolfram thesis entails the latter (but not vice versa, since the maximality thesis concerns only machines, whereas the Deutsch-Wolfram thesis concerns the behavior of all finite physical systemsâalthough any who think that every finite physical system is a computing machine will disagree; see e.g., Pitowsky 1990).
Is the Deutsch-Wolfram thesis true? This is an open question (Copeland & Shagrir 2020)âso too for the weak maximality thesis. One focus of debate is whether physical randomness, if it exists, falsifies these theses (Calude et al. 2010; Calude & Svozil 2008; Copeland 2000). But even in the case of non-random systems, speculation stretches back over at least six decades that there may be real physical processes (and so, potentially, machine-operations) whose behavior is neither computable nor approximable by a universal Turing machine. See, for example, Scarpellini 1963, Pour-El and Richards 1979, 1981, Kreisel 1967, 1974, 1982, Geroch and Hartle 1986, Pitowsky 1990, Stannett 1990, da Costa and Doria 1991, 1994, Hogarth 1994, Siegelmann and Sontag 1994, Copeland and Sylvan 1999, Kieu 2004, 2006 (see Other Internet Resources), Penrose 1994, 2011, 2016.
To select, by way of example, just one paper from this list: Pour-El and Richards showed in their 1981 article that a system evolving from computable initial conditions in accordance with the familiar three-dimensional wave equation is capable of exhibiting behavior that falsifies the Deutsch-Wolfram thesis. However, now as then, it is an open question whether these initial conditions are physically possible.
6.4.2 The âGandy argumentâ
Gandy (1980) gave a profound discussion of whether there could be deterministic, discrete systems whose behavior cannot be calculated by a universal Turing machine. The now famous âGandy argumentâ aims to show that, given certain reasonable physical assumptions, the behavior of every discrete deterministic mechanism is calculable by Turing machine. In some respects, the Gandy argument resembles and extends Turingâs argument I, and Gandy regarded it as an improved and more general alternative to Turingâs I (1980: 145). He emphasized that (unlike Turingâs argument), his argument takes âparallel working into accountâ (1980: 124â5); and it is this that accounts for much of the additional complexity of Gandyâs analysis as compared to Turingâs.
Gandy viewed the conclusion of his argument (that the behavior of every discrete deterministic mechanism is calculable by Turing machine) as relatively a priori, provable on the basis of a set-theoretic derivation that makes very general physical assumptions (namely, the four axioms mentioned in Section 3.4). These assumptions include, for instance, a lower bound on the dimensions of a mechanismâs components, and an upper bound on the speed of propagation of effects and signals. (The argument aims to cover only mechanisms obeying the principles of Relativity.) Gandy expressed his various physical assumptions set-theoretically, by means of precise axioms, which he called Principles I â IV. Principle III, for example, captures the idea that there is a bound on the number of types of basic parts (atoms) from which the states of the machine are uniquely assembled; and Principle IVâwhich Gandy called the âprinciple of local causationââcaptures the idea that each state-transition must be determined by the local environments of the parts of the mechanism that change in the transition.
Gandy was very clear that his argument does not apply to continuous systemsâanalogue machines, as he called themâand non-relativistic systems. (Extracts from unpublished work by Gandy, in which he attempted to develop a companion argument for analogue machines, are included in Copeland & Shagrir 2007.) However, the scope of the Gandy argument is also limited in other ways, not noted by Gandy himself. For example, some asynchronous algorithms fall outside the scope of Gandyâs principles (Gurevich 2012; Copeland & Shagrir 2007). Gurevich concludes that Gandy has not shown âthat his axioms are satisfied by all discrete mechanical devicesâ, and Shagrir says there is no âbasis for claiming that Gandy characterized finite machine computationâ (Gurevich 2012: 36, Shagrir 2002: 234). It will be useful to give some examples of discrete deterministic systems that, in one way or another, evade Gandyâs conclusion that the behavior of every such system is calculable by Turing machine.
First, it is relatively trivial that mechanisms satisfying Gandyâs four principles may nevertheless produce uncomputable output from computable input if embedded in a universe whose physical laws have Turing-uncomputability built into them, e.g., via a temporal variable (Copeland & Shagrir 2007). Moreover, some asynchronous algorithms fall outside the scope of Gandyâs principles (Gurevich 2012; Copeland & Shagrir 2007). Second, certain (notional) discrete deterministic ârelativistic computersâ also fall outside the scope of Gandyâs principles. Relativistic computers were described in a 1987 lecture by Pitowsky (Pitowsky 1990), and in Hogarth 1994 and Etesi & NĂ©meti 2002. The idea is outlined in the entry on computation in physical systems; for further discussion see Shagrir and Pitowsky 2003, Copeland and Shagrir 2020.
The NĂ©meti relativistic computer makes use of gravitational time-dilation effects in order to compute (in a broad sense) a function that provably cannot be computed by a universal Turing machine (e.g., the halting function). NĂ©meti and his colleagues emphasize that the NĂ©meti computer is ânot in conflict with presently accepted scientific principlesâ and that, in particular, âthe principles of quantum mechanics are not violatedâ. They suggest moreover that humans might âeven buildâ a relativistic computer âsometime in the futureâ (AndrĂ©ka, NĂ©meti, & NĂ©meti 2009: 501).
According to Gandy,
- âA discrete deterministic mechanical device satisfies principles I-IVâ (he called this âThesis Pâ; Gandy 1980: 126), and
- âWhat can be calculated by a device satisfying principles I-IV is computableâ (he labelled this âTheoremâ).
1 and 2 together yield: What can be calculated by a discrete deterministic mechanical device is (Turing-machine) computable.
However, the NĂ©meti computer is a discrete, deterministic mechanical device, and yet is able to calculate functions that are not Turing-machine computable. That is to say, relativistic computers are counterexamples to Gandyâs Thesis P. In brief, the reason for this is that the sense of âdeterministicâ implicitly specified in Gandyâs Principles (âGandy-deterministicâ) is narrower than the intuitive sense of âdeterministicâ, where a deterministic system is one obeying laws that involve no randomness or stochasticity. Relativistic computers are deterministic but not Gandy-deterministic. (For a fuller discussion, see Copeland, Shagrir, & Sprevak 2018.)
In conclusion, Gandyâs analysis has made a considerable contribution to the current understanding of machine computation. But, important and illuminating though the Gandy argument is, it certainly does not settle the question whether the Deutsch-Wolfram thesis is true.
6.4.3 Quantum effects and the âTotalâ thesis
There is a stronger form of the Deutsch-Wolfram thesis, dubbed the âTotal thesisâ in Copeland and Shagrir 2019.
The Total Thesis:
Every physical aspect of the behavior of any physical system can be calculated (to any specified degree of accuracy) by a universal Turing machine.
Logically, the Total thesis is counter-exampled by the universal Turing machine itself (assuming that the universal machine, with its indefinitely long tape, is at least a notional physical system; see Copeland & Shagrir 2020 for discussion of this assumption). This is because there is no algorithm for calculating whether a universal Turing machine halts on every given inputâi.e., there is no algorithm for calculating that aspect of the machineâs behavior. The question remains, however, whether the Total thesis is infringed by any systems that are âmore physicalâ than the universal machine. (Notice that such systems, if any exist, do not necessarily also infringe the Deutsch-Wolfram thesis, since it is possible that, even though answers to certain physical questions about the system are uncomputable, the system is nevertheless able to be simulated by a Turing machine.)
Interestingly, recent work in condensed matter quantum physics indicates thatâpossiblyâquantum many-body systems could infringe the Total thesis. In 2012, Eisert, MĂŒller and Gogolin established the surprising result that
the very natural physical problem of determining whether certain outcome sequences cannot occur in repeated quantum measurements is undecidable, even though the same problem for classical measurements is readily decidable. (Eisert, MĂŒller & Gogolin 2012: 260501.1)
This was a curtain-raiser to a series of dramatic results about the uncomputability of quantum phase transitions, by Cubitt and his group (Cubitt, Perez-Garcia, & Wolf 2015; Bausch, Cubitt, Lucia, & Perez-Garcia 2020; Bausch, Cubitt, & Watson 2021). These results concern the âspectral gapâ, an important determinant of the properties of a substance. A quantum many-body system is said to be âgappedâ if the system has a well-defined next least energy-level above the systemâs ground energy-level, and is said to be âgaplessâ otherwise (i.e., if the energy spectrum is continuous). The âspectral gap problemâ is the problem of determining whether a given many-body system is gapped or gapless.
The uncomputability results of Cubitt et al. stem from their discovery that the halting problem can be encoded in the spectral gap problem. Deciding whether a model system of the type they have studied is gapped or gapless, given a description of the local interactions, is âat least as hard as solving the Halting Problemâ (Bausch, Cubitt, & Watson 2021: 2). Moreover, this is not just a case of uncomputability in, uncomputability out. Uncomputability arises even though the initial conditions are computable (as with the notional system described in Pour-El and Richards 1981, mentioned in Section 6.4.1). Cubitt et al. emphasize:
the phase diagram is uncomputable even for computable (or even algebraic) values of its parameter Ï. Indeed, it is uncomputable at a countably-infinite set of computable (or algebraic) values of Ï. (Bausch, Cubitt, & Watson 2019: 8)
However, Cubitt admits that the models used in the proofs are somewhat artificial:
Whether the results can be extended to more natural models is yet to be determined. (Cubitt, Perez-Garcia & Wolf 2015: 211)
In short, it is an openâand fascinatingâquestion whether there are realistic physical systems that fail to satisfy the Total thesis.
7. Some Key Remarks by Turing and Church
7.1 Turing machines
Turing prefaced his first description of a Turing machine with the words:
We may compare a man in the process of computing a ⊠number to a machine. (Turing 1936 [2004: 59])
The Turing machine is a model, idealized in certain respects, of a human being calculating in accordance with an effective method.
Wittgenstein put this point in a striking way:
Turingâs âMachinesâ. These machines are humans who calculate. (Wittgenstein 1947 [1980: 1096])
It is a point that Turing was to emphasize, in various forms, again and again. For example:
A man provided with paper, pencil, and rubber, and subject to strict discipline, is in effect a universal machine. (Turing 1948 [2004: 416])
In order to understand Turingâs âOn Computable Numbersâ and later texts, it is essential to keep in mind that when he used the words âcomputerâ, âcomputableâ and âcomputationâ, he employed them not in their modern sense as pertaining to machines, but as pertaining to human calculators. For example:
Computers always spend just as long in writing numbers down and deciding what to do next as they do in actual multiplications, and it is just the same with ACE [the Automatic Computing Engine] ⊠[T]he ACE will do the work of about 10,000 computers ⊠Computers will still be employed on small calculations ⊠(Turing 1947 [2004: 387, 391])
Turingâs ACE, an early electronic stored-program digital computer, was built at the National Physical Laboratory, London; a pilot versionâat the time the fastest functioning computer in the worldâfirst ran in 1950, and a commercial model, the DEUCE, was marketed very successfully by English Electric.
7.2 Human computation and machine computation
The electronic stored-program digital computers for which the universal Turing machine was a blueprint are, each of them, computationally equivalent to a Turing machine, and so they too are, in a sense, models of human beings engaged in computation. Turing chose to emphasize this when explaining these electronic machines in a manner suitable for an audience of uninitiates:
The idea behind digital computers may be explained by saying that these machines are intended to carry out any operations which could be done by a human computer. (Turing 1950a [2004: 444])
He made the point a little more precisely in the technical document containing his design for the ACE:
The class of problems capable of solution by the machine [the ACE] can be defined fairly specifically. They are [a subset of] those problems which can be solved by human clerical labour, working to fixed rules, and without understanding. (Turing 1945 [2005: 386])
Turing went on to characterize this subset in terms of the amount of paper and time available to the human clerk.
It was presumably because he considered the point to be essential for understanding the nature of the new electronic machines that he chose to begin his Programmersâ Handbook for Manchester Electronic Computer Mark II with this explanation:
Electronic computers are intended to carry out any definite rule of thumb process which could have been done by a human operator working in a disciplined but unintelligent manner. (Turing _c_1950: 1)
It was not some deficiency of imagination that led Turing to model his L.C.M.s on what could be achieved by a human computer. The purpose for which he invented the Turing machine demanded it. The Entscheidungsproblem is the problem of finding a humanly executable method of a certain sort, and, as was explained earlier, Turingâs aim was to show that there is no such method in the case of the full first-order predicate calculus.
7.3 Church and the human computer
Turing placed the human computer center stage in his 1936 paper. Not so Church. Church did not mention computation or human computers explicitly in either of his two groundbreaking papers on the Entscheidungsproblem (Church 1936a,b). He spoke of âeffective calculabilityâ, taking it for granted his readers would understand this term to be referring to human calculation. He also used the term âeffective methodâ, again taking it for granted that readers would understand him to be speaking of a humanly executable method.
Church also used the term âalgorithmâ, saying
It is clear that for any recursive function of positive integers there exists an algorithm using which any required particular value of the function can be effectively calculated. (Church 1936a: 351)
He said further that the notion of effective calculability could be spelled out as follows:
by defining a function to be effectively calculable if there exists an algorithm for the calculation of its values. (Church 1936a: 358)
It was in Churchâs review of Turingâs 1936 paper that he brought the human computer out of the shadows. He wrote:
[A] human calculator, provided with pencil and paper and explicit instructions, can be regarded as a kind of Turing machine. It is thus immediately clear that computability, so defined [i.e., computability by a Turing machine], can be identified with (especially, is no less general than) the notion of effectiveness as it appears in certain mathematical problems ⊠and in general any problem which concerns the discovery of an algorithm. (Church 1937a: 43)
7.4 Turingâs use of âmachineâ
It is important to note that, when Turing used the word âmachineâ, he often meant not machine-in-general but, as we would now say, Turing machine. At one point he explicitly drew attention to this usage:
The expression âmachine processâ of course means one which could be carried out by the type of machine I was considering [in âOn Computable Numbersâ]. (Turing 1947 [2004: 378â9])
Thus when, a few pages later, Turing asserted that âmachine processes and rule of thumb processes are synonymousâ (1947 [2004: 383]), he is to be understood as advancing the Church-Turing thesis (and its converse), not a version of the maximality thesis. Unless his intended usage is borne in mind, misunderstanding could ensue. Especially liable to mislead are statements like the following, which a casual reader might mistake for a formulation of the maximality thesis:
The importance of the universal machine is clear. We do not need to have an infinity of different machines doing different jobs. A single one will suffice. The engineering problem of producing various machines for various jobs is replaced by the office work of âprogrammingâ the universal machine to do these jobs. (Turing 1948 [2004: 414])
In context it is perfectly clear that these remarks concern machines equivalent to Turing machines; the passage is embedded in a discussion of L.C.M.s.
Whether or not Turing would, if queried, have assented to the weak maximality thesis is unknown. There is certainly no textual evidence in favor of the view that he did so assent. The same is true of the Deutsch-Wolfram thesis and its cognates: there is no textual evidence that Turing would have assented to any such thesis.
7.5 Churchâs version of Turingâs thesis
Interestingly, the summary of Turingâs account of computability given by Church in his 1937 review was not entirely correct. Church said:
The author [Turing] proposes as a criterion that an infinite sequence of digits 0 and 1 be âcomputableâ that it shall be possible to devise a computing machine, occupying a finite space and with working parts of a finite size, which will write down the sequence to any desired number of terms if allowed to run for a sufficiently long time. (Church 1937a: 42)
However, there was no requirement proposed in Turingâs 1936 paper that Turing machines occupy âa finite spaceâ or have âworking parts of a finite sizeâ. Nor did Turing couch matters in terms of the machineâs writing down âany desired number of termsâ of the sequence, âif allowed to run for a sufficiently long timeâ. Turing said, on the contrary, that a sequence is âcomputable if it can be computed by a circle-free machineâ (Turing 1936 [2004: 61]); where a machine is circle-free if it is not one that
never writes down more than a finite number of symbols [0s and 1s]. (Turing 1936 [2004: 60])
In consequence, Churchâs version of Turingâs thesis is subtly different from Turingâs own:
Churchâs Turingâs thesis:
An infinite sequence of digits is âcomputableâ if (and only if) it is possible to devise a computing machine, occupying a finite space and with working parts of a finite size, that will write down the sequence to any desired number of terms if allowed to run for a sufficiently long time.
In so far as Church includes these three finiteness requirements (i.e., that the machine occupy a finite space, have finite-sized parts, and produce finite numbers of digits), his version of Turingâs thesis can perhaps be said to be âmore physicalâ than any of Turingâs formulations of the thesis. Churchâs finiteness requirements are in some respects reminiscent of Gandyâs idea that the states of a discrete deterministic calculating machine must be built up iteratively from a bounded number of types of basic components, the dimensions of which have a lower bound (see Section 6.4.2). Although, as explained there, Gandy imposes further requirements on a discrete deterministic calculating machine, and these go far beyond Churchâs finiteness requirements.
Notwithstanding Churchâs efforts to inject additional physical realism into the concept of a Turing machine, it isâas in Turingâs caseâunknown whether Church would, if queried, have assented to the Deutsch-Wolfram thesis or any cognate thesis. There seems to be no textual evidence either way. Church was simply silent about such matters.