A Beginner's Further Guide To Mathematical Logic
A Beginner's Further Guide To Mathematical Logic
com
www.pdfgrip.com
Raymond Smullyan
www.pdfgrip.com
Published by
British Library Cataloguing-in-Publication Data A catalogue record for this book is available from the
British Library.
On the cover, the three photos from left to right are the logicians Emil Post, Alan Turing, and Ernst
Zermelo.
All rights reserved. This book, or parts thereof, may not be reproduced in any form or by any means,
electronic or mechanical, including photocopying, recording or any information storage and retrieval
system now known or to be invented, without written permission from the publisher.
For photocopying of material in this volume, please pay a copying fee through the Copyright Clearance
Center, Inc., 222 Rosewood Drive, Danvers, MA 01923, USA. In this case permission to photocopy is not
required from the publisher.
Printed in Singapore
www.pdfgrip.com
Contents
Preface
I. Magic Sets
II. Gentzen Sequents and Some Variants
III. Craig’s Lemma and an Application
IV. A Unification
V. A Henkin-Style Completeness Proof
I. A Decision Machine
II. Variations on a Theme of Gödel
www.pdfgrip.com
III. R-Systems
IV. A Synthesis
Chapter 6. Doubling Up
I. Simple Systems
II. Standard Simple Systems
I. The B-Combinators
II. The Permuting Combinators
III. The Q-Family and the Goldfinch, G
IV. Combinators Derivable from B,T,M and I (λ-I Combinators)
References
Index
www.pdfgrip.com
Preface
Part I
Chapter 1
More on Propositional Logic
x y x × y
E E E
E O E
www.pdfgrip.com
O E E
O O O
Note the similarity of this table to the truth table for disjunction. Indeed, if we
replace E by T, O by F, and × by ∨, we get the truth table for disjunction:
x y x ∨ y
T T T
T F T
F T T
F F F
Note: Readers familiar with modern higher algebra know that a Boolean ring
is a ring with the additional properties that x + x = 0 and x2 = x for any element x
of the ring.
the reader to take it in its equivalent “negated” form, I often indicate this fact.
We also recall from The Beginner’s Guide [Smullyan, 2014] the unifying α, β
notation in propositional logic for unsigned formulas:
α α1 α2
X ∧ Y X Y
∼(X ∨ Y) ∼X ∼Y
∼(X ⊃ Y) X ∼Y
∼∼X X X
β β1 β2
∼(X ∧ Y) ∼X ∼Y
X ∨ Y X Y
X ⊃ Y ∼X Y
Problem 1. S is a Boolean truth set if and only if for every α and every formula
X the following conditions hold:
(1) α ∈ S iff α1 ∈ S and α2 ∈ S.
(2) Either X ∈ S or (∼X) ∈ S, but not both.
Exercise 1. Prove that S is a Boolean truth set iff for every β and every formula
X the following conditions hold:
(1) β ∈ S iff either β 1 ∈ S or β 2 ∈ S.
(2) Either X ∈ S or (∼X) ∈ S, but not both.
The definition of a Boolean truth set here can be shown to be equivalent to
the one given for a truth set in Chapter 6 (Propositional Tableaux) of The
Beginner’s Guide [Smullyan, 2014], where we called a set S of propositional
formulas a truth set if it was the set of all formulas true under some
interpretation I of the propositional variables. We also showed there that the
definition was equivalent to the set S satisfying the following three conditions:
for any signed formula X and any α and β :
T0: Either X or its conjugate is in S, but not both.
T1: α is in S if and only if α1 and α2 are both in S.
www.pdfgrip.com
With these facts in mind, the reader should easily see that the definitions of
(Boolean) truth set in The Beginner’s Guide and this work are equivalent. Note
that once one sees the equivalence between the definitions in the two books, it is
obvious that every Boolean truth set for propositional logic is satisfiable (since
the characterization in The Beginner’s Guide made a truth set satisfiable by
definition).
Indeed, to go a little further (to make connections between very similar
concepts a little clearer), we often define a Boolean valuation as follows. First,
by a valuation (in the context of propositional logic) is meant an assignment of
truth values t and f to all formulas. A valuation ∨ is called a Boolean valuation if
for all formulas X and Y, the following conditions hold:
B1: ∼X receives the value t under v if and only if v(X) = f [i.e. when X is false
under v]; and accordingly v(∼X) = f iff v(X) = t.
B2: X ⋀ Y receives the value t under v iff v(X) = v(Y) = t [i.e. when both X and Y
are true under the valuation v].
B3: X ∨ Y receives the value t under v iff v(X) = t or v(Y) = t [i.e. when either X is
true under v or Y is true under v, or both are true under v].
B4: X ⊃ Y receives the value t under v iff v(X) = f or v(Y) = t [i.e. when either X
is false under v or Y is true under v].
The interpretations that we used in The Beginner’s Guide in terms of
assigning truth values to all formulas by assigning truth values to the
propositional variables, and then working up through the structure of the
formulas, using the meanings of the connectives, to obtain a truth value for every
formula, are closely related to the truth values of formulas given through a
Boolean valuation. In fact a given Boolean valuation defines a unique
interpretation yielding the same truth value on all formulas in the interpretation
as in the Boolean valuation: we obtain the interpretation by looking at the truth
values the Boolean valuation assigns to the propositional variables. And if one is
given an interpretation of the propositional variables, it is clear that the
interpretation defines a Boolean valuation on the formulas, just by the way the
interpretation assigns truth values to all formulas, starting out only with truth
values on the propositional variables.
Because of this tight equivalence between Boolean valuations and
interpretations in propositional logic, we will use the words “valuation” and
www.pdfgrip.com
Consequence Relations
Consistency
Maximal Consistency
Compactness
A1: X ⊃ (Y ⊃ X).
Problem 13. To prove the above lemma, show that for any formulas X and Y, the
formula (X ⊃ Y) ⊃ (X ⊃ X) is provable. Then, taking Y ⊃ X for Y in the formula
just proved, conclude the proof of Lemma 2.
Problem 14. In preparation for the proof of the Deduction Theorem, show the
following:
(a) If S ⊢ Y, then S ⊢ X ⊃ Y.
(b) If S ⊢ X ⊃ (Z ⊃ Y) and S ⊢ X ⊃ Z, then S ⊢ X ⊃ Y.
Problem 15. Prove the Deduction Theorem as follows: Suppose S, X ⊢ Y. Then
there is a deduction Y1, …, Yn from S ∪ {X} in which Y =Yn. Now consider the
sequence X ⊃ Y1, …, X ⊃ Yn. Show that for all i ≤ n, if S ⊢ X ⊃ Yj for all j < i
then S ⊢ X ⊃ Yi. [Hint: Break the proof up into the following four cases: Yi is an
axiom of ; Yi ∈ S; Yi = X; Yi comes from two terms of the sequence Y1, …, Yi −1
by modus ponens.]
It then follows by complete mathematical induction that S ⊢ X ⊃ Yi for all i ≤
n, and so in particular, S ⊢ X ⊃ Yn, and thus S ⊢ X ⊃ Y. [Hint: In the original
sequence Y1,…, Yn, the term Yi is either an axiom of , or is X itself, or comes
from two earlier terms by modus ponens. Handle each case separately.]
We now have all the parts necessary to prove:
Theorem 3. A sufficient condition for a standard axiom system for propositional
logic to be tautologically complete is that for all X, Y, Z and α, the following
are provable:
A1: X ⊃ (Y ⊃ X)
A2: (X ⊃ (Y ⊃ Z)) ⊃ ((X ⊃ Y) ⊃ (X ⊃ Z))
A3: (a) α ⊃ α1 and (b) α ⊃ α2
www.pdfgrip.com
A4: α1 ⊃ (α2 ⊃ α)
A5: X ⊃ (∼X ⊃ Y)
A6: (∼X ⊃ X) ⊃ X
Problem 16. Prove Theorem 3 by showing that under the conditions given in
Theorem 3, the relation ⊢ is a Boolean consequence relation and thus
tautologically complete.
One can easily verify that A1–A6 are all provable in the axiom system S of
Chapter 7 of The Beginner’s Guide [Smullyan, 2014]. [One can make a tableau
proof for each of these formulas and then convert them to proofs in S by the
method explained in that chapter. The completeness proof for tableaux is not
needed for this.] We thus have a second proof of the completeness of S based on
maximal consistency, and thus the theorem:
Theorem 4. The axiom system S of Chapter 7 of The Beginner’s Guide is
tautologically complete.
Maximal consistency and Lindenbaum’s Lemma also play a key role in
alternative completeness proofs for first-order logic, to which we turn in the next
chapter.
(a) Suppose S is a Boolean truth set. We must show that in all cases for α,
the first condition of Problem 1 holds, namely that α∈ S iff α1∈ S and
α2 ∈ S.
Case 1. α is of the form X Λ Y. This is immediate.
Case 2. α is of the form ∼(X ∨ Y). Then α1 = ∼ X and α2 = ∼Y. Thus
we are to show that (∼(X ∨ Y))∈ S iff (∼X) ∈ S and (∼Y) ∈ S. Well,
• (∼(X ∨ Y)) ∈ S is true iff (X ∨ Y) ∉ S [by (4) in the definition of a
www.pdfgrip.com
consistent).
This completes the proof.
8. To show that if all finite subsets of a set S are consistent, then so is S, it
suffices to show that if S is inconsistent, then some finite subset of S is
inconsistent. Well, suppose S is inconsistent. Then for any formula X, we
have both S ⊢ X and S ⊢ ∼ X (since S ⊢ Z for every Z). Since S ⊢ X, then
F1 ⊢ X for some finite subset F1 of S [by C4]. Similarly, F2 ⊢ ∼ X for some
finite subset F2 of S. Let F = F1 ∪ F2. Then F ⊢ X and F ⊢ ∼ X [by C2],
since F1 F and F2 F). Also, X, ∼ X ⊢ Z for all Z [by C7]. Thus, for all Z, F
⊢ Z [by Problem 2], and we have shown that F is inconsistent.
9. We wish to show that every Boolean consequence relation is tautologically
complete. By Problem 8 we know that for any Boolean consequence
relation, consistency with respect to that relation is a compact property, and
so [by the denumerable compactness theorem] every consistent set is a
subset of a maximally consistent set, which must be a Boolean truth set [by
the Key Lemma]. Thus every consistent set is satisfiable (by the discussion
after the definition of a Boolean truth set), and hence every unsatisfiable set
is inconsistent. Now, suppose X is a logical consequence of S. Then S, ∼ X
is unsatisfiable, hence inconsistent. Therefore S, ∼ X ⊢ X (since S, ∼ X ⊢ Z
for every Z), and so S ⊢ X [by C8]. This concludes the proof.
10. Suppose S ⊢ X and S ⊢ X ⊃ Y. It is also true that X, X ⊃ Y ⊢ Y, since we
assume ⊢ is faithful to modus ponens. Hence S ⊢ Y [by Problem 2].
11. Proofs of the four statements:
(a) Suppose X ⊃ Y is provable and S ⊢ X. Since X ⊃ Y is provable ( ⊢ X
⊃ Y), then S ⊢ X ⊃ Y [by C2]. Thus S ⊢ X and S ⊢ X ⊃ Y ; hence S ⊢
Y [by Problem 10].
(b) Suppose X ⊃ (Y ⊃ Z) is provable, and S ⊢ X and S ⊢ Y. Since S ⊢ X
and S ⊢ X ⊃ (Y ⊃ Z) [by C2], then S ⊢ Y ⊃ Z [by Problem 10]. Since
also S ⊢ Y, then S ⊢ Z [again by Problem 10].
(c) In 11(a), we take the unit set { X} for S, and thus if X ⊃ Y is provable
and X ⊢ X, then X ⊢ Y. But X ⊢ X does hold, so that if X ⊃ Y is
provable, then X ⊢ Y.
(d) In 11(b), we take the set { X, Y } for S, and thus if X ⊃ (Y ⊃ Z) is
www.pdfgrip.com
16. We already know that conditions C1−C4 hold, and so it remains to verify
conditions C5−C8.
Re C5: (a) Since α ⊃ α 1 is provable [by A3(a)], then α ⊢ α1 [by Problem
11(c)].
(b) Since α ⊃ α2 is provable [by A3(b)], then α ⊢ α2 [by Problem
11(c)].
Re C6: Since α1 ⊃ (α2 ⊃ α) is provable [by A4], then α1, α2 ⊢ α [by
Problem 11(d)].
Re C7: Since X ⊃ (∼X ⊃ Y) is provable [by A5], then X, ∼X ⊢ Y [by
Problem 11(d)].
Re C8: [This is the only part that needs the Deduction Theorem!] Suppos
that S, ∼X ⊢ X Then, by the deduction theorem, S ⊢ ∼X ⊃ X. But
also (∼X ⊃ X) ⊃ X is provable [by A6]. Therefore S ⊢ X [by
Problem 11(a), taking ∼X ⊃ X for X and X for Y ].
www.pdfgrip.com
Chapter 2
More on First-Order Logic
I. Magic Sets
There is another approach to Completeness Theorems, the Skolem–Löwenheim
Theorem and the Regularity Theorem that has an almost magic-like simplicity. I
accordingly dub the central player a magic set.
As in propositional logic, in the context of first-order logic a valuation is an
assignment of truth values t and f to all sentences (with or without parameters).
And a valuation v is called a Boolean valuation if for all sentences (closed
formulas) X and Y, the following conditions hold:
B1: The sentence ∼X receives the value t under v if and only if v(X) = f [i.e.
when X is false under v]; and accordingly v(∼X) = f iff v(X) = t.
B2: The sentence X ∧ Y receives the value t under v iff v(X) = v(Y) = t [i.e. when
both X and Y are true under the valuation v].
B3: The sentence X ∨ Y receives the value t under v iff v(X) = t or v(Y) = t [i.e.
when either X is true under v or Y is true under v].
B4: The sentence X ⊃ Y receives the value t under v iff v(X) = f or v (Y) = t [i.e.
when either X is false under v or Y is true under v].
By a first-order valuation (in the denumerable domain of the parameters) is
meant a Boolean valuation satisfying the following two additional conditions:
F1: The sentence ∀xφ(x) is true under v if and only if for every parameter a, the
sentence φ(a) is true under v.
F2: The sentence ∃xφ(x) is true under v if and only if the sentence φ(a) is true
under v for at least one parameter a.
Let S be a set of sentences of first-order logic. We say that S is truth-
www.pdfgrip.com
(1) ∼∃xφ(x) is true iff ∼φ(a) is true for every parameter a, or, equivalently,
that ∼∃xφ(x) and ∀x∼φ(x) are logically equivalent (i.e. always have the
same truth value);
(2) ∼∀xφ(x) is true iff ∼φ(a) is true for at least one parameter a, or,
equivalently, that ∼∀xφ(x) and ∃x∼φ(x) are logically equivalent.
Now, let us again recall the uniform notation of The Beginner’s Guide, in
which we let γ be any formula of the form ∀xφ(x) or ∽∃xφ(x), and by γ(a) we
meant φ(a) or ∽φ(a) respectively. Similarly, we let δ be any formula of the form
∃xφ(x) or ∼∀xφ(x), and by δ(a) we respectively mean φ(a), ∽φ(a). Thus, in
uniform notation, under any first-order valuation v:
: The sentence γ is true under v if and only if for every parameter a, the
sentence γ(a) is true under v.
: The sentence δ is true under v if and only if the sentence δ(a) is true under v
for at least one parameter a.
The following proposition will be helpful:
www.pdfgrip.com
Proposition 1. Suppose that v is a Boolean valuation such that the following two
conditions hold (for all sentences γ and δ):
(1) if γ is true under v, then for every parameter a, the sentence γ(a) is true
under v, and
(2) if δ is true under v, then γ(a) is true under v for at least one parameter a.
Then v is a first-order valuation.
Problem 1. Prove Proposition 1.
We say that a set S of sentences of first-order logic is a Boolean truth set if
there is a Boolean valuation v such that S is the set of all sentences that are true
under v. This is equivalent to saying that for all sentences X and Y, (i) (∼X) ∈ S
iff X ∉ S; (ii) α ∈ S iff α1 ∈ S and α2 ∈ S; (iii)β ∈ S iff β1 ∈ S or β2 ∈S.
M2: For every finite set S0 of pure sentences, and every finite subset M0 of M, if
S0 is first-order satisfiable, so is S0 ∪ M0.
We shall soon prove that magic sets exist (which the reader might find
surprising!). But first you will see that one of the important features of magic
sets emerges from the following considerations:
We say that a sentence X is tautologically implied by a set of sentences S
when X is true under all Boolean valuations that satisfy S. We shall call S a
truth-functional basis for first-order logic if for any pure sentence X, the
sentence X is valid if and only if X is tautologically implied by some finite subset
S0 of S (which is equivalent to saying that there are finitely many elements X1,
…, Xn of S such that the sentence (X1∧ … ∧ Xn) ⊃ X is a tautology).
The mere existence of magic sets (even those not necessarily regular) yields
particularly simple and elegant proofs of the Skolem–Löwenheim Theorem and
of the first-order compactness theorem as a consequence of the compactness
theorem for propositional logic. We will see these proofs in a moment.
Note: But first a reminder of something from The Beginner’s Guide
[Smullyan, 2014]. It was noted there that if we consider the set of all true
sentences under a first-order valuation v, and look at the universe in which the
individuals to which predicates apply as consisting of all the parameters, we can
define an interpretation in which each n-ary predicate is true precisely on the n-
tuples of parameters on which the valuation v says it is true (and false on the rest
of n-tuples of parameters). This is called a (first-order) interpretation in the
(denumerable) domain of the parameters, and because of this we can say a set of
sentences S is first-order satisfiable in the domain of the parameters whenever
there is a first-order valuation that satisfies S. Although we are not using the
concept of an interpretation of first-order logic at the moment here, it is easy to
see that any interpretation of first-order logic simultaneously defines a first-order
valuation over the denumerable domain of the parameters.
Now suppose S is a denumerable set of pure sentences (no free variables, no
parameters) such that every finite subset of S is first-order satisfiable. Here is an
www.pdfgrip.com
also consider. Craig’s original proof of this result was quite complicated, and
many simpler proofs of this have subsequently been given. One of these which I
will give uses a variant of some axiom systems due to Gerhard Gentzen [1934,
1935], to which we now turn.
Background
sets). For this purpose, he needed a proof procedure that required that any proof
of a formula X used only subformulas of X. Such a proof procedure is said to
obey the subformula principle. Of course, tableaux using signed formulas obey
the subformula principle (and these are closely related to the Gentzen systems,
as we will see). Tableaux using unsigned formulas, almost, but not quite, obey
this principle — they use either subformulas or negations of them. [For example
∼X is not necessarily a subformula of ∼(X ∨ Y), but is the negation of one.]
Gentzen took a new symbol “→” and defined a sequent to be an expression of
the form θ → Γ, where each of θ, Γ is a finite, possibly empty, sequence of
formulas. Such an expression is interpreted to mean that if all terms of are true,
then at least one term of Γ is true. That is, under any valuation v, the sequent θ
→ Γ is true under v iff it is the case that if all terms of θ are true, then at least
one term of is Γ true. A sequent is called a tautology if it is true under all
Boolean valuations, and valid if it is true under all first-order valuations. If θ is a
non-empty sequence X1, …, Xn and Γ is a non-empty sequence Y1, …, Yk, then
the sequent X1, …, Xn → Y1, …, Yk is logically equivalent to the formula (X1 ∧
… ∧Xn) ⊃ (Y1∨ … ∨Yk). If θ is the empty sequence, then θ → Γ is written as
, where is the empty set, or more simply → Γ.
Now, if Γ is a non-empty sequence, then the sequent → Γ (which is ) is
read “if all elements of the empty set are true, then at least one element of Γ is
true.” Well, all elements of the empty set are true (since there are no elements of
the empty set), and so the sequent → Γ is logically equivalent to the proposition
that at least one element of Γ is true. Thus a sequent →Y1, …, Yk is logically
equivalent to the formula Y1 ∨ … ∨Yk.
Now, suppose that the right side Γ of the sequent θ → Γ is empty. The
sequent is then more simply written as θ →. It is interpreted to mean that if all
the terms of the sequence θ are true, then at least one element of the empty set
is true. Well, it is obviously not the case that some element of the empty set is
true, so that θ → expresses the proposition that it is not the case that all terms of
the sequence θ are true. If θ is a non-empty sequence X1, …, Xn, then the sequent
X1, …, Xn → is logically equivalent to the formula ∼ (X1 ∧ … ∧Xn).
case that at least one element set if true. Thus the sequent → is simply false
(under all valuations).
At this point, I must tell you an amusing incident: I was once giving a lecture
on Gentzen sequents to a mathematics group. I started with Gentzen sequents in
which both sides of the arrow were non-empty, and after having erased, first the
left side, and then the right side, I was left with just the arrow alone on the board.
After explaining that it meant falsehood, the logician William Craig (the
inventor of Craig’s Interpolation Lemma) was in the audience, and with his
typical sense of humor, raised his hand and asked, “And what does it mean if
you write nothing on the board?”
In more modern Gentzen type systems, the sequents are of the form U → V,
where U and V are finite sets of formulas, instead of sequences of formulas, and
this is the course we will take.
We shall first consider a Gentzen type axiom system G0 for propositional logic.
For any finite set S of formulas and any formulas X1, …, Xn, we abbreviate S ∪
{X1, …, Xn} by S, X1, …, Xn. Thus,
abbreviates
The system G0 for propositional logic has only one axiom scheme and the
eight inference rules below. Here X and Y are formulas of propositional logic,
and U and V are sets of formulas of propositional logic.
Conjunction
www.pdfgrip.com
Disjunction
Implication
Negation
The inference rules introduce the logical connectives. There are two rules for
each connective — one to introduce the connective on the left side of the arrow,
the other to introduce the connective on the right side of the arrow. As an
example of how the rules are applied, the conjunction rule C1 is that from the
sequent U, X, Y → V one can infer the sequent U, X, ∧ Y → V. As for Rule C2, it
says that from the two sequents U → V, X and U → V, Y one can infer the
sequent U → V, X ∧ Y.
It should be obvious that the axioms of G0 are tautologies, and we leave it to
the reader to verify that the inference rules preserve tautologies, in other words,
that in any application of an inference rule, the conclusion is truth-functionally
implied by the premises. It therefore follows (by mathematical induction) that
we have: All provable sequents of G0 are tautologies; thus the system G0 is
correct.
www.pdfgrip.com
We will soon prove that the system G0 is complete, i.e. that all tautologies are
provable in G0.
Let S be a non-empty set {TX1, …, TXn, FY1, …, FYk} of signed formulas. By |S|
we shall mean the sequent X1, …, Xn → Y1, …, Yk. We call |S| the corresponding
sequent of S. We leave it to the reader to verify the important fact that S is
unsatisfiable if and only if its corresponding sequent |S| is a tautology (and, for
later use when we get to Gentzen systems for first-order logic), that if the
elements of S are signed formulas of first-order logic, then S is unsatisfiable iff
|S| is valid.
Of course, if S is a set {FY1, …, FYk} in which there are no formulas signed
T, we take |S| to be the sequent → Y1, …, Yk, and if S is a set {TX1, …, TXn} in
which there are no formulas signed F, we take |S| to be the sequent X1, …, Xn
→.
More generally, for any set of signed formulas, if U is the set of formulas X
such that TX ∈ S, and V is the set of formulas Y such that FY ∈ S, then |S| is the
sequent U → V.
The system G0 in uniform notation is the following:
Axioms
Inference Rules
Problem 6. Verify that the above uniform system really is the system G0.
Proofs in Gentzen type systems are usually in tree form, unlike the linear
form of proofs in the axiom systems previously considered. Only now, unlike the
trees of tableaux, the tree must grow upwards if the user wishes to prove a
statement by analyzing it from the outside in (as must take place by the tableau
rules): the origin is at the bottom and the end points are at the top. The origin is
the sequent to be proved, and the end points are the axioms used in the proof. As
an example, here is a proof in tree form of the sequent p ⊃ q → ∼q ⊃ ∼p.
Completeness of G0
We now wish to show that all tautologies are provable in the Gentzen system G0
for propositional logic.
By a tableau proof of a sequent X1, …, Xn → Y1, …, Yk is meant a closed
tableau proof showing that the set {TX1, …, TXn, FY1, …, FYk} is unsatisfiable
(if the tableau is started with all these formulas, the tableau can be extended by
the tableau rules for propositional logic in such a way that every branch closes).
We will now show how, from a tableau proof of a sequent, we can obtain a proof
of the sequent in the Gentzen system G0. To this end, it will be useful to first
introduce, as an intermediary, another type of tableau that I will now call a block
tableau (this was called a modified block tableau in my book First-Order Logic
[Smullyan, 1968, 1995]). The analytic tableaux, which we have studied so far,
are variants of the tableaux of Evert Beth [1959], whereas the block tableaux, to
which we now turn, are variants of the tableaux of J. Hintikka [1955]. Like the
tableaux of Hintikka, the points of the tree in a block tableau proof (this time
again with the origin at the top, as in analytic tableaux) are not single formulas,
but are finite sets of formulas, which we call the blocks of the tableaux. And
www.pdfgrip.com
what we do at any stage of the construction depends solely on the end points of
the tree (rather than on the branch as a whole, as in analytic tableaux).
We will first consider block tableaux for propositional logic. By a block
tableau for a finite set K of signed formulas of propositional logic, we mean a
tree constructed by placing the set K at the origin, and then continuing according
to the following rules (here, for a set S and formulas X, Y, we abbreviate the sets
S ∪ {X} and S ∪ {X} ∪ {Y} by {S, X} or {S, X, Y} respectively):
A′:
B′:
We call a block tableau closed if each end point contains some element and
its conjugate.
Changing from an analytic tableau to a block tableau is a relatively simple
matter. To begin with, in constructing an analytic tableau, when using Rule A,
and inferring α1 from α, subjoin α2 immediately after it. That is, use Rule A in
the form:
www.pdfgrip.com
Now, to start a block tableau for a sequent X1, …, Xn → Y1, …, Yk, put the set
{TX1, …, TXn, FY1, …, FYk} at the origin, for we wish to prove that set of
formulas is unsatisfiable. This concludes stage 0 of the construction of the block
tableau corresponding to the analytic tableau that begins as follows:
Then, for each n, after completion of stage n for the two tableaux, when we
use Rule A (Rule B) in the analytic tableau, we use Rule A′ (Rule B′,
respectively) in the block tableau, and that concludes stage n + 1 of the
constructions. We continue this until both tableaux close (which must happen at
some stage, if the set of formulas {TX1, …, TXn, FY1, …, FYk} is unsatisfiable,
i.e. the sequent is a tautology). Since we know that the tableau method for
propositional logic is complete, this construction shows that the block tableau
method for propositional logic is also complete.
As an example, here is a closed analytic tableau for the sequent p ⊃ q, q ⊃ r
→ p ⊃ r:
www.pdfgrip.com
To convert a block tableau to a proof in the Gentzen system G0, just replace each
block
by the sequent X1, …, Xn → Y1, …, Yk, and the resulting tree, when turned
upside down, is a proof in the Gentzen system.
For example, from the above closed block tableau, we get the following proof
in the Gentzen system:
www.pdfgrip.com
For the Gentzen system G1, we take the axioms and inference rules of the system
G0 (only now applied to formulas of first-order logic instead of formulas of
propositional logic), and add the following inference rules:
One then converts a closed block tableau into a proof in the Gentzen system
www.pdfgrip.com
G1 as before, i.e. one replaces each block {S} by the sequent |S| and turns the
resulting tree upside down.
The System GG
(A)
(B)
(C)
(D)
The Completeness of GG
Gentzen system, a sentence might change from one side of the arrow to the
other). We now need an analytic tableau system in which this does not happen.
By an altered tableau we mean one constructed according to the following
rules:
www.pdfgrip.com
www.pdfgrip.com
In the quantification rules (above and below), the asterisk “*” means that the
parameter in the conclusion must not appear in the premise.
Here are the rules of the system altered tableaux in uniform notation (where
the α, β, γ and δ are of course unsigned formulas):
let A1 = Tα1 and let A2 = Tα2; if the latter, we let A1 = Fβ1 and let A2 = Fβ2. We
let B be any sentence of the form Tβ or Fα. If it is the former, we let B1, B2 be
Tβ1 and Tβ2 respectively; if the latter, we let B1, B2 be Fα1 and Fα2, respectively.
We let C be any sentence of the form Tγ or Fδ, and we take C(a) to be Tγ(a),
Fδ(a), respectively. We let D be any sentence of the form Tδ or Fγ, and we take
D(a) to be Tδ(a), Fγ(a), respectively. With this notation, the altered tableau rules
for first-order logic are now most succinctly expressed as follows:
Problem 7. Prove that the altered tableau system is correct and complete, and
then finish the proof of the completeness of the Gentzen system GG.
so that the existence of interpolants for all valid sequents is equivalent to the
existence of interpolants for all valid sentences X ⊃ Y. Thus we will consider
Craig’s Lemma for First-Order Logic in the equivalent form that there exist
interpolants for all valid sequents.
We will prove Craig’s Lemma for the Gentzen system GG.
It suffices to show that there is an interpolant for each of the axioms of GG, and
that for each of the inference rules of GG, if there is an interpolant for the
premise, or interpolants for the two premises, then there in an interpolant for the
conclusion.
Exercise. Prove this for the axioms and for Rules A and B (which don’t involve
the quantifiers).
Proving the claim of the exercise really proves Craig’s Interpolation Lemma
for propositional logic. Now let’s turn to Rules C and D, which do involve the
quantifiers.
The proof for Rule D is relatively simple. Suppose that X is an interpolant for
U, δ(a) → V, and that a does not occur in U, δ or V. Then X must also be an
interpolant for U, δ → V for the following reasons:
Since X is an interpolant of U, δ(a) → V, then every parameter of X occurs in
V (as well as in U, δ(a)). But since a does not occur in V (by hypothesis), then a
cannot occur in X. From this, it follows that all parameters of X occur in U, δ (as
www.pdfgrip.com
well as in V).
To see this, let b be any parameter that occurs in X. We know that b ≠ a. We
also know that b occurs in V (all parameters of X do). It remains to show that b
occurs in U, δ. If b occurs in U, we are done. If not, then b must occur in δ(a)
(since b occurs in U, δ(a)). But the parameters of δ are those of δ(a) other than a.
Since b occurs in δ(a) and b ≠ a, then b must occur in δ. This proves that all
parameters of X occur in U, δ (as well as in V). Also all predicates of X occur in
U, δ and in V since they all occur in U, δ(a) and in V, and the predicates of δ are
the same as those of δ(a).
Finally, since U, δ(a) → X is valid (as well as X → V), and a does not occur
in U, δ or in X, then U,δ → X is valid (in fact it follows from U, δ(a) → X by
Rule D). Thus X is an interpolant for the sequent U, δ → V.
The proof that if X is an interpolant for U → V, γ(a), and a does not occur in
U or in V, γ (a), then X is an interpolant for U → V, γ is similar to the preceding
proof, and is left to the reader. This takes care of Rule D.
The proof for Rule C is much trickier! Suppose that X is an interpolant for U,
γ(a) → V. Then U, γ(a) → X and X → V are of course valid, but X may fail to be
an interpolant for U, γ → V because the parameter a might occur in X, but not in
U, γ. If a does not occur in X, or if a does occur in U, γ, there is no problem. The
critical case is one in which a does occur in X, but not in U, γ. For this case, we
find another interpolant for U, γ → V as follows:
We take some variable x that does not occur in X and we let φ(x) be the result
of replacing every occurrence of a in X by x. We note that φ(a) is X itself. We
will show that ∀ xφ(x) is an interpolant for U, γ → V.
Let γ′ = ∀xφ(x). Then γ′(a) = φ(a), which is X. We already know that U, γ →
γ′(a) is valid (because it is U, γ → X), and since a does not occur in U, γ, and
obviously not in γ′ (which is ∀xφ(x)), then U, γ → γ′ is valid (it is in fact
derivable from U, γ → γ′(a) by Rule D of the system GG). Also γ′(a) → V
(which is X → V) is valid by hypothesis. Hence γ′ → V is valid (by Rule C,
taking U to be the empty set). Thus U, γ → γ′ and γ′ → V are both valid, and all
predicates and parameters of γ′ are both in U, γ and in V (since a does not occur
in γ′). Thus ∀xφ(x), which is γ′, is an interpolant for U, γ → V. This completes
the proof for the γ rule of Rule C.
As to the δ rule of Rule C, a similar proof, which we leave to the reader,
www.pdfgrip.com
When this is true, we also say that the sentences of S implicitly define the
predicate P.
We remark that this condition is equivalent to the condition that for any two
interpretations that satisfy S, if they agree on each of P1, …, Pn, then they agree
www.pdfgrip.com
This theorem is far from obvious, but Craig’s Lemma provides a neat and
elegant proof of it.
Suppose that P is implicitly definable from P1, …, Pn with respect to S. Let P′
and S′ be as previously defined, and assume we have
Let X be the conjunction of the elements of S (the order does not matter), and let
X′ be the conjunction of the elements of S′. We thus have
Since the above sentence is valid, then for any parameter a the sentence (X ∧ X′)
⊃ (Pa ≡ P′a) is valid. [We recall that S is a set of sentences without parameters
because that is one of the conditions on both explicit and implicit definitions.
Thus the sentence X ∧ X′ has no parameters.] Hence by propositional logic, the
sentence (X ∧ X′)⊃ (Pa ⊃ P′a), is valid. Thus, again by propositional logic, so is
the following sentence:
In this sentence, P does not occur in X′ ⊃ P′a and P ′ does not occur in X ∧ Pa.
By Craig’s Lemma, there is an interpolant Z for the sentence (X ∧ Pa)⊃ (X′ ⊃
P′a). All predicates and parameters of Z occur in both X ∧ Pa and in X′ ⊃ P′a.
Hence neither P nor P′ can occur in Z; in fact all the predicates of Z are among
P1, …, Pn. Also,Z contains no parameters other than a, since X and X′ contain no
parameters, as noted above. Let x be a variable that does not occur in Z, and let
www.pdfgrip.com
IV. A Unification
Let us review some of the main results that we have proved about first-order
logic in The Beginner’s Guide [Smullyan, 2014] and in this chapter.
T1: The Completeness Theorem for Analytic Tableaux. Every valid formula is
provable by the tableau method.
T2: The Skolem–Löwenheim Theorem. For any set S of formulas, if S is
satisfiable, then it is satisfiable in a denumerable domain.
T3: The Compactness Theorem. For any infinite set S of formulas, if every finite
www.pdfgrip.com
subset of S is satisfiable, so is S.
T4: The Completeness of a First-Order Axiom System.
T5: The Regularity Theorem. For every valid pure sentence X, there is a (finite)
regular set R which truth-functionally implies X.
T6: Craig’s Interpolation Lemma. If the sentence X ⊃ Y (of propositional or
first-order logic) is valid, there is a sentence Z such that X ⊃ Z and Z ⊃ Y are
both valid and, for propositional logic, every propositional variable occurring
in Z occurs in both X and Y, while, for first-order logic, every predicate and
parameter of Z occurs in both X and Y.
In my paper “A Unifying Principle in Quantification Theory” [Smullyan,
1963], I stated and proved a result that yields all of T1 − T6 as special cases. To
this we now turn. What follows is slightly different from the original formulation
of 1963.
Consider any property ϕ of sentences. A set of sentences that has property ϕ
will be called ϕ-consistent (i.e. “consistent” with having the property ϕ).
We are now going to be especially interested in certain sets of sentences that
have what we will call an analytic consistency property, and we will use the
Greek symbol Γ to signal that a property of sets is an analytic consistency
property. Thus we define a set S of sentences to be Γ-consistent — which means
the same as saying that the set S has the analytic consistency property Γ — if the
following conditions hold:
Γ0: S contains no element and its negation (or no element and its conjugate, if we
are considering signed formulas).
Γ1: For any α in S, the sets S ∪ {α1} and S ∪ { α2} are both Γ-consistent (and
hence so is S ∪{υ1, υ2}).
Γ2: For any β in S, at least one of the sets S ∪ {β1} and S ∪ {β2}is Γ-consistent.
Γ3: For any γ in S and any parameter a, the set S Γ {γ(a)} is Γ-consistent.
Γ4: For any δ in S, the set S ∪ {δ(a)} is Γ-consistent for any parameter a that
does not occur in any element of S.
Theorem U. [The Unification Theorem] For any analytic consistency property
Γ, if S is Γ-consistent, then S is satisfiable in the domain of the parameters.
One proof of Theorem U is quite similar to the completeness proof for
tableaux: Given a Γ-consistent set S, it easily follows that no tableau for S can
close (verify this!). Hence a systematic tableau for S will contain an open branch
www.pdfgrip.com
that includes all elements of S, and is a Hintikka set, and thus satisfiable in the
domain of the parameters.
Discussion. We proved Theorem U using analytic tableaux, although the
theorem makes no reference to tableaux. My 1963 proof used neither tableaux
nor König’s lemma, and is as follows: We shall consider now only the case that
S is denumerable (the construction for a finite set S is even simpler, and the
modification is left to the reader). We shall also assume that only finitely many
parameters are found in the formulas of S (if not, we can always introduce a new
denumerable group of parameters, and use them along with the parameters in the
formulas of S, in the construction ahead).
In what follows, consistent will mean Γ-consistent. Given a consistent
denumerable set of sentences S, the idea is to define an infinite sequence of
formulas whose set of terms is a Hintikka set that contains all the elements of S.
The sequence is defined in stages: at each stage we have a finite segment (X1,
…, Xk) of our desired infinite sequence, and at the next stage, we extend that
sequence to a larger finite sequence (X1, …, Xk, …, Xk+m).
First some definitions and notations: Let θ be a finite sequence (X1, …, Xk).
We shall say that θ is consistent with S when the set S ∪{X1, …, Xk} is
consistent. For any formulas Y1, …, Ym, by the expression θ, Y1, …, Ym is meant
the sequence (X1, …, Xk, Y1, …, Ym).
Now suppose we have completed the nth stage and have on hand a sequence
(X1, …, Xk), where k ≥ n. We call this sequence θn, and we assume it to be
consistent with S. We then extend θn to a finite sequence θn+1 in a manner
depending on the nature of Xn:
(a) If Xn is an α, we take θn+1 to be θn, α1, α2, sn+1.
(b) If Xn is an β, then either θn, β1 or θn, β2 is consistent with S. If the former we
take βn+1 to be θn, β1, sn+1; if the latter, we take θn+1 to be θn, β2, sn+1.
(c) If Xn is a γ, we let a be the first parameter for which γ(a) is not a term of θn,
and we take θn+1 to be θn,γ(a), γ, sn+1.
www.pdfgrip.com
Give an axiom system for first-order logic, call a set S consistent (with
respect to the system) if no element of S is refutable, and no finite set {X1, …,
Xn} of elements of S are such that the conjunction X1∧ … ∧ Xn is refutable. For
the axiom system given in The Beginner’s Guide [Smullyan, 2014], or for any
other standard system, this consistency is easily shown to be an analytic
consistency property. Hence every consistent set is satisfiable and every
unsatisfiable set is inconsistent. If now X is valid, then ∽X is unsatisfiable, and
hence inconsistent. This means that ∽X is refutable, i.e. (∽∽X) is provable, and
so is X. Thus every valid formula is provable, and the system is complete.
T5: (The Regularity Theorem).
Let us review some definitions and facts: We recall that a (finite) regular set
R means a set whose members can be arranged in a sequence
M which is both maximally Γ-consistent and E-complete, and that the set M
obtained in this way is a Hintikka set, and hence satisfiable (so that M ’s subset S
is also satisfiable).
Problem 10. Show that if M is both maximally Γ-consistent and E-complete,
then M is a Hintikka set.
Note: As previously remarked, for any synthetic consistency property Δ of
finite character, if M is both maximally Δ-consistent and E-complete, then M is
not only a Hintikka set, but even a truth set. Indeed, more generally, any
Hintikka set S having the property that for every X, either X ∊ S or (∽ X) ∊ S,
must be a truth set. We leave the proof of this as an exercise for the reader.
Now, given an analytic consistency property Γ of finite character, how do we
actually go about extending a given Γ-consistent set S of finite character to a set
M which is both maximally Γ-consistent and E-complete? (As in our proof of
Theorem U above, we will assume that the number of parameters in S is finite,
or that we have an additional denumerable number of parameters available to us,
none of which occur in any formula of S, i.e. parameters that we can just add to
the parameters in S if necessary.)
Here is Lindenbaum’s method of extending a consistent set S to a maximally
consistent set M (assuming consistency to be of finite character), which we will
extend from the conventional meaning of consistent to our meaning of Γ-
consistent (i.e. analytically consistent): We first arrange all the sentences of first-
order logic in some infinite sequence X1, X2 …, Xn, Xn+1, …. Now consider a Γ-
consistent set S. We construct an infinite sequence of sets S0, S1, …, Sn, Sn+1, …
as follows: We take S0 to be S. Then, for every positive n, assuming Sn already
defined, we take Sn+1 to be Sn if Sn ∪ {Xn} is not consistent; while if Sn ∪ {Xn}
is consistent, and if Xn is not some δ, then we take Sn+1 to be Sn ∪ {Xn}. But if
Sn ∪ {Xn} is consistent and Xn is some δ, then Sn ∪ { δ, δ(a)} is also consistent
for any parameter a new to Sn ∪ {δ} (since we are dealing with an analytic
consistency property). In this case we take Sn+1 to be Sn ∪ {δ, δ(a)}, where a is
a parameter new to Sn and to δ. The set M that results from taking the union of
all the sets S0, S1, …, Sn, Sn+1, … is both maximally consistent and E-complete.
For it is easily seen to be E-complete by construction (realize that at every stage
of the construction, there will always be a new parameter available). And the
union M of all the consistent sets Sn is consistent because we are assuming our
www.pdfgrip.com
(a) Suppose that X is a valid pure sentence. Then X is true under all first-
order valuations, and thus also true under all Boolean valuations that
satisfy M (because all such valuations are also first-order valuations by
condition M1 in the definition of a magic set). Thus X is tautologically
implied by M [by the definition of tautological implication], and hence
by some finite subset M0 of M [by Problem 8 of Chapter 6 of The
Beginner’s Guide [Smullyan, 2014]].
(b) Conversely, suppose X is a pure sentence that is tautologically implied
by some finite subset M0 of M. Then the set M0∪{∼X} is not truth-
functionally satisfiable (i.e. not satisfied by any Boolean valuation), so
that M0∪{∼X} is not first-order satisfiable (since every first-order
valuation is also a Boolean valuation). If ∼X were first-order satisfiable,
then M0 ∪ {∼X} would also be first-order satisfiable [by M2], but since
it isn’t, then ∼X is not first-order satisfiable, which means that X is
valid.
This completes the proof.
3. We must show that if M is a complete regular set of sentences, then M is a
magic set. So let M be a complete regular set of sentences.
Proof of M1: We must show that any Boolean valuation that satisfies M is
also a first-order valuation. So suppose v is a Boolean valuation that satisfies
M. By Proposition 1, to show that v is a first-order valuation, it suffices to
show that if γ is true under v, so is γ(a) for every parameter a, and that if δ is
true under v, then so is δ(a) for at least one parameter a. Well, suppose γ is
true under v. Since M is complete, γ ⊃ γ(a) is in M for every parameter a.
Thus γ ⊃ γ (a) is true under v (since every member of M is true under v).
Then γ (a) must also be true under v (since v is a Boolean valuation).
Now suppose δ is true under v. Since M is complete, δ ⊃ δ (a) is in M
for some parameter a. Thus δ ⊃ δ (a) if true under v (since every member of
M is true under v). Then δ (a) must also be true under v (since v is a
Boolean valuation).
Proof of M2: We must show that for every finite set S of pure sentences,
and every finite subset M0 of M, if S is first-order satisfiable, so is S ∪ M0.
We first show that for any finite set S of sentences (possibly with
www.pdfgrip.com
By (1) and (2) above, S1 is satisfiable, and for each i < n, if Si is satisfiable,
so is Si+1. It then follows by mathematical induction that each of the sets S1,
S2, …, Sn is satisfiable. Thus S ∪ M0, which is Sn, is satisfiable.
www.pdfgrip.com
As for completeness, one can make a systematic altered tableau for a set
in essentially the same manner as for analytic tableaux (in the instructions
given in The Beginner’s Guide [Smullyan, 2014], just replace α,β,γ,δ by
A,B,C,D respectively). If the systematic altered tableau runs on forever
without closing, then it contains at least one infinite open branch, and the
set S of the sentences on the branch satisfies the following conditions:
(0) For no X in S is it the case that both T X and F X are both in S, or that T
X and T ∼X are both in S or that F X and F ∼X are both in S.
(1) If A ∊ S, so are A1 and A2.
(2) If B ∊ S, then B1 ∊ S or B2 ∊ S.
(3) If C ∊ S, then so is C(a) for every parameter a.
(4) If D ∊ S, then so is D(a) for at least one parameter a.
Such a set S is very much like a Hintikka set, and the proof of its
satisfiability is very close to that for Hintikka set. About the only significant
difference is in the assignment of truth values to the atomic sentences. Now,
for any atomic unsigned sentence X, if T X ∊ S or F ∼X ∊ S, give X the
value truth (in which case T X and F ∼X are then true). If either T ∼X ∊ S
or F X ∊ S, then give X the value false (in which case T ∼X and F X are
then true). [No ambiguity can result, because if T X ∊ S or F ∼X ∊ S, then
neither T ∼X ∊ S nor F X ∊ S can hold, and if T ∼X ∊ S or F X ∊ S, then
neither T X ∊ S nor F ∼X ∊ S can hold.] If none of T X, T ∼X, FX, F ∼X
occur in S, then give X the value truth or falsity at will — say truth. Under
this valuation, all elements of S can be shown to be true by mathematical
induction on degrees, much as in the proof of Hintikka’s Lemma. Thus S is
satisfiable.
We now see that if the systematic altered tableau for S does not close,
then S is satisfiable, hence if S is not satisfiable, then any systematic altered
tableau for S must close. This completes the completeness proof for the
altered tableau method.
To complete the proof of the completeness of the Gentzen system GG,
we now need altered block tableaux for first-order logic, whose rules in
uniform notation are the following:
www.pdfgrip.com
(i.e. “from the sequent |S2|, to infer the sequent |S1|”). For any block tableau
rule of the form
(i.e. “from the sequents |S2| and |S3|, to infer the sequent |S1|”). Well, the
counterparts of the eight block tableau rules are precisely the eight rules of
www.pdfgrip.com
the system GG. To see this, let U be the set of unsigned sentences X such
that T X ∊ S and let V be the set of sentences Y such that F Y ∊ S. Consider,
for example, the block tableau rule
which is
whose counterpart is the sequent U → V, each end point S is closed, and the
sequence |S| is an axiom of GG:
(i) |S, T X, F X| is U, X → V, X;
(ii) |S, T X, T ∼X| is U, X, ∼X → V ;
(iii) |S, F X, F ∼X| is U → V, X, ∼X.
And since the counterparts of the block rules are rules of GG, it follows that
if we replace each point S of the tree by the sequent |S| and turn the tree
upside down, we have a proof of the sequent U → V in the system GG. This
concludes the proof of the completeness of GG.
8. It is understood that P′ is a predicate distinct from each of P, P1, …, Pn, and
that P′ occurs in no element of S, and that S′ is the result of substituting P′
for P in every element of S.
Now suppose that P is explicitly definable from P1, …, Pn with respect
to S. Let φ(x) be a formula whose predicates are among P1, …, Pn but do
not involve P and which is such that
which implies
9. (a) Suppose every unsatisfiable set has an associate. Consider any valid
sentence X. Then the unit set {∼X} is unsatisfiable, hence has an associate
R. Since R ∪ {∼X} is not truth-functionaly satisfiable, then R tautologically
implies X [since X must be clearly true in any Boolean valuation satisfying
R]. Moreover, X contains no critical parameter of R, which proves the
Regularity Theorem.
www.pdfgrip.com
(b) Conversely, assume the Regularity Theorem. Let S be a finite set {X1,
…, Xn} of sentences which is unsatisfiable. Then the sentence ∼(X1 Λ
… Λ Xn) is valid. Hence by the assumed Regularity Theorem, there is a
regular set R which tautologically implies ∼(X1 Λ … Λ Xn) and which is
such that no critical parameter of R occurs in any element of ∼(X1 Λ …
Λ Xn), and thus no critical parameter of R occurs in any element of S.
Since R tautologically implies ∼(X1 Λ … Λ Xn), then R ∪ {X1, …, Xn} is
truth-functionally unsatisfiable, i.e. R ∪ S is truth-functionally
unsatisfiable. Thus R is an associate of S.
10. We are given that M is maximally Γ-consistent and E-complete, and we
are to show that M is a Hintikka set. In what follows, “consistent” will
mean Γ-consistent. We shall say that a sentence X is consistent with M if
the set M ∪ {X} is consistent. Since M is maximally consistent, then if X is
consistent with M, then X must be a member of M (otherwise M would be
only a proper subset of the larger consistent set M ∪ {X}, contrary to M’s
maximality).
H0: No formula and its conjugate (or formula and its negation, if unsigned
formulas are in questions) can be in the set M, because M is consistent,
and that is one of the conditions on (Γ) consistency.
H1: Suppose α ∊ M. Then of course is consistent with M. Thus α1 and α2
are both consistent with M (since Γ is an analytic consistency property).
Therefore α1 and α2 are members of M. This proves that if α ∊ M, then
α1 ∊ M and α2 ∊ M.
H2: Suppose β ∊ M. Then β is consistent with M and thus either β1 is
consistent with M or β2 is consistent with M, and therefore either β1 ∊ M
or β2 ∊ M.
H3: Suppose γ ∊ M. Then γ is consistent with M, so that γ(a) is consistent
with M for every parameter a. Therefore γ(a) ∊ M for every parameter a.
H4: Suppose δ ∈ M. Then δ(a) ∈ M for at least one parameter a by the
assumption that M is E-complete.
By H1–H4, M is a Hintikka set.
www.pdfgrip.com
Part II
Chapter 3
Some Special Topics
I. A Decision Machine
We consider a calculating machine with infinitely many registers R1, R2,…,Rn,
…. We call n the index of Rn. Each register is, so to speak, in charge of a certain
property of numbers. To find out whether a given number n has a certain
property, one goes to the register in charge of that property and feeds in the
www.pdfgrip.com
number. The machine then goes into operation, and one of three things happens:
(1) The machine eventually halts and flashes a signal — say a green light —
signifying that the number does have the property, in which case we say
that the register affirms n.
(2) The machine eventually halts and flashes a different signal — say a red
light — to indicate that the number doesn’t have the property, in which case
we say that the register denies n.
(3) The machine runs on forever, without ever being able to determine whether
or not the number does have the property, in which case we say that n
stumps the register, or that the register is stumped by n.
We assume we are given a machine M that satisfies the following two
conditions:
M1: With each register R in the calculating machine is associated another register
R′ in the calculating machine called the opposer of R, which affirms those
numbers which R denies, and denies those numbers which R affirms (and
hence is stumped by those and only those numbers which stump R).
M2: With each register R in the calculating machine is associated another register
R# in the calculating machine called the diagonalizer of R, such that for any
number x, the register R# affirms x iff R affirms xx and denies x iff R denies
xx. (I.e. R# affirms x iff R affirms x * x and R# denies x iff R denies x * x, for
our special function * which has the properties listed above for it; we will
write xy for x * y from now on in this chapter.)
Recall our denumerable list of registers R1, R2,…, Rn,…. For any number n,
we let n′ be the index of (Rn)′, and let n# be the index of (Rn)#. Thus Rn′ = (Rn)′
and Rn# = (Rn)#.
We will say that two registers Ra and Rb behave the same way towards a
number n when they either both affirm n, or both deny n or are both stumped by
n. We call two registers similar if they behave the same way towards every
number. It is easy to see that for any register R, the registers R′# and R#′ are
similar, for each denies x iff R affirms xx, and each affirms x iff R denies xx. (R#′
means the same as (R′)# and R#′ means the same as (R#)′, but we will always
eliminate the parentheses if the resulting meaning is clear.)
Universal Registers
www.pdfgrip.com
Universal Registers
We shall call a register U universal if, for all numbers x and y, the register U
affirms xy iff Rx affirms y.
Contra-Universal Registers
We shall call a register V contra-universal if for all numbers x and y, the register
V affirms xy iff Rx denies y.
Creative Registers
We shall call a register C creative if, for every register R, there is at least one
number n such that C affirms n iff R affirms n.
Proposition 1.5. Any creative register can be stumped.
Proposition 1.6. Every universal register is creative, and every contra-universal
register is creative.
Note: Propositions 1.5 and 1.6 provide additional and particularly simple
proofs of Propositions 1.1 and 1.3.
For the rest of this section, we let A be the set of all numbers xy such that Rx
www.pdfgrip.com
affirms y, and let B be the set of all numbers xy such that Rx denies y.
Fixed Points
A Halting Problem
Domination
Affirmation Sets
www.pdfgrip.com
Affirmation Sets
By the affirmation set of a register R we shall mean the set of all numbers
affirmed by R. We shall call a set of numbers S an affirmation set if it is the
affirmation set of some register.
Proposition 1.12. If one of the registers is universal, then there exists an
affirmation set whose complement is not an affirmation set.
Some of the sentences are called provable and some are called refutable, and we
assume the system is consistent, by which we mean that no sentence is both
provable and refutable. A sentence is called decidable if it is either provable or
refutable, and it is called undecidable if it is neither provable nor refutable. The
system is called complete if every sentence is decidable, and is called incomplete
otherwise. We let P be the set of all provable sentences and R be the set of all
refutable sentences.
www.pdfgrip.com
For any set W of sentences, we let W0 be the set of all indices of the elements
of W. Then P0 is the set of all n such that Sn is provable, and R0 is the set of all n
such that Sn is refutable.
Set Representation
We shall say that a predicate H represents the set of all numbers n such that H(n)
is provable. Thus to say that H represents a number set A is to say that for all
numbers n, the sentence H(n) is provable iff n ∊ A. We call the set A
representable if some predicate represents it.
We shall call K a refutability predicate if it represents the set R0. Thus the
statement that K is a refutability predicate is equivalent to the condition that for
every n the sentence K(n) is provable iff Sn is refutable.
Creative Predicates
We shall call a predicate K creative if for every predicate H there is at least one
number n such that H(n) is provable iff K(n) is provable.
Proposition 2.5. If K is creative, then K(n) is undecidable for some n.
Proposition 2.6. Every provability predicate is creative, and every refutability
predicate is creative.
Fixed Points
Domination
III. R-Systems
The systems of I and those of II are really the same, only in different dress. The
common underlying theme will be fully explained in the next section of this
chapter. Meanwhile, we will consider yet another embodiment of the theme,
which has direct applications to recursion theory.
We consider a denumerable collection of number sets, which we will call “R-
sets”, and an enumeration A1, A2, …, An, … of them in some order, and another
enumeration B1, B2, …, Bn, … of them in some other order, such that for any
disjoint pair (S1, S2) of R-sets, there is some n such that S1 = An and S2 = Bn. We
shall call such a pair of enumerations an R-system if the following three
conditions hold:
R0: For every n, the set An is disjoint from Bn i.e.,
R1: Associated with each number n is a number n′ such that An′ = Bn and Bn′ = An
[thus the ordered pair (An′, Bn′) is the pair (Bn, An)].
R2: Associated with each n is a number n# such that for every number x,
(1) x ∊ An# iff xx ∊ An.
(2) x ∊ Bn# iff xx ∊ Bn.
We call an R-set Aw universal if for all numbers x and y the number xy ∊ Aw iff y
∊ Ax. We call an R-set Av contra-universal if for all numbers x and y the number
xy ∊ Av iff y ∊ Bx.
Proposition 3.2. If some R-set Aw is universal, then there is some n such that n
itself lies outside both An and Bn.
Creative R-Sets
We shall call an R-set Ac creative if for every R-set Ax there is a least one
number n such that n ∊ Ac iff n ∊ Ax.
Fixed Points
We now let A be the set of all numbers xy such that y ∊ Ax, and let B be the set of
all numbers xy such that y ∊ Bx.
We shall call a number x a fixed point of the pair (An, Bn) if x ∊ An iff x ∊ A
and x ∊ Bn iff x ∊ B. [An equivalent definition is that a number xy is a fixed
point of a pair (An, Bn) if xy ∊ An iff y ∊ Ax and xy ∊ Bn iff y ∊ Bx.]
Proposition 3.8. (a) If Aw is a universal R-set, then any fixed point of (Aw′, Bw′) =
(Bw, Aw) lies outside both Aw and Bw. (b) If Av is a contra-universal R-set, then
any fixed point of (Av, Bv) lies outside both Av and Bv.
Proposition 3.10. There is no number n such that for all numbers x and y, xy ∈
An ∪ Bn iff y ∉ Ax ∪ Bx.
R-Separable R-Sets
Domination
IV. A Synthesis
As already mentioned, the discussions and problems of sections I, II and III are
really all the same, only dressed differently. In each of the three, we have
numerical relations A(x, y) and B(x, y) such that the following three conditions
hold:
D0: A(x, y) and B(x, y) cannot both hold simultaneously for any ordered pair of
numbers (x, y).
D1: Associated with each number x is a number x′ such that for all numbers y:
(1) A(x′, y) iff B(x, y);
(2) B(x′, y) iff A(x, y).
D2: Corresponding to each number x is a number x# such that for all y:
(1) A(x#, y) iff A(x, y * y);
(2) B(x#, y) iff B(x, y * y).
• For the decision machine M of I, A(x, y) is the relation “Rx affirms y”,
and B(x, y) is the relation “Rx denies y”.
• For the mathematical system of II, A(x, y) is the relation “Hx(y) is
provable”, and B(x, y) is the relation “Hx(y) is refutable”.
• For the mathematical R-systems of III, A(x, y) is the relation “y ∈ Ax”,
and B(x, y) is the relation “y ∈ Bx”.
We shall call the pair A(x, y) and B(x, y) a duo if conditions D0, D1 and D2
hold. Any result proved about duos in general is simultaneously applicable to the
items of sections I, II, and III. Thus, in the problems that follow, for each n ≤ 12,
the solution to Problem n simultaneously proves Propositions 1 n, 2n, and 3n of
sections I, II, and III.
In this section we will again abbreviate x * y by xy and we will omit the word
“section” in front of I, II, and III. We let A be the set of all numbers xy such that
A(x, y) is true, and we let B be the set of all numbers xy such that B(x, y) is true.
It will be valuable to know that not only are A(x, y) and B(x, y) disjoint [for
every pair (x, y)], but A and B are also disjoint (contain no common element n).
This is because every n is xy for some x and y, and it is given that A(x, y) and
B(x, y) don’t both hold, which means that xy cannot be in both A and B, and thus
n cannot be in both A and B.
• For the decision machine M of I, A is the set of all numbers xy such that Rx
www.pdfgrip.com
Undecidable Numbers
Universal Numbers
We shall call a number w universal if for all x it is the case that wx ∈ A iff x ∈ A,
or, which is the same thing, for all x and y, the relation A(w, xy) holds iff A(x, y)
holds.
• For the decision machine M of I, to say that Rw is a universal register is to say
that for all x and y, Rw affirms xy iff Rx affirms y.
• For the mathematical system of II, to say that Hw is a provability predicate
(a “universal predicate for provability”) is to say that for all x and y, Hw(xy)
is provable iff Sxy is provable (which is the same as saying Hx(y) is
provable).
• For the mathematical R-systems of III, to say that w is universal is to say that
www.pdfgrip.com
Contra-Universal Numbers
Creative Numbers
We call a number c creative if for every number n there is at least one number x
such that cx ∈ A iff nx ∈ A.
• For the decision machine M of I, to say that a register Rc is creative is to say
that for every register Rn, there is at least one number x such that Rc affirms x
iff Rn affirms x.
• For the mathematical system of II, to say the predicate Hc is creative is to
say that for every predicate Rn, there is at least one number x such that Hc(x)
is provable iff Hn(x) is provable.
• For the mathematical R-systems of III, to say that c is creative is to say that,
for every number n, there is at least one number x such that x ∈ Ac iff x ∈ An.
Fixed Points
Problem 1. Prove that if w is universal, then there is at least one number x such
that wx is undecidable.
Problem 2. Prove that if some number is universal, then there is at least one
number x such that xx is undecidable.
Problem 3. Prove that if v is contra-universal, then there is at least one number x
such that vx is undecidable.
Problem 4. Prove that if some number is universal, then there is some number x
such that xx′ is undecidable. And prove that if some number is contra-universal,
then there is some number x such that xx is undecidable.
Problem 5. Prove that if c is creative, then there is at least one number x such
that cx is undecidable.
Problem 6. Prove that every universal number is creative, and that every contra-
universal number is creative.
Problem 7. Prove that every number has a fixed point.
Problem 8. Suppose w is universal and v is contra-universal. Prove:
(a) If x is any fixed point of w′, then x is undecidable, and so is wx.
(b) If x is any fixed point of v, then x is undecidable, and so is vx.
Problem 9. Suppose h is a number such that, for all numbers x:
(a) If x ∈ A, then hx ∈ A.
(b) If x ∈ B, then hx ∉ A.
Prove that hx is undecidable for some x.
Why is this a strengthening of the result of Problem 1?
Problem 10. Prove that there is no number n such that for all numbers x the
number nx is decidable if and only if x is undecidable.
www.pdfgrip.com
Domination
Let us say that a number n dominates a number m if, for every number x, if mx ∊
A, then nx ∊ A.
• For the decision machine M of I, to say that the register Rn dominates the
register Rm is to say that Rn affirms all numbers affirmed by Rm.
• For the mathematical system S of II, to say that the predicate Hn dominates
the predicate Hm is to say that that for all x, if Hm(x) is provable, so is Hn(x).
• For the mathematical R-systems of III, to say that An dominates Am is to say
that Am ⊆ An.
Set Representation
We will say that the number n represents the set S if, for all x, nx ∊ A iff x ∊ S.
• For the decision machines of I, to say that a register R represents a set S is
just to say that the set S is the affirmation set of the register R.
• For the mathematical systems S of II, to say that predicate H represents a set
S is just to say that S is the set of all numbers n such that H(n) is provable.
• For the mathematical R-systems of III, to say that the R-set An represents a
set S is just to say that y ∊ An iff y ∊ S, i.e., An = S.
1. Suppose w is universal. Then for any number x, w(w#′x) ∊ A holds iff w#′x ∊
A, which is true iff w#x ∊ B, which is true iff w(xx) ∊ B. Thus:
(1) w(w#′x) ∊ A iff w(xx) ∊ B.
If we take w#′ for x in (1), we obtain:
(2) w(w#′w#′) ∊ A iff w(w#′w#′) ∊ B.
But A and B are disjoint, so w(w#′w#′) can be in neither A nor B, which
www.pdfgrip.com
Chapter 4
Elementary Formal Systems and Recursive
Enumerability
Representability
www.pdfgrip.com
Similarly, if (E1), (E2), …, (En) are mutually independent (no two having a
common predicate), then the representable relations of (E1) ∪ … ∪ (En), are the
result of joining together those of each of (E1), …, (En) alone.
Now suppose R1, …, Rn are each formally representable over K. Since we are
assuming an unlimited stack of predicates at our disposal, we can represent R1,
…, Rn in mutually independent systems (E1), (E2), …, (En), and so we have:
(b) Existential Quantifiers. For any relation R(x1, …, xn, y) of degree two or
higher, by its existential quantification ∃yR(x1, …, xn, y) — also abbreviated
∃R — is meant the set of all n-tuples (x1, …, xn) such that R(x1, …, xn, y)
holds for at least one y:
(c) Explicit Transformation. Let n be any positive integer and consider n
variables x1, …, xn. Let R be a relation of degree k and let α1, …, αk each be
either one of the n variables x1, …, xn or a constant term (i.e. a string of
symbols of K). By the relation λx1, …, xnR(α1, …, αk) is meant the set of all
n-tuples (a1, …, an) of strings of K such that R(b1, …, bk) holds, where each
bi is defined as follows:
(1) If αi is one of the variables xj, then bi = aj.
(2) If αi is a constant c, then bi = c.
[For example, for n = 3 and constants c and d, λx1, x2, x3R(x3, c, x2, x2, d) is the
set of all triples (a1, a2, a3) such that the relation R(a3, c, a2, a2, d) holds.]
is explicitly definable from the k-ary relation R and that the operation which
assigns λx1, …, xnR(α1, …, αk) to the relation R is an explicit transformation.
(a) If R1 and R2 are of the same degree and formally representable over K, then
R1 ∪ R2 and R1 ∩ R2 are formally representable over K and of the same
degree as R1 and R2.
(b) If the relation R(x1, …, xn, y) is a relation of degree n + 1(n greater than or
equal to one) that is formally representable over K, then the relation ∃yR(x1,
…, xn, y) is a relation of degree n that is formally representable over K.
(c) If the relation R(x1, …, xk) is a relation of degree k that is formally
representable over K, then λx1, …, xnR(α1, …, αk) is a relation of degree n
that is formally representable over K (here each αi is one of the variables x1,
…, xn or a constant).
We also know by Problem ∑1, taking D for the alphabet K, that the existential
quantification of a recursively enumerable relation is recursively enumerable.
Now, every ∑1-relation is the existential quantification of a ∑0-relation, hence of
a recursively enumerable (in fact recursive) relation, and is therefore recursively
enumerable. This proves that:
www.pdfgrip.com
Problem 4. Suppose R(x1, …, xn, y) is ∑1. Show that the relation ∃yR(x1, …, xn,
y) [as a relation between x1, …, xn] is ∑1.
For any function f(x) (from numbers to numbers), we say that the function f(x)
is ∑1 if the relation f(x) = y is ∑1.
A function f which assigns to each pair (x, y) of numbers a number f(x, y) is said
to be 1-1 (which is read as “one-to-one”) if for every number z there is at most
one pair (x, y) such that f(x, y) = z. In other words, if f(x, y) = f(z, w), then x = z
and y = w. The function f is said to be onto the set N of numbers, if every number
z is f(x, y) for some pair (x, y). Thus f is 1-1 onto N iff every number z is f(x, y)
for one and only one pair (x, y).
We now want a recursive function δ(x, y) which is 1-1 onto N. [Note that
such a function provides an enumeration of all ordered pairs of positive
integers.] There are many ways to construct such a function. Here is one way:
We start with (1, 1) followed by the three pairs whose highest number is 2, in the
order (1, 2), (2, 2), (2, 1), followed by the five pairs whose highest number is 3,
in the order (1, 3), (2, 3), (3, 3), (3, 1), (3, 2), followed by … (1, n), (2, n), …, (n,
n), (n, 1), (n, 2), …, (n, n - 1), etc. We then take δ(x, y) to be the position that (x,
y) takes in this ordering, which is the number of pairs which precede (x, y) plus
1. For example,
www.pdfgrip.com
Each number z is δ(x, y) for just one pair (x, y). We let K (z) [to often be
abbreviated Kz] be x, and let L(z) [or Lz] be y. Thus Kδ(x,y) = x and Lδ(x,y) = y.
Problem 8. (a) Prove that δ(Kz, Lz) = z. (b) Prove that the functions K(x) and
L(x) are Σ0, and hence recursive.
For each n ≥ 2 we define the function δn(x1, …, xn) by the following inductive
scheme:
An obvious induction argument shows that for each n ≥ 2 the function δn(x1,
…, xn) is a 1-1 recursive function onto the set N of positive integers.
www.pdfgrip.com
For each n ≥ 2 and 1 ≤ i ≤ n, we define (x) as follows: x = δn(x1, …, xn) for just
one n-tuple (x1, …, xn), and we take (x) to be xi. Thus (δn(x1, …, xn)) = xi.
Note that the symbols 1 and 2, from which we will construct our dyadic
numerals for the positive integers, correspond to the alphabet K in the definition
of an elementary formal system given in The Beginner’s Guide [Smullyan, 2014]
and earlier here.
It is obvious that representability in an arbitrary dyadic system is equivalent
to representability in a transcribed dyadic system.
We define “TS term”, “atomic TS formula”, “TS formula”, “TS sentence” as
we did “term”, “atomic formula”, “formula”, “sentence”, only using TS
variables and TS predicates instead of variables and predicates, respectively.
We now construct the universal system (U) in the following manner: We first
extend the alphabet K7 to the alphabet K8 by adding the symbol “∧”. Then, by a
base we mean a string of the form X1, …, Xk (or a single formula X, if k = 1).
Here X and X1, …, Xk are all TS formulas. These TS formulas are called the
components of the base X1 ∧ … ∧ Xk, or of the base X. Then we extend the
alphabet K8 to the alphabet K9 by adding the symbol “ ⊢ ”, and we define a
sentence of (U) to be an expression of the form B⊢ X, where B is a base and X is
a TS sentence (a TS formula with no variables). We call the sentence B⊢ X true
if X is provable in that TDS whose axiom schemes are the components of B.
Thus X1 ∧ … ∧ Xk ⊢ X is true iff X is provable in that TDS whose axiom
schemes are X1, …, Xk.
said to represent the set of all n-tuples (a1, …, an) of numbers such that the
sentence B ⊢ Pa1, …, an is true. Thus a predicate X1 ∧ … ∧ Xk ⊢ P of degree n
represents (in U) the set of all n-tuples (a1, …, an) such that Pa1, …, an is
provable in that TDS whose axiom schemes are X1, …, Xk; thus it represents in
(U) the same relation as P represents in the TDS whose axiom schemes are X1,
…, Xk. Thus a relation or set is representable in (U) if and only if it is recursively
enumerable. In this sense, (U) is called a universal system for all recursively
enumerable relations.
We shall let T be the set of true sentences of (U), and let T0 be the set of
Gödel numbers of the members of T (using the dyadic Gödel numbering).
We now wish to show that the set T is formally representable in an
elementary formal system over the alphabet K9, and that the set T0 is recursively
enumerable. Our entire development of recursion theory is based on this crucial
fact!
We are to construct an elementary formal system W over K9 in which T is
represented. To do this, we must first note that the implication sign of W must be
different from the implication sign of transcribed systems. We will use “→” for
the implication sign of W, and use “imp” for the implication sign of transcribed
systems. Similarly, we will use the ordinary comma for the punctuation sign of
W and use “com” for the punctuation sign of transcribed systems. For variables
of W, (not to be confused with transcribed variables), we will use the letters x, y,
z, w, with or without subscripts. Predicates (not to be confused with predicates of
(U) or transcribed predicates) are to be introduced as needed.
Problem 11. Construct w by successively adding predicates and axiom schemes
to represent the following properties and relations:
N(x) : x is a number (dyadic numeral)
Acc(x) : x is a string of accents
Var(x) : x is a transcribed variable
dv (xy) : x and y are distinct TS variables
P (x) : x is a transcribed predicate
t(x) : x is a transcribed term
www.pdfgrip.com
where X0 is the Gödel number of X. Any dyadic numeral n itself has a Gödel
number n0, hence a norm nn0. We recall that our dyadic Gödel numbering has
the advantage of being an isomorphism with respect to concatenation, i.e. if n is
the Gödel number of X and m is the Gödel number of Y, then nm is the Gödel
number of XY. Therefore, if n is the Gödel number of X, then nn0 is the Gödel
number of Xn (the norm of X). Thus the Gödel number of the norm of X is the
norm of the Gödel number of X.
For any number x, we let normx be the norm of x.
Problem 12. Show that the relation normx = y is recursively enumerable.
For any set A of numbers we let be the set of all numbers whose norm is in
A.
Lemma. If A is recursively enumerable, so is .
Problem 13. Prove the above lemma.
Problem 14. Using the above lemma, now prove that if A is recursively
enumerable, then there is a Gödel sentence for A.
As we have seen, there cannot be a Gödel sentence for , and thus by
Problem 14, the set cannot be recursively enumerable. And so the set T0 is
recursively enumerable, but not recursive.
1. (a) We have already proved this for when R1 and R2 are sets (i.e. relations of
degree 1), but the proof for relations of degree n is essentially the same:
Suppose R1 and R2 are relations of degree n and that both are formally
representable over K. There they are representable in a common elementary
formal system over K. In this system let P1 represent R1 and let P2 represent
R2. Then take a new predicate Q and add the following two axioms:
(b) Let P represent the relation R(x1, …, xn, y) in (E). Take a new predicate
Q of degree n and add the following axiom scheme:
(c) Let P represent the relation R(x1, …, xk) in (E). Take a new predicate Q
and add the following axiom scheme:
since the two are complements of each other. We have seen that the explicit
transformations λ x1, …, xnR(α1,…, αk) and λ x1, …, xn , where
each αi is one of the variables x1, …, xn or a constant, are also formally
representable over K. But each of these is the complement of the other, so
we see that the collection of sets and relations solvable over K is closed
under explicit transformation as well.
3. (a) Suppose that the relation R(x1, …, xn, y) is recursively enumerable. Let D
be a dyadic system in which R is represented by the predicate P, x < y is
represented by E, and the relation “the successor of x is y” is represented by
S.
Let W (x1, …, xn, y) be the relation z ≤ y)R(x1, …, xn, z). To
represent the relation W, take a new predicate Q and add the following
two axiom schemes:
Then Q represents W.
As for the relation (∃z ≤ y)R(x1, …, xn, z) (as a relation between x1,
…, xn, y), it is equivalent to the relation ∃z(z ≤ y ∧ R(x1, …, xn, z). Let
S(x1, …, xn, y, z) be the relation z ≤ y∧R(x1, …, xn, z). It must be
recursively enumerable, because it is the intersection S1 ∩ S2 of the
relations S1 and S2 where
and
which is true iff ∃z∃ω(S(x, y, z) ∧ R(x, ω)). Since the condition S(x, y, z) ∧
R(x, ω) is Σ0, then ∃ω(S(x, y, z)R(x, ω), is Σ1, hence so is the condition
∃z∃ω(S(x, y, z)R(x, ω)) [by Problem 4].
Thus the condition f(x) = y ∧ y ∊ A is Σ1, and therefore so is ∃y(f(x) = y
∧ y ∊ A) [again by Problem 4]. Thus A′ is Σ1.
www.pdfgrip.com
6. Our goal is to prove that the relation g(x) = y is Σ1. First let R1(x, y) be the Σ0
relation (x = 1 ∧ y = 12)∨ (x = 2 ∧ y = 122) Thus R1(x, y) holds iff x is the
single digit 1 or the single digit 2 and g(x) = y.
Let R2(x, y, x1, y1) be the Σ0 relation
Call a set S of ordered pairs of numbers special, if for every pair (x, y) in S,
either R1(x, y) or (∃x1 ≤ x)(∃y1 ≤ y)((x1, y1) ∊ S ∧ R2(x, y, x1, y1)). It is
easily seen by mathematical induction on the length of x (i.e. the number of
occurrences of 1 or 2) that if S is special, then for every pair (x, y) in S, it
must be that g(x) = y.
Next we must show that conversely, if g(x) = y, then (x, y) belongs to
some special set.
Well, the x in g(x) = y must be of the form x = d1d2 … dn, where each di
is the digit 1 or the digit 2. For each i ≤ n, we let gi be the dyadic Gödel
number of di (thus, if di = 1, then gi = 12 and if di = 2, then gi = 122). Now
let Sx be the set
The set Sx is obviously special, and contains the pair (x, g(x)) [which is the
pair (d1d2 … dn, g1g2 … gn)]. Thus if g(x) = y, then (x, y) ∈ Sx and so (x, y)
belongs to a special set. Thus we have:
(1) g(x) = y iff (x, y) belongs to a special set.
We now use Lemma K. For any number z, we let Sz be the set of all pairs (x,
y) such that K(x, y, z) holds. Thus (x, y) ∈ Sz iff K(x, y, z). Since (x, y) ∈ Sz
iff K(x, y, z), then to say that Sz is special is to say that for all x and y, K(x,
y, z) implies that the following ∑0 condition — call it A(x, y, z) — holds:
To see the equivalence, it is obvious that C(z) implies P (z). To see the
converse, suppose P (z) holds. We must show that C(z) holds. Let x and y
be arbitrary numbers. We need to show that if K(x, y, z) holds, then so does
A(x, y, z). If K(x, y, z) holds, then it must also be true that x ≤ z and y ≤ z [by
the second condition in Lemma K], and therefore A(x, y, z) holds [by the
assumption P (z)]. Thus, for all x and y, K(x, y, z) ⊃ A(x, y, z), and so C(z)
holds.
This proves that C(z) is equivalent to P (z), and hence Sz is special iff
P(z) holds.
Finally, g(x) = y iff (x, y) belongs to some special set, which is true iff
(b) Since x = δ2(Kx, Lx) [by the definition of δ2 and Problem 8], we have
that and
(c) Suppose x = δn+1(x1, …, xn+1). Then [by the definition of δn+1], x =
δ(δn(x1, …, xn), xn+1). Thus Kx = δn(x1, …, xn) and Lx = xn+1.
(d) Let x = δn+1(x1, …, xn, xn+1) for some x1, …, xn+1. Then by definition
www.pdfgrip.com
But also
Now,
(e) Take Then R(x1, …, xn, y1,
…, ym) iff M(δn(x1, …, xn), δm(y1, …, ym)).
11. N(x): x is a number
P(x): x is a TS predicate
t(x): x is a TS term
F (x): x is a TS formula
S(x): x is a TS sentence
B(x): x is a TS base
12. Let D be a dyadic system with predicate G and the following axiom
schemes:
Then G represents the relation “the Gödel number of x is y”. Now add a
predicate M and the axiom scheme:
We take for n the Gödel number h of H, and obtain that H h is true iff hh0 ∈
www.pdfgrip.com
A.
However, hh0 is the Gödel number of Hh! Thus Hh is true iff its Gödel
number is in A, and if thus a Gödel sentence for A.
www.pdfgrip.com
Chapter 5
Some Recursion Theory
We use the letter T to denote the set of true sentences of the universal system
(U) and use T0 to denote the set of Gödel numbers of the sentences in T. We now
define ωn to be the set of all x such that r(n, x) ∈ T0. Thus x ∈ ωn iff r(n, x) ∈
T0.
Problem 1. (a) Prove that ωn is recursively enumerable. (b) Prove that for every
recursively enumerable set A, there is some n such that A = ωn.
www.pdfgrip.com
We shall call n an index of a set A if A = ωn. We have just shown that a set A
is recursively enumerable if and only if it has an index, and so we have
succeeded in arranging all recursively enumerable sets in an infinite sequence
ω1, ω2, …, ωn, … in such a manner that the relation ωx contains y (which is the
relation r(x, y) ∊ T0) is recursively enumerable.
We next wish to show that for each n ≥ 2, we can enumerate all recursively
enumerable relations of degree n in an infinite sequence such
that the relation is a recursively enumerable relation among the
variables x, y1, …, yn. For this, it will be useful to use the indexing of recursively
enumerable sets that we already have, along with the pairing function δ(x, y) and
the n-tupling functions δn(x1, …, xn). We simply define to be δn(x1,
…, xn) ∊ ωi. Since ωi is recursively enumerable and the function δn(x1, …, xn) is
also recursively enumerable, so is [by (a) of Problem 9 of Chapter
4]. Also, for any recursively enumerable relation R(x1, …, xn), the set A of
numbers δn(x1, …, xn) such that R(x1, …, xn) holds is recursively enumerable [by
(b) of Problem 9 of Chapter 4]. Thus the set A has some index i so that R(x1, …,
xn) holds iff δn(x1, …, xn) ∊ ωi, which is true iff . Thus Thus
every recursively enumerable relation has an index. [We also write for x ∊
ωn, since we are regarding number sets as relations of degree 1.]
For each n we let Un+1(x, y1, … yn) be the relation [as a relation
among x, y1, …, yn]. These relations are referred to as universal relations: U2 is
the universal relation for all recursively enumerable sets, U3 is the universal
relation for all recursively enumerable relations of degree 2, and so forth.
We sometimes write Rx(y1, …, yn) instead of , since the number of
arguments makes the superscript n clear. And we write U(x, y1, …, yn) instead of
Un+1(x, y1, …, yn).
www.pdfgrip.com
Iteration Theorems
The second basic tool of recursion theory is the iteration theorem, which we will
shortly state and prove in both a simple and more general form. To fully
appreciate the significance and power of these theorems, the reader should first
try solving the following exercises:
Exercise 1. We know that for any two recursively enumerable sets A and B, the
union A ∪ B is recursively enumerable. Moreover, the process of finding this
union is effective, in the sense that there is a recursive function φ(i, j) such that if
i is an index of A and j is an index of B, then φ(i, j) is an index of A ∪ B — in
other words there is a recursive function φ(i, j) such that for all numbers i and j,
ωφ(i,j) = ωi ∪ ωj. Prove this.
For any relation R(x, y), by its inverse is meant the relation satisfying the
condition (x, y) iff R(x, y).
Exercise 2. Show that there is a recursive function t(i) such that for every
number i the relation Rt(i)(x, y) is the inverse of the relation Ri(x, y), i.e. Rt(i)(x, y)
iff Ri(y, x). [This can be paraphrased: Given an index of a recursively
enumerable relation of two arguments, one can effectively find an index of its
inverse.]
For any function f(x) and number set A, by f-1(A) is meant the set of all
numbers n such that f(n) ∊ A. Thus n ∊ f-1(A) iff f(n) ∊ A.
Exercise 3. Prove that for any recursive function f(x), there is a recursive
function φ(i) such that for every number i, it is the case that
The reader who has solved the above exercises has probably had to go back to
the universal system (U) each time. The iteration theorem, to which we now
turn, will free us from this once and for all!
Theorem 1. [Simple Iteration Theorem] For any recursively enumerable
relation R(x, y) there is a recursive function t(i) such that for all numbers x and
i, x ∊ ωt(i) iff R(x, i).
One way to prove Theorem 2 is to use Theorem 1 and the fact that for any
recursively enumerable relation R(x1, …, xm, y1, …, yn) there is a recursively
enumerable relation M(x, y) such that R(x1, …, xm, y1, …, yn) iff M(δm(x1, …,
xm), δn(y1, …, yn)) [by Problem 10 (e) of Chapter 4].
Then, by the iteration theorem, there is a recursive function φ(i1, i2) such that for
all numbers x, i1, i2,
Therefore,
For Exercise 2, let R(x1, x2, y) be the recursively enumerable relation Ry(x2,
x1). It is recursively enumerable because it is explicitly defined from the
recursively enumerable relation U(x1, x2, y). Then by the simple iteration
theorem there is a recursive function φ(i) such that for all x and i, Rφ(i)(x1, x2) iff
R(x1, x2, i), which is true iff Ri(x2, x1).
Maximal Indexing
Well, believe it or not, all the above statements are true! They are all special
cases of the theorem that follows:
Consider a recursively enumerable relation R(x, y). Now consider a number n
and the set x: R(x, n), i.e. the set of all numbers x such that R(x, n) holds. This set
is recursively enumerable, hence is ωn* for some n*. Given the relation R, what
n* is depends on what n is. Wouldn’t it be curious if n could be chosen such that
n* is n itself? Well, this is indeed so:
Theorem 5. [Weak Recursion Theorem] For any recursively enumerable
relation R(x, y), there is a number n such that ωn = {x : R(x, n)} [in other words,
for all x, we have n ∊ ωn iff R(x, n)].
I will give the beginning of the proof, and leave it as a problem for the reader
to finish it.
The relation Ry(x, y), as a relation between x and y is recursively enumerable.
Hence, by the simple iteration theorem there is a recursive function t(y) such that
for all x and y, x ∊ ωt(y) iff Ry(x, y).
Now, take any recursively enumerable relation R(x, y). Then the relation R(x,
t(y)) is also a recursively enumerable relation between x and y, hence has some
index m. Thus Rm(x, y) iff R(x, t(y)). Now comes the great trick: ….
Problem 7. (a) Prove Theorem 6. (b) Now prove the other four “believe it or
nots”.
www.pdfgrip.com
But by Theorem 6, there is a number n such that ωf(n) = ωn. Thus for all x, x ∊
ωn iff x ∊ ωf(n), which is true iff R(x, n). Thus x ∊ ωn iff R(x, n).
For each of these nine properties, we can ask if it is solvable. These are
typical questions of recursion theory that have been answered by individual
methods. Rice’s theorem, to which we now turn, answers a host of such
questions in one fell swoop.
A number set A is called extensional if it contains with any index of a
recursively enumerable set all other indices of it as well — in other words, for
every i and j, if i ∊ A and ωi = ωj, then j ∊ A.
Theorem R. [Rice’s Theorem] The only recursive extensional sets are ℕ and .
We will soon see how Rice’s theorem [Rice, 1953] is relevant to the nine
properties mentioned above, but first let us go to its proof. We first need the
following:
Lemma. For any recursive set A other than ℕ and , there is a recursive
function f(x) such that for all n, n ∊ A iff f(n) ∉ A
Problem 8. (a) Prove the above lemma. [Hint: If A is neither ℕ nor , then
there is at least one number a ∊ A and at least one number b ∉ A.Use these two
numbers a and b to define the function f(x). (b) Now prove Rice’s Theorem.
[Hint: Use the function f(x) of the Lemma and then apply Theorem 6 to the
function f(x).]
Applications
Let us consider the first of the aforementioned nine properties P (n), namely that
ωn is infinite. Obviously, if ωi is infinite and ωi = ωj, then ωj is also infinite.
Thus the set of the indices of all infinite sets is extensional. Also, there is at least
one n such that ωn is infinite and at least one n such that ωn is not infinite. Hence
by Rice’s Theorem, the set of the indices of all infinite sets is not recursive.
The same argument applies to all the other eight cases, so that none of them
are solvable!
By the weak recursion theorem, for any number i, there is a number n such that
www.pdfgrip.com
For the second proof (to be finished by the reader – Problem 9 (b) below), let
S(x, y, z) be the recursively enumerable relation Rz(x, y, z). By the iteration
theorem, there is a recursive function t(y, z) such that for all y and z, ωt(y,z) = {x :
S(x, y, z)}, and thus ωt(y,z) = {x : Rz(x, y, z)}.
Now, given a recursively enumerable relation M(x, y, z), the relation M(x, y,
t(y, z)) is recursively enumerable, hence has an index h. Thus Rh(x, y, z) iff M(x,
y, t(y, z)). Then ….
Problem 9. (a) Finish the first proof. (b) Finish the second proof.
As a special case of Theorem 7, we have:
Theorem 8. [Myhill’s Fixed Point Theorem] There is a recursive function φ(y)
such that for all y, ωφ(y) = {x : Ry(x, φ(y))}.
To say that a set A is not recursively enumerable is to say that given any
recursively enumerable subset ωi of A there is a number n in A but not in ωi — a
witness, so to speak that ωi is not the whole of A. A set A is called productive if
there is a recursive function φ(x) — called a productive function for A — such
that for every number i, if ωi is a subset of A, then φ(i) ∊ A – ωi [i.e. φ(i) is in A,
but not in ωi].
A simple example of a creative set is Post’s set C, the set of all numbers x such
that x ∊ ωx. Thus for any number i, i ∊ C iff i ∊ ωi. If ωi is disjoint from C, then
i is outside both ωi and C, and therefore the identity function I(x) is a creative
function for C.
A less immediate example of a creative set is the set K of all numbers ∂(x, y)
such that x ∊ ωy. The set K was also introduced by Post [1943] and is sometimes
called the complete set. The proof that K is creative is less immediate than the
proof for C, and involves use of the iteration theorem, as we will see.
It is also correct to say that a set A is not recursively enumerable iff for every
recursively enumerable set ωi, whether a subset of A or not, there is a number j
such that j ∊ A iff j ≠ ωi — thus attesting to the fact that A ≠ ωi. By a completely
www.pdfgrip.com
productive function for A is meant a recursive function φ(x) such that for every
number i, φ(i) ∊ A iff φ(i) φ ∉ ωi. A set A is called completely productive if
there is a completely productive function for A.
A set A is called completely creative if A is recursively enumerable and there
is a completely productive function φ(x) for the complement of A; and such a
function is called a completely creative function for A. Thus a completely
creative function for A is a recursive function φ(x) such that for every number i,
φ(i) ∊ A iff φ(i) ∊ ωi.
Post’s set C of all x such that x ∊ ωx is obviously not only creative, but
completely creative, since the identity function is a completely creative function
for C. Therefore a completely creative set exists. Post apparently did not
capitalize on the fact that C is not only creative, but completely creative. I hazard
a guess that if he had capitalized on this fact, then many results in recursion
theory would have come to light sooner.
We will see later that in fact any productive set is completely productive, and
hence that any creative set is completely creative. This important result is due to
John Myhill [1955], and will be seen to be a consequence of the strong recursion
theorem.
Generative Sets
Many important results about completely creative sets do not use the fact that the
set is recursively enumerable, and so we define a set A to be generative if there
is a recursive function φ(x) — called a generative function for A — such that for
all numbers i, φ(i) ∊ A iff φ(i) ∊ ωi. Thus a generative set is like a completely
creative set except that it is not necessarily recursively enumerable. A set A is
generative iff its complement is completely productive. Our emphasis will be
on generative sets.
Many-One Reducibility
theory, but throughout this chapter, reducible will mean many-one reducible.
Theorem 9. If A is reducible to B and A is generative, then B is generative.
It will be useful to state and prove Theorem 9 in a more specific form: For
any recursively enumerable relation R(x, y), by the iteration theorem there is a
recursive function t(y) such that ωt(i) = {x : R(x, i)}. Such a function t(y) will be
called an iterative function for the relation R(x, y). We will now state and prove
Theorem 9 in the more specific form:
Theorem 9*. If f(x) reduces A to B and φ(x) is a generative function for A, and
t(y) is an iterative function for the relation f(x) ∊ ωy, then f(φ(t(x))) is a
generative function for B.
Universal Sets
For any function f(x, y) of two arguments, and for any number i, by fi is meant
the function which assigns to each x the number f(x, i). Thus fi(x) = f(x, i).
Weak Co-productivity
We will say that A is weakly co-productive under φ(x) if φ(x) is recursive and
condition C2 holds.
Problem 19. Prove the above lemma. [Hint: Let M(x, y, z) be the recursively
enumerable relation x ∊ ωy ∧x = φ(z). Then use the strong recursion theorem,
Theorem 7.]
Problem 20. Now prove Theorem 13. [Hint: Suppose φ(x) is a weakly co-
productive function for A. Let t(y) be a recursive function related to φ(x) as in the
above lemma. Then show that the function φ(t(y)) is a generative function for A.]
From Theorems 13 and 12 we have:
www.pdfgrip.com
Discussion
Recursive Isomorphism
1. (a) We showed in the last chapter that the relation g(x) = y (i.e. x0 = y) is
www.pdfgrip.com
Therefore, for any number n, the relation r(n, x) = y [as a relation between x
and y] is recursively enumerable. And so for any recursively enumerable set
A, the condition r(n, x) ∈ A [as a property of x] is recursively enumerable,
since it is equivalent to ∃ y(r(n, x) = y ∧ y ∈ A). In particular, since T0 is
recursively enumerable, the condition ∃ y(r(n, x) = y ∧ y ∈ T0) is
recursively enumerable, and this is the condition x ∈ ωn. Thus ωn is
recursively enumerable.
(b) Every recursively enumerable set A is represented in (U) by some
predicate H. Thus for every number x, x ∈ A iff Hx ∈ T. We let h be the
Gödel number of H, and so for all x, Hx ∈ T iff r(h, x) ∈ T0, which is
true iff x ∈ ωh. Thus A = ωh.
2. The relation R(x, y) is represented in (U) by some predicate H, and so for all
numbers i and x, R(i, x) iff Hicomx ∈ T, the latter of which is true iff hi0cx0
∈ T0, where h, i0, c,x0 are the respective Gödel numbers of H, i, com, x (in
dyadic notation). Thus R(i, x) iff r(hi0c,x) ∈ T0, which is true iff x ∈ ωhi0c.
We thus take t(y) to be hy0c, and so t(i) = hi0c, so that now R(i, x) iff x ∈
ωt(i).
3. Given a recursively enumerable relation R(x1, …, xm, y1, …, yn), we use the
recursively enumerable relation M(x, y), which yields, for all numbers x1, …,
xm, i1, …, in,
(1) R(x1, …, xm, i1, …, in) iff M(δm(x1, …, xm), δn(i1, …, in)). We now
apply Theorem 1 to the relation M(x, y), and so there is a recursive
function t(i) such that for every x and i we have x ∈ ωt(i) iff M(x, i). We
now take δm(x1, …, xm) for x and δn(i1, …, in) for i, and obtain
(2) δm(x1, …, xm) ∈ ωt(δn(i1,…,in)) iff M (δm(x1, …, xm), δn(i1, …, in)). By
(1) and (2) we have:
(3) R(x1, …, xm, i1, …, in) iff δm(x1, …, xm) ∈ ωt(δn(i1,…,in)) which is true
iff Rt(δn(i1,…,in))(x1, …, xm).
www.pdfgrip.com
4. We will show that our recursively enumerable enumeration ω1, ω2, …, ωn,
… is maximal. Given a recursively enumerable enumeration B1, B2, …, Bn,
…, let R(x, y) be the recursively enumerable relation x ∈ By. By the simple
iteration theorem, there is a recursive function f(x) such that for all x and i,
we have x ∈ ωf(i) iff R(x, i), which is true iff x ∈ Bi. Thus x ∈ ωf(i) iff x ∈ Bi,
and so ωf(i) = Bi.
5. Define S(x1, …, xm, y, y1, …, yn) iff Ry(x1, …, xm, y1, …, yn). The relation S
is recursively enumerable (why?), and so by the iteration theorem, there is a
recursive function φ(i,i1, …, in) such that for all x1, …, xm, i, i1, …, in:
Rφ(i,i1,…,in)(x1, …, xm) iff S(x1, …, xm, i, i1, …, in) iff Ri(x1, …, xm, i1, …,
in).
6. The great trick is to take m for y, and we thus obtain Rm(x, m) iff R(x, t(m)).
But we also have Rm(x, m) iff x ∈ ωt(m)! Thus x ∈ ωt(m) iff R(x, t(m)), and so
x ∈ ωn iff R(x, n), where n is the number t(m). [Clever, huh?]
7. (a) Given a recursive function f(x) let R(x, y) be the relation x ∈ ωf(y). It is
recursively enumerable [since it is equivalent to ∃ z(f(y) = z ∧ x ∈ ωz)], so
that by Theorem 5, there is some n such that for all x, x ∈ ωn iff R(x, n). But
R(x, n) ≡ x ∈ ωf(n), and so, for all x, x ∈ ωn iff x ∈ ωf(n). Thus ωn = ωf(n).
(b) Believe it or not:
#1 is but a special case of (a), namely when f(x) = x + 1.
#2 is just Theorem 6!
#3 is but the special case of #4 in which f(x) = x. And so we must show #4:
For f(x) a recursive function, the relation x = f(y) is recursively enumerable.
Hence by Theorem 5, there is some n such that for all x, x ∈ ωn iff x = f(n).
This means that f(n) is the one and only member of ωn, and so ωn = {f(n)}.
As for #5, let R(x, y) be the relation y ∈ ωx. This relation is recursively
enumerable, and so there is a number n such that for all x, x ∈ ωn iff R(x,
n), but R(x, n) iff n ∈ ωx, and so x ∈ ωn iff x ∈ ωx.
8. (a) We are given that A is recursive and that A ≠ ℕ and A ≠ . Since A ≠ ,
www.pdfgrip.com
then there is at least one number a such that a ∈ A, and since A ≠ ℕ there is
at least one number b such that b ∉ A. Define f(x) to be b if x ∈ A and to be
a if x ∉ A. Since A is recursive, then A and its complement Ā are both
recursively enumerable. Hence the function f(x) is recursive, since the
relation f(x) = y is equivalent to the recursively enumerable relation (x ∈ A
∧ y = b) ∨ (x ∈ Ā ∧ y = a). If x ∈ A, then f(x) ∈ A [since f(x) = b], and if x
∉ A, then f(x) ∈ A [since f(x) = a]. Thus x ∈ A iff f(x) ∉ A.
(b) Let A be a recursive set that is neither ℕ nor . Take f(x) as in the
lemma. By Theorem 6, there is some n such that ωn = ωf(n). Thus n and
f(n) are indices of the same recursively enumerable set, yet one of them
is in A and the other isn’t. Therefore A is not extensional.
9. (a) If we take t(i) for y in the statement Rt(i)(x, y) iff M(x, i, d(y)) we obtain
Rt(i)(x, t(i)) iff M(x, i, d(t(i))). But it is also true that ωd(t(i)) = {x : Rt(i)(x,
t(i))} [this is true because, for any n, ωd(n) = {x : Rn(x, n)}]. Thus ωd(t(i)) = {x
: M(x, i, d(t(i))). We now take φ(i) to be d(t(i)), yielding ωφ(i) = {x : M(x, i,
φ(i))}.
(b) As noted, Rh(x, y, z) iff M(x, y, t(y, z)). We take h for z and so Rh(x, y,
h) iff M(x, y, t(y, h)). But since ωt(y,h) = {x : Rh(x, y, h)}, it follows that
ωt(y,h) = {x : M(x, y, (t(y, h))) }. We thus take φ(y) to be t(y, h), yielding
ωφ(y) = {x : M(x, y, φ(y))}.
10. This is the case that M(x, y, z) is the relation Ry(x, z).
11. Since f reduces A to B, then for all x, f(x) ∈ B iff x ∈ A. For any number i
we take φ(t(i)) for x and we obtain:
(1) f(φ(t(i))) ∈ B iff φ(t(i)) ∈ A.
Also, φ(t(i)) ∈ A iff φ(t(i)) ∈ ωt(i) [since φ(x) is a generative function for A],
so we have that:
(2) f(φ(t(i))) ∈ B iff φ(t(i)) ∈ ωt(i).
Next, since t(y) is an iterative function for the relation f(x) ∈ ωy, then for all
x and i we have x ∈ ωt(i) iff f(x) ∈ ωi. We now take φ(t(i)) for x and obtain:
(3) φ(t(i)) ∈ ωt(i) iff f(φ(t(i))) ∈ ωi.
By (2) and (3), we see that f(φ(t(i))) 2 B iff f(φ(t(i))) ∈ ωi, and thus
f(φ(t(x))) is a generative function for B.
www.pdfgrip.com
12. For any number i, x ∈ ωi iff for all x,δ (x, i) ∈ K. Let φi be the function
that assigns to each x the number δ(x, i). Thus φi(x) = δ (x, i).
Consequently, x ∈ ωi iff φi(x) ∈ K, and so φi reduces ωi to K.
13. Suppose A is universal. Postφs set C is recursively enumerable and
generative. Since C is recursively enumerable and A is universal, then C is
reducible to A. Then, since C is generative, so is A [by Theorem 9].
14. Post’s set K is obviously uniformly universal under the recursive function
δ (x, y). Now, suppose A is universal. Then K is reducible to A under some
recursive function f(x). Then A must be uniformly universal under the
function f(δ (x, y)), because for any numbers i and x, x ∈ ωi iff δ (x, i) 2 K,
which is true iff f(δ (x, i)) ∈ A. Thus x ∈ ωi iff f(δ (x, i)) ∈ A.
15. We are given that φ(x) is a generative function for B. Thus for every
number i,
20. We want to show that for any number y, φ(t(y)) ∈ ωy iff φ(t(y)) ∈ A.
(1) Suppose φ(t(y)) ∈ ωy. Then ωy ∩ {φ(t(y))} = {φ(t(y))}. But also ωy ∩
{φ(t(y))} = ωt(y) [by the Lemma]. Then φ(t(y)) ∈ A [by condition C2,
taking t(y) for i, since A is weakly co-productive under φ(x)]. Thus if
φ(t(y)) ∈ ωy, then φ(t(y)) ∈ A.
(2) Suppose φ(t(y)) ∉ ωy. Then ωy ∩ {φ(t(y))} = . Again by the Lemma,
ωy ∩ {φ(t(y))} = ωt(y). Thus ωt(y) = . Again by condition C2, taking t(y)
for i, φ(t(y)) ∉ A. Thus if φ(t(y)) ∉ ωy, then φ(t(y)) ∉ A, or, equivalently
if φ(t(y)) ∈ A, then φ(t(y)) ∈ ωy, and thus φ(t(y)) is a generative function
for A.
Answer to the Bonus Question. Of course not! A universal set is generative and
a generative set A cannot be recursive, for if it were, its complement Ā would be
recursively enumerable, hence there would be a number n such that n ∈ A iff n ∈
Ā, which is impossible.
www.pdfgrip.com
Chapter 6
Doubling Up
In preparation for the double recursion theorems we will consider here, we note
that by the iteration theorem, there is a recursive function d such that for all y
and z, ωd(y,z) = {x : Ry(x, y, z)}. This function will play a key role, and I shall call
it the double diagonal function.
Theorem 1. [Weak Double Recursion Theorem] For any two recursively
enumerable relations M1(x, y, z) and M2(x, y, z) , there are numbers a and b such
that:
(1) ωa = {x : M1(x, a, b)};
(2) ωb = {x : M2(x, a, b)}.
Problem 1. Prove Theorem 1. [Hint: Use the double diagonal function d and
consider the relations M1(x, d(y, z) , d(z, y)) and M2(x, d(z, y) , d(y, z)).]
We now wish to show that in Theorem 1, the numbers a and b can be found as
recursive function of the indices i and j of the relations M1(x, y, z) and M2(x, y,
z). This is Theorem 3 below, which will soon be seen to be a consequence of:
Theorem 2. [Strong Double Recursion Theorem] For any recursively
enumerable relations M1(x, y1, y2, z1, z2) and M2(x, y1, y2, z1, z2), there are
www.pdfgrip.com
recursive functions φ1(i, j) and φ2(i, j) such that for all numbers i and j:
(1)
(2)
Problem 2. Prove Theorem 2. [Hint: We again use the double diagonal function
d. By the iteration theorem there are recursive functions t1(i, j) and t2(i, j) such
that for all numbers i, j, x, y, z:
(1)
(2)
As an obvious corollary of Theorem 2 we have:
Theorem 3. For any two recursively enumerable relations M1(x, y, z1, z2) and
M2(x, y, z1, z2), there are recursive functions φ1(i, j) and φ2(i, j) such that for all
numbers i and j:
(1)
(2)
A special case of Theorem 3 is:
Theorem 4. [Strong Double Myhill Theorem] There are recursive functions
φ1(i, j) and φ2(i, j) such that for all numbers i and j:
(1)
(2)
Problem 3. (a) Why is Theorem 3 a corollary of Theorem 2. (b) Why is
Theorem 4 a special case of Theorem 3?
Discussion
The proof we have given of the strong double recursion theorem involved three
applications of the iteration theorem. Actually, the theorem can be proved using
only one application of the iteration theorem, and that can be done in more than
one way.
Several workers in the field of recursion theory have found a way of deriving
the weak double recursion theorem as a consequence of the strong (single)
recursion theorem, and I have subsequently found a way of deriving the strong
www.pdfgrip.com
More Doubling Up
We shall say that an ordered pair (A1, A2) is reducible to the ordered pair (B1, B2)
under a recursive function f(x) if f(x) simultaneously reduces A1 to B1 and A2 to
B2. This means that A1 = f–1(B1) and A2 = f–1(B2); in other words, it means that
for every number x:
(1) If x ∈ A1, then f(x) ∈ B1;
(2) If x ∈ A2, then f(x) ∈ B2;
(3) If x ∉ A1 ∪ A2, then f(x) ∉ B1 ∪ B2.
If (1) and (2) above both hold, without (3) necessarily holding, we say that
(A1, A2) is semi-reducible to the ordered pair (B1, B2) under the recursive
function f(x).
We shall call a pair (U1, U2) doubly universal if every disjoint pair (A, B) of
recursively enumerable sets is reducible to (U1, U2). We shall call a pair (U1, U2)
semi-doubly universal if every disjoint pair (A, B) of recursively enumerable sets
is semi-reducible to (U1, U2).
In this chapter we will prove that a pair (A, B) is doubly generative iff it is
doubly universal.
We shall call a disjoint pair (A, B) doubly co-productive under a recursive
function g(x, y) if for every disjoint pair of recursively enumerable sets (ωi, ωj),
if ωi is disjoint from A and ωj is disjoint from B, then the number g(i, j) is
www.pdfgrip.com
Effective Inseparability
Kleene’s Construction
Proposition 1. If k(x, y) is a Kleene function for a disjoint pair (A1, A2), then
(A1, A2) is completely effectively inseparable under k(x, y).
Problem 7. Prove Theorem 5. [Hint: Use the pairing function δ(x, y). Let A1 be
the set of all numbers δ(i, j) such that δ(i, j) ∈ ωj and let A2 be the set of all
numbers δ(i, j) such that δ(i, j) ∈ ωi. Show that δ(x, y) is a Kleene function for
the pair (A1 − A2, A2 − A1). Then use the separation principle to get the
recursively enumerable sets K1 and K2.]
Problem 10. Prove that if (A, B) is called doubly generative, then A and B are
each generative sets.
The next theorem is vital:
Theorem 8. If (A, B) is completely effectively inseparable and A and B are both
recursively enumerable, then (A, B) is doubly generative.
Problem 11. Prove Theorem 8. [Hint: (A, B) is completely effectively
inseparable under some recursive function ζ(x, y). By the iteration theorem, there
are recursive functions t1(n), t2(n) such that for all x and n:
(1)
(2)
Let φ(i, j) = ζ(t1(j), t2(i)). Show that (A, B) is doubly generative under the
function φ(i, j).]
Since the Kleene pair (K1, K2) is completely effectively inseparable and K1
and K2 are both recursively enumerable, it follows from Theorem 8 that the pair
(K1, K2) is doubly generative. Thus there exists a doubly generative pair of
recursively enumerable sets.
Suppose (A1, A2) is a disjoint pair which is doubly generative under the function
g(x, y). Then, for all numbers i and j, the following three conditions hold:
C1: If ωi = ℕ and ωj = , then g(i, j) ∈ A1;
C2: If ωi = and ωj = ℕ, then g(i, j) ∈ A2;
C3: If ωi = and ωj = , then g(i, j) ∉ A1 ∪ A2.
pair (B1, B2) of recursively enumerable sets there are recursive functions t1(x),
t2(x) such that:
(1) (B1, B2) is semi-reducible to (A1, A2) under g(t1(x), t2(x));
(2) If condition C3 also holds, then (B1, B2) is reducible to (A1, A2) under g(t1(x),
t2(x)). [Hint: Use the result of Problem 16 of the previous chapter.]
Double Co-productivity
Problem 14. Prove that if (A, B) is doubly co-productive under the recursive
function g(x, y), then conditions D0, D1 and D2 hold.
We will say that (A, B) is called weakly doubly co-productive under the
recursive function g(x, y) if conditions D0, D1 and D2 hold.
1. Let m be an index of the relation M1(x, d(y, z), d(z, y)), and let n be an index
of M2(x, d(z, y), d(y, z)). Then:
(1) Rm(x, y, z) iff M1(x, d(y, z), d(z, y)),
(2) Rn(x, y, z) iff M2(x, d(z, y), d(y, z)).
In (1), take m for y and n for z, and in (2), take n for y and m for z. As a
result, we have:
(1′) Rm(x, m, n) iff M1(x, d(m, n), d(n, m)),
(2′) Rn(x, n, m) iff M2(x, d(m, n), d(n, m)).
Now, ωd(m,n) = {x : Rm(x, m, n)} = {x : M1(x, d(m, n), d(n, m))} [by (1′)];
and
ωd(n,m) = {x : Rn(x, n, m)} = {x : M2(x, d(m, n), d(n, m))} [by (2′)].
We thus take a = d(m, n) and b = d(n, m), and have:
2. We define:
where t1(i, j) and t2(i, j) are the functions mentioned in the hint and d(x, y) is
the double diagonal function. We will see that these two functions work.
By the definitions of t1 and t2, we know that for all numbers i, j, x, y, z,
(1) iff M1(x, i, j, d(y, z), d(z, y));
(2) iff M2(x, i, j, d(z, y), d(y, z)).
In (1), we take t1(i, j) for y and t2(i, j) for z, and in (2), we take t2(i, j) for y
and t1(i, j) for z, obtaining:
(1′) iff M1(x, i, j, φ1(i1, i2), φ2(i1, i2));
(2′) iff M2(x, i, j, φ1(i1, i2), φ2(i1, i2)).
Also:
(3)
(4)
From (3) and (1′), we have ωφ1(i,j) = {x : M1(i, j, φ1(i, j), φ2(i, j))}.
www.pdfgrip.com
From (4) and (2′), we have ωφ2(i,j) = {x : M2(i, j, φ1(i, j), φ2(i, j))}.
3. (a) Given the recursively enumerable relations M1(x, y, z1, z2) and M2(x, y, z1,
z2) define two new relations iff M1(x, y1, z1, z2) and
iff M2(x, y2, z1, z2). Then apply Theorem 2 to the relations
and .
(b) This follows from Theorem 3 by taking M1(x, y, z1, z2) and M2(x, y, z1,
z2) to both be the relation Ry(x, z1, z2).
4. If x ∈ A′ and x ∈ B′, we get the following contradiction: There must be
numbers n and m such that:
(1) R1(x, n).
(2) R2(x, m).
(3) If R2(x, m), then m > n [since (∀ z ≤ n) ∼R2(x, z)].
(4) If R1(x, n), then n > m [since (∀ z ≤ m) ∼R1(x, z)].
From (1) and (4) it follows that n > m. From (2) and (3) it follows that it is
also true that m > n, which is impossible.
5. We are to show that the following two characterizations of the recursive
inseparability of the disjoint pair of sets (A, B) are equivalent:
C1: There is no recursive superset of A disjoint from B.
C2: For any disjoint recursively enumerable supersets ωi of A and ωj of B,
ωi cannot be the complement of ωj.
To show that C1 implies C2, assume C1. Now suppose ωi, ωj are disjoint
supersets of A and B respectively. If ωi were the complement of ωj then ωi
would be a recursive superset of A disjoint from B, contrary to C1.
Therefore ωi is not the complement of ωj, which proves C2.
Conversely, suppose C2 holds. Let A′ be any superset of A disjoint from
B. We are to show that A′ cannot be recursive and hence C1 holds. Well, if
A′ were a recursive set ωi, its complement would be a recursively
enumerable superset ωj of B, disjoint from A, which would be contrary to
C2. Thus A′ is not recursive.
6. We are given that k(x, y) is a recursive function such that for all i and j,
(1) k(i, j) ∈ ωi − ωj implies k(i, j) ∈ A2;
www.pdfgrip.com
But
so that k(t(i), t(j)) ∈ A1 [since k(x, y) is a Kleene function for (A1, A2)], and
therefore f(k(t(i), t(j))) ∈ B1 [since (A1, A2) is semi-reducible to (B1, B2)
under f(x)]. Thus k′(i, j) ∈ B1.
This proves that if k′(i, j) ∈ ωj − ωi, then k′(i, j) ∈ B1.
By a similar argument, if k′(i, j) ∈ ωi − ωj, then k′(i, j) ∈ B2. Thus k′(x,
y) is a Kleene function for (B1, B2).
9. If (A, B) is semi-doubly universal, then by definition the Kleene pair (K1, K2)
is semi-reducible to (A, B). Hence [by Proposition 2] (A, B) is in turn a
Kleene pair, and therefore completely effectively inseparable [by
Proposition 1].
10. Suppose φ(x, y) is a doubly generative function for (A, B). Let c be any
index of the empty set, and thus for every number i, the set ωi is disjoint
from ωc. Thus φ(x, c) [as a function of x] is a generative function for A,
because for every number i, ωi is disjoint from ωc, so that φ(i, c) ∈ A iff
φ(i, c) ∈ ωi [by (1)]. Similarly, φ(c, x) [as a function of x] is a doubly
generative function for the set B.
11. The set {x : x ∈ ωn ∨ x ∈ A} is simply ωn ∪ A, and the set {x : x ∈ ωn ∨ x
∈ B} is simply ωn ∪ B. Using the t1 and t2 recommended in the hint, we
have:
(1)
(2)
Thus A and ωn are both subsets of for all n, and B and ωn are both
subsets of for all n. Also as in the hint, we take
www.pdfgrip.com
Given two numbers i and j, let k = φ(i, j). Thus k = ζ(t1(j), t2(i)). Since
(A, B) is completely effectively inseparable under ζ(x, y) and A ⊆ ωt1(j) and
B ⊆ ωt2(i) [remember, A ⊆ ωt1(n) and B ⊆ ωt2(n) for all n], and ζ(x, y) is a
completely effective inseparability function for (A, B), then
(a) k ∈ ωt1(j) iff k ∈ .
Now, consider any disjoint pair (ωi, ωj). We are to show that k ∈ A iff k
∈ ωi and k ∈ B iff k ∈ ωj, and thus that (A, B) is doubly generative under
φ(x, y).
Suppose that k ∈ A. Then k ∈ ωt1 (j) [since A ⊆ ωt1(j) for every n].
Hence k ∈ ωt2(i) [by (a)]. Thus k ∈ ωi ∪ B [since ωt2(i) = ωi ∪ B]. But k ∉
B, since B is disjoint from A. Thus k ∈ ωi. This proves that if k ∈ A then k
∈ ωi.
Conversely, suppose that k ∈ ωi. Then k ∈ ωt2(i) [since ωi ⊆ ]. Thus
k ∈ [by (a)], and since ωt1(j) = ωj ∪ A, then k ∈ ωj ∪ A. But since ωj
is disjoint from ωi we have the k ∉ ωj, and so k ∈ A. Thus if k ∈ ωi, then k
∈ A.
This proves that k ∈ A iff k ∈ ωi. The proof that k ∈ B iff k ∈ ωj is
similar.
12. We are given that (A1, A2) is doubly generative under g(x, y).
(a) Suppose that ωi = ℕ and ωj = . Of course ωi is disjoint from ωj, so that
g(i, j) ∈ ℕ iff g(i, j) ∈ A1 [and also g(i, j) ∈ iff g(i, j) ∈ A2], but since
g(i, j) ∈ ℕ, it follows that g(i, j) ∈ A1.
(b) Similarly, if ωi = and ωj = ℕ, then g(i, j) ∈ A2.
(c) If and , since is obviously disjoint from , then g(i, j) ∈ A1
iff , and iff g(i, j) ∈ A2. But since , then g(i, j)
∈ A1 and g(i, j) ∈ A2.
13. By Problem 16 of the last chapter, for any recursively enumerable set B
there is a recursive function t(y) such that for all numbers i, if i ∈ B then
, and if i ∉ B, then .
www.pdfgrip.com
Now, let (B1, B2) be any disjoint pair of recursively enumerable sets.
There are recursive functions t1(i), t2(i) such that for any number i,
(1) If i ∈ B1, then ;
(2) If i ∉ B1, then ;
(3) If i ∈ B2, then ;
(4) If i ∉ B2, then .
Now assume that conditions C1 and C2 hold between the function g(x, y)
and the pair (A1, A2).
(a) Suppose i ∈ B1. Then [by (1)]. Since B2 is disjoint from B1,
then i ∉ B2, and hence [by (4)], and thus the pair is
(ℕ, ), and so g(t1(i), t2(i)) ∈ A1. This proves that i ∈ B1 implies g(t1(i),
t2(i)) ∈ A1.
(b) The proof that i ∈ B2 implies g(t1(i), t2(i)) ∈ A2 is similar. Thus (B1, B2)
is semi-reducible to (A1, A2) under the function g(t1(x), t2(x)).
(c) Suppose that condition C3 also holds. Now suppose that i ∉ B1 and i ∉
B2. Then by (2) and (4), and , and therefore g(t1(i),
t2(i)) ∉ A1 and g(t1(i), t2(i)) ∉ A2 [by C3]. Thus g(t1(x), t2(x)) reduces
(B1, B2) to (A1, A2).
14. We are given that (A, B) is doubly co-productive under the recursive
function g(x, y).
D0: This is obvious!
D1: Suppose ωi = {g(i, j)} and . We are to show that g(i, j) ∈ A.
Consider the following conditions:
(1) ωi is disjoint from ωj;
(2) ωj is disjoint from B;
(3) ωi is disjoint from A.
If these three conditions all hold, then g(i, j) would be outside all four sets
A, B, ωi, ωj [by the definition of double co-productivity]. Hence g(i, j)
would be outside ωi, hence outside {g(i, j)}, which is not possible.
Therefore it cannot be that all three of these conditions hold, but (1) and (2)
www.pdfgrip.com
do hold, therefore (3) doesn’t hold, which means that ωi is not disjoint from
A. Thus g(i, j) is not disjoint from A, so that g(i, j) ∈ A This proves D1.
D2: The proof of D2 is symmetric to that of D1.
15. By Theorem 3 there are recursive functions t1(i, j) and t2(i, j) such that for
all numbers i and j,
(1)
Thus,
(1′)
(2′)
16. Consider any disjoint pair (ωi, ωj). Let φ(i, j) = g(t1(i, j), t2(i, j)), where
t1(y1, y2), t2(y1, y2) are as given by the lemma. Then by the lemma,
(a)
(b)
We first wish to show:
(1) If φ(i, j) ∈ ωi, then φ(i, j) ∈ A;
(2) If φ(i, j) ωj, then φ(i, j) ∈ B.
To show (1), suppose that φ(i, j) ∈ ωi. Then ωi ∩ {φ(i, j)} = {φ(i, j)}
is the set {φ(i, j)}.
Since φ(i, j) ∈ ωi and ωi is disjoint from ωj, then φ(i, j) ∉ ωj, and
therefore . Hence, by (b), . Thus is the unit
set {φ(i, j)} and is the empty set , and so by condition D1 of the
definition of “weakly doubly co-productive,” g(t1(i, j), t2(i, j)) ∈ A. Thus
φ(i, j) ∈ A, which proves (1).
The proof of (2) is similar (using D2 instead of D1).
It remains to prove the converses of (1) and (2). To this end, we first
establish:
(3) If φ(i, j) ∈ A ∪ B, then φ(i, j) ∈ ωi ∪ ωj.
We will prove the equivalent propositions that if φ(i, j) ∉ ωi ∪ ωj, then φ(i,
j) ∉ A ∪ B.
www.pdfgrip.com
Well, suppose φ(i, j) ∉ ωi ∪ ωj. Then ω(i, j) ∈ ωi and φ(i, j) ∉ ωj. Since
φ(i, j) ∉ ωi, then . Then by (a), . Similarly, since
φ(i, j) ∉ ωj, then . Hence by condition D0 of the definition of
“weakly doubly productive,”
17. (a) Suppose that A is recursively enumerable. Let M(x, y) be the recursively
enumerable relation x ∈ ωy ν x ∈ A. By the iteration theorem, there is a
recursive function t(i) such that for all i, ωt(i) = {x : M(x, i)}. Thus
(b) Suppose (A, B) is effectively inseparable under ζ(x, y) and A and B are
both recursively enumerable. Now let (ωi, ωj) be a disjoint pair such that
ωi is disjoint from A and ωj is disjoint from B. Since ωi is disjoint from
both ωj and A, then ωi is disjoint from ωj ∪ A. Since B is disjoint from
both ωj and A, then B is disjoint from ωj ∪ A. Thus ωi and B are both
disjoint from ωj ∪ A, so that ωi ∪ B is disjoint from ωj ∪ A.
Thus, ωj ∪ A and ωi ∪ B are disjoint supersets of A and B
respectively. By (a), there are recursive functions t1 and t2 such that
and . Therefore , are disjoint supersets
of A and B respectively. Because the pair (A, B) is assumed to be
effectively inseparable under ζ(x, y), ζ(t1(j), t2(i)) [which we define g(i, j)
to be] lies outside both and . But and ,
which means g(i, j) must of course lie outside both ωi and ωj (and A and
www.pdfgrip.com
Chapter 7
Metamathematical Applications
I. Simple Systems
By a simple system I shall mean a pair (P, R) of disjoint sets, together with a
function neg(x) and a function r(x, y) satisfying condition C1 below.
For any numbers h and n, we abbreviate neg(h) by h′, and r(h, n) by h(n), or
sometimes by rh(n).
C1: For any numbers h and n, h′(n) ∊ P iff h′(n) ∊ R, and h′(n) ∊ R iff h(n) ∊
P.
Intended Applications
follows: We arrange all the sentences of in an infinite sequence S1, S2, …, Sn,
… according to the magnitude of their Gödel numbers. We refer to n as the index
of the sentence Sn. We take one particular variable x and arrange all formulas
with x as their only free variable in an infinite sequence φ1(x), φ2(x), …, φn(x), …
according to the magnitude of their Gödel numbers. We refer to n as the index of
φn(x). We arrange all formulas with the two variables x and y as their only free
variables in an infinite sequence ψ1(x, y), ψ2(x, y), …, ψn(x, y), …, according to
the magnitudes of their Gödel numbers, and for any numbers n, x and y, we
define s(n, x, y) to be the index of the sentence
For the associated simple system S of , we take P to be the set of indices of
the provable sentences, and take R to be the set of indices of the refutable
sentences. For any numbers h and n, we take r(h, n) to be the index of the
sentence [i.e. the sentence resulting from substituting the number (the
name of n) for all free occurrences of x in φh(x)]. We take h′ [which is the
abbreviation of neg(h)] to be the index of the formula ∼φh(x) Of course φh′ (x) is
provable (refutable) iff φh(x) is refutable (provable). And so the items (P, R),
neg(h), and r(x, y) constitute a simple system. Consequently, whatever we prove
for simple systems in general will hold for Peano Arithmetic and other systems.
Returning to simple systems in general, in accordance with intended
applications, we might call members of P provable numbers, and members of R
refutable numbers. We will call a number decidable if it is in either P or R, and
call it undecidable otherwise. We call S complete if every number is decidable,
otherwise incomplete.
We call two numbers equivalent if whenever either one is in P, so is the other,
and if either one is in R, so is the other.
We will refer to the function r(x, y) as the representation function for sets of
the system S, and we say that a number h represents the set A of all numbers n
such that h(n) ∈ P. Thus, to say that h represents A is to say that for all n, h(n) ∈
P iff n ∈ A. We call A representable in if some number h represents it.
We say that h completely represents A iff h represents A and h′ represents the
complement of A.
We say that h defines A if, for all numbers n:
(1) n ∈ A implies h(n) ∈ P ;
www.pdfgrip.com
For any function f(x) and set A, by f–1(A) is meant the set of all n such that
f(n) ∈ A. Thus n ∈ f–1(A) iff f(n) ∈ A.
Recall that rh(n) is r(h, n) and that rh(n) = h(n), where r(x, y) is the
representation function for sets.
Problem 2. Show that h represents A iff . Show also that h contra-
represents A iff
Problem 3. Prove that if some set A is represented in S and its complement is
not, then S is incomplete.
For any set A, by A* we shall mean the set of all numbers h such that h(h) ∈
A.
Problem 4. Prove that neither the complement of P*, nor the complement
of R*, is representable in S.
By the diagonalization of a number h we shall mean the number h(h). By the
skew diagonalization of h we shall mean the number h(h′).
Discussion
We define d(x) to be r(x, x), which is x(x). We note that for any set A the set A* is
d–1(A).
Problem 7. Prove that for any sets A and B:
(a) If A ⊆ B, then f–1(A) ⊆ f–1(B).
(b) If A is disjoint from B, then f–1(A) is disjoint from f–1(B).
(c) f–1( ) is the complement of f–1(A).
In particular, taking d(x) for f(x),
(a1) If A ⊆ B, then A* ⊆ B *.
(a2) If A is disjoint from B, then A* is disjoint from B *.
(a3)
Admissible Functions
We shall call a function f(x) admissible if for every number h there is a number k
such that for every n the number k(n) is equivalent to h(f(n)).
www.pdfgrip.com
Separation
Given a disjoint pair (A, B), we say that a number h separates A from B (in ) if
h represents a superset of A and contra-represents some superset of B.
We say that h exactly separates A from B if h both represents A itself and
contra-represents B.
The following problem is key.
Problem 8. Suppose f(x) is admissible. Prove the following:
(a) If A is representable, so is f–1(A). If A is contra-representable, so is f–1(A).
(b) If a disjoint pair (A, B) is exactly separable in S, then so is the pair (f–1A), f–
1(B)).
Normal Systems
We shall call S a normal system if the diagonal function d(x) is admissible. This
means that for every number h there is a number — which we shall denote h# —
such that for all n the numbers h#(n) is equivalent to h(d(n)).
Problem 9. Suppose is normal. If A is representable in , does it necessarily
follow that A* is representable in ?
Problem 10. Show that for any normal system, neither nor is representable.
Fixed Points
Problem 13. Show that no superset A of R*, with A disjoint from P*, can be
definable in . [Hint: Show that if R* ⊆ A, and h defines A in , then h is in both
A and P*(and hence that A is not disjoint from P*).]
The word “undecidable” has two very different meanings in the literature, which
might sometimes cause confusions. On the one hand, a sentence of a system is
called undecidable in the system if it is neither provable nor refutable in the
system. On the other hand, the system as a whole is called undecidable if the set
of provable sentence is not recursively solvable (or: under an effective Gödel
numbering, the set of Gödel numbers of the provable sentences is not recursive).
We shall thus call a simple system undecidable if the set P is not recursive.
The two senses of undecidable, though different, have the following
important link:
Theorem 1. If is undecidable (but standard), then is incomplete (some
number is undecidable in , i.e. outside both P and R).
Recursive Inseparability
B), if the pair (f–1(A), f–1(B)) is recursively inseparable, then the pair (A, B) is
also recursively inseparable. Conclude that if (P*, R*) is recursively inseparable,
so is (P, R).
We next wish to prove:
Theorem 3. If every recursive set is definable in then the pair (P*, R*) is
recursively inseparable, and so is the pair (P, R).
Problem 21. Prove Theorem 3. [Hint: Use Problem 13.]
Recall that, given a disjoint pair (A, B), we say that a number h separates A from
B (in ) if h represents a superset of A and contra-represents some superset of B.
And we say that h exactly separates A from B if h both represents A and contra-
represents B.
We next wish to prove the following theorem, which has the interesting
corollaries that follow it:
Theorem 4. If some disjoint pair (A1, A2) is recursively inseparable in , but
separable in , then the pair (P, R) is recursively inseparable.
Corollary 1. If some recursively inseparable pair is separable in , then is
undecidable.
Corollary 2. If some recursively inseparable pair is separable in , then is
incomplete (assuming S is standard).
Note: Corollary 2 is essentially Kleene’s symmetric form of Gödel’s theorem
[Kleene, 1952]. Kleene proved this for a particular pair of recursively
inseparable sets, but the proof works for any recursively inseparable pair of
recursively enumerable sets.
Problem 22. Prove Theorem 4.
Rosser Systems
We shall call a Rosser system if for any two recursively enumerable sets A and
B, there is a number h that separates A − B from B − A. [If A and B are disjoint,
then of course such an h also separates A from B.]
We shall call a Rosser system an exact Rosser system, if for any two disjoint
www.pdfgrip.com
The following proposition is key for this, as well as for some further results.
Proposition 3. If is an effective Rosser system, then for any recursively
enumerable relations R1(x, y), R2(x, y), there is a number h such that for any n:
Problem 24. Prove Proposition 3. [Hint: Apply the weak double recursion
theorem to the relations R1 (x, Π(y, z)), R2(x, Π(y, z)), where Π(x, y) is a Rosser
function for .]
Now we can prove:
Theorem 6. If is an effective Rosser system, then is an exact Rosser system.
Problem 25. Prove Theorem 6. [Hint: Let R1(x, y) be the relation x ∈ A ∨y(x) ∈
R, and let R2(x, y) be the relation x ∈ B ∨ y(x) ∈ P. Recall that y(x) = r(y, x).
Apply Proposition 3 to R1(x, y) and R2(x, y).]
We say that S has the Rosser fixed point property if for any two recursive
functions f1(x) and f2(x), there is a number h such that for all n for which ωf1(n)
is disjoint from ωf2(n):
(1) h(n) ∈ ωf1(n) implies h(n) ∈ P ;
(2) h(n) ∈ ωf2(n) implies h(n) ∈ R.
Theorem 7. Every effective Rosser system has the Rosser fixed point property.
Problem 26. Prove Theorem 7. [Hint: Given recursive functions f1(x) and f2(x),
apply Proposition 3 to the relations y(x) ∈ ωf1(x) and y(x) ∈ωf2(x).]
Uniform Incompleteness
By the principal part of a simple system S we mean the pair (P, R).
We call a simple system S ′ an extension of a simple system S if P ⊆ P′ and R
⊆ R′, where (P′, R′) is the principal part of S′, and the representation function
www.pdfgrip.com
r(x, y) and the function neg (x) are the same in both S and S′.
Now consider an infinite sequence = 1, 2, …, n, …, where each Sn+1 is an
extension of Sn. Such a sequence is called an effective sequence, or a recursively
enumerable sequence, if the relations x ∈ Py and x ∈ Ry are both recursively
enumerable. The system is called uniformly incompletable if, for any
recursively enumerable sequence 1, 2, …, n,…, of extensions of , there is a
number h such that, for every n, the number h(n) is an undecidable number of n
(i.e. h(n) is outside Pn ∪ Rn).
Before leaving this chapter, I should say a little more about the relation of
simple systems to the first-order arithmetical systems in standard formalization
of Tarski [1953].
Consider such a first-order system and its associated simple system S, as
defined much earlier in this chapter.
is called a Rosser system, more specifically a Rosser system for sets, iff its
associated simple system S is a Rosser system, as defined here, which is
equivalent to the condition that for any pair of recursively enumerable sets A and
B there is a formula φh(x) such that for all n, if n ω A − B, then φh( ) is provable
(in ), and if n ∈ B − A, then φh( ) is refutable in . [In terms of the associated
simple system S, if n ∈ A − B, then h(n) ∈ P and if n ∈ B − A, then h(n) ∈ R.]
The system is called an exact Rosser system (for sets) if for any pair of disjoint
recursively enumerable sets A and B, there is a predicate H such that for all n, n
∈ A iff H(n) is provable, and n ∈ B iff H(n) is refutable.
In the literature, a function f(x) is said to be definable if there is a formula
Ψ(x, y) such that for all n and m:
www.pdfgrip.com
Ƒ is called a Rosser system for binary relations if for any two recursively
enumerable relations R1(x, y), R2(x, y), there is a formula Ψ(x, y) that separates
the relation R1(x, y)Λ ∼ R2(x, y) from R2(x, y) Λ ∼ R1(x, y). Sheperdson’s
Theorem is that every consistent Rosser system for binary relations is an exact
Rosser system for sets. This appears to be incomparable in strength with the
Putnam–Smullyan Theorem, which is that every consistent Rosser system in
which all recursive functions of one argument are definable is an exact Rosser
system for sets.
The proof of Shepherdson’s Theorem (which can be found in [Shepherdson,
1961]) is not very different from my proof that every effective Rosser system is
an exact Rosser system. Indeed, this theorem and its proof borrowed heavily
from Shepherdson’s Theorem and proof.
The statements (a1), (a2), and (a3) clearly follow from statements (a), (b),
and (c) when f = d(x), because d-1(A) is the set A* for any set A.
8. Suppose f(x) is admissible. Given h, let k be such that k(n) is equivalent to
h(f(n)) for all n.
(a) We assert that if h represents a set A, then k represents f-1(A), and that if
h contra-represents A, then k contra-represents f-1(A).
To prove the former, suppose h represents A. Then n ∈ f−1(A) iff f(n) ∈
A, which is true iff h(f(n)) ∈ P [since h represents A], which is true iff k(n)
∈ P. Thus, for all n, n ∈ f-1(A) iff k(n) ∈ P, and thus k represents f-1(A).
The proof that if h contra-represents A, then k contra-represents f-1(A)
is similar (just replace “P “ by “R” and “represents” by “contra-
represents”).
(b) Suppose h exactly separates the set A from the set B. Then h represents
www.pdfgrip.com
To apply the general case to (1) and (2): Since the function r(x, y)
is recursive, so is the diagonal function d(x) [which is r(x, x)], and
since A* = d-1(A), it follows from (1′) and (2′) that if A is recursively
enumerable (recursive), so is A*.
15. We must show that (a) every set representable in is recursively
enumerable, and that (b) every set definable in is recursive.
(a) Suppose h represents A. Then for all n, n ∈ A iff r(h, n) ∈ P. Let f(x) =
r(h, x) Since the function r(x, y) is recursive, so is the function f(x).
Thus for all, n ∈ A iff f(n) ∈ P, and so A = f−1(P). Since P is
recursively enumerable, and f(x) is a recursive function, then f−1(P) is
recursively enumerable [by Problem 14], and thus A is recursively
enumerable.
(b) If A is completely representable [which is equivalent to A being
definable, by Problem 1(a)], then A and are both representable, hence
both recursively enumerable, and so A is then recursive.
16. (a) Suppose that A is a recursively enumerable set that is not recursive, and
that A is representable in . Since A is not recursive, its complement is not
recursively enumerable, hence not representable [since only recursively
enumerable sets are representable in , by Problem 15]. Thus A is a
representable set whose complement is not representable. Hence is
incomplete [by Problem 3].
(b) This follows from (a) and the fact that there exists a recursively
enumerable set that is not recursive, such as Post’s complete set K [if all
recursively enumerable sets were representable in , K would also have
to be recursively enumerable].
17. Suppose is undecidable and effective. Then the set P is not recursive, and
so its complement is not recursively enumerable. Yet the set R is
recursively enumerable [since is effective], and so R is not the
www.pdfgrip.com
is recursively inseparable.
22. Suppose (A1, A2) is disjoint and recursively inseparable, and that h
separates A1 from A2 in . Then h represents some superset B1 of A1 and
contra-represents some superset B2 of A2. In this situation B1 and B2 must
be disjoint [because n ∈ B1 iff h(n) ∈ P and n ∈ B2 iff h(n) ∈ R, but P and
R are disjoint]. Since (A1, A2) is recursively inseparable, and A1 ⊆ B1 and
A2 ⊆ B2, then the pair (B1, B2) is recursively inseparable (verify this!).
Now, by Problem 2, B1 = (P) and B2 = (R), where r(x, y) is the
representation function for sets. Thus the pair ( (P), (R)) is
recursively inseparable, and since the function rh(x) is recursive [because
r(x, y) is], then the pair (P, R) is recursively inseparable [by Problem 20].
Now, Corollary 1 follows because if (P, R) is recursively inseparable,
then of course P is not recursive, so that is undecidable.
Corollary 2 then follows from Corollary 1 and Theorem 1 (if is
effective but undecidable, then is incomplete).
23. (a) To prove Proposition 1, suppose K1 is separable from K2 in and let h
effect the separation. Then h represents some superset A1 of K1 and contra-
represents some superset A2 of K2. Since P is disjoint from R, then A1 is
disjoint from A2. Of course, h exactly separates A1 from A2.
Since (K1, K2) is effectively inseparable, so is the pair (A1, A2)
[verify this!]. Since is an effective system, then A1 and A2 are
recursively enumerable sets. Thus (A1, A2) is an effectively inseparable
pair of recursively enumerable sets, hence is doubly universal [by
Theorem 1. of Chapter 6]. Thus h exactly separates the doubly universal
pair (A1, A2).
(b) To prove Proposition 2, suppose that some doubly universal pair (U1,
U2) is exactly separable in and that all recursive functions of one
argument are admissible in . Let (A, B) be a disjoint pair of
recursively enumerable sets that we wish to exactly separate in .
Thus we have:
(1) [(p1 ∨ q2) ∧ ∼(p2 ∧ q1)] ⊃q1;
(2) [(p2 ∨ q1) ∧ ∼(p1 ∧ q2)] ⊃q2;
(3)∼ (p1 ∧ p2);
(4)∼ (q1 ∧ q2).
We are to infer that p1 ≡ q1 and p2 ≡ q2.
(a) First we show that p1 ⊃ q1. Well, suppose p1. Then p1 ∨ q2 is true.
Thus (1) is equivalent to
Since q2 is false, it follows from (2′′) that p1 is true. This proves that q1
⊃ p1. Thus p1 ≡ q1.
The proof that p2 ≡ q2 is similar (just interchange p1 with p2 and q1
with q2).
Note: This proof is a modification of a proof of a result of John
Shepherdson.
26. Suppose is an effective Rosser system. Given two recursive functions
f1(x) and f2(x) let R1(x, y) be the relation y(x) ∈ ωf1(x) and let R2(x, y) be
www.pdfgrip.com
Part III
Chapter 8
Beginning Combinatory Logic
Fixed Points
For any elements A, B and C, we say that C combines A with B if Cx = A(Bx) for
every element x. We say that the composition condition holds if for every A and
B, there is some C that combines A with B.
Egocentric Elements
Agreeable Elements
www.pdfgrip.com
Compatible Elements
We will say that an element A is compatible with B if there are elements x and y
such that Ax = y and By = x. When such a pair of elements (x, y) exists, it is
called a cross point of the pair (A, B).
Problem 6. Prove that if conditions C1 and C2 of Theorem 1 hold, then any two
elements A and B are compatible, i.e. any two elements have a cross point.
Smug Elements
Fixation
Narcissistic Elements
for every x and y. This K plays a key role in combinatory logic, as we shall see.
Note: Why the alternative word “kestrel”? Well, in my book To Mock a
Mockingbird [Smullyan, 1985], I used talking birds as combinators. If one calls
out the name of a bird y to a bird x the bird x calls back the name of some bird,
and this bird I call xy. Thus xy is the bird named by x upon hearing the name of
y. I gave bird names to the standard combinators of the literature, usually names
whose first letter is the first letter of the combinator in the literature. For
example, I call the bird representing the standard combinator K by the name
www.pdfgrip.com
We shall call a combinator L a lark if for all x and y, the following condition
holds:
www.pdfgrip.com
The lark was not a standard combinator in the literature prior to the publication
of my To Mock a Mockingbird, but it is so now. Several papers have been
written about it. The lark plays a key role in my treatment of combinatory logic.
One nice thing about the lark is that without assuming conditions C1 and C2,
the mere existence of a lark ensures that every element has a fixed point.
Problem 23. Prove this.
Problem 24. Prove that no element can be both a lark and a kestrel.
Problem 25. Prove that if a lark L is narcissistic, then L must be a fixed point of
every element x.
Problem 26. Prove that no kestrel can be a fixed point of a lark.
However, it is possible for a lark to be a fixed point of a kestrel.
Problem 27. Show that if a lark L is a fixed point of a kestrel, then L must be a
fixed point of every element.
An interesting thing about larks is that the existence of a lark L, without any
other information, is enough to guarantee the existence of an egocentric element.
In fact, we can write down an expression using only the letter L and parentheses
that names an egocentric element. The shortest expression I have been able to
find uses twelve occurrences of the letter L. Is there a shorter one? I raised this
as an open problem in To Mock a Mockingbird, and I don’t know if this problem
has been solved yet.
Problem 28. Given a lark L, prove that there is an egocentric element, and then
write down the name of such an element, using only the letter L and parentheses.
Discussion
Sages
An element θ will be called a sage if its existence not only guarantees that every
element has a fixed point, but θ is also such that when applied to any element x,
it yields a fixed point of x. Thus, θ is a sage, if, for every x, θx is a fixed point of
x, i.e., x (θx) = θx for every element x.
In the classic literature of combinatory logic, sages are called fixed point
combinators. I introduced the word sage in my book To Mock a Mockingbird,
and the term sage has subsequently become standard in the literature, so I will
use the term sage here. Much study has been devoted to sages, and in a later
chapter, I will show many ways of deriving a sage from various other
combinators. For now, let me just state that if contains a lark L and a mocker
M, and if the composition law holds, them contains a sage.
Problem 30. Prove this.
Note: It follows from the last problem that if contains a lark L and an
identity combinator I, and obeys the composition condition, the contains a
sage, because in that case must also contain a mocker M by Problem 30.
www.pdfgrip.com
with B. Thus for all y, we have Cy = A(By). We are also given that there is a
mocker M, and so C has some fixed point y. Thus y = Cy and Cy = A(By),
yielding y = A(By). Now take x to be By, so that y = Ax and x = By. Thus Ax
= y and By = x.
7. Of course it is: Suppose x is a fixed point of A, so that Ax = x. Now take y to
be x itself. Then Ax = y and Ay = x, which means that A is smug.
8. We are given that the composition condition C1 holds and that there is a
smug element A. Thus there are elements x and y such that Ax = y and Ay =
x. Since Ax = y, we can substitute Ax for y in the equation Ay = x, yielding
A(Ax) = x. Now let B be an element that combines A with itself. Thus Bx =
A(Ax), and since A(Ax) = x, we have Bx = x. Thus B has the fixed point x.
9. Of course not! If A is fixed on x and A is fixed on y, then for any element z
we see that Az = x and Az = y so that x = y.
10. It is obviously (b) that is true. Suppose x is fixated on y. Then xz = y for
every z. Hence, taking y for z, we see that xy = y, so that y is a fixed point
of x.
11. It is obviously (a) that is true. Suppose A is narcissistic. Then Ax = A for
every x, so of course AA = A.
12. As given, the statement is obviously true: If A is narcissistic, then Ax = A
and Ay = A. Thus Ax and Ay are both equal to A, so that Ax = Ay.
13. Yes, it does: If A is narcissistic, then Ax = A, so that Axy = Ay. But also Ay
= A, yielding Axy = A.
14. Suppose A is narcissistic. Then by the last problem Axy = A. Thus (Ax)y =
A for every y, which means that Ax is narcissistic.
15. Suppose Kx = x (i.e that x is a fixed point of K). Then Kxx = xx. But also
Kxx = x (since Kxy = x for every y). Thus Kxx is equal to both xx and to x,
so that xx = x, which means that x is egocentric.
16. Suppose Kx is egocentric. Thus Kx(Kx) = Kx. Also Kx(Kx) = x (since Kxy
= x for every y). Thus Kx(Kx) is equal to both Kx and x, so that Kx = x.
17. Suppose Kx is a fixed point of K. Then Kx is egocentric by Problem 15.
Hence x is a fixed point of K by Problem 16.
18. Suppose Kx = Ky. Now, Kxz = x for every z. Since Kx = Ky, it follows that
Kyz = x for every z. But also Kyz = y. Hence x = y.
19. We will show that if Kx = K, then for any elements y and z, it must be the
www.pdfgrip.com
If K were a lark, then Kxy = x(yy) for all x and y, so that for x = K and y = K,
we would have:
For (1) and (2) it would follow that K = K(KK), violating the fact that for no
x can K = Kx (by Problem 19). Thus K cannot be a lark.
25. Suppose that a lark L is narcissistic. Then Lxy = L for every x and y by the
solution to Problem 13. Taking Lx for y, we obtain Lx(Lx) = L. But Lx(Lx)
is a fixed point of x [by the solution to Problem 23]. Thus L is a fixed
point of x.
26. Suppose that LK = K (i.e. that K is a fixed point of L). Then LKK = KK.
But LKK = K(KK) (since L is a lark), so that K(KK) = KK. Then by the
cancellation law, taking KK for x and K for y, we see that KK = K, which
means that K is egocentric, which is contrary to Problem 20. Therefore K
cannot be a fixed point of L.
27. If the lark L is a fixed point of the kestrel K, then KL = L. So, for any x,
KLx = Lx. But KLx = L, since K is a kestrel. So we have Lx = L for all x,
and L is narcissistic. Thus L is a fixed point of every element x by Problem
25.
28. We first show that if x is any fixed point of LL, then xx must be egocentric.
www.pdfgrip.com
Well, suppose that LLx = x. Now LLx = L(xx) [since L is a lark]. Thus x
= L(xx). Therefore xx = L(xx)x, which is xx(xx). Thus xx is egocentric.
We can actually compute a fixed point of LL, in terms of just L: As we
saw in the solution of Problem 23, for any x, a fixed point of x is Lx(Lx).
Taking LL for x, we see that a fixed point of LL is L(LL)(L(LL)), and
therefore an egocentric element is L(LL)(L(LL)) repeated, which is:
29. (1) Suppose I is agreeable. Then for any x there is some y such that Iy = xy.
But Iy = y, so that y = xy, and thus y is a fixed point of x. Conversely,
suppose that every element x has a fixed point y. Then xy = y. But y = Iy, so
that xy = Iy, and thus I agrees with x on y.
(2) Suppose that every pair of elements is compatible. Now consider any
element A. Then A is compatible with I. Thus there are elements x and y
such that Ax = y and Iy = x. Since Iy = x, then y = x (since y = Iy). Thus
Ax = y and y = x and it follows that Ax = x, which means that x is a fixed
point of A. Thus every element has a fixed point, and consequently I is
agreeable by (1).
(3) Since Ix = x for every x, it follows that II = I, which means that I is
egocentric. If I were narcissistic, that would mean that for every x, Ix =
I. But also Ix = x. Hence if I were narcissistic, then we would have x = I
for every element x, contrary to the fact that has more than one
element.
30. Take M to be LI. Well, LIx = I(xx) = xx. Thus LI is a mocker.
31. We will see that any element θ that combines M with L must be a sage.
We already know that Lx(Lx) is a fixed point of x (by the solution to
Problem 23). Now, Lx(Lx) = M(Lx), and consequently M(Lx) is a fixed
point of x. Now, suppose that combines M with L. Then θx = M(Lx), and so
θx is a fixed point of x. Thus θ is a sage.
www.pdfgrip.com
Chapter 9
Combinatorics Galore
I. The B-Combinators
Some Derivatives of B
From B alone, one can derive combinators D, B1, E, B2, D1B+, D2, , which are
defined by the following conditions:
(a) Dxyzw = xy(zw), the dove
(b) B1xyzw = x(yzw)
(c) Exyzwv = xy(zwv), the eagle
(d) B2xyzwv = x(yzwv)
(e) D1xyzwv = xyz(wv)
(f) B+xyzw = x(y(zw))
(g) D2xyzwv = x(yz)(wv)
(h) xy1y2y3z1z2z3 = x(y1y2y3)(z1z2z3).
All these combinators, including B, are called compositors, since they serve
to introduce parentheses. The standard ones are the bluebird B and D, which I
call the dove. We shall also have good use for E, which I call the eagle.
Problem 4. (1) Derive these combinators from B. [Hints: Do the following in
order:
(a) Express D in terms of B.
(b) Express B1 in terms of B and D (which can then be reduced to an expression
in terms of B alone).
(c) Express E in terms of B and B1.
(d) Express B2 in terms of B and E.
(e) Express D1 in terms of B and B1, or, alternatively, in terms of B and D.
(f) Express B+ in terms of B and D1.
(g) Interestingly, D2 can be expressed in terms of D alone!
(h) Express in terms of E alone!]
(2) Using mathematical induction, show that for each positive integer n, there is
a combinator Bn, derivable from B, satisfying the condition
Note that by this definition, the bluebird B also has the name B0, and B1 and B2
www.pdfgrip.com
are the same as defined above (where different variables were used in the
definition).
Some Relations
Given a permuting combinator A satisfying the condition Axyz = abc, where abc
is some permutation of xyz, by A* is meant the combinator satisfying the
condition:
Thus C*, R*, F*, V* are combinators defined by the following conditions:
Thus C**, R**, F**, V** are combinators defined by the following conditions:
Problem 16. Show that C**, R**, F**, V** are all derivable from B and T.
www.pdfgrip.com
The problem is far simpler than it might appear. One can handle all four cases
in one fell swoop!
Vireos Revisited
you cannot have a quixotic bird without a quizzied bird, nor a quizzied bird
without a quixotic bird, or if there isn’t such a Chinese proverb, then there
should be!
Problem 21. What is the sense behind such a proverb?
The quirky bird, Q3, the quacky bird, Q4, the quintessential bird, Q5, and the
quivering bird, Q6, are defined by the following conditions:
Problem 25. We have seen that Q is derivable from B and T. But also, B is
derivable from Q and T. Prove this. [It is not obvious.]
Problem 26. One can derive C from Q and T more easily than from B and T, in
fact, with an expression of only four letters. How can this be done?
Problem 27. Derive G from B and T, or from combinators already derived from
B and T.
Problem 29. There are two interesting ways of deriving W′ from B, T and M.
One way is to first derive W′ from M2 and the robin R, and then reduce that
expression to B and T, which should yield an expression of five letters. Another
way is to first derive W′ from B, T and M2, which when reduced to B, T and M
yields a different expression, one that also contains five letters. The reader
www.pdfgrip.com
Show that these are all derivable from B, T and M, in fact from W and B.
I have found good use for a combinator H (dubbed the hummingbird in To Mock
a Mockingbird). The combinator H is defined by the condition
Problem 33. Show that H is derivable from B, C and W, and thus from B, M and
T.
Problem 34. One can also derive W from H and R, or, more neatly, from C, H
and R. How? [Hint: First derive W′.]
Larks Revisited
We recall the lark L defined by the condition Lxy = x(yy). The lark can be
derived from B, T and M in many ways.
Problem 35.
(a) Show that L is derivable from B, R and M or from B, C and M, and then
reduce the latter expression to one in terms of B, M and T.
(b) Show that L is derivable from B and W. This fact is rather important.
www.pdfgrip.com
Problem 40. We shall have good use for some of the combinators P, P1 , P2 and
P3 satisfying the following conditions:
These can be derived from B, M and T, in fact from B, Q and W (and hence from
B, C and W). How?
Problem 41. Show that M is derivable from P and I.
We will later need combinators Φ, Φ2, Φ3, Φ4, defined by the following
conditions:
www.pdfgrip.com
Problem 42. (a) Derive Φ from B and S. (b) Show by induction that for each n,
there is a combinator Φn derivable from B and S satisfying the equation:
1. To begin with, let us note that if B is present, then the composition condition
must hold, since for any elements x and y, an element that combines x with y
is Bxy [since (Bxy)z is Bxyz, which is x(yz)].
We recall from the solution of Problem 1 of the last chapter that if C is
any element that combines x with M, then CC is a fixed point of x (we
stated this result as Theorem 1*). Well, BxM combines x with M, and so
BxM(BxM) is a fixed point of x. But BxM(BxM) is also equal to M(BxM), by
the definition of M [as applied to the first M in the expansion of M(BxM)].
So M(BxM) is also a fixed point of x (for every x!). We will use this result
several times in this chapter and the next.
2. Since B is present, the composition condition holds. Then by the solution of
Problem 2 of the last chapter, any fixed point of M is egocentric. Now,
M(BMM) is a fixed point of M [by the previous problem of this chapter,
taking M for x]; hence M(BMM) is egocentric. Let us double-check: To
reduce clutter, let A be BMM. We are to show that MA is egocentric. Well,
MA = M(BMM) = BMM(BMM). But BMM(BMM) = M(M(BMM)), since B is
a bluebird. Then we have M(M(BMM)) = M(MA) = MA(MA). Thus MA =
MA(MA), and MA is egocentric.
3. By Problem 20 (b) of the last chapter, any fixed point of a kestrel K is
narcissistic. Since M(BKM) is a fixed point of K [by the solution to Problem
1], it must be narcissistic [as the reader can check directly, by showing that
M(BKM)x = M(BKM)].
4. (1) (a) We take D to be BB. Let us check that this works: BBxy = B(xy).
Hence BBxyz = B(xy)z. Then
www.pdfgrip.com
Thus,
Hence
Hence,
(2) We will show by mathematical induction that we can define all the
combinators
5. Actually, any fixed point A of T will commute with every element x. For
suppose T A = A. Then for every element x, we see that Ax = T Ax = xA.
Thus A commutes with x.
6. Take R to be BBT. Then
We might note that in terms of B and T, the expression BCR for the finch
has 12 letters (since C has 8 and R has 3), while using the expression
B(RRR)R for the finch gives us 13 letters in terms of B and T; using the
expression BC(CC) for the finch, we get 25 letters in terms of B and T.
As for F = ETT ET, as it stands, when E is replaced by B(BBB), we get
an expression of 11 letters, but the expression can be reduced to only 8
letters in terms of B and T, as we will now see:
Hence,
23. Q4 is derivable from C and Q3, and Q3 is derivable from C and Q4:
Thus,
34. First we take W′ to be HR. Thus HRxy = Rxyx = yxx = W′xy. Then take W
www.pdfgrip.com
Chapter 10
Sages, Oracles and Doublets
Sages Again
The Oracle O
Double Sages
A pair (y1, y2) is called a double fixed point of a pair (x1, x2) if x1y1y2 = y1 and
x2y1y2 = y2. We say the system has the weak double fixed point property if every
pair (x1, x2) has a double fixed point. We say that the system had the strong
double fixed point property if there is a pair (θ1, θ2), called double sage pair —
or, more briefly, a double sage — such that for every x and y, the pair (θ1xy,
θ2xy) is a double fixed point of (x, y), i.e.
www.pdfgrip.com
There are many ways to construct double sages from the combinators B, M and
T, several of which can be found in Chapter 17 of my book Diagonalization and
Self-Reference [Smullyan, 1994], and one of which we shall now consider.
Others, which I will leave as exercises, have solutions which can be found in the
book just mentioned.
The existence of such an N obviously implies that the system has the weak
double fixed point property.
Problem 29. Why?
Before considering the construction of a double sage, let us see how a nice
combinator N can be ultimately derived from B, T and M. To this end, we will
use a combinator n defined by the condition:
Problem 30.
(a) Derive n from B, C and W or from combinators derivable from them. [It is
easiest to derive n directly from C, W, W*, V* and Φ3.]
(b) Now show that any fixed point of n (such as M(Ln)) is a nice combinator.
Problem 31. Show that from N, C, W and W1 one can construct combinators θ1
and θ2 such that (θ1, θ2) is a double sage.
Double Oracles
We will call a pair (A1, A2) a double oracle if every double fixed point of (A1,
A2) is a double sage. It turns out that every double sage is a double fixed point of
any double oracle (A1, A2).
www.pdfgrip.com
18. In the equation Uxy = y(xxy), just substitute U for x, and we have UUy =
y(UUy). Thus UU is a sage.
19. QQW is an oracle, since QQW xy = W (Qx)y = Qxyy = y(xy) = Oxy.
20. We take CB for Q in QQW from the previous problem, which shows us that
CB(CB)W is an oracle.
21. SI is an oracle, since SIxy = Iy(xy) = y(xy) = Oxy.
22. PI is an oracle, since PIxy = y(Ixy) = y(xy) = Oxy.
23. See the solution to Problem 25 below.
24. See the solution to Problem 26 below.
25. BOM is a Turing combinator, since
θx = O(θx) (they are both equal to x(θx)). Since x = Oθx for every term x, it
follows that θ = Oθ, which means θ that θ is a fixed point of O!
29. In the equation Nzxy = z(Nxxy)(Nyxy) substitute x for z to obtain:
(1) Nxxy = x(Nxxy)(Nyxy).
Next, in the equation Nzxy = z(Nxxy)(Nyxy) substitute y for z to obtain:
(2) Nyxy = y(Nxxy)(Nyxy).
By (1) and (2), the pair (Nxxy, Nyxy) is a double fixed point of (x, y).
30. (a) We take n to be C(V*Φ3W (W2C)). We leave it to the reader to verify that
Thus (A1, A2) is a double sage, and (O1, O2) is a double oracle, as we were
asked to show.
Before proving that every double sage is a double fixed point of (O1,
www.pdfgrip.com
O2), let us note that if A1xy = A2xy for all x and y, then A1 = A2, because
from A1xy = A2xy for all x and y it follows that A1x = A2x for all x, and
hence that A1 = A2.
Now, suppose that (θ1, θ2) is a double sage. Thus:
(1) θ1xy = x(θ1xy)(θ2xy),
(2) θ2xy = y(θ1xy)(θ2xy).
Also,
(1′) O1θ1θ2xy = x(θ1xy)(θ2xy),
(2′) O2θ1θ2xy = y(θ1xy)(θ2xy).
Therefore, θ1xy = O1 θ1 θ2xy and θ2xy = O2 θ1 θ2xy.
Then θ1 = O1 θ1 θ2 and θ2 = O2 θ1 θ2, which means that (θ1, θ2) is a
double fixed point of (O1, O2).
www.pdfgrip.com
Chapter 11
Complete and Partial Systems
When t is a term constructed only from variables that are among x1, x2, …, xn,
and our applicative system contains an element A (no element of contains
variables) such that Ax1x2 … xn = t holds, then we say that A is a combinator of
the applicative system and also say that the combinator A with arguments x1,
x2, …, xn realizes the term t. The equation Ax1x2 … xn = t is called the defining
equation of the combinator A.
It turns out that all constants in any applicative system are combinators,
because in applicative systems the meaning of constants is always specified by a
defining equation. And the systems are also always defined so that every term
www.pdfgrip.com
alphabetical order of the names of the variables.] In any case, you can see that, if
we can prove what we have claimed, then in any applicative system including at
least S and K as constants, there is a combinator A for which
Combinatorial Completeness
Now we wish to prove our principal claim, namely that given any ordered list of
variables to be taken as arguments for a combinator, and any term t built only
from variables included in the list of arguments, there is a combinator with the
given arguments (in the order specified in the list) that is derivable from just S
and K (no variables) which realizes the term t.
To begin with, the identity combinator I is derivable from S and K: just take I
to be SKK, for SKKx = Kx(Kx) = x = Ix. Thus it suffices to show that every
combinator is derivable from S, K and I. This is what we will now do.
Let us first mention the following induction principle, which we will call the
term induction principle: Consider a set S whose members are variables or
constants or both, and let S+ be the set of all terms built from elements of S (and
parentheses). To show that all terms in S+ have a certain property P, it suffices to
show that all elements of S have the property, and that for any terms t1 and t2 of
S+, if t1 and t2 have property P, then so does (t1t2). This principle is as self-
evident as mathematical induction, and can in fact be derived from mathematical
induction on the number n of occurrences of elements of S in the term.
Let us call a simple term if it is built only from variables and S, K, and I. In
the algorithm we are about to present, we will assume we will be starting with a
simple term W built only from variables, but we will successively turn it into a
simple term built from S, K, and I and variables. In the end, we will have built a
term from that is built only of S, K, and I (by having eliminated all the
variables in a specified ordered list of arguments for a combinator , a list that
includes all the variables in ). And if x1, …, xn is the list of variables specified
as arguments for the combinator, it will be the case that x1 … xn = , where the
in question is the original one containing only variables (all of which are
www.pdfgrip.com
Now, consider any simple term and consider a variable x, which may or
may not be in . We shall call a term x an x-eliminate of if the following
three conditions hold:
(a) The variable x does not occur in x.
(b) All variables in x are in .
(c) The equation = holds.
We wish to first show that every simple term has an x-eliminate x that is
derivable from S, K and I and variables in excluding the variable x. We define
x by the following inductive rules:
(1) If is x itself, we take x to be I.
(2) If is any term in which x does not occur, we take x to be K .
(3) If is a compound term in which x occurs, we take ( )x to be .
Note that for any possible simple term, one and only one of the three rules will
apply. When rule (3) is applied, it will require finding the x-eliminates of the two
parts of the compound term.
When we called the scheme of rules for x-elimination we just introduced
inductive, we meant that if we have a simple term from which we wish to
eliminate the variable x we operate repeatedly on the components of our term
working from outside inwards looking for x-eliminates of the components of the
term, until we have eliminated x from all the subterms and finally the whole
term.
Let us consider an example. Let’s see if we can derive from S, K and I a
combinator A of one argument x with the defining equation Ax = (xx). (Thus we
are checking whether the mocker/mockingbird is derivable from S, K and I.) We
want to see if eliminating x, the only variable in the term (and the only variable
in the list of required arguments for the combinator) will give us a term in S, K
and I. Well, (xx) is a compound term in which x occurs, so rule (3) tell us that the
x-eliminate of (xx) consists of the starling S placed in front of the x-eliminates of
the two parts of the compound term, which are both x. By rule (1), the x-
eliminate of x is I. So the x-eliminate of (xx) is SII. Let us check that the three
conditions which an x-eliminate must satisfy actually hold here: (a) The variable
x does not occur in SII; (b) All variables in SII are in (xx) [this holds vacuously,
www.pdfgrip.com
since there are no variables in SII]; (c) SIIx = Ix(Ix) = xx. (Of course xx is just
(xx) with its outer parentheses suppressed for easier reading.)
Now let us prove that for any simple term, the method of x-eliminates we
have described does result in a term that satisfies the conditions on an x-
eliminate.
Proposition 1. is an x-eliminate of . In other words:
(1) The variable x does not occur in .
(2) All variables of occur in .
(3) The equation x = holds.
This proposition is proved by term induction.
Problem 2. (i) Show that if is x, then is an x-eliminate of .
(ii) Show that if is any term in which x does not occur, then is an x-
eliminate of .
(iii) Show that if and are x-eliminates of and respectively, so is
Now let be a term built only from variables among x1, x2, …, xn. Suppose
we want a combinator A with arguments x1x2 … xn derivable from only S, K and
I (using no variables) and satisfying the condition Ax1 … xn = . Well, we take
A to be , which is an x1, x2, …, xn eliminate of .
All variables of A occur in , hence are among x1, x2, …, xn, yet none of the
variables x1, x2, …, xn are in A, which means that A contains no variables at all.
Since A has been formed by a variable-elimination scheme which introduces
only the combinators S, K, and I, A must be built only from S, K, and I. And the
condition Ax1x2 … xn = holds.
We would like to point out that, given a term t in S, K and I (that is, including
terms in which variables are liberally strewn throughout) and an ordered list of
variables x1, x2, …, xn including all the the variables in t, the method of
eliminates described here can also be used, if one wishes, to transform any such
term t into an expression Wx1x2 … xn in which W is a combinator built only with
S, K and I. (To see this one only has to look once more through the method and
the proof that the method works to see that it applies as well in this situation.)
Now let us consider an example of how to find a cancellative combinator with
more than one variable using the method of eliminates: Suppose we want a
combinator A satisfying the condition Axy = y. Thus A has the two arguments x,
y and the term we want to use to define A does not include x. Well, we want A to
be a y, x eliminate of y, i.e. an x-eliminate of a y-eliminate of y. Thus we must
first find a y-eliminate of y, and that is simply I. We then want an x-eliminate of
I, which is KI. Thus our solution should be KI. Let us check: KIx = I, so that
KIxy = Iy = y.
Now for a permuting combinator: How can we derive the thrush T from S, K
and I? [Txy = yx]. We have to find an x-eliminate of a y-eliminate of yx. Well, a
y-eliminate of y is I, and a y-eliminate of x is Kx, and so a y-eliminate of yx is
SI(Kx). We now need an x-eliminate of SI(Kx) = (SI)(Kx). Well, an x-eliminate
of SI is K(SI), and an x-eliminate of Kx is simply K (by Problem 3), and so an x-
eliminate of SI(Kx) is S(K(SI))K. Thus S(K(SI))K should be our solution. Let us
check:
hummingbird. In the first column of the table it is immediately after the Greek
letter λ (and before the period) that the creators of the table list the arguments of
the combinator in question, since those arguments cannot be determined simply
from the constant-free term used to define the combinator. For the goldfinch, the
first column entry in the goldfinch row, λabcd.ad(bc), tells us that the
combinator referred to in this row of the table has the arguments a, b, c and d,
and the defining equation is Gabcd = ad(bc). The next two column entries for
the row tells us that the common symbol for this combinator is G and that it is
commonly called the goldfinch. The next column entry on the row tells us that
this combinator can be derived from B, C and W as BBC. And then the final
column entry for the goldfinch row tells us the goldfinch can be derived from S
and K alone by the term
There are two results of combinatory logic known as the First and Second Fixed
Point Theorems. The second one will be dealt with in the next chapter. The first
will be considered now. To motivate this, the reader should try the following
exercise:
Exercise. (a) Find a combinator Γ that satisfies the condition Γxy = x(Γy). (b)
Find a combinator Γ that satisfies the condition Γxy = Γyx.
The solutions of (a) and (b) are special cases of the First Fixed Point
Theorem.
In what follows t(x, x1, …, xn) is a term in which x, x1, …, xn are precisely the
variables that occur in the term. Recall that all realizable terms are built up only
from variables (and parentheses).
Theorem F1. [First Fixed Point Theorem] For any realizable term t(x, x1, …, xn)
there is a combinator Γ satisfying the condition Γx1 … xn = t(Γ, x1, …, xn).
Remarks. This theorem may be a bit startling, since the satisfying condition for
Γ involves Γ itself! This is related to self-reference or recursion, as will be seen
in the next chapter.
Problem 5.
www.pdfgrip.com
(1) Prove Theorem F1. [There are two different proofs of this. One involves a
fixed point of a combinator that realizes the term t(x, x1, …, xn). The other
uses a combinator that realizes the term t((xx), x1, …, xn).]
(2) Now find solutions to (a) and (b) of the exercise stated before Theorem F1.
Doubling Up
Theorem F1F1. [First Double Fixed Point Theorem] For any two realizable
terms t1(x, x1, …, xn) and t2(x, x1, …, xn) there are combinators Γ1, Γ2 such that:
(1) Γ1x1 … xn = t1(Γ2, x1, …, xn).
(2) Γ2x1 … xn = t2(Γ1, x1, …, xn).
λ-I Combinators
C and I and variables. If a nice term has no variables, then it is a combinator, and
will accordingly be called a nice combinator. Thus a nice combinator is a
combinator built only from S, B, C and I.
We define an x-eliminate of a term as before and we must now show that if
is any nice term, then has a nice x-eliminate, provided that x actually occurs
in ! Well, for this purpose we must redefine our set of recursive rules to
produce the appropriate x-eliminate .
1. If = x, we take to be I. [Thus xx = I.]
2. For a compound term in which x occurs, it occurs either in or in or in
both.
(a) If x occurs both in and , we take to be .
(b) If x occurs in but not in , we take ( to be .
(c) If x occurs in but not in , then:
(c1) if consists of x alone, we take to be . [Thus , if
x does not occur in .]
(c2) if V ≠ x (but x occurs in and not in ), we take to be .
Note that this set of rules, unlike the previous set of rules for x-elimination,
contains no rule for eliminating x when the original term you wish to eliminate x
from is a term in which x doesn’t occur. But this causes no problem in our
current non-cancellative combinator situation. If our original term contains no
occurrence of x, we would have no reason to even start off on an x-elimination,
because we only eliminate variables from the list of arguments for the desired
combinator, and if the original term contains no x, there would be no x in the
argument list, so we wouldn’t try to do x-elimination at all. So we are only
concerned with doing an x-elimination either on an x standing alone, which is
covered by Rule (1), or on a compound term in which x occurs in either or
or both. If we do find after eliminating x from such a compound term (and we
are told by the rules as to what to do when x doesn’t occur in one of the terms)
that there are no longer any x’s in the compound term that result, we just know
that we have the final x-eliminate for the term (or subterm) we are working on.
Proposition 3. For any nice term in which x occurs, the term is a nice x-
eliminate of .
Problem 7. Prove Proposition 3.
We leave it to the reader to show that for a nice term , if variables x and y
www.pdfgrip.com
Notice that this time we are proving Proposition 3 for every term built up
from S, B, C, I and variables (we said any nice term). I pointed out earlier we
also could have done this for terms built from S and K and variables with the
earlier method of eliminates and the same proofs. But since terms built up only
from variables are also terms built up from S, B, C, I and possibly variables, we
have the following immediate consequence of Proposition 3:
Corollary. For every constant-free term t and every list of possible arguments
x1, x2, …, xn for a combinator, if the variables occurring in the term t are x1, x2,
…, xn too, there is a combinator A in the class of variable-free terms built from
S, B, C and I such that Ax1x2 … xn = t (i.e. there is a non-cancellative
combinator A that realizes the term t).
This proves that S, B, C and I form a basis for the class of all λ-I combinators.
It is known that no proper subset of {S, B, C, I} is a basis for the class of λ-I
combinators, but J. B. Rosser found a curious two-element basis for the class of
λ-I combinators, namely I and a combinator J defined by the condition:
B, T, I Combinators
The class of all combinators derivable from B, T and I has been studied by
Rosser [1936] in connection with certain logical systems in which duplicative
combinators like M, W, S, J have no place. In To Mock a Mockingbird I showed
that these three combinators can be replaced by only two, namely I and the
combinator G (the goldfinch), which we recall is defined by the condition:
www.pdfgrip.com
Whether this discovery was new or not, I honestly don’t know. Now let us derive
B and T from G and I.
Problem 8.
(a) First derive from G and I the combinator Q3 (which, we recall, is defined as
satisfying the condition Q3xyz = z(xy)).
(b) Then derive the cardinal C from G and I. [Cxyz = xzy.]
(c) From C and I, we can obtain T. [Alternatively, we can get T from Q3 and I].
(d) From C we can get the robin R [Rxyz = yzx], and then from C, R and Q3, we
can get the queer combinator Q [Qxyz = y(xz)], and then from Q and C we
can obtain B.
1. t is a term if and only if there is a finite sequence t1, …, tn = t such that for
each i ≤ n, either ti is a variable or constant, or there are numbers j1 and j2
both less than i such that ti = (tj1, tj2).
2. We will show that no matter which rule applies to our simple term , then
will be an x-eliminate of .
(i) Let us first consider the case that is x itself. Then, by rule (1),
. We must thus show that I is an x-eliminate of . Well, since there
are no variables in the term I, then of course the variable x does not occur in
I, and it is vacuously true that all variables in I (of course there are none)
occur in . Also, Ix = x, and so I is an x-eliminate of .
(ii) Now we will show that if is any term in which x does not occur,
then, by rule (2) is an x-eliminate of . Clearly, x does not occur in
. Secondly, the variables of are variables of , so that all variables of
are variables of . Thirdly, . Thus is indeed an x-eliminate
of .
(iii) Finally, we consider the composite term in which x does occur,
so that rule (3) applies. Suppose that is an x-eliminate of , and that
is an x-eliminate of . We must show that is an x-eliminate of .
www.pdfgrip.com
(a) Since x does not occur in (since is an x-eliminate of ) and x does not
occur in ), and x does not occur in S, it follows that x does not occur in
.
(b) Every variable y in must occur in , because y either occurs in
or or both, which means it occurs in or or both. So it occurs in the
compound term .
(c) Lastly, we must show that . Well,
all x, x1, …, xn
We thus take AA to be Γ!
Note to the reader: I believe that virtually all fixed points and recursion
theorems in recursion theory and in combinatory logic, as well as the
construction of Gödel self-referential sentences, are simply elaborate
variations on this trick of the above second proof. In all cases, one has an
equation containing Ax of the left side of an equal sign, and xx on the right
side. Taking A for x, we end up with AA on both sides.
Now here is the solution of the Exercise:
We now let (Γ1, Γ2) be a cross point of (A2, A1) [not of (A1, A2) but of (A2,
A1)]. Thus, A2Γ1 = Γ2 and A1Γ2 = Γ1. Then
(1′) Γ1x1 … xn = A1Γ2x1 … xn = t1(Γ2, x1, …, xn).
(2′) Γ2x1 … xn = A2Γ1x1 … xn = t2(Γ1, x1, …, xn).
For the second proof, given terms t1(x, x1, …, xn) and t2(x, x1, …, xn), let A1
and A2 be combinators satisfying the following conditions:
(1) A1xyx1 … xn = t1(xyx, x1, …, xn).
(2) A2xyx1 … xn = t2(xyx, x1, …, xn).
Then,
(1′) A1A2A1x1 … xn = t1(A2A1A2, x1, …, xn).
(2′) A2A1A2x1 … xn = t2(A1A2A1, x1, …, xn).
We thus take Γ1 = A1A2A1 and Γ2 = A2A1A2.
7. We are given that is a nice term and that x occurs in . At each stage of
the construction of , no new constant was introduced other than S, B, C or
I, so that is certainly nice.
We leave it to the reader, using term induction, to prove that x does not
occur in and that all variables in occur in . It remains to show that
. We do this by induction.
1 If = x, then is I and . So . It remains to
show the statement for when we have a compound term .
2(a) Suppose that x is in both and , and that and . We
are to show that . Well,
so that
2(b) For the case when x occurs in but not in , we assume by the induction
hypothesis that . Now, in this case,
so that
www.pdfgrip.com
(c2) In the case that (and when x does not occur in ), we assume
that and we must show that . Well, in this
case , so that
Chapter 12
Combinators, Recursion and the Undecidable
In this chapter, we work with the complete system: all combinators derivable
from S and K. We shall now see how combinatory logic is related to
propositional logic, recursion theory, and the undecidable.
Logical Combinators
We now want two combinators t and f to stand for truth and falsity respectively.
Several workable choices are possible, and the one we take is due to Henk
Barendregt [1985]. We take t to be the kestrel K, and for f we take the
combinator KI, so that in what follows t is an abbreviation for K, while f is an
abbreviation for KI. We shall call t and f propositional combinators. We note
that for any combinators x and y, whether propositional combinators or not, txy =
x and fxy = y, which turns out to be a technical advantage.
There are only two propositional combinators t and f (i.e. K and KI). We now
use the letters p, q, r, s as standing for arbitrary propositional combinators,
rather than propositions, and we call p true if p = t, and false if p = f. We define
the negation ∼ p of p by the usual conditions: ∼t = f and ∼f = t; we define the
conjunction p ∧ q by the truth-table conditions: t ∧ t = t, t ∧ f = f, f ∧ t = f, f ∧ f
= f. We define the disjunction p ∨ q, the conditional p ⊃ q, and the bi-
conditional p ≡ q similarly. We now want combinators that can realize these
operations.
Problem 1. Find combinators N (negation combinator), c (conjunction
combinator), d (disjunction combinator), i (implication combinator), and e
(equivalence or bi-conditional combinator), satisfying the following conditions:
(a) Np = ∼ p,
(b) cpq = p ∧ q,
www.pdfgrip.com
(c) dpq = p ∨ q,
(d) ipq = p ⊃ q,
(e) epq = p ≡ q.
We now see that by combining the above combinators in various ways, we
have combinators that can do propositional logic, in the sense that they can
compute compound truth tables.
Arithmetic Combinators
We shall now see how the arithmetic of the natural numbers can be embedded in
combinatory logic.
Each natural number is represented by a combinator denoted (not to be
confused with the Peano numeral designating n). There are several schemes of
doing this, one of which is due to Church [1941]. The scheme we will use is
again due to Henk Barendregt [1985]. Here the vireo V plays a major role [Vxyz
= zxy]. We let σ (read as “sigma”) be the successor combinator Vf (which is
V(KI)). We take to be the identity combinator I, take to be ; take to be
,… take …. The first thing that needs to be shown is that the
combinators, , are all distinct.
Problem 2. Show that if m ≠ n, then . [Hint: First show that for every
, where n+ is an abbreviation for n + 1. Then show that if , then
. Finally, show, for all m and n and all positive k, that
The combinators are called numerical combinators. A combinator A
is called an arithmetical combinator of type 1 if, for every n, there is some m
such that . We call A an arithmetical combinator of type 2 if, for every n
and m, there is some number p such that . In general, A is called an
arithmetical combinator of type n if, for all numbers x1, …, xn, there is a number
y such that .
We now wish to show that there is an addition combinator, which we will
denote ⊗, such that , and a multiplication combinator, which we
will denote ⊗, such that and an exponential combinator such
that .
We need some items in preparation for all this:
www.pdfgrip.com
We have vital need of a combinator Z, called the zero tester, such that
and for any positive n.
Problem 4. Find such a combinator Z.
Problem 5. Does there exist a combinator A such that , but , for
positive?
We now consider some recursion properties.
Problem 6. Prove that for any combinator A and any number k, there is a
combinator Γ satisfying the following two conditions:
Hints:
1. C2 is equivalent to the following condition: For any positive n,
Property Combinators
Relational Combinators
tuples (k1, …, kn), if R(k1, …, kn) holds, then , and if R(k1, …, kn)
doesn’t hold, then .
Problem 10. Show that there is a relational combinator g that computes the
relations x > y (x is greater than y). [Hint: This relation is uniquely determined
by the following conditions:
1. If x = 0, then x > y is false.
2. If x ≠ 0 and y = 0, then x > y is true.
If x ≠ 0 and y ≠ 0, then x > y is true if and only if x − 1 > y − 1 is true.]
Functional Combinators
Consider a function f(x) of one argument, i.e. an operation that assigns to each
number n a number devoted by f(n). A combinator A will be said to realize the
function f(x) if, for every number n, the condition holds.
Problem 11. Show that if functions f(x) and g(x) are both realizable, so is the
function f(g(x)), i.e. the operation that assigns to each number n the number
f(g(n)).
Problem 12. Suppose R(x, y) is computable and that f(x) and g(x) are both
realizable. Show that the following are computable:
(a) The relation R(f(x), y).
(b) The relation R(x, g(y)).
(c) The relation R(f(x), g(y)).
Problem 13. Show that for any regular relational combinator A of degree 2,
there is a combinator A1 such that for all numbers n and m,
(1) If , then .
(2) If , then .
Problem 14. Now prove the Minimization Principle: for every regular relational
combinator A of degree 2, there is a combinator A′, called a minimizer of A, such
that for every number n, , where k is the smallest number for which
. [Hint: Use the A1 of Problem 13 and the cardinal C [Cxyz = xzy].]
For any numbers n and m, by n * m we shall now mean the number which, when
written in ordinary base 10 notation, consists of n (in base 10 notation) followed
by m (in base 10 notation). For example
Under each symbol I have written its Gödel number. The Gödel number of a
compound expression (a complex term, or a sentence) is the number obtained by
replacing S by 1, K by 2, …, = by 5, and reading the result as a number in base
10. For example, the Gödel number of KS(= is 2135).
We let be the set of true sentences, and let be the set of the Gödel
numbers of the true sentences. The question now is whether the set is
combinatorially computable. Is there a combinator Γ such that for every if
n ∊ , and if n ∉ ?
The question is of staggering importance, since any formal mathematical
question can be reduced to whether a certain number belongs to . The question
www.pdfgrip.com
Gödel Numerals
Let me first point out that there are two different schemes for constructing terms,
which involve two different ways of introducing parentheses. One scheme
consists of building terms according to the following two rules:
(1) S and K are terms.
(2) If X and Y are terms, so is (XY).
The second scheme replaces (2) by:
(2′) If X and Y are terms, so is (X)(Y).
We shall continue to use the first scheme.
Now, for any number n, the numeral , like any other term, has a Gödel
number. We let n# be the Gödel number of the numeral . For example, is the
combinator I, which in terms of S and K is ((SK)K), whose Gödel number is
3312424. Thus 0# = 3312424.
As for 1#, it is the Gödel number of the numeral , and , where σ is the
combinator Vf, and Vf, when reduced to S and K, is a horribly long expression,
and consequently has an unpleasantly long Gödel number, which I will
abbreviate by the letter “s”. Thus, in what follows, s is the Gödel number of (the
term whose abbreviation is) σ, and so 1# is the Gödel number of , which is 3
* s * 0# * 4.
Next, 2# uses the Gödel number of , and the Gödel number of that term is
3 *s * 1# * 4. Similarly 3# = 3 * s * 2# * 4. And so on ….
We now need to show that the function f(n) = n# is combinatorially realizable,
i.e. that there is a combinator δ such that holds (for every n).
Problem 17. Show that there is such a combinator δ.
Normalization
www.pdfgrip.com
Normalization
For any expression X, by ⌈X⌉ is meant the numeral that designates the Gödel
number of X; we might call ⌈X⌉ the Gödel numeral of X. By the norm of X is
meant X⌈X⌉, i.e. X followed by the Gödel numeral of X. If n is the Gödel number
of X, then n# is the Gödel number of ⌈X⌉, so n * n# is the Gödel number of X⌈X⌉,
i.e. of the norm of X.
We now want a combinator Δ, called a normalizer, such that
holds, for all n.
Problem 18. Exhibit such a combinator Δ.
Problem 19. Which, if either, of the following statements is true?
(a) Δ⌈X⌉ = X⌈X⌉
(b) Δ⌈X⌉ = ⌈X⌈X⌉⌉
The normalizer can do some amazing things, as you will see!
Representability
We shall say that a combinator Γ represents a number set A if for every number
n, the equation holds iff n ∊ A. Thus Γ represents the set of all n such that
holds. (Note that the combinatorial computability of a number set A by a
combinator Γ that we worked with earlier requires more than representability, for
in the case of combinatorial computability if n ∉ A then it must be the case that
, while for representability, it is only necessary that .)
Problem 22. Suppose Γ computes A. Which, if either, of the following
statements are true?
(1) A is representable.
(2) The complement of A is representable.
Recall that for any function f(x) and set A, by f−1(A) is meant the set of all
numbers n such that f(n) ∊ A. Thus n ∊ f−1(A) iff f(n) ∊ A.
Problem 23. Prove that if the function f(x) is realizable, and if A is
representable, then f−1(A) is representable.
Gödel Sentences
Discussion
We have just seen that the complement of the set is not representable. What
about the set ? Is it representable? Yes, it is. As suggested earlier, combinatory
logic can be formalized. The following is an axiom system for the full
combinatory logic:
Symbols: S, K, (,), =
Terms: As previously defined recursively
Sentences (or Equations): Expressions of the form X = Y, where X and Y are
terms.
Axiom Schemes: For any terms X, Y, Z:
(1) SXY Z = XZ(Y Z)
(2) KXY = X
(3) X = X.
Inference Rules
R 1 : From X = Y to infer Y = X.
R 2 : From X = Y to infer XZ = YZ.
R 3 : From X = Y to infer ZX = ZY.
R 4 : From X = Y and Y = Z to infer X = Z.
It is relatively easy to construct an elementary formal system in which one
can successively represent the set of terms, sentences, and provable sentences.
Thus the set is formally representable, from which it easily follows that the set
is recursively enumerable. And it is well known that recursive enumerability
is the same thing as representability in combinatory logic. Thus the set is
indeed representable.
The set is another example of a recursively enumerable set that is not
recursive.
www.pdfgrip.com
For this purpose, we take d to be Tt, where T is the thrush [Txy = yx].
Then dxy = xty (verify!). And dpq = p ∨ q (verify all four cases!).
(d) Take the implication combinator i to be Rt, where R is the robin [Rxyz =
yzx]. Then ixy = xyt, because ixy = Rtxy = xyt. Then ipq = p ⊃ q
www.pdfgrip.com
(verify!).
(e) We want the equivalence combinator e to be such that
Then epq has the same values as p ≡ q. Verify, recalling that we have
shown above that Nt = f, that Nf = t, and that, for all x and y, txy = x and
fxy = f.
2. Suppose for some n. Thus Consequently,
Thus IK = KI, so that K = KI contrary to Problem 19
of Chapter 8.
Next, we must show that if then To do so, we instead
show the equivalent fact that if then .
Well, suppose that . Then or so that
Applying V here, we see that which implies that
[again since it is always true that fxy = y]. This proves that if
then
Now, for any positive k, we saw at the beginning of this solution that
After what we have just proved, we now see as well that it follows
from that This proves that if m ≠
n, then (because if m ≠ n then for some positive k, m = n + k or n = m
+ k).
3. Since we want a combinator P such that We take P to be
Tf, where T is the thrush [Txy = yx]. Then
Voila!
4. We take Z to be Tt, where T is the thrush [Txy = yx]. Recall that and σ =
Vf.
(1) Thus
(2) Thus
5. Yes, the zero tester Z is such an A.
www.pdfgrip.com
if for n positive
Now, by Problem 5, is if and is A(Γ(Px)) if for
n positive. We thus want a combinator Γ which satisfies the condition:
Thus the combinator we want, and which exists by the fixed point
www.pdfgrip.com
12. Let Γ compute the relation R(x, y), and let A1 realize f(x) and A2 realize
g(x). Then:
(a) BΓA1 computes the relation R(f(x),y), because:
which is t if R(n, g(m)) holds, and is f if R(n, g(m)) doesn’t hold. Thus
BCDΓ A2 computes R(x, g(y)).
(c) Let S(x, y) be the relation R(x, g(y)), which is computed by BCDΓ A2.
Then by (a) S(fx), y), which is R(f(x), g(y)) is computed by B(BCDΓ
A2)A1.
Alternatively, let δ(x, y) be the relation R(f(x), y). It is computed by
BΓ A1, according to (a). Then R(f(x), g(y)) is δ(x, g(y)), hence is
computed by BCD(BΓA1)A2, according to (b).
13. Given a regular relational combinator A of degree 2, by the fixed point
theorem, there is a combinator A1 satisfying the condition:
15. Let R(x, y) be the relation 10y > x. We must first find a combinator A that
computes the relation 10y > x.
We let R1(x, y) be the relation R(y, x), i.e. the relation 10x > y. If we find
a combinator A1 that computes the relation R1 (x, y), then CA1 will compute
www.pdfgrip.com
And so
19. It is (b) that is true. Let n be the Gödel number of X. Then so that
and since we have:
Since n is the Gödel number of X, then n#, which we first defined to be the
Gödel number of , must also be the Gödel number of since
#
Thus n*n is the Gödel number of X and so:
(2)
(c) By combining (1) and (2), we see (b), i.e. that
20. We take X to be BA where Δ is the normalizer. By the definition of
the bluebird B, we have BAΔ = A(Δ Consequently,
24. This is but a special case of Problem 23: Let f(x) = x*52. By Problem 16,
f(x) is realizable by the combinator If A is representable, so is f–1(A)
[by Problem 23], and f–1(A) is the set of all n such that f(n) ∈ A, i.e. the set
of all n such that (n*52) ∈ A.
25. Suppose A is representable. Let A′ be the set of all n such that the number
n*52 is in A. Then A′ is representable [by Problem 24]. Let Γ represent the
set A′. By the fixed point theorem, there is some X such that We
will show that the sentence X = t is a Gödel sentence for A.
Let n be the Gödel number of X. Then Hence and
therefore which is true iff n ∈ A′ [since represents A′].
And n ∈ A′ iff (n * 52) ∈ A. Thus iff (n * 52) ∈ A. But n * 52 is
the Gödel number of the sentence X = t. Thus the sentence X = t is a Gödel
sentence for A.
26. There cannot be a Gödel sentence for the complement of for such a
sentence X would be true if and only if its Gödel number n was not in
which means that X would be true iff its Gödel number was not the Gödel
number of a true sentence, and this is impossible. [Looked at it another
way, every sentence is a Gödel sentence for the set and no sentence can
be a Gödel sentence for both a set A and its complement (why?). Hence no
sentence can be a Gödel sentence for the complement of .]
Since there is no Gödel sentence for the complement of then the
complement of is not representable [by Problem 25], and therefore the set
is not computable [by Problem 22].
www.pdfgrip.com
Afterword
Where to Go from Here
This book and its predecessor have only scratched the surface of the collection
of topics subsumed under the name of “Mathematical Logic.” We have done
only the beginnings of the fields of recursion theory and combinatory logic, and
have done virtually nothing in model theory, proof theory, and other logics such
as modal logic, intuitionistic logic, relevance logic, and others.
I strongly suggest that you next turn your attention to set theory and the
continuum problem. Let me tell you a wee bit about this, hoping to whet your
appetite.
One purpose of the subject known as axiomatic set theory is to develop all
mathematics out of the notions of logic (the logical connectives and the
quantifiers) together with the notion of an element being a member of a set of
elements. We have used the symbol “∈” for “is a member of” and “x ∈ A” for “x
is a member of the set A”.
A pioneer in the development of axiomatic set theory was Gottlob Frege
[1893]. His system had, in addition to axioms of first-order logic, just one axiom
of set theory, namely that for any property P, there is a unique set consisting of
those and only those things that have the property P. Such a set is written {x :
P(x)}, which is read “the set of all x’s having the property P.
This principle of Frege is sometimes referred to as the abstraction principle,
or the unlimited abstraction principle. One can define the identity relation x = y
in terms of set inclusion as “x and y belong to the same sets”
Frege’s abstraction principle has the marvelous advantage
of allowing us to obtain just about all the sets necessary for mathematics, for
example, the following:
P1: The empty set , which is{x :∼ (x = x)}.
P2: For any two elements a and b, the set {a, b} whose members are just a and b.
www.pdfgrip.com
P4: For any set a, the union ∪a, i.e. the set of all elements that are members of
an element of a, thus ∪a = {x: ∃y(y ∈ a ∧ x ∈ y)}.
Despite the useful things Frege’s system can do, it has an extremely serious
drawback: it is inconsistent! This was pointed out to Frege in a letter by Bertrand
Russell [1902], who observed that according to Frege’s abstraction principle,
there would exist the set A of all sets which are not members of themselves (A =
{x: x ∉ x}). Thus, for any set x, one would have x ∈ A iff x ∉ x Taking A for x,
we would have A ∈ A iff A ∉ A, which is a contradiction! Thus Frege’s system
is inconsistent.
Frege was broken-hearted over Russell’s discovery, and felt that his whole
life’s work had been in vain. Here he was wrong: His work was salvaged by
Zermelo and others, and is really the basis for Zermelo–Fraenkel set theory
(frequently abbreviated as ZF), which is one of the most significant systems of
set theorem in existence, the other of similar rank being the system in Principia
Mathematica of Whitehead and Russell [1910].
What Zermelo [1908] did was to replace Frege’s abstraction principle by the
following, known as the limited abstraction principle, which is that for any
property P of sets and any set a, there exists the set of all the elements of a
having property P. Thus for any set a, there exists the set {x: P(x)∧ x ∈a}. This
principle appears to be free of any possible contradiction. However, Zermelo had
to take the existence of the sets {a,b}, and ∪a as separate axioms. He also
took an additional axiom known as the axiom of infinity, which we will discuss
later.
On the basis of the Zermelo axioms on hand, one can now derive the positive
integers. First, let us note that the set {a, b} is well-defined even if a and b are
the same element, and then {a, a} is simply the set {a}. Well, Zermelo took 0 to
be the empty set ; 1 to be the set whose only element is 0; 2 to be the
set whose only element is 1, and so forth. Thus the Zermelo natural numbers are
etc. Thus for each natural number n, n+1 is{n}.
Later, Von Neumann [1923] proceeded differently. In his scheme, each
natural number is the set of all lesser natural numbers. Thus 0 is . and 1 is { }
www.pdfgrip.com
as with Zermelo, but 2 is now not {1}, but rather the set {0, 1}; 3 is the set {0, 1,
2}; …, n+1 is the set {0, 1, …, n}. This is the scheme now generally adopted,
since it generalizes in the infinite case to the important sets called ordinals,
which we will soon discuss. Note that in Zermelo’s scheme, each natural number
other than 0 contains just one element, whereas in Von Neumann’s scheme, each
number n contains exactly n elements.
Now that we have the natural numbers, how do we define what it means to be
a natural number? That is, we want a formula N(x) using just the logical
connectives and quantifiers and the single predicate symbol ∈ such that for any
set a, the sentence N(a) is true if and only if a is a natural number. Well, for any
set a, we define a+ as a ∪ {a}, i.e. the set whose elements are those of a,
together with a itself. Thus
We now define a set a to be inductive if 0 ∈ a, and if, for all sets b, if b ∈ a, then
b+ a. We then define a set x to be a natural number if it belongs to all inductive
sets. The principle of mathematical induction follows immediately from the very
definition of a natural number. Indeed, the five Peano Postulates follow from the
definition of “natural number”.
However, we don’t yet have any guarantee that there is such a thing as the set
of all natural numbers. Zermelo had to take this as an additional axiom, and this
is known as the axiom of infinity. The set of all natural numbers is denoted “ω”.
Let us recall the Peano Postulates:
1. 0 is a natural number.
2. If n is a natural number, so is n+.
3. There is no natural number such that n+ = 0.
4. If n+ = m+, then n = m.
5. [Principle of Mathematical Induction] Every inductive set contains all the
natural numbers.
All these five postulates are easy consequences of the definition of natural
number. The axiom of infinity is not necessary to prove the Peano Postulates!
This fact appears to not be completely well known. I recall that when I once
announced these facts at a lecture, a very eminent logician later told me that he
www.pdfgrip.com
After that finally come ω·4, ω·5, …, ω · ω, … ωω, …. We can keep going.
The ordinals so far considered are all denumerable. But there exist non-
denumerable ordinals as well. Indeed, with the axiom of
substitution/replacement (of Zermelo–Fraenkel set theory), for every set x, there
is an ordinal α of the same size as x (i.e. α can be put into 1-1 correspondence
www.pdfgrip.com
with x).
An ordinal is called a successor ordinal if it is α+[α∪{α}] for some ordinal α.
Ordinals other than 0 and successor ordinals are called limit ordinals. All natural
numbers other than 0 are successor ordinals. The first limit ordinal is ω. The
ordinals ω + 1, ω + 2, …, ω+n are all successor ordinals, and ω·2 (the set of all
of them) is the next limit ordinal. The ordinals ω·3, ω·4, …, ω· n are all limit
ordinals, and so is ω·ω. Actually, an infinite ordinal is a successor ordinal if and
only if it is of the form +n, for some limit ordinal α and some positive integer n.
An ordinal α is said to be less than an ordinal β(in symbols, α < β) if α ∈ β.
Thus each ordinal is the collection of all lesser ordinals.
Rank
Given a scheme that assigns to each ordinal α a set denoted S, for any ordinal α,
by Uα<λ Sα is meant the union of all the sets S, where α < λ. By an important
result known as the transfinite recursion theorem, one can assign to each ordinal
α a set Rα such that for every ordinal α and every limit ordinal λ, the following
three conditions hold:
(1)
(2) Rα+ = [the set of all subsets of Rα].
(3) Rα = Uα<λ Rλ.
A set is said to have rank if it is a member of some Rα, and by the rank of
such a set x is meant the least ordinal α such that x ∈ Rα.
Do all sets have rank? This is so iff the ’ relation is well founded, i.e. iff there
exists no infinite sequence x1, x2, … xn, … such that for each n, the set xn+1 is a
member of xn. This condition is taken as another axiom of set theory, and is
known as the axiom of foundation. It rules out sets without rank.
The Axiom of Choice (frequently abbreviated AC) in one form is that for any
non-empty set a of non-empty sets, there is a function C (a so-called choice
function for a) that assigns to each member x of a an element of x. [The function,
www.pdfgrip.com
Constructible Sets
Consider a sentence X whose constants are all in a set a. X is said to be true over
a if it is true when the quantifiers are interpreted as ranging over all elements of
a, i.e. it is true when is read as “for all x in a” and “∃x” is read as “for some
x in a”. A formula φ(x) whose constants are all in a is said to define over a the
set of all elements k such that φ(k) is true over a, and a subset b of a is called
definable over a (more completely, first-order definable over a) if it is defined
over a by some formula whose constants are all in a.
www.pdfgrip.com
The definition of the Mα’s differs from that of the Rα’s only in condition (2):
Whereas Rα+1 consists of all subsets of Rα+1, the set Mα+1 consists of only those
subsets of Mα that are definable over Mα.
It is not enough to say it is undecidable in ZF. The fact remains that for every
infinite set a, either there is a set b intermediate in size between a and or
there isn’t, and we want to know which. As already, indicated, Gödel
conjectured that there is such a set b for some infinite set a, and that when more
is known about sets, we will see that Cantor’s conjecture is wrong. [Realize that
for any non-empty finite set a, there are always sets b with more elements than a
and less elements than .]
The whole study of the independence of GCH in ZF is fascinating beyond
belief, and the reader who studies this is in for a real treat!
www.pdfgrip.com
References
Barendregt, Henk, The Lambda Calculus — Its Syntax and Semantics, Studies in
Logic and the Foundations of Mathematics Vol. 103, North-Holland, 1985.
Beth, Evert, The Foundations of Mathematics. A Study in the Philosophy of
Science, North-Holland, 1959.
Braithwaite, Reginald, Kestrels, Quirky Birds and Hopeless Egocentricity, ebook
published by http://leanpub.com, 2013.
Church, Alonzo, The Calculi of Lambda Conversion, Princeton Univ. Press,
1941.
Cohen, Paul J., “The independence of the continuum hypothesis,” Proceedings
of the National Academy of Sciences of the United States of America 50(6):
1143–1148, 1963.
Cohen, Paul J., “The independence of the continuum hypothesis, II,”
Proceedings of the National Academy of Sciences of the United States of
America 51(1): 105–110, 1964.
Craig, William, “Three uses of the Herbrand–Gentzen theorem in relating model
theory and proof theory,” The Journal of Symbolic Logic 22(3): 269–285,
1957.
Davis, Martin, Computability and Unsolvability, McGraw-Hill, 1958; Dover,
1982.
Ehrenfeucht, Andrzej and Feferman, Solomon, “Representability of recursively
enumerable sets in formal theories,” Arch. Fur Math. Logick und
Grundlagenforschung 5: 37–41, 1960.
Fraenkel, Abraham Bar-Hillel, Yehoshua and Levy Azriel, Foundations of Set
Theory, North-Holland, 1973 (originally published in 1958). (Fraenkel’s
final word on ZF and ZFC, according to Wikipedia.) Frege, Gottlob,
Grundgesetze der Arithmetic, Verlag Hermann Pohle, Vol. I/II, 1893.
Partial translation of volume I, The Basic Laws of Arithmetic, by M. Furth,
Univ. of California Press, 1964.
Gentzen, Gerhard, “Untersuchungen über das logische Schliessen I,”
Mathematische Zeitschrift 39(2): 176–210, 1934.
Gentzen, Gerhard, “Untersuchungen über das logische Schliessen II,”
www.pdfgrip.com
Index
Ê combinator, 186
E combinator (eagle), 186
E-complete set of sentences, 54–56
eagle combinator E, 186
effective (and recursively enumerable) sequence of simple systems, each of
which is an extension of the preceding system, 159
www.pdfgrip.com
n combinator, 208
N combinator (nice combinator), 208
n-tupling function δn (x1, …, xn), 98
narcissistic element, in combinatory logic, 176
natural number (numerical) combinators, 234
neg(x), requisite function of a simple system, 149
negated form of an iff statement, 7, 8
negation H′ of a sentence H in a mathematical system , 74
negation of Generalized Continuum Hypothesis (GCH) is consistent with ZF
(Cohen), 258, 260
nice combinator N, 208
non-cancellative (λ-I) combinator, 216
non-denumerable ordinal, 257
non-standard formula, 44
norm of a string of symbols of the alphabet of the universal system U, 102
norm of an S, K term, 241
normal simple system, 153
normalizer of a natural number combinator, 241
numerical (natural number) combinators, 234, 235
numerical relation A(x, y) in a synthesis system, 79
numerical relation B(x, y) in a synthesis system, 79
www.pdfgrip.com
proof of the Compactness Theorem for first-order logic using magic sets, 28
proof of the completeness of an axiom system for first-order logic by use of the
Unification Theorem, 52
proof of the completeness of propositional logic based on maximal consistency,
7–15
proof of the completeness of standard axiom systems for first-order logic by
Henkin’s method, 54–56
proof of the completeness of the method of analytic tableaux by use of the
Unification Theorem, 51
proof of the efficacy of the method of eliminates required to obtain a combinator
derived from S and K, 220, 221
proof of the efficacy of the method of eliminates required to obtain a combinator
derived from S, B, C and I, 225, 226
proof of the Regularity Theorem by use of the Unification Theorem, 52, 53
proof of the Regularity Theorem using magic sets, 28
proof of the Skolem–Löwenheim Theorem by use of the Unification Theorem,
51
proof of the Skolem–Löwenheim Theorem using magic sets, 28
proof of the Unification Theorem by means other than tableaux or König’s
Lemma, 50, 51
proof of the Unification Theorem using tableaux, 50
proof that the set of Gödel numbers of trecursively separable pair of setshe true
sentences of U is not recursive, 102, 103
proof that the set of true sentences T of the universal system U is formally
representable and that the set T0 is recursively enumerable, 101, 102
property combinator, 236
propositional logic, 3–22, 35–38
propositional logic combinators, 233, 234
provability predicate in a mathematical system , 74
provable number of a simple system, 150
provable sentence of a mathematical system , 73
provable sentence of an elementary formal system, 90
punctuation sign of an elementary formal system, 89
pure sentence of first-order logic, 25
Putnam, Hillary, 157, 160
Putnam–Smullyan Theorem, 160
s and K as constants from which all other combinators can be derived, 215–223
S, B, C, and I are a basis for all λ-I combinators, 224–226
S, K term, 239
sage constructions, 205–208
sage element, in a combinatory logic, 179
satisfaction of a set of sentences by a valuation, 25
satisfiability of first-order sentences in the domain of the parameters, 24
Schöfinkel, Moses, 173
Second Fixed Point Theorem, for combinators, 242
semi-doubly universal ordered pair of sets, 135, 138
semi-reducibility of one ordered pair of sets to another ordered pair of sets, 135,
138, 139
sentence of a mathematical system , 73
sentence of an elementary formal system, 90
sentence of the universal system U, 100
separability of a pair of R-sets, 78
separation of a disjoint pair of number sets in a simple system, 153
separation of two binary relations by a formula in a first-order system, 160
Separation Principle, 136, 138
sequent (Gentzen), 29–46
sequent corresponding to a set of signed formulas, 33
set A in synthesis system, 79
set B in synthesis system, 79
set operations, similarity with logical connectives, 3, 4
Shepherdson’s Theorem, 161
Shepherdson, John, 161
similar registers, 70
Simple Iteration Theorem, 115, 116
simple system, 149–161
simple system associated with a system of first-order arithmetic, 150
simple term, in an S, K combinatorial system, 218
skew diagonalization of a number in a simple system, 151
Skolem–Löwenheim Theorem, 23, 27, 28, 49, 51
smug element, in a combinatory logic, 175
www.pdfgrip.com