sbook
sbook
Non-Classical
Anita Wasilewska
January 9, 2021
2
For my daughter Agatha
Light of my life
4
Contents
5
3.5 Many Valued Semantics: Lukasiewicz, Heyting, Kleene, Bohvar . 117
3.6 M Tautologies, M Consistency, and M Equivalence of Languages 135
3.6.1 M Tautologies Verification Methods . . . . . . . . . . . . 135
3.6.2 M Consistency . . . . . . . . . . . . . . . . . . . . . . . . 137
3.6.3 M Equivalence of Languages . . . . . . . . . . . . . . . . 144
3.7 Homework Problems . . . . . . . . . . . . . . . . . . . . . . . . . 146
6
6.5 Gentzen Sequent Systems GL, G, LK . . . . . . . . . . . . . . . 273
6.5.1 Gentzen Sequent Systems GL and G . . . . . . . . . . . 273
6.6 GL Soundness and Completeness . . . . . . . . . . . . . . . . . . 287
6.7 Original Gentzen Systems LK, LI
Completeness and Hauptzatz Theorems . . . . . . . . . . . . . . 293
6.8 Homework Problems . . . . . . . . . . . . . . . . . . . . . . . . . 312
7
Completeness of Classical Predicate Logic 411
9.1 Reduction Predicate Logic to Propositional Logic . . . . . . . . . 412
9.1.1 Henkin Method . . . . . . . . . . . . . . . . . . . . . . . . 418
9.2 Proof of Completeness Theorem . . . . . . . . . . . . . . . . . . . 429
9.3 Deduction Theorem . . . . . . . . . . . . . . . . . . . . . . . . . 440
9.4 Some other Axiomatizations . . . . . . . . . . . . . . . . . . . . . 445
9.5 Homework Problems . . . . . . . . . . . . . . . . . . . . . . . . . 450
8
Chapter 1
Introduction: Paradoxes
and Puzzels
Until recently, till the end of the 19th century, mathematical theories used to be
built in an intuitive or axiomatic way. In other words, they were based either
intuitive ideas concerning basic notions of the theory - ideas taken from the
reality - or on the properties of these notions expressed in systems of axioms.
The historical development of mathematics has shown that it is not sufficient
to base theories on an intuitive understanding of their notions only. This fact
became especially obvious in set theory. The basic concept of this theory, set, is
certainly taken from reality, for there we come across many examples of various
sets, all of which are finite. But in mathematics it is also necessary to consider
infinite sets, such as the set of all integers, the set of all rational numbers, the
set of all segments, the set of all triangles.
By a set, we mean intuitively, any collection of objects. For example, the set of
all even integers or the set of all students in a class. The objects that make up a
set are called its members (elements). Sets may themselves be members of sets
for example, the set of all sets of integers has sets as its members. Most sets are
not members of themselves; the set of all students, for example, is not a member
of itself, because the set of all students is not a student. However, there may
be sets that do belong to themselves - for example, the set of all sets. However,
a simple reasoning indicates that it is necessary to impose some limitations on
the concept of a set.
Russell, 1902 Consider the set A of all those sets X such that X is not a
member of X. Clearly, by definition, A is a member of A if and only if
9
A is not a member of A. So, if A is a member of A, the A is also not a
member of A; and if A is not a member of A, then A is a member of A.
In any case, A is a member of A and A is not a member of A.
This paradox arises because the concept of set was not precisely defined and
was too liberally interpreted. Russell noted the self-reference present in his
paradox (and other paradoxes, two of them stated below) and suggested that
every object must have a definite non-negative integer as its type. Then an
expression x is a member of the set y is meaningful if and only if the type of
y is one greater than the type of x. So, according to the theory of types, it is
meaningless to say that a set belongs to itself, there can not be such set A, as
stated in the Russell paradox.
The paradoxes concerning the notion of a set are called logical paradoxes (anti-
nomies). Two of the most known (besides the Russell’s) logical paradoxes are
Cantor and Burali-Forti antinomies. Both were stated at the end of 19th cen-
tury. The Cantor paradox involves the theory of cardinal numbers, Burali-Forti
paradox is the analogue to Cantor’s in the theory of ordinal numbers. They will
make real sense only to those already familiar with both of the theories, but we
will state them here because they do have an intuitive meaning and had very
important consequences.
The cardinal number cardX of a setX intuitively corresponds, in a case of finite
sets, to a number of elements of the set X. cardX is formally defined to be
the set of all sets Y that are equinumerous with X ( i.e., for which there a
one-to-one correspondence between X and Y ). We define cardX ≤ cardY to
mean that X is equinumerous with a subset of Y ; by cardX < cardY we mean
cardX ≤ cardY and cardX 6= cardY . Cantor proved that if P(X) is the set
of all subsets of X, then cardX < cardP(X). The cardinal numbers behave
similarly to natural numbers in many cases, in particular Schröder- Berstein
proved that if cardX ≤ cardY and cardY ≤ cardX, then cardX = cardY .
The ordinal numbers are the numbers assigned to sets in a similar way as cardinal
numbers but they deal with ordered sets.
Cantor, 1899 Let C be the universal set - that is, the set of all sets. Now, P(C)
is a subset of C, so it follows easily that cardP(C) ≤ cardC. On the other
hand, by Cantor theorem, cardC < cardP(C), so also cardC ≤ cardP(C)
and by Schröder- Berstein theorem we have that cardP(C) = cardC, what
contradicts cardC < cardP(C).
Burali-Forti, 1897 Given any ordinal number, there is a still larger ordinal
number. But the ordinal number determined by the set of all ordinal
numbers is the largest ordinal number.
10
is successful, but difficult in practice and has certain other drawbacks as well.
A different criticism of the logical paradoxes is aimed at their assumption that
the notion of a set is defined in such a way that, for every property P (x),
there exists a corresponding set of all objects x that satisfy P (x). If we reject
this assumption, then the logical paradoxes are no longer derivable. Russell’s
Paradox then simply proves that there is no set A of all sets that do not belong
to themselves; the paradoxes of Cantor and Burali-Forti show that there is no
universal set and no set that contains all ordinal numbers.
It became obvious that the paradoxes described above, as well as many similar
ones occurring in intuitive set theory cannot be avoided by referring to intuition
only. The solution looked for was to characterize the intuitive concept of set by
a suitable set of axioms. If we do so, we obtain an axiomatic set theory without
such antinomies.
The first such axiomatic set theory was invented by Zermello in 1908. However,
in 1922 Fraenkel pointed out some shortcomings of Zermello’s axiomatizations
and proposed improvements. The result is called Zermello-Frankel set theory
ZF, or ZFC, where C stands for axiom of choice. Today ZFC is the standard
form axiomatic set theory. ZFC is considered the most common foundation of
mathematics sufficient to include all actual mathematical theories.
A more radical interpretation of the paradoxes has been advocated by Brouwer
and his intuitionist school. They refuse to accept the universality of certain
basic logical laws, such as the law of excluded middle: A or not A. Such a law,
they claim, is true for finite sets, but it is invalid to extend it to all sets. It
means that the intuitionists’ view of the concept of infinite set differs from that
of most mathematicians. Intuitionists reject the idea of infinite set as a closed
whole. They look upon an infinite set as something which is constantly in a
state of formation. It means that, for example, the set of all positive integers
is not looked upon as a closed whole. It is infinite in a sense that to any given
finite set of positive integers it is always possible to add one more positive
integer. The notion of the set of all subsets of the set of all positive integers is
not regarded meaningful. Obviously, intuitionists’ view-point on the meaning
of the basic logical and set-theoretical concepts used in mathematics is different
from that of most mathematicians in their research. The basic difference lies
in the interpretation of the word exists. For example, let P (n) be a statement
in the arithmetic of positive integers. For the mathematicians the sentence the
exists n, such that P(n) is true if it can be deduced (proved) from the axioms
of arithmetic by means of classical logic. If the mathematician proves it, this
does not mean that he is able to indicate a method of construction of a positive
integer n such that P (n) holds. On the contrary, for the intuitionist the sentence
the exists n, such that P(n) is true only if he is able to construct a number n
such that P (n) is true. In the intuitionists’ universe we are justified in asserting
the existence of an object having a certain property only if we know an effective
method for constructing or finding such an object. The paradoxes are, in this
case, not derivable (or even meaningful), but so are many theorems of everyday
11
mathematics, and for this reason, intuitionism has found few converts among
mathematicians. But, because of its constructive flavor, it has found some
applications in computer science and this is the reason to discuss some of it
here. An exact exposition of the basic ideas of intuitionism is outside the range
of this book, but we will study intuitionists logic, which is a sort of reflection of
intuitionists ideas formulated in formalized deductive system.
As we can see, the axiomatic method was the first step leading to greater pre-
cision in the construction of mathematical theories. In intuitive mathematical
theories the borderline between that which is obvious and that which requires
proof is not exact. In axiomatic theories a system of primitive notions is as-
sumed which are characterized by a set of axioms. Other notions can be defined
by means of the primitive notions. All statements which are consequences of
the axioms are called theorems of the theory. All properties of any notion of
the theory which are not expressed in the axioms, require proof.
For some time this degree of exactness in the construction of theories seemed
sufficient. However, it turned out that the assumption of a consistent set of
axioms does not prevent the occurrence of another kind of paradoxes, called
semantic paradoxes.
For instance, let us consider the arithmetic based on the well known system
of axioms due to Peano (to be discussed in chapter 11) and let’s consider the
following simple reasoning.
Berry, 1906 Let A denote the set of all positive integers which can be defined
in the English language by means of a sentence containing at most 1000
letters. The set A is finite since the set of all sentences containing at most
1000 letters is finite. Hence, there exist positive integers which do not
belong to A. The sentence:
n is the least positive integer which cannot be defined by means of a sen-
tence of the English language containing at most 1000 letters
contains less than 1000 letters and defines a positive integer n. Therefore n
belongs to A. On the other hand, n does not belong to A by the definition
of n. This contradicts the first statement.
It is obvious that the reason for this paradox is that in its construction we used
some notions (e.g the English language, letters, sentences) which do not belong
to pure arithmetic. Usually we do not introduce definitions like the above in
mathematics. The paradox resulted entirely from the fact that we did not say
precisely what notions and sentences belong to the arithmetic and what notions
and sentences concern the arithmetic, examined as a fix and closed deductive
system. Intuitively we conceive the arithmetic to be a set of sentences expressing
certain properties of positive integers and of other notions defined by means of
the notion of integer. For example, a sentence: for every integer n, n2 ≤ 0
certainly belongs to arithmetic.
12
On the other hand we can also talk about the arithmetic. That is, assuming
that all sentences in the arithmetic are formulated in the English language, we
can formulate sentences concerning the problem how many integers we defined
in the arithmetic by means of at most 1000 letters. However, such sentences
about the arithmetic do not belong to arithmetic. They belong to another
theory, which examines the arithmetic as a new subject of investigation. This
theory is usually called meta-arithmetic. In particular, the Berry sentence does
not belong to arithmetic; it belongs to meta-arithmetic and the paradox arises
from the luck of distinction between a theory (language) and its meta-theory
(metalanguage).
For a similar reason in well defined theory the following paradoxes can not
appear.
Löb, 1955 Let A be any sentence. Let B be a sentence: If this sentence is true,
then A. So, B asserts: If B is true then A. Now consider the following
argument: Assume B is true. Then, by B, since B is true, A is true. This
argument shows that, if B is true, then A. But this is exactly what B
asserts. Hence, B is true. Therefore, by B, since B is true, A is true. Thus
every sentence is true.
In these cases the paradox arises because the concepts of ” I am true”, ” this
sentence is true, ” I am lying”” should not occur in the language (theory). It
belong to a metalanguage (meta-theory).
The Liar Paradox is a corrected version of a following paradox stated in antiquity
by a Cretan philosopher Epimenides, 600 BC.
Epimenides, a Cretan said: All Cretans are liars. If what he said is true, then,
since Epimenides is a Cretan, it must be false. Hence, what he said is
false. Thus, there must be some Cretan who is not a liar.
Note that the conclusion that there must be some Cretan who is not a liar is not
logically impossible, so we do not have a genuine paradox. However, the fact
that the utterance by Epimenides of the false sentence could imply the existence
of some Cretan who is not a liar is rather unsettling.
It follows from above semantic paradoxes that in order to exclude them from an
axiomatic theory, it is necessary to describe its language precisely, i.e. the set
of sentences of the theory and the set of signs used to build these sentences. In
13
this way we avoid contradictions caused by a collision between the theory and
meta-theory, that is, by including meta-theoretical statements in the theory.
This inclines us to introduce still greater precision in the construction of math-
ematical (and others) theories and leads to the concept of formalized theories,
in which not only the properties of primitive notions are given in an exact way
by means of a set of axioms, but also the language of the theory is precisely
defined. The formalization of the language also gives the following advantage:
it permits us to describe precisely the logical means assumed in the theory, i.e.
to define the process of deduction precisely.
In formalized mathematical theories, e.g. in formalized arithmetic and formal-
ized set theory, the paradoxes as those mentioned above cannot be constructed.
On the other hand, a mathematician (or a computer scientist) following good in-
tuitions in every -day investigations does not reach a contradiction even though
the language of the theory and the logical means he employs are not precisely
described. This is explained by the fact that his investigations can always in
practice be repeated in a corresponding formalized theory. Thus he avoids
practical difficulties of formalized theories, the formalized language of of which
is complicated and very inconvenient in every day practice. Consequently, in
mathematical practice we build theories axiomatically but always in such a way
that they can be easily formalized, i.e. that all the reasonings can be repeated in
a corresponding formalized theory. However, the formalization of the language
and the logical means are necessary if we intend to develop the meta-theory of
a given practical theory because only in this way such important notions as the
existence of a proof of a given statement or the set of all theorems of the theory
can be precisely defined. In practical, non-formalized axiomatic theories those
notions are far from this precision.
Whatever approach one takes to the paradoxes, it is always necessary first to
examine the language of logic and mathematics or given domain of computer
science, to see what symbols may be used, to determine the way ways in which
these symbols are put together to form terms, formulas, sentences, and proofs,
and to find out what can and cannot be proved if certain axioms and rules of
inference are assumed.
This is the basic task of mathematical logic, and, until it is done, there is no
basis of talking about foundations of logic, mathematics or computer science.
This approach is already almost a century old - the first formalized theory was
built by Frege in 1891. The deep and devastating results of Gödel, Tarski,
Church, Rosser, Kleene and many others followed. They created what is called
a modern mathematical logic and have earned for it its status as an independent
branch of mathematics.
14
1.2 Computer Science Puzzles
Logical and semantical paradoxes have lead the mathematicians to the devel-
opment of a modern classical logic as an independent domain of mathematics.
They have also, as we could see, led to the development of the intuitionistic logic
as rival to the classical one. The classical and intuitionistic logic differ on the
interpretation of the meaning of the word exists, but also, as we will see later,
in the interpretation of logical implication, i.e. the truthfulness of the sentences
of the form if A then B is decided differently in both logics.
In 1918, an American philosopher, C.I. Lewis proposed yet another interpreta-
tion of lasting consequences, of the logical implication. In an attempt to avoid,
what some felt, the paradoxes of implication (a false sentence implies any sen-
tence) he created a modal logic. The idea was to distinguish two sorts of truth:
necessary truth and mere possible (contingent) truth. A possibly true sentence
is one which, though true, could be false.
More recently, modal logic has become much-used in computer science as a tool
for analyzing such notions as knowledge, belief, tense.
The logics other than the classical propositional or predicate logic are usually
called non-standard logics. The use of classical logic in computer science is
known, indisputable, and well established.The existence of PROLOG and Logic
Programming as a separate field of computer science is the best example of
it. But the non-standard logics have been imported into a great many areas
of computer science and, in particular into the research about the specification
and verification of programs, the analysis of behavior of distributed systems and
into almost all areas of artificial intelligence. Even in Logic Programming, once
we start to talk about logic programming programs we run immediately into
some non-standard logics.
Modal logic, for example, has been employed in form of Dynamic logic (Harel
1979) to facilitate the statement and proof of properties of programs.
Temporal Logics were created for the specification and verification of concurrent
programs Harel, Parikh, 1979, 1983), for a specification of hardware circuits
Halpern, Manna and Maszkowski, (1983), to specify and clarify the concept of
causation and its role in commonsense reasoning (Shoham, 1988).
Intuitionistic logic, in the form of Martin-Löf’s theory of types (1982), provides
a complete theory of the process of program specification, construction, and
verification. A similar theme has been developed by Constable (1971) and
Beeson (1983).
The great strength of dynamic and temporal logics relates to their expressive
power. In such systems it is possible to express properties of programs in an
elegant and natural way. This is in large part due to enriched language of such
logics over the classical predicate calculus. In the case of intuitionistic logic
15
the motivation for their employment, as it was mentioned before, is different.
The proponents of intuitionistic logic and mathematics claim that constructive
mathematics is, generally, a more appropriate framework for computer science
than classical logic and mathematics.
Fuzzy logic, Many valued logics were created and developed to reasoning with
incomplete information. Most expert systems are forced to take decisions when
not all the facts pertaining to decision are available. In such context it is natural
to employ logics which, unlike classical logic, are suited to reasoning with such
incomplete information.
The development of different logics and the applications of logic to different
areas of computer science or even artificial intelligence only is beyond the scope
of our investigations.
We present some of the most known motivations (computer science puzzles),
which played a similar role in the development of the reasoning about knowl-
edge in distributed systems and artificial intelligence, as logical and semantical
paradoxes played in the development of logic and foundations of mathematics.
Grey, 1978. Halpern, Moses, 1984 Two divisions of an army are camped
on two hilltops overlooking a common valley. In the valley awaits the
enemy. It is clear that if both divisions attack the enemy simultaneously
they will win the battle, whereas if only one division attacks it will be
defeated. The divisions do not initially have plans for launching an attack
on the enemy, and the commanding general of the first division wishes to
coordinate a simultaneous attack (at some time the next day). Neither
16
general will decide to attack unless he is sure that the other will attack
with him. The generals can only communicate by means of a messenger.
Normally, it takes a messenger one hour to get from one encampment
to the other. However, it is possible that he will get lost in the dark or,
worst yet, be captured by the enemy. Fortunately, on this particular night,
everything goes smoothly. How long will it take them to coordinate an
attack?
17
if necessary in the face of new evidence. If our computer programs are to act
intelligently, they will need to be similarly flexible.
A large portion of the work in artificial intelligence (AI) on reasoning or deduc-
tion involves the development of formal systems that describe this process.
The most usual example of such a flexible inference is the following flying birds
example.
Reiter, 1987 Consider a statement Birds fly. Tweety, we are told, is a bird.
From this, and the fact that birds fly, we conclude that Tweety can fly.
This conclusion, however is defeasible: Tweety may be an ostrich, a pen-
guin, a bird with a broken wing, or a bird whose feet have been set in
concrete.
The inference here is non-monotonic: on learning a new fact (that Tweety has
a broken wing), you are forced to retract your conclusion that he could fly. This
original conclusion didn’t follow logically (in a sense if classical logic) from the
facts that birds typically fly and that Tweety is a bird; it followed from these
facts together with the assumption that Tweety is a typical bird. When we learn
more about Tweety, we may discover that this assumption is unjustified.
It means, by a non-monotonic reasoning (logics) we understand reasonings (log-
ics) in which the introduction of a new information (facts) can invalidate old
theorems.
The inference described here is also called a default reasoning.
It means, by default reasoning we mean the drawing of plausible inferences from
less-then-conclusive evidence in the absence of information to the contrary.
Consider now the following example.
Moore, 1983 Consider my reason for believing that I do not have an older
brother. It is surely not that one of my parents once casually remarked,
You know, you don’t have any older brothers, nor have I pieced it together
by carefully sifting other evidence. I simply believe that if I did have an
older brother I would know about it; therefore, since I don’t know of any
older brothers of mine, I must not have any.
This type of reasoning is not a form of default reasoning at all; it rather seems
to be more like reasoning about one’s own knowledge or belief. Hence it is
called an auto-epistemic reasoning, i.e. the reasoning about one’s own beliefs or
knowledge.
The auto-epistemic reasoning is intended to model the reasoning of an ideally
rational agent reflecting upon his beliefs or knowledge. Logics which describe it
are called auto-epistemic logics.
18
In addition to application to the understanding of common-sense reasoning, non-
monotonic reasoning (non-monotonic logics) has been shown to be important in
other areas. There are applications to logic programming, to planning and rea-
soning about action, and to automated diagnosis. As the formal work matures,
increasing effort is being devoted to applying the improved understanding to
the solution of practical problems.
We end this introduction by McCarthy discussion of a much used in AI puzzle
Missionaries and Cannibals, as a proof of a need of another ”stretch ” from
classical logic.
Three missionaries and three cannibals come to the river. A rowboat that
seats two is available. If the cannibals ever outnumber the mission-
aries on either bank of the river, the missionaries will be eaten. How
shall they cross the river?
19
What bridge? you say. No bridge is mentioned in the statement of the
problem. And this dunce replies, Well, they don’t say there isn’t a bridge.
You look at the English and even at the translation of the English into
first order logic, and you must admit that they don’t say there is no bridge.
So you modify the problem to exclude the bridges and pose it again, and
the dunce proposes a helicopter, and after you exclude that, he proposes
a winged horse or that the others hang onto the outside of the boat while
two row.
You now see that while a dunce, he is an inventive dunce. Despairing of
getting him to accept the problem in the proper puzzler’s spirit, you tel
him the solution. To your further annoyance, he attacks your solution on
the grounds that the boat might have a leak or lack oars. After you rectify
that omission from the statement of the problem, he suggests that a sea
monster may swim up the river and may swallow the boat. Again you are
frustrated, and you look for a mode of reasoning that will settle his hash
once and for all.
20
7. Give an example of some non-standard logics.
8. Describe a difference between classical and intuitionistic logic.
9. Give two examples of Computer Science Puzzles.
10. What a default reasoning is? Give an example.
21
22
Chapter 2
Introduction to Classical
Logic
.
Logic builds symbolic models of our world. It builds them in such a way as to
be able to describe formally the ways we reason in and about it. It also poses
questions about correctness of such models and develops tools to answer them.
Classical Logic was created to describe the reasoning principles of mathematics
and hence reflects the ”black” and ”white” qualities of mathematics; we expect
from mathematical theorems to be always either true or false and the reasonings
leading to them should guarantee this without any ambiguity. It hence admits
only two logical values and is sometimes called a two-valued logic.
The models we build are based on a principle that the language in which we
reason uses sentences. These sentences are built up from basic assertions about
the world using special words or phrases like ”not”, ”not true” ”and”, ”or”, ”
implies”, ”if ..... then”, ”from the fact that .... we can deduce”, ” if and only
if”, ”equivalent”, ”every”, ”for all”, ”any”, ”some”,” exists”. Basically, it is the
behavior of these words we want to study. Most of these words and phrases
have accepted intuitive meanings and we want our models to formalize these
meanings. To do so we first define a notion of a symbolic language and then
define a formal meaning of its symbols, called semantics.
We use symbols: ¬, for ”not”, ”not true”, ∩ for ”and”, ∪ for ”or”, ⇒ for ”
implies” , ”if ..... then”, ”from the fact that... we can deduce”, and a symbol
⇔ for ” if and only if”, ”equivalent”. We call these symbols propositional con-
nectives. There are other symbols for propositional connectives and there are
other propositional connectives as well that we will introduce later.
We use symbols: a, b, c, p, r, q, . . . , with indices, if necessary to represent the
23
basic assertions, called propositions. Hence we call the symbols a, b, c, p, r, q, . . .
propositional variables.
We use symbols: ∀ for ”every”, ”any”, and ∃ for ”some” ,” exists”. The symbols
∀, ∃ are called quantifiers.
Restricting our attention to the role of propositional connectives yields to what is
called propositional logic with the a propositional language and a propositional
semantics as its basic components. This is a quite simple model to justify,
describe and develop and we will devote first few chapters to it. We do it both
for its own sake, and because it provides a good background for developing and
understanding more difficult logics to follow.
Consideration and study of the role of propositional connectives and quantifiers
leads to what is called a predicate logic with its predicate language and semantics.
This is a much more complicated model and we will develop and study it in full
formality in chapters following the introduction and examination of the formal
propositional logic model.
In this chapter we provide motivation for and description of both propositional
and predicate languages, and discuss their semantics.
24
variables that represent the the basic assertions called propositions. Histori-
cally, we define propositions as basic, declarative sentences (assertions) that can
always be evaluated as true or false. For example, a statement: ” 2 + 2 = 4”
is a proposition as we assume that it is a well known and agreed upon truth.
A statement: ”2 + 2 = 5” is also a classical proposition (false). A statement:
2 + n = 5 according to the historical definition is not a proposition; it might
be true for some n, for example n=3, false for other n, for example n= 2, and
moreover, we don’t know what n is. Sentences of this kind are called proposi-
tional functions. We treat propositional functions within propositional model
as propositions and represent them by the propositional variables.
Similar examples can be found in natural language rather then in mathematical
language. For example we tend to accept a statement: ”The earth circulates the
sun” as a proposition while a statement: ”Ann is pretty”, even if we accept it
as a proposition by assuming that is always has exactly one logical value, could
also be treated as ambiguous; Ann may be found pretty by some people and
not pretty by others. If we try to improve the situation by saying for example:
”Ann seems to be pretty”, ” I am sure Ann is pretty” or even ”I know that Ann
is pretty” the ambiguity increases rather then decreases.
To deal with these and other ambiguities many non-classical logics were and
are being invented and examined by philosophers, computer scientists, and even
by mathematicians. We will present and study some of them later. Nevertheless
we accept all these and similar statements within classical propositional model
as propositions and represent them by the propositional variables.
Observe that one can think about a natural language as a set W of all words
and sentences based on a given alphabet A. This leads to a simple, abstract
model of a natural language NL as a pair
N L = (A, W).
Some natural languages share the same alphabet, some have different alphabets.
All of them face serious problems with a proper recognition and definitions of
accepted words and complex sentences. We do not want the symbolic languages
to share the same difficulties. We define their components precisely and in such
a way that their recognition and correctness will be easily decided. In order
to distinguish them from natural languages we call their words and sentences
formulas and denote the set of all formulas by F. We call a pair
a symbolic language.
We distinguish two categories of symbolic languages: propositional and predi-
cate. We first define the propositional language. The definition of the predicate
language, with its much more complicated structure will follow.
25
Definition 2.1
By a propositional language L we understand a pair
L = (A, F), (2.2)
where A is called propositional alphabet, and F is called a set of all well formed
propositional formulas of L.
26
These are not all formulas we can obtain in the second recursive step. Moreover,
as the recursive process continue we obtain a countably infinite set of all non-
atomic formulas.
Remark that we put parenthesis within the formulas in a way to avoid ambigu-
ity. The expression a ∩ b ∪ a is ambiguous. We don’t know whether it represents
a formula (a ∩ b) ∪ a, or a formula a ∩ (b ∪ a).
Exercise 2.1
Consider a following set
S = {¬a ⇒ (a ∪ b), ((¬a) ⇒ (a ∪ b)), ¬(a ⇒ (a ∪ b)), ¬(a → a)}.
1. Determine which of the elements of S are, and which are not well formed
formulas (wff ) of L = (A, F).
2. For any A 6∈ F re-write it as a correct formula and write in the natural
language what it says.
Solution
The formula ¬a ⇒ (a ∪ b) is not a well formed formula. A correct formula is
(¬a ⇒ (a ∪ b)). The corrected formula says: ”If a is not true , then we have a
or b ”. Another correct formula is ¬(a ⇒ (a ∪ b)). This corrected formula says:
”It is not true that a implies a or b ”.
The formula ((¬a) ⇒ (a ∪ b)) is not correct; (¬a) 6∈ F. The correct formula is
(¬ ⇒ (a∪b)). The formula ¬(a ⇒ (a∪b)) is correct. The formula ¬(a → a) 6∈ F
as the connective → does not belong to the language L. It is a correct formula of
another propositional language; the one that uses a symbol → for implication.
Exercise 2.2
Given a sentence S
”If a natural number a is divisible by 3, then from the fact that a in not divisible
by three we can deduce that a is divisible by 5.”
Write a formula corresponding to the sentence S.
Solution
First we write our sentence in a more ”logical way” as follows:
” If a natural number a is divisible by 3, then (if not(a is divisible by three) then
a is divisible by 5). We denote the sentence: ”a natural number a is divisible
by 3 ” by a, and the sentence ”a is divisible by 5” by b, and we rewrite our
sentence as: ”If a, then (if not a, then b)”.
We replace expressions If ... then and not by symbols ⇒ and ¬, respectively
and we follow the definition of the set of formulas to obtain a formula
27
(a ⇒ (¬a ⇒ b))
Exercise 2.3
Write following natural language statement:
”One likes to play bridge or from the fact that the weather is good we conclude
the following: one does not like to play bridge or one likes swimming.”
as a formula of L = (A, F).
Solution
First we identify the needed components of the alphabet A as follows.
Propositional variables: a, b, c.
a denotes statement: One likes to play bridge, b denotes a statement: the weather
is good, c denotes a statement: one likes swimming.
Connectives: ∪, ⇒, ∪.
Then we write the formula of L as (a ∪ (b ⇒ (¬a ∪ c))).
Exercise 2.4
Given a formula (a ∩ (¬a ∪ b)).
Write 2 natural language sentences which correspond to this formula.
Solution
Let propositional variables a, b denote sentences 2+2 = 4 and 2 > 1, respectively.
In this case the corresponding sentence is:
2 + 2 = 4 and we have that 2 + 2 6= 4 or 2 > 1.
If we assume that the propositional variables a, b denote sentences 2 > 1 and
2 + 2 = 4, respectively, then the corresponding natural language statement is:
28
2 > 1 and we have that 2 6> 1 or 2 + 2 = 4.
The symbols for connectives used in book are not the only used in mathematical,
logical, or computer science literature.
Other symbols employed for these most important propositional connectives are
listed in the table below.
The first of these systems of notation is the closest to ours and is drawn mainly
from the algebra of sets and lattice theory. The second comes from the Polish
logician J. Lukasiewicz. In this notation the binary connectives precede the
formulas and are not inserted between them; this enables us to dispense with
parenthesis; Lukasiewicz’s notation is usually called the Polish notation and it
is a parenthesis-free notation. The third was used by D. Hilbert. The fourth
comes from Peano and Russell, while the fifth goes back to Schröder and Pierce.
29
tives. We define their semantics, i.e. their definitions in terms of two logical
values, and give a motivation justifying the definitions as follows.
We denote a statement A is false by A = F , what stands for the logical value
of a formula A is F. We denote a statement A is true by A = T , what stands
for the logical value of a formula A is T.
Negation motivation and definition.
In accordance with the intuition, the negation of a true formula is a false formula,
and the negation of a false formula is a true formula. Moreover, the logical value
of ¬A depends on the logical values of A in a way which can be express in the
form of the following table.
Negation Table
A ¬A
T F (2.3)
F T
A B (A ∩ B)
T T T
T F F (2.4)
F T F
F F F
30
Disjunction Table
A B (A ∪ B)
T T T
T F T (2.5)
F T T
F F F
As in the case of the other connectives, the logical value of a disjunction depends
only on the logical values of its factors.
Implication motivation and definition.
The symbol ⇒ is used instead of the statements of the form if A, then B, A
implies B, and is called an implication connective. The formula (A ⇒ B) and
is called an implication and A is called its antecedent, B is called its consequent.
The semantics of the implication needs some discussion. In everyday language
the implication statement if A, then B is interpreted to mean that B can be in-
ferred from A. This interpretation differs from that given to it in mathematics,
and hence in classical semantics. The following example explains the meaning
of the statement if A, then B as understood in mathematics. It hence justifies
our semantics for the implication.
The above implication (2.6) is true for any natural number, hence, in par-
ticular, for 2,3,6.
Thus the following propositions are true:
31
Thus one case remains to be examined, namely that in which the antecedent of
an implication is a true statement, and the consequent is a false statement.
For example consider the statement:
If 6 DIVIDES 12, then 6 DIVIDES 5.
In accordance with arithmetic of natural numbers, this statement is interpreted
as false.
The above examples justifies adopting the following semantics for the impli-
cation ⇒. An implication (A ⇒ B) is interpreted to be a false statement if
and only if its antecedent A is a true statement and its consequent is a false
statement. In the remaining cases such an implication is interpreted as a true
statement.
We expressed it in a form of the following table.
Implication Table
A B (A ⇒ B)
T T T
T F F (2.10)
F T T
F F T
A B ¬A (A ∩ B) (A ∪ B) (A ⇒ B) (A ⇔ B)
T T F T T T T
T F F F T F F (2.12)
F T T F T T F
F F T F F T T
The table (2.12) indicates that the logical value of of propositional connectives
depends only on the logical values of its factors; i.e. it is independent of the
32
formulas A, B. We write the table in a ”formula in depended” form as a set of
the following equations.
¬T = F, ¬F = T ;
(T ∩ T ) = T, (T ∩ F ) = F, (F ∩ T ) = F, (F ∩ F ) = F ;
(T ∪ T ) = T, (T ∪ F ) = T, (F ∪ T ) = T, (F ∪ F ) = F ; (2.13)
(T ⇒ T ) = T, (T ⇒ F ) = F, (F ⇒ T ) = T, (F ⇒ F ) = T ;
(T ⇔ T ) = T, (T ⇔ F ) = F, (F ⇔ T ) = F, (T ⇔ T ) = T.
We use the above set of equations (2.13) to evaluate logical values of formulas.
Example 2.1
Given a formula (A ⇒ (¬A ∩ B)), such that logical values of its basic compo-
nents, i.e. the propositional formulas A, B are: A=T, and B=F. We calculate
the logical value of the formula (A ⇒ (¬A ∩ B)) by substituting the logical values
for the formulas A, B and applying the equations (2.13) as follows.
(T ⇒ (¬T ∩ F )) = (T ⇒ (F ∩ F )) = (T ⇒ F ) = F.
Exercise 2.5
Given a formula A: (((a ∪ b) ∩ ¬c) ⇒ a). Evaluate the logical value of A for the
following sets of logical values of its basic components, i.e. for the propositional
variables a, b: 1. a=T, b=F, c=F, and 2. a=F, b=T, c=T.
Solution
1. Let a=T, b=F, c=F. We evaluate the logical value of A as follows.
Extensional Connectives
We observe that our connectives are such that the logical value of a given formula
build by means of its connectives depends only of logical values of its factors.
Connectives with this property are called extensional. We hence adopt the
following definition.
33
Definition 2.2
We call a propositional connective extensional if the logical value of a given
formula build by means of this connective depends only of logical values of its
factors.
Fact 2.1
All connectives ¬, ∪, ∩, ⇒, and ⇔ are extensional.
A 51 A 52 A ¬A 54 A
T F T F T (2.14)
F F F T T
34
All Binary Connectives
A B (A◦1 B) (A ∩ B) (A◦3 B) (A◦4 B)
T T F T F F
T F F F T F
F T F F F T
F F F F F F
A B (A ↓ B) (A◦6 B) (A◦7 B) (A ↔ B)
T T F T T T
T F F T F F
F T F F T F
F F T F F T
(2.15)
A B (A◦9 B) (A◦10 B) (A◦11 B) (A ∪ B)
T T F F F T
T F T T F T
F T T F T T
F F F T T F
A B (A◦13 B) (A ⇒ B) (A ↑ B) (A◦16 B)
T T T T F T
T F T F T T
F T F T T T
F F T T T T
Functional Dependency
Fact 2.2 All two-valued propositional connectives and in particular our connec-
35
tives ¬, ∪, ∩, ⇒, and ⇔ are a functionally dependent.
Fact 2.3
The alternative negation connective ↑, and the joint negation. ↓ suffice, each
of them separately, to define all propositional connectives , whether unary or
binary.
Fact 2.4
No propositional connective other than ↑ and ↓ suffices to define all the remain-
ing connectives.
Example 2.2
Definition of negation ¬ in terms of ↑.
This is an interesting example as it shows that one can define a one argument
connective in terms of a two argument connective.
Let’s now look at Sheffer’s alternative negation connective ↑.
Alternative Negation ↑
A B (A ↑ B)
T T F
T F T (2.16)
F T T
F F T
We now write the table (3.3) in the ”formula independed” form of the following
equations.
(T ↑ T ) = F, (T ↑ F ) = T, (F ↑ T ) = T, (F ↑ F ) = T (2.17)
¬A = (A ↑ A) (2.18)
36
and call it a definition of ¬ in terms of ↑. We verify its correctness of of by
building the table below.
A ¬A (A ↑ A) computation (A ↑ A)
T F (T ↑ T ) = F F (2.19)
F T (T ↑ T ) = F T
The table shows that the logical value of a formula ¬A is the same as logical
value of a formula (A ↑ A), for any logical value their basic component A can
take, i.e. that our definition (2.18) is correct.
Example 2.3
Definition of conjunction ∩ in terms of ↑.
Observe now that the Sheffer’s connective table (3.3) looks as a negation of the
conjunction table (2.4). It means that the logical value a formula (A ∩ B) is
the same as logical value of a formula ¬(A ↑ B), for all logical values of A and
B. We write it as
(A ∩ B) = ¬(A ↑ B). (2.20)
We have just proved the formula (2.18) to be true for any formula and hence
for the formula ¬(A ↑ B), i.e. we get that ¬(A ↑ B) = (A ↑ B) ↑ (A ↑ B). The
formula (2.24) becomes (A ∩ B) = (A ↑ B) ↑ (A ↑ B).
We call the equality
(A ∩ B) = (A ↑ B) ↑ (A ↑ B) (2.21)
the definition of conjunction in terms of negation and Sheffer’s connective.
Let’s now examine the Lukasiewicz’ s joint negation connective ↓. The formula
A ↓ B is read: neither A nor B. As it is a special connective we re-write its
truth table separately.
Joint Negation ↓
A B (A ↓ B)
T T F
T F F (2.22)
F T F
F F T
We now write the table (3.30) in an ”formula independed” form of the following
equations.
(T ↓ T ) = F, (T ↓ F ) = F, (F ↓ T ) = F, (F ↓ F ) = T (2.23)
37
Observe that T ↓ T = F and F ↓ F = T . This means that logical value of a
formula (A ↓ A) is the same as logical value of a formula ¬A, for any logical
value the formula A can take. We write it as
¬A = (A ↓ A) (2.24)
and call it a definition of ¬ in terms of ↓. We verify its correctness of of by
building the table below.
A ¬A (A ↓ A) computation (A ↓ A)
T F (T ↓ T ) = F F (2.25)
F T (F ↓ F ) = T T
The table shows that the logical value of a formula ¬A is the same as logical
value of a formula (A ↓ A), for any logical value their basic component A can
take, i.e. that our definition (2.24) is correct.
Exercise 2.6
Prove that the equality
defines ∪ in terms of ↓.
Solution
To prove the correctness of the equation (2.26) we construct a table below.
A B (A ∪ B) ((A ↓ B) ↓ (A ↓ B))
T T T ((T ↓ T ) ↓ (T ↓ T )) = (F ↓ F ) = T
T F T ((T ↓ F ) ↓ (T ↓ F )) = (F ↓ F ) = T (2.27)
F T T ((F ↓ T ) ↓ (F ↓ T )) = (F ↓ F ) = T
F F F ((F ↓ F ) ↓ (F ↓ F )) = (T ↓ T ) = F
The table shows that the logical value of a formula (A∪B) is the same as logical
value of a formula ((A ↓ B) ↓ (A ↓ B)), for any logical value the formulas can
take depending of logical values of their basic components A, B, i.e. that our
definition (2.26) is correct.
38
propositional formulas that must be always true because of their syntactical
structure without reference to the meaning of the propositions they represent.
Such formulas are called propositional tautologies.
Example 2.4
Given a formula (A ⇒ A). Lets now evaluate its logical value for all possible
logical values of its basic component A, i.e. for A=T, and A=F. We put our
calculation in a form of a table below.
A (A ⇒ A) computation (A ⇒ A)
T (T ⇒ T ) = T T (2.28)
F (F ⇒ F ) = T T
Example 2.5
We construct a truth table for a formula (A ⇒ B) as follows.
A B (A ⇒ B) computation (A ⇒ B)
T T (T ⇒ T ) = T T
T F (T ⇒ F ) = F F (2.29)
F T (F ⇒ T ) = T T
F F (F ⇒ F ) = T T
The logical value of the formula (A ⇒ B) is F for A=T and B=F what means
that it is not a propositional tautology. We put these ideas in a form of the
following definition.
Definition 2.3
For any formula A of a propositional language L, we say that A is a propositional
tautology if and only if the logical value of A is T (we write it A=T) for all
possible logical values of its basic components. We write
|= A
to denote that A is a tautology.
39
by constructing the table (11.27). We leave the proofs of others as an easy
exercise.
Excluded Middle
|= (¬A ∪ A) (2.32)
One of the most frequently used classical tautologies are the laws of detachment
for implication and equivalence. The implication law was already known to the
Stoics (3rd century BC) and a rule of inference, based on it is called Modus
Ponens, so we use the same name here.
Modus Ponens
|= ((A ∩ (A ⇒ B)) ⇒ B) (2.33)
Detachment
|= ((A ∩ (A ⇔ B)) ⇒ B) (2.34)
|= ((B ∩ (A ⇔ B)) ⇒ A)
Mathematical and not only mathematical theorems are usually of the form of
an implication, so we will discuss some terminology and more properties of
implication.
40
Each of the following pairs of implications: a simple and an opposite, and a
converse and a contrary are equivalent, i.e. the following formulas are tautolo-
gies:
converse
(A =⇒ B) (B =⇒ A)
contrapositive
contrary contrary
Equivalent implications are situated at the vertices of one and the same diag-
onal. It follows from the contraposition laws that to prove all of the following
implications: (A ⇒ B), (B ⇒ A), (¬A ⇒ ¬B), (¬B ⇒ ¬A), it suffices to prove
any pairs of those implications which are situated at one and the same side of
the square, since the remaining two implications are equivalent to those already
proved to be true.
Consider now the following tautology:
|= ((A ⇔ B)) ⇔ ((A ⇒ B) ∩ (B ⇒ A))). (2.36)
The above tautology (2.36) says that in order to prove a theorem of a form of
(A ⇔ B) it suffices to prove two implications: the simple one (A ⇒ B) and the
converse one (B ⇒ A). Conversely, if a formula (A ⇔ B) is a theorem, then
the implications (A ⇒ B) and (B ⇒ A) are also theorems.
In other words, B is then a necessary condition for A, and at the same time B
is a sufficient condition for A. Accordingly, we say that a theorem of the form
41
of a formula (A ⇔ B) is often formulated as: ” B is necessary and sufficient
condition for A”.
Other laws developed by the Stoics are the hypothetical syllogism and modus
tollendo ponens. We present them here in a form of logical tautology, not as
the rule of reasoning as it was developed. The relationship between those two
approaches is quite obvious and will be discussed in detail in the proof theory
chapter.
Hypothetical syllogism
|= (((A ∪ B) ∩ ¬B) ⇒ A)
Here are some other tautologies with a history centuries old. First is called
Duns Scotus Law after an eminent medieval philosopher who lived at the turn
of the 13th century. Second is called Clavius Law , after Clavius, a Euclid
commentator who lived in the late 16th century. The reasonings based on this
law were already known to Euclid, but this type of inference became popular in
scholarly circles owing to Clavius, hence the name. The third is called Frege
Laws after G. Frege who was first to give a formulation of the classical propo-
sitional logic as a formalized axiomatic system in 1879, adopting the second of
them as one of his axioms.
Duns Scotus
|= (¬A ⇒ (A ⇒ B)) (2.39)
Clavius
|= ((¬A ⇒ A) ⇒ A) (2.40)
Frege
|= (((A ⇒ (B ⇒ C)) ∩ (A ⇒ B)) ⇒ (A ⇒ C)) (2.41)
Double Negation
|= (¬¬A ⇔ A) (2.42)
42
Next set of tautologies deal with apagogic proofs which are the proofs by
reductio ad absurdum. The method of apagogic proof consists in negating
the theorem which is to be proved. If the assumption that the theorem is
false yields a contradiction, then we conclude that the theorem is true. The
correctness of this reasoning is based on the following tautology.
Reductio ad Absurdum
The proofs based on the application of the laws of contraposition (2.35) are
also classed as apagogic. Instead of proving a simple theorem (A ⇒ B) we
prove the opposite theorem (¬B ⇒ ¬A), which is equivalent to the simple
one. The following two tautologies, also called laws of contraposition, are used,
respectively, when the hypothesis or the thesis of the theorem to be proved is
in the form of a negation.
43
Laws of Contraposition (2)
Conjunction
|= ((A ∩ B) ⇒ A), |= ((A ∩ B) ⇒ B),
|= (((A ⇒ B) ∩ (A ⇒ C)) ⇒ (A ⇒ (B ∩ C))),
|= (((A ⇒ B) ∩ (C ⇒ D)) ⇒ ((A ∩ C) ⇒ (B ∩ D))),
|= (A ⇒ (B ⇒ (A ∩ B))).
Disjunction
|= ((A ⇒ (A ∪ B)), |= ((B ⇒ (A ∪ B)),
|= (((A ⇒ B) ∩ (B ⇒ C)) ⇒ ((A ∪ B) ⇒ C)),
|= (((A ⇒ B) ∩ (C ⇒ D)) ⇒ ((A ∪ C) ⇒ (B ∪ D))).
Here are some more important and frequently used equivalence tautologies,
called also the equivalence laws.
Idempotence
|= ((A ∩ A) ⇔ A), |= ((A ∪ A) ⇔ A),
Associativity
|= (((A ∩ B) ∩ C) ⇔ (A ∩ (B ∩ C))),
|= (((A ∪ B) ∪ C) ⇔ ((A ∪ (B ∪ C))).
Commutativity
Distributivity
De Morgan
44
Implication
|= ((A ⇒ B) ⇔ (¬A ∪ B)). (2.48)
Negation of Implication
|= (¬(A ⇒ B) ⇔ (A ∩ ¬B)).
Negation of Equivalence
Double Negation
|= (¬¬A ⇔ A). (2.49)
|= (((A ∩ B) ⇒ C) ⇔ (A ⇒ (B ⇒ C))).
De Morgan laws (2.47) are named after A. De Morgan (1806 - 1871), an English
logician, who discovered analogous laws for the algebra of sets. They stated that
for any sets A,B the complement of their union is the same as the intersection
of their complements, and vice versa, the complement of the intersection of two
sets is equal to the union of their complements. The laws of the propositional
calculus were formulated later, but they are usually also called De Morgan Laws.
Definition 2.4
By a predicate language L we understand a triple
45
Alphabet A
The components of A are as follows.
1. Propositional connectives: ¬, ∩, ∪, ⇒, ⇔.
2. Quantifiers: we adopt two quantifiers; ∀ (for all, the universal quantifier)
and ∃ (there exists, the existential quantifier).
In a case of the classical logic it is possible to adopt only one quantifier and
to define the other in terms of it and propositional connectives. But the two
quantifiers express better the common intuition, so we assume that we have two
of them.
3. Parenthesis: ( and ).
4. Variabes: we assume that we have, as we did in the propositional case a
countably infinite set VAR of variables. The variables now have a different
meaning than they had in the propositional case. We hence call them variables,
or individual variables to distinguish them from the propositional variables. We
also denote denote them by different symbols, namely by letters x, y, z, ..., with
indices, if necessary. We express it by writing V AR = {x1 , x2 , ....}.
5. Constants: the constants represent in ”real life” concrete elements of sets.
We denote constants by by c, d, e..., with indices, if necessary. We assume that
we have a countably infinite set C = {c1 , c2 , ...} of constants.
6. Predicate symbols: the predicate symbols represent ”real life” relations. We
denote them by P, Q, R, ... with indices, if necessary. We use symbol P for the
set of all predicate symbols. We assume that P it countably infinite.
In ”real life” we write symbolically x < y to express that element x is smaller
then element y according to the two argument order relation <. In our predicate
language L we represent the relation < as a two argument predicate P ∈ P and
write P (x, y), where now x, y are individual variables from the set VAR.
Mathematical statements n < 0, 1 < 2, 0 < m are represented in L by
P (x, c1 ), P (c, c3 ), P (c1 , y), respectively. Here c1 , c2 , c3 are any constants and
x, y any variables.
7. Function symbols: the function symbols represent ”real life” functions. We
denote function symbols by f, g, h, ..., with indices, if necessary. We use symbol
F for the set of all function symbols. We assume that the set F is countably
infinite.
Set T of terms
Terms are expressions built out of function symbols and variables. They describe
how we build compositions of functions. We define the set T of terms recursively
as follows.
1. All variables are terms.
46
2. All constants are terms.
3. For any function symbol f representing a function on n variables, and any
terms t1 , t2 , ..., tn , the expression f (t1 , t2 , ..., tn ) is a term.
4. The set T of terms is the smallest set that fulfills the conditions 1. - 3.
Consider a ”real life” function given by a formula sin(x + y). It is a composition
of two functions defined by formulas sinx and x + y. The sin is one argument
function and we represent it as term f (x) for f ∈ F. The + is a two argument
function and we represent it as a term g(x, y) for g ∈ F. The ”real life” function
sin(x+y) is hence represented by a term f (g(x, y)), where x, y are any individual
variables from the set VAR. Observe that to obtain the predicate language
representation of for example x + y we can first write the real two argument
function formula x + y as +(x, y) and then replace the addition symbol + by
any two argument function symbol g ∈ F and get the term g(x, y).
Here are some more terms of L.
h(c1 ), f (g(c, x)), g(f (f (c)), g(x, y)), f1 (c, g(x, f (c))), g(g(x, y), g(x, h(c))) ....
Set F of formulas
Formulas are now expressions built out of elements of the alphabet A and the
set T of terms. We denote them, as in propositional case by A, B, C, ..... with
indices, if necessary. We build them, as before in recursive steps, the fist of
them says as in the propositional case: all atomic formulas are formulas. The
atomic formulas are the simplest formulas as the propositional variables were
in the case of propositional language. We define them as follows.
Definition 2.5 An atomic formula is any expression of the form R(t1 , t2 , ..., tn )
where R is any predicate R ∈ P and t1 , t2 , ..., tn are terms, i.e. t1 , t2 , ..., t∈ T.
Q(c), Q(x), Q(g(x1 , x2 )), R(c, d), R(x, f (c)), R(g(x, y), f (g(c, z))), .....
47
Definition 2.6
The set F of formulas of L is the smallest set meeting the following conditions.
1. All atomic formulas (definition 10.2) are formulas;
2. If A, B are formulas, then ¬A, (A ∩ B), (A ∪ B), (A ⇒ B), (A ⇔ B) are
formulas;
3. If A is a formula, then ∀xA, ∃xA are formulas for any variable x ∈ V AR.
R(c, d), ∃yR(y, f (c)), R(x, y), (∀xR(x, f (c)) ⇒ ¬R(x, y)),
R(c1 , c2 ), R(x, y), (R(y, d) ⇒ R(a, z)), ∃xR(x, y), ∀yR(x, y), ∃x∀yR(x, y).
Observe that the formulas ∃yP (y), (∀x(P (x) ⇒ ∃yQ(x, y))) are closed. We
call a close formula a sentence.
48
Example 2.6
Consider atomic formulas: P (y), Q(x, c), R(z), P1 (g(x, y), z). Here are some
non atomic formulas formed out of them.
1. (P (y) ∪ ¬Q(x, c)) ∈ F. This is an open formula A with two free variables
x,y. We denote A this as formula A(x, y).
2. ∃x(P (y) ∪ ¬Q(x, c)) ∈ F. We write x to denote that x is a bound variable.
The variable y is free. This is a formula B with one free variable y. We denote
B as a formula B(y).
3. ∀y(P (y) ∪ ¬Q(x, c)) ∈ F. The variable y is bound, the variable x is free.
We denote this formula by for example A1 (x).
4. ∀y∃x(P (y) ∪ ¬Q(x, c)) ∈ F has no free variables. It is a closed formula
called also a sentence.
Exercise 2.7
Given the following formulas of L:
P (x, f (c, y)), ∃cP (x, f (c, y)), ∀xf (x, P (c, y)), ∃xP (x, f (c, y)) ⇒ ∀yP (x, f (c, y)).
1. Indicate whether they are, or are not well formed formulas of F. For those
which are not in F write a correct formula. 2. For each correct, or corrected
formula identify all components: connectives, quantifiers, predicate and function
symbols, and list all its terms. 3. For each formula identify its s free and bound
variables. State which are open and which are closed formulas (sentences), if
any.
Solution
Formula A1 = P (x, f (c, y)).
It is a correct atomic formula. P is a 2 argument predicate symbol, f is a
2 argument function symbol, c is a constant. We write it symbolically: P ∈
P, f ∈ F, c ∈ C. It is an open formula with two free variables x,y. We denote
it by A1 (x, y). It has no bound variables.
Formula A2 = ∃cP (x, f (c, y)).
It is a not a correct formula, i.e. ∃cP (x, f (c, y)) 6∈ F. The expression ∃c has no
meaning because c is a constant, not a variable.
The corrected formulas are: B1 = ∃xP (x, f (c, y)), B2 = ∃yP (x, f (c, y)), and
formulas B = ∃zP (z, f (c, y)) for any variable z different then x and y.
None of the correct formulas are open. Variable y is free in B1 = B1 (y), variable
x is free in B2 = B2 (x), both variables x and y are free in all formulas B =
B(x, y). All formulas are nether close, nor open. The terms appearing in any
of them are the same as in A1 = P (x, f (c, y)) and are: x, y, c, f (c, y).
49
Formula A3 = ∀xf (x, P (c, y)).
It is a not a correct formula, i.e. ∀xf (x, P (c, y)) 6∈ F. The function symbol f in
front f (x, P (c, y)) indicate a term and terms are not formulas. Moreover, the
atomic formula P (c, y) can’t be put inside a term!
Formula A4 = ∃xP (x, f (c, y)) ⇒ ∀yP (x, f (c, y))
It is a not a correct formula. The correct formula is A = (∃xP (x, f (c, y)) ⇒
∀yP (x, f (c, y))). It has two free variables x and y and we write it as A = A(x, y).
We often use logic symbols, while writing mathematical statements in a more
symbolic way. For example, mathematicians to say ”all natural numbers are
greater then zero and some integers are equal 1” often write
∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1,
or even as
(∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1).
Observe that none of the above symbolic statements, not even the last one, are
formulas of the predicate language. These are mathematical statements written
with mathematical and logic symbols. They are written with different degrees
of ”logical precision”, the last being, from a logician point of view, the most
precise.
Our goal now is to ”translate ” mathematical and natural language statement
into correct logical formulas, i.e. into formulas of the predicate language L. Let’s
start with some observations about the statements above.
The quantifiers in ∀x∈N and ∃y∈Z used in all of them are not the one used in
logic. In our language L we use only quantifiers ∀x and ∃y, for any variables
x, y ∈ V AR. The quantifiers ∀x∈N , ∃y∈Z are called quantifiers with restricted
domain. The first is restricted to the domain of natural numbers, the second
to the integers. The restriction of the quantifier domain can, and often is given
by more complicated statements. For example we say ”for all x > 2” and
write ∀x>2 , or we say ”exists x > 2 and at same time x + 2 < 8” and write
symbolically ∃(x>2∩x+2<8) . We introduce the quantifiers with restricted domain
into our predicate logic language by expressing them within the language L as
follows.
Definition 2.7
The quantifiers ∀A(x) , ∃A(x) are called quantifiers with restricted domain,
or restricted quantifiers, where A(x) ∈ F is any formula with any free vari-
able x ∈ V AR.
50
A formula ∀A(x) B(x) stands for a formula ∀x(A(x) ⇒ B(x)) ∈ F. We write
it symbolically as
∀A(x) B(x) ≡ ∀x(A(x) ⇒ B(x)). (2.51)
51
Exercise 2.8
Given a mathematical statement S written with logical symbols
(∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1)
Solution
Step 1. The basic statements in S are: x ∈ N, x ≥ 0, y ∈ Z, y = 1. The
relations are: ∈ N, ∈ Z, ≥, =. We use one argument predicate symbols N, Z
for ∈ N, ∈ Z, respectively. We use two argument predicate symbols G for ≥,
and E for =. There are no functions. We have two constant symbols c1 , c2 for
numbers 0 and 1, respectively.
Step 2. We write N (x), Z(x) for x ∈ N, x ∈ Z, respectively. G(x, c1 ) for x ≥ 0
and E(y, c2 ) for y = 1. These are all atomic formulas.
Step 3. The statement S becomes a restricted quantifiers formula:
Example 2.7
Given a mathematical statement S written with logical symbols
(∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1)
52
Exercise 2.9
Here is a mathematical statement S:
”For all real numbers x the following holds: If x < 0, then there is a natural
number n, such that x + n < 0.”
1. Re-write S as a symbolic mathematical statement SF that only uses mathe-
matical and logical symbols. 2. Translate the symbolic statement SF into to a
corresponding formula A ∈ F of the predicate language L.
Solution
The symbolic mathematical statement SF is : ∀x∈R (x < 0 ⇒ ∃n∈N x + n < 0).
We write R(x) for x ∈ R , N(y) for n ∈ N , and atomic formula L(x, c) for the
basic statement x < 0. We write f(x,y) for the function +(x, n) and a constant
c for the number 0. We write atomic formula L(f(x,y), c) for x + n < 0. The
symbolic statement SF becomes ∀R(x) (L(x, c) ⇒ ∃N (y) L(f (x, y), c)). The corre-
sponding formula A ∈ F is ∀x(N (x) ⇒ (L(x, c) ⇒ ∃y(N (y) ∩ L(f (x, y), c))).
There are various kinds of non-mathematical statements, that obviously cannot
be justified on the basis of propositional logic. Consider for example a statement
”Any friend of Mary is a friend of John and Peter is not John’s friend. Hence
Peter is not May’s friend. ”
Intuitively, what it says is always true, but translating it it into a propositional
language we get a formula ((a ∩ ¬b) ⇒ ¬c) that can be false. The validity of the
reasoning described by the statement follows from a more complexed structure
provided by the predicate language. We will discuss the notion of validity of
predicate language formulas, i.e. a semantics for predicate logic later. Natural
language statements and reasoning with them also play a special role in creation
of non-classical logics and in Artificial Intelligence research and applications.
Exercise 2.10
Translate a natural language statement S: ”Any friend of Mary is a friend of
John and Peter is not John’s friend. Hence Peter is not May’s friend.” into a
formula A ∈ F of the predicate language L.
Solution
1. We identify the basic relations and functions (if any) and translate them into
atomic formulas.
We have only one relation of ”being a friend”. It is a two argument relation. We
write atomic formula F(x, y) for ”x is a friend of y”. We use constants m, j, p
for Mary, John, and Peter, respectively. We have the following atomic formulas:
F(x, m) for ”x is a friend of Mary”, F(x, j) for ”x is a friend of John”, F(p, j)
for ”Peter is a friend of John”.
53
2. Statement ”Any friend of Mary is a friend of John” translates into a re-
stricted quantifier formula ∀F (x,m) F (x, j). Statement ”Peter is not John’s
friend” translates into ¬F (p, j), and ”Peter is not May’s friend” translates into
¬F (p, m).
3. Restricted quantifiers formula for S is
Here are simple steps we follow in order to perform translations from natural
language to the symbolic predicate language L. They are similar to the steps we
used in the translations of mathematical formulas nevertheless we voice them
separately and call them rules of translation.
Rules of translation to L.
1. Identify the basic relations and functions (if any) and translate them into
atomic formulas.
2. Identify propositional connectives and use symbols ¬, ∪, ∩, ⇒, ⇔ for them.
3. Identify quantifiers. Restricted ∀A(x), ∃A(x) and non-restricted ∀x, ∃x.
4. Use the symbols from 1. - 3. and restricted quantifiers transformation rules
(8.18) and (8.19) to write A ∈ F of the predicate language L.
Example 2.8
Given a natural language statement S: ”For any bird one can find some birds
that are white.” The translation of S into a formula of the predicate language L
is
∀x(B(x) ⇒ ∃x(B(x) ∩ W (x))).
54
the second appearance of B(x) is bounded by the quantifier ∃x. Let’s re-write
the formula A using x to indicate this fact
In this case, and in the similar cases we can apply a predicate logic law of
quantifiers, called Rename Variables Law to our formula A and get a formula
B that is logically equivalent to A. It means that the formula B states exactly
the same what A states but is written in a more comprehensible form:
We will discuss and study Laws of Quantifiers in the next section. There is
another important law, one of the Distributivity Laws that allows us to transform
B into a formula ∀x∃y(B(x) ⇒ (B(y) ∩ W (y))). We express it as the following
example.
Example 2.9
Given a natural language statement S: ”For any bird one can find some birds
that white.” The translation of S into a formula of the predicate language L is
Exercise 2.11
Translate into L a natural language statement
S: ” Some patients like all doctors.”
Solution.
1. Atomic formulas: P(x), D(x), L(x, y). We write one argument predicate
P(x) for ” x is a patient”, one argument predicate D(x) for ” x is a doctor”, and
two argument predicate L(x,y) for ” x likes y”.
2. There is no propositional connectives in S.
3. Restricted quantifiers: ∃P (x) for ”some patients ” and ∀D(x) for ”all doctors”.
Observe that we can’t write L(x, D(y)) for ”x likes doctor y”. D(y) is a predicate,
not a term and hence L(x, D(y)) is not a formula. We have to express the
statement ” x likes all doctors y” in terms of restricted quantifiers and predicate
L(x,y) only. The statement ” x likes all doctors y” means ” all doctors y are
liked by x”, i.e. ”for all doctors y, x likes y”. This translates to ∀D(y) L(x, y)
and the statement S translates to ∃P (x) ∀D(x) L(x, y).
4. By the transformation rules we get the following translation of S into L.
55
Translations to Logic in Artificial Intelligence
Example 2.10
AI formulas corresponding to a statement
S: ”For every student there is a student that is an elephant.”
are as follows.
1. Restricted quantifiers AI formula:
∀Student(x) ∃Student(x) Elephant(x).
2. Non-restricted quantifiers AI formula :
∀x(Student(x) ⇒ ∃x(Student(x) ∩ Elephant(x))).
3. Re-name variables AI formula:
∀x(Student(x) ⇒ ∃y(Student(y) ∩ Elephant(y))).
4. AI formula after applying the the Distributivity Laws:
∀x∃y(Student(x) ⇒ (Student(y) ∩ Elephant(y))).
56
The predicate symbols P, R, Student, Elephant denote in all cases one argument
predicates but AI predicate symbols Student, Elephant (of a slightly different
language than L) impose a particular meaning called the intended interpre-
tation. The predicate symbols P, R and any elements of the set of all predicate
symbols P of L.
Exercise 2.12
Translate a natural language statement ”Any friend of Mary is a friend of
John and Peter is not John’s friend. Hence Peter is not Mary’s friend.” into a
formula A of the predicate AI language (of your choice).
Solution
Statement ”Any friend of Mary is a friend of John” translates into a restricted
quantifier AI formula ∀F riend(x,M ary) F riend(x, John).
Statement ”Peter is not John’s friend” translates into ¬F riend(P eter, John),
and ”Peter is not Mary’s friend” translates into ¬F riend(P eter, M ary).
Restricted quantifiers AI formula for S is ((∀F riend(x,M ary) F riend(x, John) ∩
¬F riend(P eter, John)) ⇒ ¬F riend(P eter, M ary)).
The AI formula is ((∀x(F riend(x, M ary) ⇒ F riend(x, John)) ∩
¬F riend(P eter, John)) ⇒ ¬F riend(P eter, M ary)).
The AI formulas are very useful, as they ”read” as natural language statements
but it is very important to remember that they do not carry any meaning, as the
natural language statements do to the reader. An atomic formula Friend(Peter,
John) is just an atomic formula of a symbolic AI language as P (c, d) is in L.
We assign a meaning to them i.e. their semantics in a separate step as we did
in the propositional case. The first step in this process is an assignment of an
interpretation in a non-empty set U of the predicate, functional and constant
symbols. Each symbol can have many interpretations in a given set and we can
can define the interpretations an many sets. The AI intended interpretation
of the two argument predicate named Friend and constants Peter, John is
to define the set U and a relation Friend. This relation must hold between
elements Peter, John and other elements of U in a way we want to define what
”friendship” means in the set U. This is called in AI a conceptualization.
57
definition of the predicate tautology. We also list and discuss the most used and
useful tautologies and equational laws of quantifiers.
The formulas of the predicate language L have meaning only when an interpre-
tation is given for the symbols. We define the interpretation I in a set U 6= ∅
by interpreting predicate, functional symbols as a concrete relation, function
defined in the universe U, and constants symbols as elements of the set U. The
set U is called the universe of the interpretation I. These two items specify a
model structure for L. We write it as a pair M = (U, I).
Given a formula A of L, and the model structure M = (U, I). Let’s denote by
AI a statement written with logical symbols determined by the formula A and
the interpretation I in the universe U. When A is a closed formula, it means it
is a sentence, formula without free variables, AI represents a proposition that
is true or false. When A is not a sentence it contains free variables and may
be satisfied (i.e. true) for some values in the universe U and not satisfied (i.e.
false) for the others. Lets look at few simple examples.
Example 2.11
Let A be a formula ∃xP (x, c) and consider a model structure M1 = (N, I1 ).
The universe of the interpretation I1 is the set N of natural numbers and we
define I1 as follows: we interpret the predicate P as relation < and the constant
c as number 5, i.e we put PI1 := andcI1 : 5.
M1 |= ∃xP (x, c)
Example 2.12
Consider now a model structure M2 = (N, I2 ) and the formula A: ∃xP (x, c).
We interpret now the predicate P as relation < in the set N of natural numbers
and the constant c as number 0, i.e. we put PI2 :< and cI2 : 0.
58
AI2 is obviously a false mathematical statement. We say that the formula A:
∃xP (x, c) is false under the interpretation I2 in M2 or that A is false in M2 .
We write it symbolically as
M2 6 |= ∃xP (x, c)
Example 2.13
Consider now a model structure M3 = (Z, I3 ) and the formula A: ∃xP (x, c).
We define an interpretation I3 in the set of all integers Z exactly as the inter-
pretation I1 , i.e. we put PI3 :< and cI3 : 0.
In this case we get AI3 : ∃x∈Z x < 0 and AI3 is obviously a true mathematical
statement. The formula A is true under the interpretation I3 in M3 (A is
satisfied, true in M3 ). We write it symbolically as
Example 2.14
Consider the following formulas: 1. A1 : R(x, y), 2. A2 : ∀yR(x, y), 3. A3 :
∃x∀yR(x, y). We define a model structure M = (N, I) where R is interpreted
as a relation ≤ defined in the set N of all natural numbers, i.e. we put RI :≤.
59
A rigorous definition of the predicate tautology is provided in a later chapter on
Predicate Logic. We construct it in the following steps.
1. We first define formally the notion of interpretation I of symbols of calL in
a set U 6= ∅ i.e. the model structure M = (U, I) for the predicate language L.
2. Then we define formally a notion ” a formula A of L a is true (valid) in
M = (U, I)”. We write it symbolically
M |= A
and call the model structure M = (U, I) a model for A.
3. We define a notion ”A is a predicate tautology” as follows.
Definition 2.8
For any formula A of predicate language L,
A is a predicate tautology (valid formula) if and only if M |= A for all
model structures M = (U, I) for L.
Definition 2.9
For any formula A of predicate language L,
A is not a predicate tautology if and only if there is a model structure
M = (U, I) for L, such that M 6|= A.
We call such model structure M a counter-model for A.
The definition 2.9 says: to prove that A is not a predicate tautology one has to
show a counter- model. It means one has to show a non-empty set U and define
an interpretation I, such that we can prove that AI is a false.
We use terms predicate tautology or valid formula instead of just saying a tau-
tology in order to distinguish tautologies belonging to two very different lan-
guages. For the same reason we usually reserve the symbol |= for propositional
case. Sometimes symbols |=p or |=f are used to denote predicate tautologies,
where ”p” stands for ”predicate” and ”f” stands ”first order”. The predicate
tautologies are also called laws of quantifiers and we will use both terms for
them.
Here are some examples of predicate tautologies and counter models for formulas
that are not tautologies.
For any formula A(x) with a free variable x:
|=p (∀x A(x) ⇒ ∃x A(x)). (2.53)
Observe that (2.53) represents an infinite number of formulas. It is a tautology
for any formula A(x) of L with a free variable x.
The inverse implication to (2.53) is not a predicate tautology.
60
6|=p (∃x A(x) ⇒ ∀x A(x)) (2.54)
To prove (2.54) we have to provide an example of a concrete formula A(x) and
construct a counter-model M = (U, I) for the formula F : (∃x A(x) ⇒ ∀x A(x)).
Let A(x) be an atomic formula P (x, c). We take as M = (N, I) for N set of
natural numbers and PI :<, cI : 3. The formula F becomes an obviously false
mathematical statement FI : (∃n∈N n < 3 ⇒ ∀n∈N n < 3).
Observe that we have to be very careful when we deal with quantifiers with
restricted domain. The most basic predicate tautology (2.53) fails when we
use the quantifiers with restricted domain.
Example 2.15
Show that
6|=p (∀B(x) A(x) ⇒ ∃B(x) A(x)). (2.55)
FI is a false because the statement n < 0 is false for all natural numbers and
F ⇒ B is a true implication for any logical value of B, so ∀n∈N (n < 0 ⇒ n > 0)
is a true statement and ∃n∈N (n < 0 ∩ n > 0) is obviously false.
Restricted quantifiers law corresponding to the predicate tautology (2.53)
is:
|=p (∀B(x) A(x) ⇒ (∃x B(x) ⇒ ∃B(x) A(x))). (2.57)
We remind that (2.57) means that corresponding proper formula of L obtained
by the restricted quantifiers transformations rules (8.18), (8.19) is a predicate
tautology, i.e.
.
Another basic predicate tautology called a dictum de omni law is: For any
formulas A(x) with a free variable x ∈ V AR,
61
|=p (∀x A(x) ⇒ A(y)), (2.59)
where y ∈ V AR and A(y) is a result of substitution of y for all free occurrences
of x in A(x) (if any) and y is free for x in A(x), what means that no occurrence
of a variable y becomes a bound occurrence in A(y). Restricted quantifiers law
corresponding to the dictum de omni law (2.59) is:
|=p (∀B(x) A(x) ⇒ (B(y) ⇒ A(y))), (2.60)
where y ∈ V AR satisfies the same condition as in (2.59).
Observe that we say A is restricted quantifiers law, or A is restricted quantifiers
tautology as a shorthand to formally saying that a formula obtained from A by
the transformations rules (8.18), (8.19) is a predicate tautology.
A more general version of (2.59) is:
|=p (∀x A(x) ⇒ A(t)), (2.61)
where t is a term and A(t) is a result of substitution of t for all free occurrences
of x in A(x) and t is free for x in A(x), what means that no occurrence of a
variable in t becomes a bound occurrence in A(t).
Here is another important tautology, called a generalization law.
|=p (A(x) ⇒ ∀x A(x)). (2.62)
The next important laws are the Distributivity Laws.
1. Distributivity of existential quantifier over conjunction holds only on one
direction, namely the following is a predicate tautology.
62
2. Distributivity of universal quantifier over disjunction holds only on one di-
rection, namely the following is a predicate tautology for any formulas A(x), B(x)
with a free variable x.
The inverse implication is not a predicate tautology, i.e.there are formulas A(x), B(x)
with a free variable x. such that
It means that we have to find a concrete formula A(x), B(x) ∈ F and a model
structure M = (U, I) that is a counter- model for the formula
Take M = (R, I) where R is the set of real numbers, and A(x), B(x) be atomic
formulas Q(x, c), R(x, c). We define QI :≥, RI :<, cI : 0. The formula F
becomes an obviously false mathematical statement
The most frequently used laws of quantifiers have a form of a logical equivalence,
symbolically written as ≡. This not a new logical connective. This is a very
useful symbol. It says that two formulas always have the same logical value,
hence it can be used in the same way we use the equality symbol =. Formally
we define it as follows.
Definition 2.10
For any formulas A, B ∈ F of the predicate language L,
Definability
For any formula A(x) ∈ F with a free variable x,
63
Renaming the Variables
Let A(x) be any formula with a free variable x and let y be a variable that does
not occur in A(x). Let A(x/y) be a result of replacement of each occurrence of
x by y, then the following holds.
Alternations of Quantifiers
Let A(x, y) be any formula with a free variables x and y.
We also define the notion of logical equivalence ≡ for the formulas of the —tex-
titpropositional language (definition 2.1) and its semantics.
Definition 2.11
For any formulas A, B ∈ F of the propositional language L,
64
Moreover, we prove that any substitution of propositional tautology by a for-
mulas of the predicate language is a predicate language tautology. The same
holds for the logical equivalence. In particular, we transform the propositional
Implication and Double Negation tautologies (2.48), (2.49) into the following
predicate equivalences.
For any formulas A, B of the predicate language L,
(A ⇒ B) ≡ (¬A ∪ B), (2.82)
¬¬A ≡ A (2.83)
We use (2.82) and (2.83) to prove the following De Morgan Laws for restricted
quantifiers.
Restricted De Morgan
For any formulas A(x), B(x) ∈ F with a free variable x,
¬∀B(x) A(x) ≡ ∃B(x) ¬A(x), ¬∃B(x) A(x) ≡ ∀B(x) ¬A(x). (2.84)
Here is a poof of first equality. The proof of the second one is similar and is left
as an exercise.
¬∀B(x) A(x) ≡ ¬∀x (B(x) ⇒ A(x)) ≡ ¬∀x (¬B(x)∪A(x)) ≡ ∃x ¬(¬B(x)∪A(x))
≡ ∃x (¬¬B(x) ∩ ¬A(x)) ≡ ∃x (B(x) ∩ ¬A(x)) ≡ ∃B(x) ¬A(x)).
If B is a formula such that B does not contain any free occurrence of x, then
the following logical equivalences hold for any formulas A(x), B(x), C(x).
∀C(x) (A(x) ∪ B) ≡ (∀C(x) A(x) ∪ B), (2.87)
65
∃C(x) (A(x) ∩ B) ≡ (∃C(x) A(x) ∩ B), (2.88)
∀C(x) (A(x) ⇒ B) ≡ (∃C(x) A(x) ⇒ B), (2.89)
∀C(x) (B ⇒ A(x)) ≡ (B ⇒ ∀C(x) A(x)). (2.90)
The proofs are similar to the proof of the restricted de Morgan Laws.
The similar generalization of the other Introduction and Elimination Laws (11.43),
(11.44), (11.47), (10.39) fails. We can easily follow Example 2.15 and construct
proper counter-models proving the following.
Example 2.16
The restricted quantifiers version of (11.43) is the following.
(a) If Mr. Smith is happy, Mrs. Smith is not happy, and if If Mr. Smith
is not happy, Mrs. Smith is not happy.
66
(b) If John doesn’t know logic, then if he knows logic, he was born in the
12th century.
(c) If from the fact that all sides of a triangle ABC are equal we can
deduce that all angles of the triangle ABC are equal and all angles
of the triangle ABC are not equal, then all sides of a triangle ABC
are equal.
(d) If it is not the fact that a line L is parallel to a line M or a line P is
not parallel the line M, then the line L is not parallel to the line M
or the line P is parallel the line M.
(e) If a number a is divisible by 3 and by 5, then from the fact that it is
not divisible by 3, we can deduce that it is also not divisible by 5.
2. For each of the following formulas write 3 corresponding natural language
sentences.
1. Determine which of the elements of S are, and which are not well
formed formulas (wff) of L = (A, F).
2. If A ∈ S is not a formula, i.e if A 6∈ F re-write it as a correct formula
and write in the natural language what it says.
4. Write a full definition of a propositional language that uses Hilbert set of
connectives. Give four examples of well form formulas of this language.
List next to them corresponding formulas of our propositional language
L.
5. Write a full definition of a propositional language L that uses Lukasiewicz
set of connectives. Give 4 examples of well form formulas of this language.
Give 4 examples of well form formulas of this language. List next to them
corresponding formulas of our propositional language L.
Propositional Semantics
1. Given a formula A: (((a ∩ b) ∪ ¬c) ⇒ b). Evaluate the logical value of A for
the following sets of logical values of its basic components, i.e. variables
a, b: 1. a=T, b=F, c=F and 2. a=F, b=T, c=T.
67
2. Given a formula A: (((a ⇒ ¬b) ∪ b) ⇒ a). Evaluate the logical value of A
for all possible logical values of its variables.
Propositional Tautologies
3. Show that ”If a number is divisible by 3 and by 5, then from the fact that
it is not divisible by 3, we can deduce that it is also not divisible by 5” is
always a true statement.
68
(b) If the temperature and air pressure remained constant, there was a
rain.The temperature did remain constant. Therefore, if there was a
rain then the air pressure did not remain constant.
(c) If a = 0 or b = 0, then ab = 0. But ab 6= 0. Hence a 6= 0 or b 6= 0.
(d) If a = 0 and b = 0, then ab = 0. But ab 6= 0. Hence a 6= 0 or b 6= 0.
A1 = R(x, y, g(c, x)), A2 = ∃xP (x, f (x, y)), A3 = ∃dR(x, y, g(c, d)),
A4 = ∀z(f (x, P (c, y)), A5 = ∃yP (x, f (c, y)) ∪ ∀yP (x, f (c, y)).
(a) Indicate whether they are, or are not well formed formulas of F. For
those which are not in F write a correct formula.
(b) For each correct, or corrected formula identify all components: con-
nectives, quantifiers, predicate and function symbols, and list all its
terms.
(c) For each formula identify its s free and bound variables. State which
are open and which are closed formulas (sentences), if any.
69
(a) Anyone who is lazy can’t learn logic.
(b) Some people are happy only if they sing.
(c) John likes everybody who does not like Mary.
(d) Everybody with green eyes likes John.
5. For each of the following formulas (some with restricted quantifiers) write
two corresponding natural language sentences.
Predicate Semantics
1. For each of the formulas and each model structure M indicate for what
values the formula is satisfied (if it contains free variables) or whether M
is its model or counter-model (if it is a closed formula. i.e. a sentence).
Formulas are:
(a) P (f (x, y), c)
(b) P (x, y) ⇒ P (y, x)
(c) ∀x∀y∀z((P (x, y) ∩ P (y, z)) ⇒ P (x, z))
(a) Model structure is M = (N, I), for N set of natural numbers and
PI :=, gI : +, fI : multiplication, and cI : 0, dI : 1.
Formulas are:
A1 : ∀x∃y(P (x, g(y, y)) ∪ P (x, g(g(y, y), d)))
A2 : ∀x∀y(P (f (x, y), c) ⇒ (P (x, c) ∪ P (y, c)))
A3 : ∃y P (g(y, y), d)
(b) Model structure is M = (Z, I), for Z set of integers and PI :=, fI : +,
70
Formulas are:
A1 : ∀x∀y P (f (x, y), f (y, x)))
A2 : ∀x∀y P (f (x, y), y)
3. Prove that the following formulas are not predicate tautologies, i.e. find
for each of them a counter-model M.
5. Use proper Equational Laws for Quantifiers to prove that the following
Restricted Introduction and Elimination Laws hold for any formulas A(x),
B(x), C(x), and B, where B does not contain any free occurrence of x.
(a) ∃C(x) (A(x) ∩ B) ≡ (∃C(x) A(x) ∩ B)
(b) ∀C(x) (A(x) ⇒ B) ≡ (∃C(x) A(x) ⇒ B)
(c) ∀C(x) (B ⇒ A(x)) ≡ (B ⇒ ∀C(x) A(x))
71
72
Chapter 3
Propositional Semantics:
Classical and Many Valued
73
and even two languages with only one binary propositional connectives, denoted
usually by ↑ and ↓, respectively, i.e languages L{↑} , L{↓} all share the same
semantics characteristic for the classical propositional logic.
The connectives have well established symbols and names, even if their semantics
can differ. We use names negation, conjunction, disjunction, implication and
equivalence (biconditional) for ¬, ∩, ∪, ⇒, ⇔, respectively. The connective ↑ is
called alternative negation and A ↑ B reads: not both A and B. The connective
↓ is called joint negation and A ↓ B reads: neither A nor B.
Other most common propositional connectives are probably modal connectives
of possibility and necessity. Standard modal symbols are for necessity and
♦ for possibility. We will also use symbols C and I for modal connectives of
possibility and necessity, respectively.
A formula CA, or ♦A reads: it is possible that A , A is possible, and a formula
IA, or A reads: it is necessary that A, A is necessary.
A motivation for notation C and I arises from topological semantics for modal
S4 and S5 logics. C becomes equivalent to a set closure operation, and I becomes
equivalent to a set interior operation.
The symbols ♦, C and , I are not the only symbols used for modal connectives.
Other symbols include N for necessity and P for possibility. There is also
a variety of modal logics created by computer scientists, all with their set of
symbols and motivations for their use and their semantics. The modal logics
extend the classical logic and hence their language is for example L{,♦,¬,∩,∪,⇒} .
Knowledge logics also extend the classical logic. Their languages add to the
classical connectives a new knowledge connective, often denoted denoted by K.
The formula KA reads: it is known that A , A is known. The language of a
knowledge logic is for example L{K, ¬,∩,∪,⇒} .
Autoepistemic logics use a believe connective, often denoted by B. The formula
BA reads: it is believed that A. They also extend the classical logic and hence
their language is L{B, ¬,∩,∪,⇒} .
Temporal logics add temporal connectives to the set of classical propositional
connectives. For example some of them use connectives (operators, as they are
often called) F, P, G, and H to denote the following intuitive readings. F A reads
A is true at some future time, P A reads A was true at some past time, GA
reads A will be true at all future times, and HA reads A has always been true
in the past. In order to take account of this variation of truth-values over time,
some formal semantics were created, and many more will be created.
It is possible to create connectives with more then one or two arguments, but
we allow here only one and two argument connectives, as logics which will be
discussed here use only those two kind of connectives.
We adopt the following definition, common to all propositional languages con-
74
sidered in our propositional logics investigations.
1. Alphabet A
The alphabet A = V AR ∪ CON ∪ P AR, where VAR, CON, PAR are all disjoint
sets and VAR, CON are non-empty sets. VAR is countably infinite and is called
a set of propositional variables; we denote elements of VAR by a, b, c, ... etc,
(with indices if necessary).
CON is a finite set of propositional connectives, P AR is a set of auxil-
iary symbols. We assume that P AR 6= ∅ and contains two elements (, ) called
parentheses, i.e. P AR = {(, )}. The set PAR may be empty, for example of a
case of Polish notation, but we assume that it contains two parenthesis as to
make the reading of formulas more natural and uniform.
2. Set F of formulas
The set F is built recursively from the elements of the alphabet A, i.e. F ⊆ A∗ ,
where A∗ is the set of all finite sequences (strings) form from elements of A and
is defined as follows.
The set F of all formulas of a propositional language LCON is the smallest
set, such that the following conditions hold:
(1) V AR ⊆ F;
75
(2) if A ∈ F, 5 ∈ C1 i.e 5 is an one argument connective, then 5A ∈ F;
The elements of the set V AR ⊆ F are called atomic formulas. The set F is
also called a set of all well formed formulas (wff) of the language LCON .
The alphabet A is countably infinite and consequently the set A∗ of all finite
sequences of elements of A is also countably infinite. By definition, F ⊆ A∗ ,
hence the set F is also countably infinite. We state as separate fact.
Fact 3.1 For any propositional language LCON = (A, F), the set F of for-
mulas is countably infinite. We hence consider here only infinitely countable
languages.
Observation 3.1
When defining a language LCON we choose not only the propositional connec-
tives but also the symbols denoting them.
For example, L1 = L{¬} and L2 = L{∼} are two different propositional lan-
guages both with negation as the only connective.
The choice of appropriate well established symbols for logical connectives de-
pends on a personal preferences of books’ authors and creators of different logics.
One can find a variety of them in the literature. We presented some historical
choices in the chapter 2.
Example 3.1
Let L1 = L{¬} and L2 = L{∼} . The formulas of both languages L1 , L2 are
propositional variables or multiple negations of of a propositional variable.
The strings a, ¬b, ¬¬b, ¬¬¬a are well formed formulas of L1 . The corresponding
formulas of L2 are a, ∼ b, ∼∼ b, ∼∼∼ a.
Observe that the strings (¬a), ¬, ¬(¬a), ¬(a), (∼ a), ¬, ∼ (∼ a) ∼ (a) are not
well formed formulas of neither of the languages L1 , L2 .
We adopt the general definition of the set F of formulas of LCON to for example
the language L{∼} as follows.
Example 3.2
76
The set F of all formulas of a propositional language L{∼} is the smallest
set, such that the following conditions hold:
(1) V AR ⊆ F (atomic formulas);
(2) if A ∈ F, then ∼ A ∈ F.
Example 3.3
Consider now LCON for the set of connectives CON = {¬}∪{⇒}, where ¬ ∈ C1
and ⇒ ∈ C2 . It means that we defined a language L = L{¬,⇒} .
By the initial recursive step we get for any a ∈ V AR, a ∈ F. By the recursive
step and its repetition we get for example that ¬a ∈ F, ¬¬a ∈ F, ¬¬¬a ∈ F, ...
etc., i.e. get all formulas from the the example 5.10 language L1 . But also we
also get that (a ⇒ a), (a ⇒ b), ¬(a ⇒ b), (¬a ⇒ b), ¬((a ⇒ a) ⇒ ¬(a ⇒ b))....
etc. are all in F and infinitely many others.
Observe that (¬(a ⇒ b))), a ⇒ b, (a ⇒) are not in F.
Example 3.4
Consider L = LCON for C1 = {¬, P, N }, C2 = {⇒}. If we understand P , N
as a possibility and necessity connectives, the obtained language is called a
modal language with only negation as non-modal connective.
The set of formulas F of L contains all formulas from example 5.11, but also
formulas N a, ¬P a, P ¬a, (N ¬b ⇒ P a), ¬P ¬a, ((N ¬b ⇒ P a) ⇒ b), .... etc.
We adopt the general definition of the set F of formulas of LCON to for example
the modal language L{¬,P,N,⇒} as follows.
Example 3.5
The set F of all formulas of a propositional language L{¬,P,N,⇒} is the
smallest set, such that the following conditions hold:
(1) V AR ⊆ F (atomic formulas);
(2) if A ∈ F, then ¬A, P A, N A ∈ F;
(3) if A, B ∈ F, then (A ⇒ B) ∈ F.
77
For any connectives 5 ∈ C1 and ◦ ∈ C2 ,
5 is called a main connective of 5A ∈ F and
◦ is a main connective of (B ◦ C) ∈ F.
Observe that it follows directly from the definition of the set of formulas that
for any formula C ∈ F, exactly one of the following holds: C is atomic, or there
is a unique formula A and a unique unary connective 5 ∈ C1 , such that C
is of the form 5A, or here are unique formulas A and B and a unique binary
connective ◦ ∈ C2 , such that C is (A ◦ B). We have hence proved the following.
Observation 3.2
For any formula A ∈ F, A is atomic or has a unique main connective.
Example 3.6
The main connective of (a ⇒ ¬N b) is ⇒. The main connective of N (a ⇒
¬b) is N . The main connective of ¬(a ⇒ ¬b) is ¬ The main connective of
(¬a ∪ ¬(a ⇒ b)) is ∪.
Definition 3.3
We define a notion of direct a direct sub-formula as follows: 1. Atomic
formulas have no direct sub-formulas. 2. A is a direct sub-formula of a formula
5A, where 5 is any unary connective. 3. A, B are direct sub-formulas of a
formula (A ◦ B) where ◦ is any binary connective.
Observation 3.3
For any formula A, A is atomic or has exactly one or two direct sub-formulas
depending on its main connective being unary or binary, respectively.
Example 3.7
The formula (¬a ∪ ¬(a ⇒ b)) has exactly ¬a and ¬(a ⇒ b) as direct sub-
formulas.
Definition 3.4
We define a notion of a sub-formula of a given formula in two steps. 1. For
any formulas A and B, A is a proper sub-formula of B if there is sequence of
formulas, beginning with A, ending with B, and in which each term is a direct
sub-formula of the next. 2. A sub-formula of a given formula A is any proper
sub-formula of A, or A itself.
78
The formula (¬a ∪ ¬(a ⇒ b)) has ¬a and ¬(a ⇒ b) as direct sub-formula. The
formulas ¬a and ¬(a ⇒ b) have a and (a ⇒ b) as their direct sub-formulas,
respectively. The formulas ¬a, ¬(a ⇒ b), a and (a ⇒ b) are all proper sub-
formulas of the formula (¬a ∪ ¬(a ⇒ b)) itself. Atomic formulas a and b are
direct sub-formulas of (a ⇒ b). Atomic formula b is a proper sub-formula of ¬b.
Example 3.8
The set of all sub-formulas of
The degree of (¬a ∪ ¬(a ⇒ b)) is 4. The degree of ¬(a ⇒ b)) is 2. The degree
of ¬a is 1. The degree of a is 0.
Note that the degree of any proper sub-formula of A must be one less than the
degree of A. This is the central fact upon mathematical induction arguments
are based. Proofs of properties formulas are usually carried by mathematical
induction on their degrees.
Example 3.9
Given a formula A : (¬I¬a ⇒ (¬Ca ∪ (Ia ⇒ ¬Ib))).
1. The language to which A belongs is a modal language L{¬,C,C,∪,∩,⇒} with
the possibility connective C and necessity connective C. Both of them are one
argument connectives.
2. The main connective of A is ⇒, the degree of A is 11.
3. All sub-formulas of A of the degree 0 are the atomic formulas a, b. All
sub-formulas of A of the degree 1 are: ¬a, Ca, Ia, Ib.
CON = C0 ∪ C1 ∪ C2 .
79
The definition of the set F of all formulas of the language LCON contains now
an additional recursive step and goes as follows.
The set F of all formulas of the language LCON with propositional constants is
the smallest set built from the signs of the alphabet A, i.e. F ⊆ A∗ , such that
the following conditions hold:
Example 3.10
Let L = L{T,¬,∩} , i.e. C0 = {V }. Atomic formulas of L are all variables and
the symbol T .
The language admits formulas that involve the symbol T like T, ¬T , (T ∩ a),
(¬a ∩ ¬T ), ¬(b ∩ T ), etc... We might interpret the symbol T as a symbol of truth
(statement that is always true).
Here are some exercises and examples dealing with the formal definition of
propositional languages, syntactical correctness, and their expressiveness.
Exercise 3.1
Given a language L = L{¬, C,I,∪,∩,⇒} and the following set S.
Determine which of the elements of S are, and which are not well formed for-
mulas of L. If A ∈ S is not a correct formula write its corrected version. For
each correct or corrected formula determine its main connective, its degree and
write what it says in the natural language.
Solution
1. C¬a ⇒ (a ∪ b) is not a well formed formula. The corrected formula is
(C¬a ⇒ (a ∪ b)). Its main connective is ⇒ and the degree is 4. The corrected
formula says: If negation of a is possible, then we have a or b.
Another corrected formula is C(¬a ⇒ (a ∪ b)). Its main connective is C, the
degree is 4. The corrected formula says: It is possible that not a implies a or b.
80
2. (C(¬a ⇒ (a ∪ b))) is not a well formed formula. The correct formula is
C(¬a ⇒ (a ∪ b)). The main connective is C, the degree is 4. The formula
C(¬a ⇒ (a ∪ b)) says: It is possible that not a implies a or b .
3. The formula C¬(a ⇒ (a ∪ b)) is a correct formula. The main connective is
C, the degree is 4. The formula says: the negation of the fact that a implies a
or b is possible.
Exercise 3.2
Given a set S of formulas:
Define a formal language LCON to which to which all formulas in S belong, i.e.
a language determined by the set S.
Solution
Any propositional language LCON is determined by its set of connectives. The
connectives appearing in the formulas of the set S are: ⇒, ¬b, , ♦ and ∪. Hence
the required language is L{¬,,♦,∪,⇒} .
Exercise 3.3
Write down a set S1 all sub-formulas of the ♦((a ∪ ¬a) ∩ b), a set S2 all proper
sub-formulas of ¬(a ⇒ (b ⇒)).
Solution
The set S1 of all sub-formulas of ♦((a ∪ ¬a) ∩ b) is
S1 = {♦((a ∪ ¬a) ∩ b), ((a ∪ ¬a) ∩ b), (a ∪ ¬a), ¬a, b, a}
a, b are atomic sub-formulas, and ♦((a ∪ ¬a) ∩ b) is not a proper sub-formula.
The set S2 of all proper sub-formulas of ¬(a ⇒ (b ⇒ c)) is
S2 = {(a ⇒ (b ⇒ c)), (b ⇒ c), a, b, c}.
Exercise 3.4
Write the following natural language statement S:
”From the fact that it is possible that Anne is not a boy we deduce that it is not
possible that Anne is not a boy or, if it is possible that Anne is not a boy, then
it is not necessary that Anne is pretty.”
in the following two ways.
1. As a formula A1 ∈ F1 of a language L{¬, , ♦, ∩, ∪, ⇒} .
81
Solution
1. We translate the statement S into a formula A1 of the modal language
L{¬, , ♦, ∩, ∪, ⇒} as follows.
Propositional variables are: a, b. The variable a denotes statement Anne is a
boy and b denotes a statement Anne is pretty.
Propositional modal connectives are: , ♦. The connective ♦ reads it is
possible that, and reads it is necessary that.
Translation: the formula A1 is (♦¬a ⇒ (¬♦¬a ∪ (♦¬a ⇒ ¬b))).
2. We translate our statement into a formula A2 of the language L{¬, ∩, ∪, ⇒}
as follows.
Propositional variables are: a, b. The variable a denotes statement it is possible
that Anne is not a boy and b denotes a statement it is necessary that Anne is
pretty. Translation: the formula A2 is (a ⇒ (¬a ∪ (a ⇒ ¬b))).
Exercise 3.5
Write the following natural language statement S:
”For all natural numbers n ∈ N the following implication holds: if n < 0, then
there is a natural number m, such that it is possible that n + m < 0, or it is
not possible that there is a natural number m, such that m > 0”
in the following two ways.
1. As a formula A1 of a language L{¬, ∩, ∪, ⇒} .
Solution
1. We translate the statement S into a formula A1 of the language L{¬, ∩, ∪, ⇒}
as follows.
Propositional variables are: a, b. The variable a denotes statement For all
natural numbers n ∈ N the following implication holds: if n < 0, then there is
a natural number m, such that it is possible that n + m < 0. The variable b
denotes statement it is not possible that there is a natural number m, such that
m > 0. Translation: the formula A1 is (a ∪ ¬b).
2. We translate the statement S into a formula A2 of a language L{¬, , ♦, ∩, ∪, ⇒}
as follows. Propositional variables are: a, b. The variable a denotes statement
For all natural numbers n ∈ N the following implication holds: if n < 0, then
there is a natural number m, such that it is possible that n + m < 0. The
variable b denotes statement there is a natural number m, such that m > 0.
Translation: the formula A2 is (a ∪ ¬♦b).
82
3.2 Extensional Semantics M
Given a propositional language LCON , the symbols for its connectives always
have some intuitive meaning. A formal definition of the meaning of these sym-
bols is called a semantics for the language LCON . A given language can have
different semantics but we always define them in order to single out special for-
mulas of the language, called tautologies, i.e. formulas of the language that is
always true under the given semantics.
We introduced in Chapter 2 a notion of a classical propositional semantics, dis-
cussed its motivation and underlying assumptions. The assumption was that
we consider only two logical values. The other one was that all classical propo-
sitional connectives are extensional. We have also observed that in everyday
language there are expressions such as ”I believe that”, ”it is possible that”,
” certainly”, etc.... and they are represented by some propositional connectives
which are not extensional. Non-extensional connectives do not play any role
in mathematics and so are not discussed in classical logic and will be studied
separately.
The extensional connectives are defined intuitively as such that the logical value
of the formulas form by means of these connectives and certain given formulas
depends only on the logical value(s) of the given formulas. We adopt a following
formal definition of extensional connectives for a propositional language L and
of an extensional semantic for L.
83
tensional.
Defining a semantics for a given propositional language means more then defin-
ing propositional connectives. The ultimate goal of any semantics is to define
the notion of its own tautology. In order to define which formulas of LCON we
want to to be tautologies under a given semantics M we assume that the set LV
of logical values of M always has a distinguished logical value, often denoted
by T for ”absolute” truth. We also can distinguish, and often we do, another
special value F representing ”absolute” falsehood. We will use these symbols
T, F. We may also use other symbols like 1, 0 or others. The value T serves to
define a notion of a tautology (as a formula always ”true”).
Extensional semantics share not only the similar pattern of defining their connec-
tives (definition 3.6), but also the method of defining the notion of a tautology.
We hence define a general notion of an extensional semantics (definition 3.7) as
sequence of steps leading to the definition of a tautology. Here are the steps.
Step1: we define all connectives of M as specified by definition 3.6.
Step 2: we define the main component of the definition of a tautology, namely
a function v that assigns to any formula A ∈ F its logical value from VL. It is
often called a truth assignment and we will use this name.
Step 3: given a truth assignment v and a formula A ∈ F, we define what does
it mean that v satisfies A, i.e. that v is a model for A under semantics M.
Step 4: we define a notion of tautology as follows: A is a tautology under seman-
tics M if and only if all truth assignments v satisfy A, i.e. all truth assignments
v are models for A.
84
Definition 3.7 (Extensional Semantics)
A formal definition of an extensional semantics M for a given language LCON
consists of specifying the following steps defining its main components.
Step 1: we define a set LV of logical values and its distinguished value T, and
define all connectives of LCON to be extensional;
Step 2: we define notion of a truth assignment and its extension;
Step 3: we define notions of satisfaction, model, counter model;
Step 4: we define notion of a tautology under the semantics M.
What differs one semantics from the other is the choice of the set LV of logical
values and definition of the the connectives of LCON , i.e. the components
defined in the Step1. The definitions for the Steps 2 and 3, 4 are modification
of the definitions established for the classical case and they are as follows.
Step 1: we follow the definition 3.6 to define the connectives of M.
Step 2 : we define a function called truth assignment and its extension in terms
of the propositional connectives as defined in the Step 1. We use the term M
truth assignment and M truth extension to stress that it is defined relatively to
a given semantics M.
85
We use names ”M truth assignment”, ”M truth extension” to stress that we
define them for the set of logical values of M and moreover, that the extension
of v connects the formulas of the language with the connectives as defined by
the semantics M.
|=M A.
86
We also say that A is M tautology if and only if all truth assignments v are
M models for A.
Observe that directly from definition 3.11 we get the following equivalent form
of the definition 3.12.
Definition 3.13
For any formula A ∈ F,
A is a M tautology if and only if v ∗ (A) = T , for all truth assignments v,
v : V AR −→ LV .
Obviously, when we develop a logic by defining its semantics we want the se-
mantics to be such that the logic has a non empty set of its tautologies. We
stress that fact by putting it in a form of the following definition.
Definition 3.14
Given a language LCON and its extensional semantics M (definition 3.7), we
say that the semantics M is well defined if and only if its set MT of all
tautologies (3.2) is non empty, i.e. when
MT 6= ∅ (3.3)
We follow the definitions and pattens established here first in section 3.3. We
use them to define and discuss in details the classical propositional semantics.
Definitions and short discussions of some of the many-valued semantics follow
next in section 3.5. Many valued logics had their beginning in the work of
Lukasiewicz (1920). He was the first to define a 3- valued extensional semantics
for a language L{¬,∩,∪,⇒} of classical logic, and called it a three valued logic
for short. The other logics, now of historical value followed and we will discuss
some of them. In particular we present a Heyting 3-valued semantics as an
introduction to the definition and discussion of first ever semantics for the intu-
itionistic logic and some modal logics. It was proposed by J.C.C McKinsey and
A. Tarski in 1946-48 in a form of cylindrical algebras, now called pseudo-boolean
algebras, or Heyting algebras. The semantics in a form of abstract algebras are
called algebraic models for logics. It became a separate field of modern logic.
The algebraic models are generalization of the extensional semantics, hence the
importance of this section. It can me treated as an introduction to algebraic
models for logics. It will be discussed again in chapter7.
87
3.3 Classical Semantics
Definition of connectives
¬ T F ∩ T F ∪ T F
F T T T F T T T
F F F F T F
88
⇒ T F ⇔ T F
T T F T T F
F T T F F T
Exercise 3.6
Prove that the above connectives formulas are correct, i.e. that they define the
same classical connectives as defined in Step 1.
Solution
This is a problem of proving equality of functions that are given the same names.
We have to show that the use of the same names: ¬, ∪, ∩, ⇒, ⇔ for them is
justified. The equality of functions is defined as follows.
Definition 3.15
Given two sets A, B and functions f, g, such that f : A −→ B and g : A −→ B.
We say that the functions f, g are equal and write it f = g if and only if f(x)
= g(x) for all elements x ∈ A.
The negation definition is the same in both cases. We prove that the two
conjunctions and two disjunctions functions are the equal by comparing both
89
definitions (3.4) and (3.5). We verify now the correctness of the implication
function formula. Consider two functions ⇒: {T, F } × {T, F } −→ {T, F } and
h : {T, F } × {T, F } −→ {T, F }, where ⇒ is the classical implication defined
by definition (3.4) and h is defined by the definition (3.5), i.e. by the for-
mula h(x, y) = ∪(¬x, y). Observe that we have already proved that functions
∪ and ¬ are equal in both cases. We prove that ⇒ = h by evaluating that
⇒ (x, y) = h(x, y) = ∪(¬x, y), for all (x, y) ∈ {T, F } × {T, F } as as follows.
T ⇒ T = T and h(T, T ) = ¬T ∪ T = F ∪ T = T yes.
T ⇒ F = F and h(T, F ) = ¬T ∪ F = F ∪ F = F yes.
F ⇒ F = T and h(F, F ) = ¬F ∪ F = T ∪ F = T yes.
F ⇒ T = T and h(F, T ) = ¬F ∪ T = T ∪ T = T yes.
This proves the correctness of the implication formula ⇒ (x, y) = ∪(¬x, y). We
write it as x ⇒ y = ¬x ∪ y and call it a formula defining implication in
terms of disjunction and negation. We verify the correctness of the equivalence
formula ⇔ (x, y) = ∪(⇒ (x, y), ⇒ (y, x)) in a similar way.
We have just proved in Exercise 3.6 that the implication ⇒ is definable in terms
of ∪ and ¬ under classical semantics as it is a composition of ∪ and ¬ defined
by the formula ⇒ (x, y) = ∪(¬x, y). The classical equivalence is definable in
terms of ⇒ and ∩ by the formula ⇔ (x, y) = ∪(⇒ (x, y), ⇒ (y, x)).
90
Proving the property of functional dependency under a given semantics M con-
sists of identifying a proper subset CON0 of the set CON of connectives, such
that each connective ◦ ∈ CON − CON0 is definable (definition 3.16) in terms
of connectives from CON0 . This is usually a difficult, and often impossible task
for many semantic. We prove now that it holds in the classical case.
Theorem 3.1
The set of connectives of the language L{¬, ∪, ∩, ⇒, ⇔} is functionally depen-
dent under the classical semantics.
Proof
Let’s take a set {¬, ∪}. We have already proved in Exercise 3.6 that the impli-
cation ⇒ and is definable in terms of ∪ and ¬ by the formula x ⇒ y = ¬x ∪ y.
The conjunction is defined by easy verification, similar to the one in Exercise
3.6, by a formula x ∩ y = ¬(¬x ∪ ¬y). By Exercise 3.6, the equivalence formula
is definable in terms of ⇒ and ∩ by the formula x ⇔ y = (x ⇒ y) ∩ (y ⇒ x).
The final formula for for the equivalence is x ⇔ y = (¬x ∪ y) ∩ (¬y ∪ x).
There are many ways to prove this theorem, it means there are many ways to
choose a proper subset CON0 of the set {¬, ∪, ∩, ⇒, ⇔} that defines all other
connectives. Here are the choices.
Proof
We list all required definability formulas, including the formulas developed in
the proof of Theorem 3.1. An easy verification of their correctness is left as an
exercise.
1. Definability in terms of ⇒ and ¬.
x ∩ y = ¬(x ⇒ ¬y), x ∪ y = ¬x ⇒ y, x ⇔ y = ¬((x ⇒ y) ⇒ ¬(y ⇒ x)).
2. Definability in terms of ∩ and ¬.
x ∪ y = ¬(¬x ∩ ¬y), x ⇒ y = ¬(x ∩ ¬y), x ⇔ y = ¬(x ∩ ¬y) ∩ ¬(y ∩ ¬x).
3. Definability in terms of ∪ and ¬.
x ⇒ y = ¬x ∪ y, x ∩ y = ¬(¬x ∪ ¬y) x ⇔ y = (¬x ∪ y) ∩ (¬y ∪ x).
There are two other important classical binary connectives denoted by ↑ and
↓. The connective ↑ was discovered in 1913 by H.M. Sheffer, who called it
alternative negation. Now it is often called a Sheffer’s connective. A formula
(A ↑ B) reads: not both A and B. The connective ↓ was discovered in 1920 by
J. Lukasiewicz and named joint negation. The formula (A ↓ B) reads: neither
A nor B. They are defined as follows.
91
Alternative Negation is a function ↑: {T, F } × {T, F } −→ {T, F } such that
T ↑ T = F, T ↑ F = T, F ↑ T = T, F ↑ F = T .
Joint Negation is a function ↓: {T, F } × {T, F } −→ {T, F } such that
T ↓ T = F, T ↓ F = F, F ↓ T = F, F ↓ F = T .
↑ T F ↓ T F
T F T T F F
F T T F F T
Theorem 3.3
All connectives of a language L{¬, ∪, ∩, ⇒, ⇔, ↑, ↓} are definable in terms of ↑,
and also separately in terms of ↓.
Proof
Definability formulas of ¬ and ∩ in terms of ↑ are the following.
¬x = x ↑ x, x ∩ y = (x ↑ y) ↑ (x ↑ y) (3.6)
92
is building semantics for any of a non-classical logic.
We define now and examine the components in the Step 2 of the definition 3.7.
We start with the basic notion of the truth assignment. We adopt the exten-
sional semantics M definition 3.8 to the classical case as follows.
The function v defined above is called the truth assignment because it can be
thought as an assignment to each variable (which represents a logical sentence)
its logical value of T(ruth) of F(alse). Observe that the domain of the truth
assignment is the set of propositional variables, i.e. the truth assignment is
defined only for atomic formulas.
We now extend the truth assignment v from the set of atomic formulas to the
set of all formulas F in order define formally the assignment of a logical value
to any formula A ∈ F.
The definition of the truth extension of the truth assignment v to the set F
follows the definition 3.8 for the extensional semantics M .
The symbols on the left-hand side of the equations represent the connectives in
their natural language meaning. The symbols on the right-hand side represent
93
connectives in their classical semantics meaning defined by the classical con-
nectives defined by the classical Truth Tables.
Observe that we did not specify v(x) of any x ∈ V AR−{a, b}, as these values do
not influence the computation of the logical value of the formula A. We say: ”v
such that” as we consider its values for the variables a and b only. Nevertheless,
the domain of the truth assignment v is always is the set of all variables VAR
and we have to remember that.
94
We define now and examine the components in Step 3 of the definition 3.7.
We adopt the extensional semantics M definitions 3.10, 3.11, and 3.12 to the
classical case as follows.
Exercise 3.7
Let A be a formula ((a ⇒ b) ∪ ¬a)) and v be a truth assignment
v : V AR −→ {T, F }, such that v(a) = T, v(b) = F , and v(x) = F for all
x ∈ V AR − {a, b}. Show that v 6|= ((a ⇒ b) ∪ ¬a)).
Short-hand Evaluation
Given any formula A ∈ F and any truth assignment v : V AR −→ {T, F }.
1. We write the value of v only for variables appearing in the formula in A.
In our case we write: a = T , b = F for v(a) = T, v(b) = F .
2. Replace all variabes in A by their respective logical values.
In our case we replace a by T and b by F in the formula A ((a ⇒ b) ∪ ¬a)). We
get an equation ((T ⇒ F ) ∪ ¬T ).
3. Use use the connectives definition, in this case the definitionTTables to eval-
uate the logical value of the equation obtained in the step 2.
In our case we evaluate ((T ⇒ F ) ∪ ¬T ) = (F ∪ F ) = F .
4. Write your answer in one of the forms: v |= A, v 6|= A or ”v satisfies A”, ” v
falsifies A”
95
In our case v falsifies A and write v 6|= ((a ⇒ b) ∪ ¬a)).
Example 3.11
Let A be a formula ((a ∩ ¬b) ∪ ¬c) and v be a truth assignment v : V AR −→
{T, F }, such that v(a) = T, v(b) = F, v(c) = T , and v(x) = T for all
x ∈ V AR − {a, b, c}. Using the the short-hand notation we get ((T ∩ ¬F ) ∪
¬T ) = ((T ∩ T ) ∪ F ) = (T ∪ F ) = T . It proves that v satisfies the formula A
and we write v |= ((a ∩ ¬b) ∪ ¬c).
The truth assignment from the Example 3.11 is a model for the formula
((a∩¬b)∪¬c) and the truth assignment from the Exercise 3.7 is a counter-model
for the formula ((a ⇒ b) ∪ ¬a)).
The second uses the notion of satisfaction and model and the fact that in any
extensional semantic the notions ” v satisfies A” and ”v is a model for A” are
interchangeable. It is stated as follows.
We write symbolically
|= A
for the statement ”A is a tautology”.
96
Remark 3.1
We use the symbol |= A only for classical tautology. For all other extensional
semantics M we must use the symbol |=M A and say ” A is a tautology under
a semantics M, or to say in short ”A is a M semantics tautology”.
We usually use the definition 3.24 to express that a formula in not a tautology,
i.e. we say that a formula is not a tautology if it has a counter model. To stress
it we put it in a form of a following definition.
Definition 3.25
For any formula A ∈ F,
A is not a tautology if and only if A has a counter model;
i.e. when there is a truth assignment v : V AR −→ {T, F }, such that v 6|= A.
6|= A.
This ends the formal definition of classical semantics that follows the pattern
for extensional semantics established in the definition 3.7.
There is a large number of basic and important tautologies listed and discussed
in Chapter 2. We assume that the reader is familiar, or will familiarize with
them if needed. We will refer to them and use them within our book. Chapter
2 also provides the motivation for classical approach to definition of tautolo-
gies as ways of describing correct rules of our mathematical reasoning. It also
contains an informal definition of classical semantics and discusses a tautology
verification method. We have just defined formally the classical semantics. Our
goal now is to prove formally that the notion of classical tautology is decidable
(Theorem 3.9) and to prove correctness of the tautology verification method
presented in Chapter 2. Moreover we present here other basic tautology verifi-
cation methods and prove their correctness.
97
definition 3.23 we have to examine all truth assignments v : V AR −→ {T, F }.
If they all evaluate to T, we proved that |= A. If at least one evaluates to F, we
found a counter model and proved 6|= A. The verification process is decidable,
if the we have only a finite number of v to consider. So now all we have to
do is to count how many truth assignments there are, i.e. how many there are
functions that map the set V AR of propositional variables into the set {T, F } of
logical values. In order to do so we need to introduce some standard notations
and some known facts. For a given set X, we denote by |X| the cardinality of
X. In a case of a finite set, it is called a number of elements of the set. We
write |X| = n to denote that X has n elements, for n ∈ N . We have a special
names and notations to denote the cardinalities of infinite set. In particular
we write |X| = ℵ0 and say ” cardinality of X is aleph zero,” for any countably
infinite set X. We write |X| = C and say ” cardinality of X is continuum” for
any uncountable set X that has the same cardinality as Real numbers.
Definition 3.26
For any A ∈ F, let V ARA be a set of all propositional variables appearing
in A. Any function vA : V ARA −→ {T, F }, is called a truth assignment
restricted to A.
Example 3.12
Let A = ((a ⇒ ¬b) ∪ ¬c). The set of variables appearing in A is V ARA =
{a, b, c}. The truth assignment restricted to A is any function vA : {a, b, c} −→
{T, F }.
Definition 3.27
Given a formula A ∈ F and a set V ARA of all propositional variables appearing
in A. Any function vA : V ARA −→ {T, F }, such that v |= A ( v 6|= A) is
called a restricted model (counter model) for A.
98
We use the following particular case of Theorem 3.4 to count, for any formula
A, possible truth assignment restricted to A, i.e. all possible restricted models
and counter models for A.
99
Assume vA |= A for all vA : V ARA −→ {T, F }. Take any v : V AR −→ {T, F },
as V ARA ⊆ V AR, any v : V AR −→ {T, F } is an extersion of some vA , i.e.
v(a) = vA (a) for all a ∈ V ARA . By Truth Extension Definition 3.19 we get
that v ∗ (A) = vA ∗ (A) = T and v |= A. This ends the proof.
Directly from Theorem 3.7 and the above Theorem 3.8 we get the proof of
the correctness and decidability of the Truth Tables Method, and hence the
decidability of the notion of classical propositional tautology.
The table (3.9) shows that all w : V ARA −→ {T, F } are restricted models for
A and hence by Theorem 3.9 we proved that |= (a ⇒ (a ∪ b)) and T 6= ∅.
100
Observe that the table (3.9) proves that the formula 6|= ((a ⇒ ¬b) ∪ ¬c).
Moreover we have proved that the condition (3.3) of the definition 3.14 is fulfilled
and the classical semantics is well defined. We put it as a separate statement.
Fact 3.2
The classical semantics is well defined.
The complexity of the truth table methods grows exponentially. Impossible for
humans to handle formulas with more then few variables, and cumbersome for
computers for formulas with a great number of variables, In practice, if we need,
we use often much shorter and more elegant tautology verification methods pre-
sented below.
Exercise 3.8
Follow the Proof by Contradiction Method and examine whether
|= (a ⇒ (a ∪ b)).
Solution
We use a short-hand notation.
Assume that 6|= (a ⇒ (a ∪ b)). It means that (a ⇒ (a ∪ b)) = F for some truth
assignment v. By definition of implication ⇒ we have that
(a ⇒ (a ∪ b)) = F if and only if a = T and (a ∪ b) = F.
From a = T and (a ∪ b) = F we get (T ∪ b) = F . This is a contradiction with
the definition of disjunction ∪. Hence we proved |= (a ⇒ (a ∪ b)).
Exercise 3.9
Use the Proof by Contradiction Method to decide whether
|= ((a ∪ b) ⇒ a).
101
and we get that the truth assignment v is such that v ∗ ((a ∪ b)) ⇒ v(a) = F .
By definition implication ⇒ we have that v ∗ ((a ∪ b)) ⇒ v(a) = F if and only
if v(a) ∪ v(b) = T and (a) = F . From (a) = F and v(a) ∪ v(b) = T we get
that F ∪ v(b) = T . This is possible for any v : V AR −→ {T, F }, such that
v(b) = T . This proves that any truth assignment v : V AR −→ {T, F }, such that
(a) = F, v(b) = T is a counter model for ((a∪b) ⇒ a) , i.e. that 6|= ((a∪b) ⇒ a).
Substitution Method
We define and prove the correctness of a method, called Substitution Method
that allows us to obtain new tautologies from formulas already proven to be
tautologies.
We can use the same reasoning as we used in the solution to the Exercise 3.8
that proved |= (a ⇒ (a ∪ b)) to prove that, for example the formulas
A(a/A1 , b/A2 )
102
Let A ∈ F be a formula and V ARA = {a1 , a2 , ...an } be the set of all proposi-
tional variables appearing in A. We will denote it by A(a1 , a2 , ...an ).
Given a formula A(a1 , a2 , ...an ), and A1 , ...An be any formulas. We denote by
A(a1 /A1 , ..., an /An )
the result of simultaneous replacement (substitution) in A(a1 , a2 , ...an ) the vari-
ables a1 , a2 , ...an by formulas A1 , ...An , respectively.
Theorem 3.10
For any formulas A(a1 , a2 , ...an ), A1 , . . . , An ∈ F,
If |= A(a1 , a2 , ...an ) and B = A(a1 /A1 , ..., an /An ), then |= B.
Proof. Let B = A(a1 /A1 , ..., an /An ). Let b1 , b2 , ...bm be all those propositional
variables which occur in A1 , ...An . Given a truth assignment v : V AR −→
{T, F }, any values v(b1 ), v(b2 ), ...v(bm ) defines the logical value of A1 , ...An , i.e.
v ∗ (A1 ), ...v ∗ (An ) and, in turn, v ∗ (B).
Let w : V AR −→ {T, F } be a truth assignment such that w(a1 ) = v ∗ (A1 ), w(a2 ) =
v ∗ (A2 ), ...w(an ) = v ∗ (An ). Obviously, v ∗ (B) = w∗ (A). Since A is a proposi-
tional tautology, w∗ (A) = T , for all possible w, hence v ∗ (B) = w∗ (A) = T for
all truth assignments w and B is also a tautology.
Fact 3.3
For any A, B ∈ F, |= ((A ∪ B) ⇒ A).
Generalization Method
Now let’s look at the task of finding whether the formulas (3.10), (3.11) are
tautologies from yet another perspective. This time we observe that both of
them are build in a similar way as a formula (A ⇒ (A ∪ B)), for A = ((a ⇒
b) ∩ ¬c), B = ¬d in (3.10) and for A = ((a ⇒ b) ∩ ¬c), B = ((a ⇒ ¬e)) in
(3.11).
It means we represent, if it is possible, a given formula as a particular case of
some much simpler general formula. Hence the name Generalization Method.
We then use Proof by Contradiction Method or Substitution Method to examine
whether the representation of the given formula is /is not a tautology.
In this case, we prove, for example Proof by Contradiction Method by that
|= (A ⇒ (A ∪ B)), for any formulas A, B ∈ F and get, as a particular cases for
A, B that that both formulas (3.10), (3.11) are tautologies.
103
Let’s assume that there are formulas A, B ∈ F 6|= (A ⇒ (A ∪ B)). This means
that (A ⇒ (A ∪ B)) = F for some truth assignment v. This holds only when
A = T and (A ∪ B) = F , i.e. (T ∪ B) = F . This is a contradiction with the
definition of ∪. So |= (A ⇒ (A ∪ B)) for all A, B ∈ F.
Exercise 3.10
Show that v |= (¬((a ∩ ¬b) ⇒ ((c ⇒ (¬f ∪ d)) ∪ e)) ⇒ ((a ∩ ¬b) ∩ (¬(c ⇒
(¬f ∪ d)) ∩ ¬e))), for all v : V AR −→ {T, F }.
Solution
Observe that we really have to prove that |= (¬((a ∩ ¬b) ⇒ ((c ⇒ (¬f ∪ d)) ∪
e)) ⇒ ((a ∩ ¬b) ∩ (¬(c ⇒ (¬f ∪ d)) ∩ ¬e))). We can hence use any of our
tautology verification methods. In this case V ARA = {a, b, c, d, e, f }, so there
are 26 = 64 restricted truth assignments to consider. Much too many to apply
the Truth Table Method. Our formula is also far too complicated to guess a
simple tautology from which we could obtain it by the Substitution Method.
The Proof by Contradiction Method is less complicated, but before we apply
it let’s look closer at the sub-formulas of our formula and patterns they form
inside the formula it, i.e. we try to apply the Generalization Method first.
Let’s put B = (a ∩ ¬b), C = (c ⇒ (¬f ∪ d)), D = e. We re-write our formula
in a general form as (¬(B ⇒ (C ∪ D)) ⇒ (B ∩ (¬C ∩ ¬D))) and prove that for
all B, C, D ∈ F,
We use Proof by Contradiction Method, i.e. we assume that there are formulas
B, C, D ∈ F, such that
This means that there is a truth assignment v, such that (we use short-hand
notation) (¬(B ⇒ (C ∪ D)) ⇒ (B ∩ (¬C ∩ ¬D))) = F . By definition of
implication it is possible if and only if ¬(B ⇒ (C ∪ D)) = T and (B ∩ (¬C ∩
¬D)) = F , i.e. if and only if
(B ⇒ (C ∪D)) = F and (B ∩(¬C ∩¬D)) = F Observe that (B ⇒ (C ∪D)) = F
if and only if B = T, C = F, D = F. We now evaluate the logical value of
(B ∩ (¬C ∩ ¬D)) for B = T, C = F, D = F , i.e. we compute (B ∩ (¬C ∩ ¬D)) =
(T ∩ (¬F ∩ ¬F )) = (T ∩ (T ∩ T )) = T . This contradicts that we must have
(B ∩ (¬C ∩ ¬D)) = F . This proves that for all B, C, D ∈ F
104
and that all truth assignments are models for (¬((a ∪ b) ⇒ ((c ⇒ d) ∪ e)) ⇒
((a ∪ b) ∩ (¬(c ⇒ d) ∩ ¬e))).
T = {A ∈ F : |= A.}
Example 3.13
The following formulas are contradictions
C = {A ∈ F : =| A.}
Following the proof of Theorem 3.10 we get similar theorem for contradictions,
and hence a a proof of correctness of the Substitution Method of constructing
new contradictions.
Theorem 3.11
For any formulas A(a1 , a2 , ...an ), A1 , . . . , An ∈ F,
If A(a1 , a2 , ...an ) ∈ C and B = A(a1 /A1 , ..., an /An ), then B ∈ C.
105
Directly from the Theorem 3.11 we get the following.
Example 3.14 For any formulas A, B ∈ F, the following formulas are contra-
dictions
Observe, that there are formulas which neither in T nor in C, for example
(a ∪ b). Any truth assignment v, such that v(a) = F, v(b) = F falsifies (a ∪ b)
and it proves that it is not a tautology. Any truth assignment v, such that
v(a) = T, v(b) = T satisfies (a ∪ b), what proves that it is not a contradiction.
Next important notions for any logic are notions of consistency, inconsistency of
the sets of formulas and the independence of a formula from the set of formulas.
We adopt the following definitions.
Definition 3.31
A truth truth assignment v : V AR −→ {T, F } is model for the set G ⊆ F of
formulas if and only if v |= A for all formulas A ∈ G. We denote it by
v |= G.
S
The restriction vG of the model v to the domain V ARG = A∈G V ARA is
called a restricted model for G.
Exercise 3.11
Find a model and a restricted model for a set
Solution
Let v be a truth assignment v : V AR −→ {T, F }. By the defininition 3.31,
v |= {((a ∩ b) ⇒ b), (a ∪ b), ¬a} if and only if v ∗ (((a ∩ b) ⇒ b)) = T, v ∗ ((a ∪ b) =
T ), and v ∗ (¬a) = T . Observe that |= ((a ∩ b) ⇒ b), so we have to find v, such
that v ∗ ((a ∪ b)) = T, v ∗ (¬a) = T . This holds if and only if v(a) = F and
F ∪ v(b) = T , i.e. if and only if v(a) = F and v(b) = T . This proves that
any v such that v(a) = F and v(b) = T is a model for G, and G has only one
restricted model. We put it as a separate fact.
106
Fact 3.4
S
Given G = {((a ∩ b) ⇒ b), (a ∪ b), ¬a}, we have that V ARG = A∈G V ARA =
{a, b} and vG : {a, b} −→ {T, F }, such that vG (a) = F and vG (b) = T is a
unique restricted model for G.
Observation 3.4
For some sets G ⊆ F , V ARG can be infinite. For example, for G = V AR we
have that V ARG = V AR and the notions of model and restricted model are the
same.
Definition 3.32
A set G ⊆ F is called consistent if and only if there is v : V AR −→ {T, F },
such that v |= G.
Otherwise the set G is called inconsistent.
Example 3.15
The set G1 = {((a ∩ b) ⇒ b), (a ∪ b), ¬a} is consistent as v : V AR −→ {T, F },
such that v(a) = F and v(b) = T is the model for G1 .
Example 3.16
The set G1 = {((a ∩ b) ⇒ b), (a ∩ ¬a), ¬a} is a finite inconsistent set as it
contains a formula (a ∩ ¬a) ∈ C.
The set G2 = V AR ∪ {¬a} for some a ∈ V AR, is an infinite inconsistent set
as it contains a certain variable a and its negation ¬a.
Of course the most obvious example of an infinite consistent set is the set T of
all tautologies, and of an infinite inconsistent consistent set is the set C of all
contradictions.
Definition 3.33
107
A formula A ∈ F is called independent from a set G ⊆ F if and only if
the sets G ∪ {A} and G ∪ {¬A} are both consistent. I.e. when there are truth
assignments v1 , v2 such that
Exercise 3.12
Show that a formula A = ((a ⇒ b) ∩ c) is independent from the set G =
{((a ∩ b) ⇒ b), (a ∪ b), ¬a}.
Solution
We define two truth assignments v1 , v2 : V AR −→ {T, F } such that v1 |=
G ∪ {(a ⇒ b) ∩ c)} and v2 |= G ∪ {¬(a ⇒ b) ∩ c)} as follows. We have just proved
(Exercise 3.11) that any v : V AR −→ {T, F }, such that v(a) = F, v(b) = T
is a model for G. Take as v1 any truth assignment such that v1 (a) = v(a) =
F, v1 (b) = v(b) = T, v1 (c) = T. We evaluate v1 ∗ (A) = v1 ∗ (((a ⇒ b) ∩ c)) =
(F ⇒ T ) ∩ T = T . This proves that v1 |= G ∪ {A}. Take as v2 any truth
assignment such that, v2 (a) = v(a) = F, v2 (b) = v(b) = T, v2 (c) = F . We
evaluate v2 ∗ (¬A) = v2 ∗ (¬(((a ⇒ b) ∩ c)) = T ∩ T = T . This proves that
v2 |= G ∪ {¬A}. It ends the proof that formula A is independent from G.
Exercise 3.13
Show that a formula A = (¬a ∩ b) is not independent from G = {((a ∩ b) ⇒
b), (a ∪ b), ¬a}.
Exercise 3.14
Given a set G = {a, (a ⇒ b)}.
Find a formula A that is independent from G.
Solution
Observe that truth assignment v such that v(a) = T, v(b) = T is the only
restricted model for G. So we have to come up with a formula A such that
there are two different truth assignments, v1 , v2 such that v1 |= G ∪ {A} and
v2 |= G ∪ {¬A}. Let’s think about as simple a formula as it could be, namely
108
let’s consider A = c, where c any propositional variable (atomic formula) dif-
ferent from a and b. G ∪ {A} = {a, (a ⇒ b), c} and any truth assignment v1 ,
such that v1 (a) = T, v1 (b) = T, v1 (c) = T is a model for G ∪ {c}. Likewise for
G ∪ {¬c} = {a, (a ⇒ b), ¬c}. Any v2 such that v2 (a) = T, v2 (b) = T, v2 (c) = F
is a model for G ∪ {¬c}. This proves that we have found the formula A = c that
is independent from G.
Exercise 3.15
Find an infinite number of formulas that are independent from G = {((a ∩
b) ⇒ b), (a ∪ b), ¬a}.
Solution
First we have to find all v : V AR −→ {T, F } such that v |= {((a ∩ b) ⇒
b), (a ∪ b), ¬a}, i.e such that (shorthand notation) ((a ∩ b) ⇒ b) = T, (a ∪
b) = T, ¬a = T . Observe that |= ((a ∩ b) ⇒ b), so we have to consider only
(a ∪ b) = T, ¬a = T . This holds if and only if a = F and (F ∪ b) = T , i.e. if
and only if a = F and b = T. This proves that that vG such that vG (a) = F
and vG (b) = T is the only one restricted model for G. All possible models for
G must be extensions of vG . We define a countably infinite set of formulas (and
their negations) and corresponding extensions v of vG (restricted to to the set
of variables {a, b}) such that v |= G as follows.
Observe that all extensions of v of vG have as domain the infinitely countable set
V AR = {a1 , a2 , . . . , an . . . . }. We take as the infinite set of formulas in which
every formula is to be proved independent of G the set of atomic formulas
109
Another extended list of tautologies and their discussion is presented in Chapter
2.
As the next step we define notions of a logical equivalence and an equivalence of
languages. We prove that all of the languages
are equivalent under classical semantics and hence can be used (and are) as
different languages for classical propositional logic.
We generalize these notions to the case of any extensional semantics M in sec-
tion 3.6. We also discuss and examine some particular many valued extensional
semantics and properties of their languages in section 3.5.
Some Tautologies
For any A, B ∈ F, the following formulas are tautologies.
Disjunction, Conjunction
Contraposition (1)
Contraposition (2)
110
Double Negation
(¬¬A ⇔ A), (3.16)
Logical Equivalences
Logical equivalence is a very useful notion to use when we want to obtain new
formulas or new tautologies, if needed, on a base of some already known in a
way that guarantee preservation of the logical value of the initial formula. For
any formulas A, B, we say that are logically equivalent if they always have the
same logical value. We write it symbolically as A ≡ B. We have to remember
that the symbol ” ≡” not a logical connective. It is a metalanguage symbol for
saying ”A, B are logically equivalent”. This is a very useful symbol. It says
that two formulas always have the same logical value, hence it can be used in
the same way we use the equality symbol ” =.” Formally we define it as follows.
Observe that the following property follows directly from the definition 3.34.
Property 3.1
For any formulas A, B ∈ F,
A ≡ B if and only if |= (A ⇔ B)
For example we write the laws of contraposition (3.17), (3.18), and the law of
double negation (3.19) as logical equivalences as follows.
E - Contraposition (2)
E - Double Negation
¬¬A ≡ A. (3.19)
Logical equivalence is a very useful notion when we want to obtain new formu-
las, or tautologies, if needed, on a base of some already known in a way that
111
guarantee preservation of the logical value of the initial formula.
B1 = A1 (A/B).
If A ≡ B, then A1 ≡ B1 . (3.20)
Proof
By the logical equivalence Definition 3.34 proving our theorem statement 8.44
is equivalent to proving that the implication
Example 3.17
Let A1 = (C ∪ D) and B = ¬¬C. By E - Double Negation equivalence (3.19)
we have that ¬¬C ≡ C. Let B1 = A1 (C/B) = A1 (C/¬¬C) = (¬¬C ∪ D). By
the Equivalence Substitution Theorem 3.12
(C ∪ D) ≡ (¬¬C ∪ D).
Equivalence of Languages
112
The next set of equivalences, or corresponding tautologies, correspond the notion
of definability of connectives discussed in section 3.3. For example, a tautology
(A ⇒ B) ≡ (¬A ∪ B) (3.22)
Observation 3.5 The direct proof of this and other Definability of Connectives
Equivalences presented here follow from the definability formulas developed in
the the proof of the Definability of Connectives Theorem 3.2, hence the names.
We are using the logical equivalence notion, instead of the tautology notion, as
it makes the manipulation of formulas much easier.
The equivalence 3.22 allows us, by the force of Theorem 3.12 to replace any
formula of the form (A ⇒ B) placed anywhere in another formula by a formula
(¬A ∪ B) while preserving their logical equivalence. Hence we can use the
equivalence (3.22) to transform a given formula containing implication into an
logically equivalent formula that does contain implication (but contains negation
and disjunction).
We usually use the equation 3.22 to transform any formula A of language con-
taining implication into a formula B of language containing disjunction and
negation and not containing implication at all, such that A ≡ B.
Example 3.18
Let A = ((C ⇒ ¬B) ⇒ (B ∪ C)).
We use equality (3.22) to transform A into a logically equivalent formula not
containing ⇒ as follows.
((C ⇒ ¬B) ⇒ (B ∪ C)) ≡ (¬(C ⇒ ¬B) ∪ (B ∪ C))) ≡ (¬(¬C ∪ ¬B) ∪ (B ∪ C))).
It means that for example that we can, by the Theorem 3.12 transform any
formula A of the language L1 = L{¬,∩,⇒} into a logically formula B of the
language L2 = L{¬,∩,∪} . In general, we say that we can transform a language
L1 into a logically equivalent language L2 if the following condition C1 holds.
C1: for any formula A of L1 , there is a formula B of L2 , such that A ≡ B.
Example 3.19
Let A = (¬A ∪ (¬A ∪ ¬B)). We also can use, in this case, the equivalence 3.22
as follows.
113
(¬A ∪ (¬A ∪ ¬B)) ≡ (¬A ∪ (A ⇒ ¬B)) ≡ (A ⇒ (A ⇒ ¬B)).
It means we eliminated disjunction from A by replacing it by logically equivalent
formula containing implication only.
Observe, that we can’t always use the equivalence (3.22) to eliminate any dis-
junction. For example, we can’t use it for a formula A = ((a ∪ b) ∩ ¬a).
In order to be able to transform any formula of a language containing disjunc-
tion (and some other connectives) into a language with negation and implication
(and some other connectives), but without disjunction we need the following log-
ical equivalence.
Example 3.20
Consider a formula A = (a ∪ b) ∩ ¬a).
We use equality (3.23) to transform A into its logically equivalent formula not
containing ∪ as follows: ((a ∪ b) ∩ ¬a) ≡ ((¬a ⇒ b) ∩ ¬a).
In general, we use the equality 3.23 and Theorem 3.12 to transform any formula
C of the language L2 = L{¬,∩,∪} into a logically equivalent formula D of the
language L1 = L{¬,∩,⇒} . In general, the following condition hols.
C2: for any formula C of L2 , there is a formula D of L1 , such that C ≡ D.
The languages L1 and L2 for which we the conditions C1, C2 hold are logically
equivalent and denote it by L1 ≡ L2 .
We put it in a general, formal definition as follows.
Example 3.21
To prove the logical equivalence L{¬,∪} ≡ L{¬,⇒} we need two definability equiv-
alences (3.22 ) and (3.23), and the Theorem 3.12.
114
Exercise 3.16
To prove the logical equivalence L{¬,∩,∪,⇒} ≡ L{¬,∩,∪} we needed only the de-
finability equivalence (3.22).
Solution
The equivalence (3.22) proves, by Theorem 3.12 that for any formula A of
L{¬,∩,∪,⇒} there is B of L{¬,∩,∪} that equivalent to A, i.e. condition C1 holds.
Any formula A of language L{¬,∩,∪} is also a formula of L{¬,∩,∪,⇒} and of course
A ≡ A, so both conditions C1 and C2 of definition 3.35 are satisfied.
Exercise 3.17
Show that L{¬,∩} ≡ L{¬,⇒} .
Solution
The equivalence of languages holds by Theorem 3.12, Observation 3.5, and the
following two logical equalities. Definability of Conjunction in terms of
implication and negation and Definability of Implication in terms of con-
junction and negation:
(A ∩ B) ≡ ¬(A ⇒ ¬B) (3.24)
(A ⇒ B) ≡ ¬(A ∩ ¬B). (3.25)
Exercise 3.18
Show that L{¬,∩} ≡ L{¬,∪} .
Solution
Similarly, it is true by Theorem 3.12, Observation 3.5, and the following two log-
ical equalities. Definability of disjunction in terms of negation and conjunction
and definability of conjunction in terms of negation and disjunction:
Exercise 3.19
Show that L{¬,∩} ≡ L{↑} and L{¬,∩} ≡ L{↑}
115
Proof
We use the proof of Theorem 3.3 to prove the following definability equivalences
of ¬ and ∩ in terms of ↑:
¬A ≡ (A ↑ A), (A ∩ B) ≡ (A ↑ B) ↑ (A ↑ B) (3.29)
Idempotent
(A ∩ A) ≡ A, (A ∪ A) ≡ A,
Associativity
Commutativity
(A ∩ B) ≡ (B ∩ A), (A ∪ B) ≡ (B ∪ A),
Distributivity
De Morgan Laws
Negation of Implication
116
Consider a tautology A: |= ((¬(A ⇒ B) ⇒ ¬A) ⇒ (A ⇒ B)).
We know by (3.22) that (A ⇒ B) ≡ (¬A ∪ B). By Theorem 3.12, if we replace
(A ⇒ B) by (¬A ∪ B) in A, the logical value of A will remain the same and
((¬(A ⇒ B) ⇒ ¬A) ⇒ (A ⇒ B)) ≡ ((¬(¬A ∪ B) ⇒ ¬A) ⇒ (¬A ∪ B)). Now
we use de Morgan Laws and Double Negation Laws and by Theorem 3.12 we
get ((¬(A ⇒ B) ⇒ ¬A) ⇒ (A ⇒ B)) ≡ ((¬(¬A ∪ B) ⇒ ¬A) ⇒ (¬A ∪ B)) ≡
(((¬¬A ∩ ¬B) ⇒ ¬A) ⇒ (¬A ∪ B)) ≡ (((A ∩ ¬B) ⇒ ¬A) ⇒ (¬A ∪ B)).
This proves that
|= (((A ∩ ¬B) ⇒ ¬A) ⇒ (¬A ∪ B)).
Exercise 3.20
Prove using proper logical equivalences that
(i) ¬(A ⇔ B) ≡ ((A ∩ ¬B) ∪ (¬A ∩ B)),
(ii) ((B ∩ ¬C) ⇒ (¬A ∪ B)) ≡ ((B ⇒ C) ∪ (A ⇒ B)).
Solution (i)
¬(A ⇔ B)≡(3.28) ¬((A ⇒ B) ∩ (B ⇒ A))≡de M organ (¬(A ⇒ B) ∪ ¬(B ⇒
A))≡(3.32) ((A ∩ ¬B) ∪ (B ∩ ¬A))≡commut ((A ∩ ¬B) ∪ (¬A ∩ B)).
Solution (ii)
((B ∩ ¬C) ⇒ (¬A ∪ B))≡(3.23) (¬(B ∩ ¬C) ∪ (¬A ∪ B))≡de M organ ((¬B ∪ ¬¬C) ∪
(¬A ∪ B))≡(3.19) ((¬B ∪ C) ∪ (¬A ∪ B))≡(3.23) ((B ⇒ C) ∪ (A ⇒ B)).
Many valued logics in general and 3-valued logics in particular is an old object
of study which has its beginning in the work of a Polish mathematician Jan
Leopold Lukasiewicz in 1920. He was the first to define a 3 - valued semantics
for the language L{¬,∩,∪,⇒} of classical logic, and called it a three valued logic
for short. He left the problem of finding a proper axiomatic proof system for it
(i.e. the one that is complete with respect to his semantics) open. The same
happened to all other 3 - valued semantics presented here. They were also first
called 3 valued logics and this terminology is still widely used. Nevertheless, as
these logics were defined only semantically, i.e. defined by providing a semantics
for their languages we call them just semantics (for logics to be developed), not
logics. Existence of a proper axiomatic proof system for a given semantics and
proving its completeness is always a next open question to be answered (when
it is possible). A process of creating a logic (based on a given language) always
is three fold: we define semantics, create axiomatic proof system and prove
117
completeness theorem that established a relationship between semantics and
proof system.
The first of many valued logics invented were first presented in a semantical
form only for other components to be developed later. We can think about the
process of their creation as inverse to the creation of Classical Logic, Modal
Logics, the Intuitionistic Logic which existed as axiomatic systems longtime be-
fore invention of their formal semantics.
There has been many of proposals relating both to the intuitive interpretation
of this third value ⊥. If T is the only designated value, the third value ⊥ cor-
responds to some notion of incomplete information, like undefined or unknown
and is often denoted by the symbol U or I. If, on the other hand, ⊥ corresponds
to inconsistent information, i.e. its meaning is something like known to be both
true and false then corresponding semantics takes both T and the third logical
value ⊥ as designated. In general, the third logical value denotes a notion of
”unknown”, ”uncertain”, ”undefined”, or even can express that ”we don’t have
a complete information”, depending on the context and motivation for the logic
we plan to develop. In all of presented here semantics we take T as designated
value, i.e. T is the value that defines the notion of satisfiability and tautology.
Lukasiewicz Semantics L
Motivation
Lukasiewicz developed his semantics (called logic) to deal with future contin-
118
gent statements. According to him, such statements are not just neither true
nor false but are indeterminate in some metaphysical sense. It is not only that
we do not know their truth value but rather that they do not possess one. Intu-
itively, ⊥ signifies that the statement cannot be assigned the value true of false;
it is not simply that we do not have sufficient information to decide the truth
value but rather the statement does not have one.
Step 1: L Connectives
¬ ⊥=⊥, ¬T = F, ¬F = T.
x ∩ y = min{x, x}.
x ∪ y = max{x, y}
119
∩ F ⊥ T
¬ F ⊥ T
F F F F
T ⊥ F
⊥ F ⊥ ⊥
T F ⊥ T
∪ F ⊥ T ⇒ F ⊥ T
F F ⊥ T F T T T
⊥ ⊥ ⊥ T ⊥ ⊥ T T
T T T T T F ⊥ T
Step 4: L Tautology
120
We define, for any A ∈ F, A is a L tautology if and only if v ∗ (A) = T for
all v : V AR −→ {F, ⊥, T }. We also say that A is a L tautology if and only if
all truth assignments v : V AR −→ {F, ⊥, T } are L models for A. We write the
statement ” A is a L tautology” symbolically as
|=L A.
As a next step we define, as we did in the case of classical semantics the notions
of restricted truth assignment and restricted models, (Definitions 3.26, 3.27) i.e.
we have the following.
Any function vA : V ARA −→ {F, ⊥, T }, such that vA |=L A ( vA 6 |=L A)
is called a restricted L model ( L counter model) for A, where V ARA is the set
of all propositional variables appearing in A. We call the function vA , a truth
assignment restricted to A, or restricted truth assignment for short.
We prove, in the same way we proved Theorem 3.8 in Section 3.3, the following
theorem that justifies the correctness of the truth tables L tautologies verifica-
tion method.
Directly from Theorem 3.13 we get that the notion of L propositional tautology
is decidable, i.e. that the following holds.
We just proved (Theorem 3.14) the correctness of the truth table tautology ver-
ification method for L semantics stated as follows.
121
process and give answer: 6|=L A. Otherwise we continue. If all truth assignments
evaluate A to T , we give answer: |=L A.
Consider, for example, a formula A: (a ⇒ a). There are 31 = 3 possible re-
stricted truth assignment v : {a} −→ {F, ⊥, T }. We list them, and evaluate
their value on the formula A in a form of an extended truth table as follows.
Moreover (3.35) proves that the condition (3.3) of the definition 3.14 is fulfilled
and the L semantics is well defined. We put it as a separate fact.
Fact 3.5
The Lukasiewicz semantics L is well defined.
As a next step we can adopt all other classical tautology verification methods
from Section 3.3. It is a quite straightforward adaptation and we leave it a san
exercise. Moreover it works for all of many valued semantics presented here, as
does the Decidability Theorem 3.14.
When defining and developing a new logic the first question one asks is how it
relates and compares with the classical case, it means with the classical logic. In
case of new semantics (logics defined semantically) we describe this relationship
in terms of respective sets of tautologies.
Let LT, T denote the sets of all L and classical tautologies, respectively.
Theorem 3.15
The following relationship holds between classical and L tautologies:
LT 6= T and LT ⊂ T. (3.36)
Proof
Consider a formula (¬a ∪ a). It is obviously a classical tautology. Take any
truth assignment v : V AR −→ {F, ⊥, T } such that v(a) =⊥. By definition we
have that v ∗ (¬a ∪ a) = v ∗ (¬a) ∪ v ∗ (a) = ¬v(a) ∪ v(a) = ¬ ⊥ ∪ ⊥=⊥ ∪ ⊥=⊥.
122
This proves that v is a L counter-model for (¬a ∪ a) and hence 6|=L (¬a ∪ a).
This proves LT 6= T.
Observe now that if we restrict the values of functions defining L connectives to
the values T and F only, we get the functions defining the classical connectives.
It is directly visible when we compare the L and classical connectives truth
tables. This means that if v ∗ (A) = T for all v : V AR −→ {F, ⊥, T }, then
v ∗ (A) = T for any v : V AR −→ {F, T } and for any A ∈ F, i.e. LT ⊂ T.
Exercise 3.21
Use the fact that v : V AR −→ {F, ⊥, T } is such that v ∗ ((a ∩ b) ⇒ ¬b) =⊥
under L semantics to evaluate v ∗ (((b ⇒ ¬a) ⇒ (a ⇒ ¬b)) ∪ (a ⇒ b)). Use
shorthand notation.
Solution
Observe that ((a ∩ b) ⇒ ¬b) =⊥ in two cases.
c1: (a ∩ b) =⊥ and ¬b = F .
c12: (a ∩ b) = T and ¬b =⊥ .
Consider c1. We have ¬b = F , i.e. b = T , and hence (a ∩ T ) =⊥ if and only
if a =⊥. We get that v is such that v(a) =⊥ and v(b) = T . We evaluate (in
short hand notation) v ∗ (((b ⇒ ¬a) ⇒ (a ⇒ ¬b)) ∪ (a ⇒ b)) = (((T ⇒ ¬ ⊥) ⇒
(⊥⇒ ¬T )) ∪ (⊥⇒ T )) = ((⊥⇒⊥) ∪ T ) = T .
Consider c2. We have ¬b =⊥, i.e. b =⊥, and hence (a∩ ⊥) = T what is
impossible, hence v from case c1 is the only one, and v ∗ (((b ⇒ ¬a) ⇒ (a ⇒
¬b)) ∪ (a ⇒ b)) = T.
L4 Semantics
123
Exercise 3.22
Here are 3 simple problems.
1. Write down L4 Connectives Truth Tables.
2. Give an example of a L4 tautology.
3. We know that the formula ((a ⇒ b) ⇒ (¬a ∪ b)) is a classical tautology, i.e.
|= ((a ⇒ b) ⇒ (¬a ∪ b)). Verify whether |=L4 ((a ⇒ b) ⇒ (¬a ∪ b)).
Solution 1.
Here are L4 Connectives Truth Tables.
∩ F ⊥1 ⊥2 T
F F F F F
¬ F ⊥1 ⊥2 T
⊥1 F ⊥1 ⊥1 ⊥1
T ⊥1 ⊥2 F
⊥2 F ⊥1 ⊥2 ⊥2
T F ⊥1 ⊥2 T
∪ F ⊥1 ⊥2 T ⇒ F ⊥1 ⊥2 T
F F ⊥1 ⊥2 T F T T T T
⊥1 ⊥1 ⊥1 ⊥2 T ⊥1 ⊥1 T T T
⊥2 ⊥2 ⊥2 ⊥2 T ⊥2 ⊥2 ⊥2 T T
T T T T T T F ⊥1 ⊥2 T
Solution 2.
Observe that by definition of L4 implication we get x ⇒ x = T for all x ∈
{F, ⊥1 , ⊥2 , T }. Hence v ∗ ((a ⇒ a)) = v(a) ⇒ v(a) = T for all v, what proves
|=L4 (a ⇒ a).
Solution 3.
We use the Proof by Contradiction Method (section 3.3) to verify whether
|=L4 ((a ⇒ b) ⇒ (¬a ∪ b)). Observe that it applied to any situation, as its
correctness is based on our classical reasoning. Assume that 6|=L4 ((a ⇒ b) ⇒
(¬a∪b)). Let v : V AR −→ {F, ⊥1 , ⊥2 , T }, such that v ∗ (((a ⇒ b) ⇒ (¬a∪b))) 6=
T. Observe that in L4 semantics, for any formula A ∈ F, v ∗ (A) 6= T gives us
three possibilities v ∗ (A) = F, v ∗ (A) = ⊥1 , or v ∗ (A) = ⊥2 to consider ( as
opposed to one case in classical case). It is a lot of work, but still less then listing
and evaluating 42 = 16 possibilities of all restricted truth assignment. Moreover,
our formula is a classical tautology, hence we know that it evaluates in T for all
combinations of T and F. A good strategy is to examine first some possibilities
124
of evaluating variables a, b for combination of ⊥1 , ⊥2 with hope of finding a
counter model. So let’s v be a truth assignment such that v(a) = v(b) = ⊥1 .
We evaluate v ∗ ((a ⇒ b) ⇒ (¬a ∪ b)) = ((⊥1 ⇒ ⊥1 ) ⇒ (¬⊥1 ∪ ⊥1 )) = (T ⇒
(⊥1 ∪ ⊥1 )) = (T ⇒ ⊥1 ) = ⊥1 . This proves that v is a counter-model for our
formula. Observe that the v serves also as a L counter model for A when we
put ⊥1 =⊥ and so we get
|= ((a ⇒ b) ⇒ (¬a ∪ b)), 6|=L4 ((a ⇒ b) ⇒ (¬a ∪ b)), 6|=L ((a ⇒ b) ⇒ (¬a ∪ b))
Obviously, any v such that v(a) = v(b) = ⊥2 is also a counter model for A, as
v ∗ ((a ⇒ b) ⇒ (¬a ∪ b)) = ((⊥2 ⇒ ⊥2 ) ⇒ (¬⊥2 ∪ ⊥2 )) = (T ⇒ (⊥2 ∪ ⊥2 )) =
(T ⇒ ⊥2 ) = ⊥2 . We leave it as an exercise to find all possible counter models
for A.
Heyting Semantics H
Motivation
We discuss here the semantics H because of its connection with intuitionistic
logic. The H connectives are such that they represent operations in a cer-
tain 3 element algebra, historically called a 3 element pseudo-boolean algebra.
Pseudo-boolean algebras were created by McKinsey and Tarski in 1948 to pro-
vide semantics for the intuitionistic logic. The intuitionistic logic, the most
important rival to the classical logic was defined and developed by its inventor
Brouwer and his school in 1900s as a proof system only. Heyting provided its
first axiomatization which everybody accepted. McKinsey and Tarski proved
the completeness of the Heyting axiomatization with respect to their pseudo
boolean algebras semantics. The pseudo boolean algebras are also called Heyt-
ing algebras in his honor and so is our semantics H.
The other type of models, called Kripke models were defined by Kripke in 1964
and were proved later to be equivalent to the pseudo-boolean models. They are
very general and serve as a method of defining not extensional semantics for
125
various classes of logics. That includes semantics for a great number of modal,
knowledge, belief logics, and many new logics developed and being developed
by computer scientists.
H Connectives
⇒ F ⊥ T
¬ F ⊥ T
F T T T
T F F
⊥ F T T
T F ⊥ T
126
to denote that a formula A is an H tautology.
Theorem 3.18
Let HT, LT, T denote the sets of all H, L, and classical tautologies, respectively.
Then the following relationship holds.
HT 6= LT, HT 6= T, and HT ⊂ T. (3.39)
Proof
A formula (¬a ∪ a) a classical tautology and not an H tautology. Take any
truth assignment v : V AR −→ {F, ⊥, T } such that v(a) =⊥. We evaluate
is v ∗ ((¬a ∪ a) = ¬ ⊥ ∪ ⊥= F ∪ ⊥=⊥ This proves that (¬a ∪ a) 6∈ HT and
hence HT 6= T. Directly from the definition of H connectives we get that if
we restrict the values of the functions defining them T and F only, we get the
functions defining the classical connectives. Hence for any formula A ∈ TH we
have that A ∈ TH and LT ⊂ T. A formula (¬¬a ⇒ a) is a L tautology and
not an H tautology by easy evaluation as presented in example 3.23 and (3.40).
This proves HT 6= LT.
Exercise 3.23
We know that v : V AR −→ {F, ⊥, T } is such that v ∗ ((a ∩ b) ⇒ (a ⇒ c)) =⊥
under H semantics.
Evaluate v ∗ (((b ⇒ a) ⇒ (a ⇒ ¬c)) ∪ (a ⇒ b)). You can use a short hand
notation.
127
Solution
By definition of H connectives we have that for any v, v ∗ ((a∩b) ⇒ (a ⇒ c)) =⊥
if and only if a ∩ b) = T and (a ⇒ c) =⊥ if and only if a = T, b = T and
(T ⇒ c) =⊥ if and only if c =⊥. Hence v ∗ ((a ∩ b) ⇒ (a ⇒ c)) =⊥ if and only
if a = T, b = T, c =⊥ . We evaluate v ∗ (((b ⇒ a) ⇒ (a ⇒ ¬c)) ∪ (a ⇒ b)) =
(((T ⇒ T ) ⇒ (T ⇒ ¬ ⊥)) ∪ (T ⇒ T )) = ((T ⇒ (T ⇒ F )) ∪ T ) = T .
Exercise 3.24
We know that the following formulas are basic classical tautologies
The H semantics can serve as a tool of proving that some formulas are not
intutionistic tautologies, but it is not a universal one
Example 3.22
We know that the classical tautology (¬(a ∩ b) ⇒ (¬a ∪ ¬b)) is not intuitionistic
tautology, but nevertheless |=H (¬(a ∩ b) ⇒ (¬a ∪ ¬b)).
Proof
We use the Proof by Contradiction Method (section 3.3.1) and shorthand no-
tation. Assume that 6|=H (¬(a ∩ b) ⇒ (¬a ∪ ¬b)). Let v : V AR −→ {F, ⊥, T }
such that v ∗ ((¬(a ∩ b) ⇒ (¬a ∪ ¬b))) 6= T . We have to consider two cases:
c1 v ∗ ((¬(a ∩ b) ⇒ (¬a ∪ ¬b))) =⊥ and c2 v ∗ ((¬(a ∩ b) ⇒ (¬a ∪ ¬b))) = F .
If we get a contradiction in both cases we have proved |=H (¬(a∩b) ⇒ (¬a∪¬b)).
128
Contradiction with the definition of ⊥. This proves that case c1 always leads
to contradiction.
We can of course also use the Truth Tables Method that involves listing and
evaluating all of 23 = B restricted truth assignments.
Kleene Semantics K
Motivation
In Kleene’s semantics the third logical value ⊥, intuitively, represents undecided.
Its purpose is to signal a state of partial ignorance. A sentence a is assigned a
value ⊥ just in case it is not known to be either true of false.
For example, imagine a detective trying to solve a murder. He may conjecture
that Jones killed the victim. He cannot, at present, assign a truth value T or
F to his conjecture, so we assign the value ⊥, but it is certainly either true of
false and ⊥ represents our ignorance rather then total unknown.
K Connectives
We adopt the same language as in a case of classical, Lukasiewicz’s L, and Heyt-
ing H semantics, i.e. L = L{¬,⇒,∪,∩} .
¬ ⊥=⊥, ¬F = T, ¬T = F.
129
The K implication is defined by the same formula as the classical, i.e.
x ⇒ y = ¬x ∪ y. (3.41)
⇒ F ⊥ T
F T T T
⊥ ⊥ ⊥ T
T F ⊥ T
We write
|=K A
to denote that A is a K tautology. We prove, in the same way as in case of L
semantics the following theorems that justify truth table method of verification
and decidability theorem for K.
Directly from Theorem 3.19 we get that the notion of K propositional tautology
is decidable, i.e. that the following holds.
130
We write
KT = {A ∈ F : |=K A}
to denote the set of all K tautologies. The following establishes relationship
between L, K, and classical tautologies.
Theorem 3.21
Let LT, T, KT denote the sets of all L, classical, and K tautologies, respec-
tively. Then the following relationship holds.
Proof
Obviously |= (a ⇒ a) and also by (3.35) |=L (a ⇒ a). Consider now any
v such that v(a) =⊥. We evaluate in K semantics v ∗ (a ⇒ a) = v(a) ⇒
v(a) =⊥⇒⊥=⊥. This proves that 6 |=K (a ⇒ a) and hence the first two rela-
tionships in (3.42) hold. The third one follows directly from the the fact that,
as in the L case, if we restrict the functions defining K connectives to the values
T and F only, we get the functions defining the classical connectives.
Exercise 3.25
We know that formulas ((a ∩ b) ⇒ a), (a ⇒ (a ∪ b)), (a ⇒ (b ⇒ a)) are classical
tautologies. Show that none of them is K tautology.
Solution Consider any v such that v(a) = v(b) =⊥. We evaluate (in short
hand notation) v ∗ (((a ∩ b) ⇒ a) = (⊥ ∩ ⊥) ⇒⊥=⊥⇒⊥=⊥⇒⊥⊥, v ∗ ((a ⇒
(a ∪ b))) =⊥⇒ (⊥ ∪ ⊥) =⊥⇒⊥=⊥, and v ∗ ((a ⇒ (b ⇒ a))) = (⊥⇒ (⊥⇒⊥
) =⊥⇒⊥=⊥. This proves that v such that v(a) = v(b) =⊥ is a counter model
for all of them. We generalize this example and prove that in fact a similar truth
assignment can serve as a counter model for not only any classical tautology,
but also for any formula A of L{¬,⇒,∪,∩} .
Theorem 3.22
For any formula A ∈ F, 6|=K A, i.e. the set of all K tautologies is empty. We
write it as
KT = ∅.
Proof
We show that a truth assignment v : V AR −→ {F, ⊥, T }, such that v(a) =⊥
for all a ∈ V AR is a counter model for any A ∈ F. We carry the proof the by
mathematical induction over the degree d(A) of the formula A.
Base Case: n=1 i.e. d(A) = 1. In this case we have that A = ¬a for any
a ∈ V AR, or A = (a ◦ b) for a, b ∈ V AR, ◦ ∈ {∪, ∩, ⇒}.
131
We evaluate: v ∗ (A) = v ∗ (¬a) = ¬v ∗ (a) = ¬ ⊥=⊥, v ∗ (a ◦ b) = v ∗ (a) ◦ v ∗ (b) =⊥
◦ ⊥=⊥ . This proves that the Base Case holds.
Inductive assumption: v ∗ (B) =⊥ for all B such that d(B) = k and 1 ≤ k < n.
Inductive thesis: v ∗ (A) =⊥ for any A such that d(A) = n.
Let A be such that d(A) = n. We have two cases to consider.
Case 1. A = ¬B, so d(B) = n − 1 < n. By inductive assumption v ∗ (B) =⊥.
Hence v ∗ (A) = v ∗ (¬B) = ¬v ∗ (B) = ¬ ⊥=⊥ and inductive thesis holds.
Case 2. A = (B ◦ C) for B, C ∈ F, ◦ ∈ {∪, ∩, ⇒} (and d(A) = n). Let
d(B) = k1 , d(C) = k2 . Hence d(A) = d(B ◦ C) = k1 + k2 + 1 = n. We
get that k1 + k2 = n − 1 < n. From k1 + k2 < n we get that k1 < n and
k2 < n. Hence by inductive assumption v ∗ (B) =⊥ and v ∗ (C) =⊥. We evalu-
ate: v ∗ (A) = v ∗ (B ◦ C) = v ∗ (B) ◦ v ∗ (C) =⊥ ◦ ⊥=⊥. This ends the proof.
Observe that the theorem 3.22 does not invalidate relationships (3.42). They
become now perfectly true statements
LT 6= ∅, T 6= ∅, and ∅ ⊂ T.
But when we develop a logic by defining its semantics we must make sure for
semantics to be such that it has a non empty set of its tautologies. The semantics
K is an example of a correctly and carefully defined semantics that is not well
defined in terms of the definition 3.14. We write is as separate fact.
Fact 3.6
The Kleene semantics K is not well defined.
Bochvar semantics B
Motivation
132
sentences are neither true of false but rather paradoxical or meaningless. The
semantics follows the principle that the third logical value, denoted now by m
is in some sense ”infectious”; if one one component of the formula is assigned
the value m then the formula is also assigned the value m.
Bohvar also adds an one argument assertion operator S that asserts the logical
value of T and F , i.e. S F = F , S T = T and it asserts that meaningfulness is
false, i.e S m = F .
Language LB
The language of B semantics differs from all previous languages in that it con-
tains an extra one argument assertion connective S added to the usual set
{¬, ⇒, ∪, ∩} of the language L = L{¬,S,⇒,∪,∩} of all previous semantics.
LB = L{¬,S,⇒,∪,∩} . (3.43)
The set LV of logical values is {T, m, F }. T is the distinguished value.
B Connectives
We define the connectives of LB the functions defined in the set {F, mT } by the
following truth tables.
∩ F m T ∪ F m T
¬ F m T
F F m F F F m T
T m F
m m m m m m m m
T F m T T T m T
⇒ F m T
S F m T
F T m T
F F T
m m m m
T F m T
For all other steps of definition of B semantics we follow the standard way estab-
lished for extensional M semantics, we did in all previous cases. In particular
we define the notion of B tautology as follows.
Definition 3.38
A formula A of LB is a B tautology if and only if v ∗ (A) = T , for all
133
v : V AR −→ {F, m, T }, i.e. if all variable assignments v are B models for A.
We write
|=B A
to denote that A is an B tautology.
We, prove, in the same way as for all previous logics semantics, the following
theorems that justify the truth table method of verification and decidability for
B tautologies.
Let denote by FB the set of formulas of the language LB and by BT the set of
all B tautologies:
BT = {A ∈ FB : |=B A}.
Which formulas (if any) are the B tautologies is more complicated to determine
then in the case previous semantics because we have the following Fact 3.7.
Fact 3.7
For any formula A ∈ FB which do not contain a connective S, i.e. for any
formula A of the language L{¬,⇒,∪,∩} , 6|=B A.
134
3.6 M Tautologies, M Consistency, and M Equiv-
alence of Languages
The classical truth tables verification method a and classical decidability theo-
rem hold in a proper form in all of L. H, K and B semantics as it was discussed
separately for each of them. We didn’t discuss other classical tautologies veri-
fication methods of substitution and generalization. We do it now in a general
and unifying way for a special case of an extensional M semantics, namely for
any semantics M with a finite set of logical values.
Example 3.23
Consider a formula (¬¬a ⇒ a) and H semantics. We evaluate
135
Example 3.24
Consider a formula (¬¬a ⇒ a) and L semantics. We evaluate
We also proved that the set HT of all H tautologies is different from the set set
LT of all L tautologies, i.e.
LT 6= HT (3.44)
Directly from Theorem 3.25 and the above we get that the notion of M propo-
sitional tautology is decidable, i.e. that the following holds.
Substitution Method
The Substitution Method allows us to obtain, as in a case of classical semantics
new M tautologies from formulas already proven to be M tautologies. The
theorem 3.27 and its proof is a straightforward modification of the classical
proof (theorem 3.27) and we leave it as an exercise to the reader. It assesses
the validity of the substitution method. In order to formulate and prove it we
first remind of the reader of needed notations.
Let A ∈ F be a formula and V ARA = {a1 , a2 , ...an } be the set of all proposi-
tional variables appearing in A. We will denote it by A(a1 , a2 , ...an ). Given a
136
formula A(a1 , a2 , ...an ), and A1 , ...An be any formulas. We denote by
Theorem 3.27
For any formulas A(a1 , a2 , ...an ), A1 , . . . , An ∈ F,
If |=M A(a1 , a2 , ...an ) and B = A(a1 /A1 , ..., an /An ), then |=M B.
We have proved (exercise 3.24) that the formula D(a) = (¬¬a ⇒ a) is L tau-
tology. By the above theorem 3.27 we get that D(a/A) = (¬¬A ⇒ A) is also L
tautology for any formula A ∈ F. We hence get the following.
Fact 3.8
For any A ∈ F, |=L (¬¬A ⇒ A).
M Generalization Method
In this method we represent, if it is possible, a given formula as a particular
case of some simpler general formula. Hence the name Generalization Method.
We then use other methods to examine the simpler formula thus obtained.
Exercise 3.26
Prove that
|=L (¬¬(¬((a ∩ ¬b) ⇒ ((c ⇒ (¬f ∪ d)) ∪ e)) ⇒ ((a ∩ ¬b) ∩ (¬(c ⇒ (¬f ∪
d)) ∩ ¬e))) ⇒ (¬((a ∩ ¬b) ⇒ ((c ⇒ (¬f ∪ d)) ∪ e)) ⇒ ((a ∩ ¬b) ∩ (¬(c ⇒
(¬f ∪ d)) ∩ ¬e)))).
Solution
Observe that our formula is a particular case of a more general formula (¬¬A ⇒
A) for A = (¬((a ∩ ¬b) ⇒ ((c ⇒ (¬f ∪ d)) ∪ e)) ⇒ ((a ∩ ¬b) ∩ (¬(c ⇒ (¬f ∪
d)) ∩ ¬e))) and by fact 3.8 our formula is proved to be L tautology.
3.6.2 M Consistency
One of the most important notions for any logic are notions of consistency and
inconsistency. We introduced and discussed them in case of classical semantics
in section 3.3. We formulate them now for any M extensional semantics and
examine them in cases of L and H semantics.
137
Consider LCON and let S =
6 ∅ be any non empty set of formulas of LCON . Let
M be an extensional semantics for LCON . We adopt the following definitions.
Exercise 3.27
Prove that the set
G = {((a ∩ b) ⇒ b), (a ∪ b), ¬a}
is L, H, and K consistent.
Fact 3.9
For any non empty set G of formulas of a language L{¬,⇒,∪,∩} ,
if G is consistent under classical semantics, then it is L, H, and K consistent.
138
Exercise 3.28
Give an example of an infinite set G of formulas of a language LB = L{¬,S,⇒,∪,∩}
that is L, H, K and B consistent.
Solution
Observe that for the set G to be considered to be L, H, K consistent its formulas
must belong to the sub language L{¬,⇒,∪,∩} of the language LB . Let’s take, for
example a set
G = {(a ∪ ¬b) : a, b ∈ V AR}.
G is infinite since the set V AR is infinite. Consider any v : V AR −→ {F, m, T }
or v : V AR −→ {F, ⊥, T } such that v(a) = T, v(b) = F , we have v ∗ (a ∪ b) =
v(a) ∪ v(b) = T ∪ T = T in all semantics L, H, K and B. This proves that G is
L, H, K and B consistent.
Exercise 3.29
Prove that the set
G = {(a ∩ ¬a) : a ∈ V AR}
is L, H, K, and B inconsistent..
Solution
We know that the set G is classically inconsistent, i.e. v ∗ ((a ∩ ¬a)) 6= T for all
v : V AR −→ {F, T } under classical semantics. It also holds for We have to
show that it also holds for L, H, K and B semantics when we restrict the func-
tions defining their connectives to the values T and F only. In order to prove
inconsistency under L, H, K, semantics we have to show that v ∗ ((a ∩ ¬a)) 6= T
for all v : V AR −→ {F, ⊥, T } under the respective semantics, i.e. we have to
evaluate additional case v(a) =⊥ in all of them. Observe that negation ¬ is de-
fined in all of them as ¬ ⊥=⊥, and v ∗ ((a ∩ ¬a)) =⊥ ∩¬ ⊥=⊥ ∩ ⊥=⊥6= T . This
proves that G is L, H, and K inconsistent. The case of B semantics is similar,
except that now we consider all v : V AR −→ {F, m, T } and the additional
case is v(a) = m. By definition ¬m = m and v ∗ ((a ∩ ¬a)) = m ∩ m = m 6= T .
Exercise 3.30
Give an example of sets G1 , G2 containing some formulas that include the S
connective of the language LB = L{¬,S,⇒,∪,∩} such that G1 is B consistent and
G2 is B inconsistent
139
Solution
There are many such sets G, here are just two simple examples.
Definition 3.42
Let M be an extensional semantics for LCON . We say that a formula A is a
M contradiction if it doesn’t have a M model.
Example 3.25
A formula (Sa ∩ S¬a) of LB = L{¬,S,⇒,∪,∩} is a B contradiction.
Proof
Assume that there is v, such that v |= (Sa ∩ S¬a), i.e. v ∗ ((Sa ∩ S¬a)) = T if
and only if (shorthand notation) Sa = T and S¬a = T . But Sa = T if and
only of a = T . In this case S¬T = SF = F 6= T. This contradiction proves that
such v does not exist, i.e. that for all v, v 6|= (Sa ∩ S¬a).
Here is a simple problem asking to create your own, specific M semantics ful-
filling certain specifications. This semantics is different from all of previous
semantics defined and examined. We also ask to examine some of its properties,
including M consistency and M inconsistency. We provide an example two dif-
ferent semantics. We encourage the reader to come up with his/hers own and
to write down formally its full definition according to definition 3.7 as it was
done in the case of L semantics.
140
Review Problem
Part 2. Define formally, following all steps of the defnition 3.7, a 3 valued
extensional semantics LK for the language L{¬, L, ∪, ⇒} under the following
assumptions.
s1 We assume that the third value is denoted by ⊥ is intermediate between
designated value T and F, i.e. that F <⊥< T.
s2 We model a situation in which one ”likes” only truth, represented by T; i.e.
in which
LT = T, L ⊥= F, LF = F.
Part 3.
1. Verify whether the formulas A1 and A2 from the Part 1. have a model/
counter model under your semantics LK. You can use shorthand notation.
2. Verify whether the following set G is LK consistent. You can use shorthand
notation.
G = {La, (a ∪ ¬Lb), (a ⇒ b), b }.
Part 1 Solution
1. We translate the statement into a formula A1 ∈ F1 of a language L{¬, L, ∩, ∪, ⇒}
as follows.
141
Propositional variables: a, b, where a denotes statement: play bridge, b denotes
a statement: the weather is good.
A2 = (a ∪ (b ⇒ (¬a ∪ c))).
Part 2 Solution 1
Here is a simple LK semantics. We define the logical connectives by writing
functions defining connectives in form of the truth tables and skipping other
points of the definition 3.7. We leave it to the reader as an exercise to write
down a full definition according to the definition 3.7.
LK Semantics 1
L F ⊥ T ¬ F ⊥ T
F F T T F F
∩ F ⊥ T ∪ F ⊥ T ⇒ F ⊥ T
F F F F F F ⊥ T F T T T
⊥ F ⊥ ⊥ ⊥ ⊥ T T ⊥ T ⊥ T
T F ⊥ T T T T T T F F T
We verify whether the condition s3 is satisfied, i.e. whether |=LK (LA ∪ ¬LA)
by simple evaluation. Let v : V AR −→ {F, ⊥, T } be any truth assignment. For
any formula A, v ∗ (A) ∈ {F, ⊥, T } and LF ∪ ¬LF = LF ∪ ¬LF = F ∪ ¬F ∪ T =
T, L ⊥ ∪¬L ⊥= F ∪ ¬F = F ∪ T = T, LT ∪ ¬LT = T ∪ ¬T = F ∪ T = T.
Part 2 Solution 2
Here is another simple LK semantics. Writing, yet again, a full definition is left
to the reader as an exercise.
142
LK Semantics 2
The logical connectives are the following funcions in the set {F, ⊥, T }, where
{F <⊥< T }. We define ¬F = T, ¬ ⊥= T, ¬T = F and, as by s2, LT =
T, L ⊥= F, LF = F. We define, for any x, y ∈ {F, ⊥, T }
From the above definition we can see the LK satisfies the requirement s3 that
especially |=LK (LA ∪ ¬LA) since for any truth assignment v, no matter what
values v ∗ (LA) and v ∗ (¬LA) are, the combination of them by ∪ will always be T .
Part 3
1. Verify whether the formulas A1 and A2 from the Part 1. have a model/
counter model under your semantics LK. You can use shorthand notation.
Solution 1
A model for A1 = (La ∪ (b ⇒ (¬La ∪ L¬a))) under LK semantics 1 is any v,
such that v(a) = T . By easy evaluation, A1 does not have no counter model,
i.e. |=LK A1 . Also any v, such that v(a) = T is a model for A1 as we have
v ∗ (A2 ) = T ∪ v ∗ ((b ⇒ (¬a ∪ c))) = T by definition of ∪.
Solution 1
G is LK consistent under semantics 1 because any v, such that v(a) = T, v(b) =
T is a LK model for G under semantics 1 by straightforward evaluation.
Solution 2
Consider any v, such that v(a) = v(b) = T . We evaluate: v ∗ (La) = LT =
T, v ∗ ((a ∪ ¬Lb)) = T ∪ F = T, v ∗ (a ⇒ b)) = T ⇒ T = T. This proves
v |=LK G, i.e. G is consistent.
Part 3
3. Give an example on an infinite, LK consistent set of formulas of the language
L{¬, L, ∩, ∪, ⇒} . Some formulas must contain the connective L.
143
Solution
The infinite set G = {La : a ∈ V AR} is consistent under both LK semantics,
as any v, such that v(a) = T we get v ∗ (La) = LT = T by s2.
The infinite set G = {(La ∪ (b ∩ L¬c)) : a, b, c ∈ V AR} is consistent under the
semantics 2 by its definition of ∪. Any v, such that v(a) = T is its model.
A ≡M B
Exercise 3.31
The classical logical equivalence (A ∪ B) ≡ (¬A ⇒ B) holds for all formulas A,
B and is defining ∪ in terms of negation and implication. Show that it does not
hold under L semantics, i.e. that there are formulas A, B, such that
(A ∪ B) 6≡L (¬A ⇒ B)
Solution
Consider a case when A = a and B = b. By definition 3.43 we have to show
v ∗ ((a ∪ b)) 6= v ∗ ((¬a ⇒ b)) for some v : V AR → {F, ⊥, T }. Observe that
v ∗ ((a ∪ b)) = v ∗ ((¬a ⇒ b)) for all v : V AR → {F, T }. So we have to check only
truth assignments that involve ⊥. Let v be any v such that v(a) = v(b) =⊥.
We evaluate v ∗ ((a ∪ b) =⊥ ∪ ⊥=⊥ and v ∗ ((¬a ⇒ b)) = ¬ ⊥⇒⊥= F ⇒⊥= T .
This proves that (a ∪ b) 6≡L (¬a ⇒ b).. and hence we have proved (A ∪ B) 6≡L
(¬A ⇒ B).
144
We proved that the classical equivalence defining disjunction in terms of nega-
tion and implication can’t be used for the same goal in L semantics. It does not
mean that we can’t define L disjunction in terms of L implication. In fact, we
prove by simple evaluation that the following holds.
Fact 3.10
The L disjunction is definable in terms of L implication only, i.e. for any
formulas A, B ∈ F
(A ∪ B) ≡L ((A ⇒ B) ⇒ B).
B1 = A1 (A/B).
If A ≡M B, then A1 ≡M B1 .
Example 3.27
Let A1 = (a ⇒ (¬a ∪ b)) and consider a sub formula A = (¬a ∪ b) of A1 . By
Fact 3.10, (¬a ∪ b) ≡L ((¬a ⇒ b) ⇒ b). Take B = ((¬a ⇒ b) ⇒ b) and let
B1 = A1 (A/B) = A1 ((¬a ∪ b)/((¬ ⇒ b) ⇒ b)) = (a ⇒ ((¬a ⇒ b) ⇒ b)). By the
M Equivalence Theorem 3.28
M Equivalence of Languages
Definition 3.44
Given two languages: L1 = LCON1 and L2 = LCON2 , for CON1 6= CON2 . We
say that L1 and L2 are M logically equivalent and denote it by
L 1 ≡M L 2
145
if and only if the following conditions C1, C2 hold.
C1 For any formula A of L1 , there is a formula B of L2 , such that A ≡M B,
C2 For any formula C of L2 , there is a formula D of L1 , such that C ≡M D.
L{¬,⇒} ≡L L{¬,⇒,∪}
Solution
Condition C1 holds because any formula of L{¬,⇒} is a formula of L{¬,⇒,∪} .
Condition C2 holds because the Fact 3.10 equivalence (A ∪ B)≡L ((A ⇒ B) ⇒
B) and the Theorem 3.28.
1. ((a ↑ b) ↑ (a ↑ b) ↑ a)
2. (a ⇒ ¬b) ⇒ ¬a
3. ♦(a ⇒ ¬b) ∪ a, ♦(a ⇒ (¬b ∪ a), ♦a ⇒ ¬b ∪ a
4. (¬♦a ⇒ ¬a), (¬♦a ⇒ ¬a), ¬♦(a ⇒ ¬a)
5. ((a ∪ ¬K¬a)), KK(b ⇒ ¬a), ¬K(a ∪ ¬a)
146
6. (B(a ∩ b) ⇒ Ka), B((a ∩ b) ⇒ Ka)
7. G(a ⇒ b) ⇒ Ga ⇒ Gb), a ⇒ HF a, F F a ⇒ F a
12. (¬(a ∩ b) ∪ a)
147
6. Find definability formula defining ⇒ in terms of ↓.
7. Define ∩ in terms of ⇒ and ¬.
8. Find definability formula defining ∩ in terms of ↓ alone.
9. Given a formula A: (((a ∩ b) ∪ ¬c) ⇒ b). Evaluate (do not use shorthand
notation) v ∗ (A) for truth assignments v : V AR −→ {T, F } such that
(i)v(a) = T, v(b) = F, v(c) = F, v(x) = T for all x ∈ V AR − {a, b, c},
(ii)v(a) = F, v(b) = T, v(c) = T, v(x) = F for all x ∈ V AR − {a, b, c}.
10. Given a formula A: (((a ⇒ ¬b) ∪ b) ⇒ a). Evaluate (use shorthand
notation) v ∗ (A) for all truth assignments restricted to A.
11. Given a formula A: (((a ↓ ¬b) ∪ b) ↑ a). Evaluate (do not use shorthand
notation) v ∗ (A) for truth assignments v : V AR −→ {T, F } such that
(i) v(a)=T, v(b)=F, v(c) =F for all c ∈ V AR − {a, b},
(ii) v(a)=F, v(b)=T, v(c) =T for all c ∈ V AR − {a, b}.
(iii) List all restricted models and counter-models for A.
Write the following natural language statement From the fact that it is
possible that 2 + 2 6= 4 we deduce that it is not possible that 2 + 2 6= 4 or, if
it is possible that 2 + 2 6= 4, then it is not necessary that you go to school.
as a formula . A ∈ F of a language L = L{¬,∩,∪,⇒} .
(i) Find a restricted model v for the formula A.
(ii) Find 3 models w of A such that v ∗ (A) = w∗ (A) the for v from (i).
How many of such models exist?
(iii) Find all models, counter-models (restricted) for A. Use shorthand
notation.
(iv) Is A ∈C?, is A2 ∈T? Justify your answers.
12. Given v : V AR −→ {T, F } such that v ∗ ((¬a ∪ b) ⇒ (a ⇒ ¬c)) = F .
Evaluate: v ∗ (((b ⇒ a) ⇒ (a ⇒ ¬c)) ∪ (a ⇒ b)).
13. Show that all of the truth assignments v1 , v2 , v3 defined below are models
for the formula A : ((a ∩ ¬b) ∪ ¬c).
v1 : V AR −→ {T, F }, is such that v1 (a) = T, v1 (b) = F, v1 (c) = T,
and v1 (x) = F , for all x ∈ V AR − {a, b, c};
v2 : V AR −→ {T, F } is such that v2 (a) = T, v2 (b) = F, v2 (c) = T,
v2 (d) = T, and v2 (x) = F for all x ∈ V AR − {a, b, c, d};
v3 : V AR −→ {T, F } is such that v3 (a) = T, v3 (b) = F, v3 (c) = T,
v3 (d) = T, v3 (e) = T , and v3 (x) = F , for all x ∈ V AR − {a, b, c, d, e}.
14. Prove that for any formula A ∈ F, if A has a model (counter- model),
then it has uncountably many models (counter-models). More precisely,
as many as there are real numbers. Hint Use the Counting Functions
Theorem 3.4.
148
15. Use Generalization Method to determine whether
|= (¬((a ∪ b) ⇒ ((c ⇒ d) ∪ e)) ⇒ ((a ∪ b) ∩ (¬(c ⇒ d) ∩ ¬e))).
18. Use Truth Table and Substitution Methods to prove |= (¬¬A ⇔ A).
19. Use Truth Table and Substitution Methods to prove to prove the Reductio
ad Absurdum tautology ((¬A ⇒ (B ∩ ¬B)) ⇒ A).
20. Use Proof by Contradiction Method to prove the Exportation and Impor-
tation tautology (((A ∩ B) ⇒ C) ⇔ (A ⇒ (B ⇒ C))).
21. For the formulas listed below determine whether they are tautologies or
not. If a formula is not a tautology list its counter-model (restricted). Use
shorthand notation.
(i) A1 = (¬(a ⇒ (b ∩ ¬c)) ⇒ (a ∩ ¬(b ∩ ¬c)))
(ii) A2 = ((a ∩ ¬b) ⇒ ((c ∩ ¬d) ⇒ (a ∩ ¬b)))
(iii) A3 = (¬(A ∩ ¬B) ∪ (A ∩ ¬B))
22. Find all models and a counter-model restricted to G (if exist) for the
following sets G of formulas. Use shorthand notation.
(i) S1 = {a, (a ∩ ¬b), (¬a ⇒ (a ∪ b))}
(ii) S2 = {(a ⇒ b), (c ∩ ¬a), b}
(iii) S3 = {a, (a ∩ ¬b), ¬a, c}
27. Find an infinite number of formulas that are independent from a set
G = {(a ⇒ (a ∪ b)), (a ∪ b), ¬b, (c ⇒ b)}. Use shorthand notation.
29. Give an example of an infinite set G and an infinite set of formulas inde-
pendent from it.
149
Equivalence of Languages
4. Prove, using proper logical equivalences (list them at each step) that
(i) ¬(A ⇔ B) ≡ ((A ∩ ¬B) ∪ (¬A ∩ B)).
(ii) ((B ∩ ¬C) ⇒ (¬A ∪ B)) ≡ ((B ⇒ C) ∪ (A ⇒ B)).
150
3. Extend B, semantics to a language containing the equivalence connective.
Are your semantics well defined as by definition 3.14?
4. Let v : V AR −→ {F, ⊥, T } be any v, such that v ∗ ((a ∪ b) ⇒ (a ⇒ c)) =⊥
under H semantics. Evaluate v ∗ (((b ⇒ a) ⇒ (a ⇒ ¬c)) ∪ (a ⇒ b)).
5. Verify which of the classical tautologies (3.12) are, and which are not L
tautologies.
6. Verify which of the classical tautologies (3.13) are, and which are not L
tautologies.
7. Give an example of 3 formulas
8. For each of 3-valued logic semantics presented in this chapter, find 5 clas-
sical tautologies that are tautologies of that logic.
9. Examine the notion of definability of connectives as defined in section 3.3,
definition 3.16 for L semantics. semantics.
12. Given a set G = {((a ∩ b) ⇒ b), (a ∪ b), a}. Verify whether G is consistent
under L semantics.
13. Given a language L{¬,⇒,∪,∩} . We define: A formula A ∈ F is called M
independent from a set G ⊆ F if and only if the sets G ∪ {A} and
G ∪ {¬A} are both M consistent. I.e. when there are truth assignments
v1 , v2 such that v1 |=M G ∪ {A} and v2 |=M G ∪ {¬A}.
Given a set G = {((a ∩ b) ⇒ b), (a ∪ b), a}.
(i) Find a formula A that is L independent from a set G.
(ii) Find a formula A that is H independent from a set G.
(iii) Find an infinite number of that are L independent from a set G.
(iv) Find an infinite number of that are H independent from a set G.
14. By exercise 3.31 the classical logical equivalence (A ∪ B) ≡ (¬A ⇒ B)
does not hold under L semantics, i.e. that there are formulas A, B, such
that (A ∪ B) 6≡L (¬A ⇒ B). Show 3 formulas A,B such that it does hold
for L semantics, i.e. such that are formulas A, B, such that (A ∪ B) ≡L
(¬A ⇒ B).
151
152
Chapter 4
Proof systems are built to prove, construct formal proofs of statements formu-
lated in a given language formulated in a given language. First component of
any proof system is hence its formal language L. Proof systems can be thought
as an inference machine with special statements, called provable statements, or
theorems being its final products. The starting points are called axioms of the
proof system. We distinguish two kinds of axioms: logic LA and specific SA.
When building a proof system for a given language and its semantics i.e. for a
logic defined semantically we choose as a set of logical axioms LA some subset of
tautologies, i.e. statements always true. This is why we call them logical axioms.
A proof system with only logic axioms LA is also called logic proof systems,
or just proof systems for short. If we build a proof system for which there is
no known semantics, as it has happened in the case of classical, intuitionistic,
and modal logics, we think about the logical axioms as statements universally
true. We choose as axioms (finite set) the statements we for sure want to be
universally true, and whatever semantics follows they must be tautologies with
respect to it. Logical axioms are hence not only tautologies under an established
semantics, but they also guide us how to establish a semantics, when it is yet
unknown.
For the set of specific axioms SA we choose these formulas of the language that
describe our knowledge of a universe we want to prove facts about. They are not
universally true, they are true only in the universe we are interested to describe
and investigate. This is why we call them specific axioms. A proof system with
logical axioms LA and specific axioms SA is called a formal theory based on a
proof system with logic axioms LA.
The inference machine is defined by a finite set of rules, called inference rules.
153
The inference rules describe the way we are allowed to transform the information
within the system with axioms as a staring point. The process of this transfor-
mation is called a formal proof. The provable formulas for which we have a
formal proof are called consequences of the axioms, or theorem, or just simple
provable formulas. We use proof systems not only to be able to build formal
proofs in them, but also to search for proofs of given statements of their the
language. We distinguish special proof systems for which it is possible to define
a mechanical method for determining, given any statement of A, but which also
generates a proof, is called syntactically decidable or automatically decidable, or
an automated system
When building a proof system we choose not only axioms of the system, but also
specific rules of inference. The choice of rules is often linked, as was the choice of
axioms, with a given semantics. We want the rules to preserve the truthfulness
of what we are proving from axioms via the rules. Rules with this property
are called sound rules and the system a sound proof system. The notion of
truthfulness is always defined by a given propositional, or predicate language
L semantics M. Rules of inference can be sound under one semantics and not
sound under another. When developing a proof system S the first goal is prove
a theorem, called Soundness Theorem about its relationship with its semantics
M. It states that the following holds for any formula A of the language L of the
system S. If a formula A is provable from logical axioms LA of S only, then A
is a tautology under the semantics M.
A proof system can be sound under one semantics, and not sound under the
other. For example a set of axioms and rules sound under classical logic seman-
tics might not be sound under intuitionistic semantics, H, L, K semantics, or
others. This is why we talk about proof systems for classical logic, intuitionistic
logic, for modal logics etc. In general there are many proof systems that are
sound under a given semantics, i.e. there are many sound proof systems for a
given logic semantically defined. We present some examples at the end of the
chapter. Given a proof system S with logical axioms LA that is sound under a
given semantics M . Let TM be a set of all tautologies defined by the semantics
M , i.e. TM = {A : |=M A}. A natural questions arises: are all tautologies
defined by the semantics M , provable in the system S that is sound under the
semantics M . The positive answer to this question is called a completeness
property of the system S. Because we ask the completeness property question
for sound systems only we put it in a form of a theorem called a Completeness
Theorem for a proof system S, under a semantics M . It states that the following
holds for any formula A of the language L of the system S. A formula A is
provable in S if and only if A is a tautology under the semantics M . We write
it symbolically as: `S A if and only if |=M A. The Completeness Theorem is
composed from two parts: the Soundness Theorem and the completeness part
that proves the completeness property of a sound system.
Proving the Soundness Theorem for S under a semantics M is usually a straight-
forward and not a very difficult task. We first prove that all logical axioms are
154
M tautologies, and then that all inference rules of the system preserve the notion
of the M truth ( M model). Proving the completeness part of the Completeness
Theorem is always a crucial and very difficult task.
We will study two proofs of the Completeness Theorem for classical propositional
Hilbert style proof system in chapter 5, and a constructive proofs for automated
theorem proving systems for classical logic the chapter 6.
Observe that we formulated all these basic theorems linking semantics and syn-
tax (provability) in a general manner. As we first consider propositional lan-
guages (chapters 5, 6, 7) and hence we use proof systems for propositional logics
as examples. The case of predicate logics will be discussed in chapters 8, 9, 10,
11.
4.1 Syntax
In this section we formulate a definition of a proof system S by specifying and
defining and all its components. We define a notion of a formal proof in a
given proof system, and give simple examples of different proof systems. When
defining a proof system S we specify, as the first step, its formal language L.
When It can be a propositional, or a predicate language. It is a first component
of the proof system S. Given a set F of well formed formulas, of the language
L, we often extend this set, and hence the language L to a set E of expressions
build out of the language L, and some additional symbols, if needed. It is a
second component of the proof system S. Proof systems act as an inference
machine, with provable expressions being its final products. This inference
machine is defined by setting, as a starting point a certain non-empty, proper
subset LA of E, called a set of logical axioms of the system S. The production
of provable formulas is to be done by the means of inference rules. The inference
rules transform an expression, or finite string of expressions, called premisses,
into another expression, called conclusion. At this stage the rules don’t carry
any meaning - they define only how to transform strings of symbols of our
language into another string of symbols. This is a reason why investigation of
proof systems is called syntax or syntactic investigation as opposed to semantcal
methods, which deal with semantics of the language and hence of the proof
system. The syntax- semantics connection within proof systems is established
by Soundness and Completeness theorems and will be discussed in detail in the
section 4.2.
155
formulas; E is a set of expressions of S; LA ⊆ E is a non empty set of logical
axioms of the system; R is a finite set of rules of inference.
156
4. The set R of rules of inference of S
We assume that the proof system S contains a finite number of inference rules.
We assume that each rule has a finite number of premisses and one conclusion.
We also assume that one can effectively decide, for any inference rule, whether
a given string of expressions form its premisses and conclusion or do not, i.e.
that all rules r ∈ R are primitivvely recursive.
We put it in a formal definition as follows.
P1 ; P2 ; .... ; Pm
(r) .
A
Semantical Link. We want the rules of inference to preserve truthfulness i.e. to
be sound under the semantics M.
Formal Proofs in S
Given a proof system S = (L, E, LA, R). Final products of a single or multiple
use of the inference rules of S, with logical axioms LA taken as a starting point
are called provable expressions of the system S. A single use of an inference rule
is called a direct consequence. A multiple application of rules of inference with
axioms taken as a starting point is called a formal proof. Formal definitions are
as follows.
157
Definition 4.3 (DirectConsequence)
A conclusion of a rule of inference is called a direct consequence of its premisses.
I.e. for any rule of inference r ∈ R, if (P1 , ...Pn , C) ∈ r, then C is called a direct
consequence of P1 , ...Pn by virtue of r.
E1 ∈ LA, En = E,
`S E
B
S1 = (L{P, ⇒} , F, {(A ⇒ A)}, (r)
) (4.2)
PB
where A, B are any formulas. Observe that even the system S1 has only one
axiom, it represents an infinite number of formulas. We call such axiom an
axiom schema.
Consider now a system S2
B
S2 = (L{P,⇒} , F, {(a ⇒ a)}, (r) ), (4.3)
PB
where a ∈ V AR is any variable (atomic formula) and B ∈ F is any formula.
Observe that the system S2 also has only one axiom similar to the axiom of S1 ,
both systems have the same rule of inference but they are very different proof
systems. For example a formula ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))) is an
158
axiom of the system S1 for A = (P a ⇒ (b ⇒ c) but is not an axiom of the
system S2 , as this systems permits axioms of the form: (a ⇒ a) for a being a
propositional variable.
Exercise 4.1 Let S1 , S2 be proof systems (7.2), (7.3), respectively. Show that
A1 = (a ⇒ a), A2 = P (a ⇒ a).
axiom rule (r) application
for B = (a ⇒ a)
159
A2 = P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))),
rule (r) application
A3 = P P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))),
rule (r) application
A4 = P P P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))).
rule (r) application
It is not a proof in S2 . Moreover
Observe that even if the set of axioms and the inference rules of the proof system
are primitively recursive it doesn’t mean that the notion of ”provable expres-
sion” is also primitively recursive, i.e. that there always will be an effective,
mechanical method (effective procedure) for determining, given any expression
A of the system, whether there is a proof of A. We define the following notions
Observe that the above notion of decidability of the system S does not require
us to find a proof, it requires only a mechanical procedure of deciding whether
there is, or there is no such a proof. We hence introduce a following notion.
The existence of prove systems for classical logic and mathematics that are
syntactically decidable or syntactically semi-decidable was stated (in a differ-
ent form) by German mathematician David Hilbert in early 1900 as a part of
what is called Hilbert’s program. The main goal of Hilbert’s program was to
provide secure foundations for all mathematics. In particular it addressed the
problem of decidability; it said that here should be an algorithm for deciding
the truth or falsity of any mathematical statement. Moreover, it should use
only ”finitistic” reasoning methods. Kurt Gdel showed in 1931 that most of the
goals of Hilbert’s program were impossible to achieve, at least if interpreted in
160
the most obvious way. Nevertheless, Gerhard Gentzen in his work published in
1934/1935 gave a positive answer to existence of syntactical decidability. He
invented proof systems for classical and intiutionistic logics, now called Gentzen
style formalizations. They formed a basis for development of Automated The-
orem Proving area of mathematics and computer science. We will study the
Gentzen style formalizations in chapter ??.
Automated proof systems are also called automated theorem proving systems,
Gentzen style formalizations, syntactically decidable systems and and we use all
of these terms interchangeably.
Example 4.1 Any complete Hilbert style proof system for classical proposi-
tional logic is an example of a decidable, but not syntactically decidable proof
system. We conclude its decidability from the Completeness Theorem (to be
proved in next chapter) and the decidability of the notion of classical tautology
(proved in chapter 3).
Example 4.2 The Gentzen style proof systems for classical and intuiionistic
propositional logics presented in chapter ??, are examples of proof systems that
are of both decidable and syntactically decidable.
Fact 4.1
The systems proof systems S1 and S2 defined by (7.2) and (7.3), respectively
are syntactically decidable.
Proof Let’s now to think how we can search for a proof in S2 of a formula
P P ((P a ⇒ (b ⇒ c)).
If P P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))) had the proof, the only last step
in this proof would have been the application of the rule (r) to the formula
P P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))). This formula, in turn, if it had
the proof, the only last step in its proof would have been the application of
the rule r to the formula P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))). And again,
this one could be obtained only from the formula ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒
(b ⇒ c))) by the virtue of the rule r. Here the search process stops; the rule r
puts P in front of the formulas, hence couldn’t be applied here. The formula
161
((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))) isn’t an axiom of S2 , what means that
the only possible way of finding the proof has failed, i.e. we have proved that
6`S1 P P P ((P a ⇒ (b ⇒ c)) ⇒ (P a ⇒ (b ⇒ c))).
The above example of proof search in S2 defines the following an effective,
automatic Procedure S1 , S2 of searching for a proof of our formula in both
our proof systems. If the search ends with an axiom, we have a proof, if it
doesn’t end with an axiom it means that the proof does not exists. We have
described it, as an example, for one particular formula. It can be easily extended
to any formula A of L{P,⇒} as follows.
Procedure S1 , S2
Step : Check the main connective of A.
If main connective is P , it means that A was obtained by the rule r.
Erase the main connective P .
Repeat until no P left.
If the main connective is ⇒,check if a formula A is an axiom.
If it is an axiom , STOP and YES, we have a proof.
If it is not an axiom , STOP and NO, proof does not exist.
Observe also, that the systems S1 and S2 are such that we can easily describe
a general form of their provable formulas defined by (4.1) as PS = {E ∈ E :
`S E}. Namely we have the following.
PS1 = {P n (A ⇒ A) : n ∈ N, A ∈ F},
PS2 = {P n (a ⇒ a) : n ∈ N, a ∈ V AR},
The proof systems S1 and S2 are very simple, indeed. Here is an example of
another two, similar but slightly more complex proof systems.
Consider two proof systems S3 and S4 of the language L{∪,¬} with the set of
expressions E = F and is defined as follows.
(A ∪ ¬A)
S3 = (L{∪,¬} , F, {(A∪¬A)}, (r) , for any A, B ∈ F ). (4.4)
(B ∪ (A ∪ ¬A))
162
B
S4 = (L{∪,¬} , F, {(A∪¬A)}, (r) , for any A, B ∈ F ), (4.5)
(B ∪ (A ∪ ¬A))
Exercise 4.2 Given proof systems S3 and S4 defined by (4.4), (7.40), respec-
tively.
1. Describe the sets PS3 , PS4 of provable formulas of S3 and S4 .
2. Decide whether is it true/ false that PS3 = PS4 . If yes, prove it, if not, give
an example of a formula A such that A ∈ PS4 and A 6∈ PS3 , or vice versa.
Solution 1.
Let’s first describe the set of provable formulas of both systems. Consider proof
system S3 . Obviously, for any formula A ∈ F, (A ∪ ¬A), as it is the axiom. It
constitutes a proof of length 1 A1 = (A ∪ ¬A)and we have that
`S3 (A ∪ ¬A).
One application of the inference rule (r) to axiom(A ∪ ¬A) gives us a proof
A1 = (A ∪ ¬A), A2 = ((A ∪ ¬A) ∪ (A ∪ ¬A)), and hence
The application of the rule (r) to the already proven above formula A2 give us
the proof A1 = (A ∪ ¬A), A2 = ((A ∪ ¬A) ∪ (A ∪ ¬A)), A3 = (((A ∪ ¬A) ∪ (A ∪
¬A)) ∪ (A ∪ ¬A)), and
It is easy to see that all provable formulas of S3 will be of the form of the proper
disjunction of the axiom of S3 , what we denote as follows:
[
PS3 = { (A ∪ ¬A)n : A ∈ F}, (4.6)
n∈N
163
A1 , A2 , A3 , A4 by inserting axiom (C ∪ ¬C) (or axiom (A ∪ ¬A), if we wish
as A3 step of the proof. We have to remember that the definition 4.4 of the
formal proof allows us to insert an axiom in any place within the proof. A1 =
(A ∪ ¬A), A2 = (B ∪ (A ∪ ¬A)), A3 = (C ∪ ¬C), A4 = (A ∪ (C ∪ ¬C)) and
hence
Multiple application of the rule (r) in S4 means its application to multiple forms
of the axiom. Finally it is clear that we can only construct formal proofs of all
possible formulas of the form (B ∪ (A ∪ ¬A)), and of course of a form of any
axiom (proofs of the length 1) (A ∪ ¬A) for A, B being all possible formulas.
Remark that by saying A, B ∈ F we do not say that A 6= B, that we do not
exclude that case A = B. In particular case we have that
Hence
Solution 2.
D = n∈N (A ∪ ¬A)n .
S
We prove now that PS3 ⊆ PS4 . S Let D ∈ PS3 . By (4.6)
D = n∈N (A ∪ ¬A)n = ( n∈N (A ∪ ¬A)n−1 ∪ (A ∪
S
Observe that
S by definition n−1
¬A)) and n∈N (A ∪ ¬A) is a formula of L{∪,¬} . We can denote it by B. We
have proved in (4.7) that for any B ∈ F, `S4 (B ∪ (A ∪ ¬A)). But by definition
D = (B ∪ (A ∪ ¬A)), hence we proved that D ∈ PS4 .This ends the proof.
This proves that ((a ∪ ¬b) ∪ (a ∪ ¬a)) ∈ PS4 . Obviously 6 `S3 ((a ∪ ¬b) ∪ (a ∪ ¬a))
and ((a ∪ ¬b) ∪ (a ∪ ¬a)) 6∈ PS3 . We have proved that the proof systems S3 and
S4 defined by (4.4), (7.40) are such that PS3 ⊆ PS4 and PS3 6= PS4 .
Consider now a following proof system S5 .
164
Exercise 4.3
Given proof systems S5 defined by (7.41).
1. Find a formal proof of a formula ¬(A ⇒ (A ∪ B)) in S5 , i.e. show that
`S5 ¬(A ⇒ (A ∪ B)).
2. Find a formal proof of a formula ¬((a ∪ ¬b) ⇒ ((a ∪ ¬b) ∪ (a ∪ ¬a))).
Solution
1. We construct a formal proof B1 , B2 , B3 , B4 of the formula ¬(A ⇒ (A ∪ B))
as follows. We write comments next to each step of the proof.
B1 = (A ⇒ (A ∪ B)) axiom, B2 = (A ⇒ (A ∪ B)) axiom,
B3 = ((A ⇒ (A ∪ B)) ∪ ¬(A ⇒ (A ∪ B))) rule (r1) application to B1 and B2 ,
B4 = ¬(A ⇒ (A ∪ B)) rule (r2) application to B3
for A = (A ⇒ (A ∪ B)) and B = ¬(A ⇒ (A ∪ B)).
Observation 4.1 Observe that the formula ¬((a ∪ ¬b) ⇒ ((a ∪ ¬b) ∪ (a ∪ ¬a)))
is a particular case of the formula ¬(A ⇒ (A ∪ B)) for A = (a ∪ ¬b) and
B = (a ∪ ¬a) and its proof is just a particular case of the proof constructed in
case 1.
165
4.1.1 Consequence Operation
Given a proof system S = (L, E, LA, R). While proving expressions we often
use some extra information available, besides the axioms of the proof system.
This extra information is called hypotheses in the proof.
E1 , E2 , . . . En
E1 ∈ LA ∪ Γ, En = E
When the set of hypothesis Γ is a finite set and Γ = {B1 , B2 , ..., Bn }, then we
write
B1 , B2 , ..., Bn `S E
instead of {B1 , B2 , ..., Bn } `S E. The case when Γ is an empty set i.e. when
Γ = ∅ is a special one. By the definition of a proof of E from Γ, ∅ ` E means
that in the proof of E only logical axioms LA of S were used. We hence write
it as we did before
`S E
to denote that E has a proof from the empty set Γ. The set of all expressions
provable from Γ (and logical axioms LA in S is denoted by PS (Γ), i.e.
166
When discussing properties of provability in proof systems we often use a notion
of a consequence operation. In order to follow this tradition we call provable
expressions from Γ in S consequences of Γ. The set of all expressions provable
is then called the set of all consequences from Γ. We observe that when talking
about consequences of Γ in S, we define in fact a function which to every set
Γ ⊆ E assigns a set of all its consequences. We denote this function by CnS
and adopt the following definition.
167
(t3) if F1 ⊆ C(F2 ) and F2 ⊆ C(F3 ) , then F1 ⊆ C(F3 ), transitivity.
We say that the consequence operation C has a finite character if additionally
it satisfies the following condition t4.
(t4) if a formula B ∈ C(F ), then there exists a finite set F0 ⊆ F , such
that B ∈ C(F0 ). finiteness.
The monotonicity condition (t2) and transitivity condition (t3) are often re-
placed by the following conditions (t2’), (t3’), respectively.
For any formula B ∈ F, any any sets F, F 0 , ∈ 2F ,
Theorem 4.1
The Tarski definition 4.10 is equivalent with definitions where one, or both con-
ditions (t2), (t3) are replaced respectively by conditions (t2’), (t3’)given by equa-
tions (4.13) and (4.14).
Proof We prove the equivalency of conditions (t1) - (t3) and (t1) - (t3’). We
leave the proof of the other equivalency to the reader.
Assume (t3). By substituting
in (t3) we obtain
C(C(F )) ⊆ C(F ).
On the other hand, it follows from (t1) and (t2)
which with the previous inclusion gives (t3’). Conversely, suppose that (t3’) is
satisfied. If F2 ⊆ C(F3 ), then by (t2) we obtain C(F2 ) ⊆ C(C(F3 )). By (t3’)
C(C(F3 )) = (C(F3 ), hence C(F2 ) ⊆ (C(F3 ) and we proved (t3).
168
Definition 4.11 (Deductive System)
Given a formal language L = (A, F) and a Tarski consequence C (definition
4.10). A system
D = (L, C)
is called a Tarski deductive system for the language L.
Theorem 4.2
Given a proof system S = (L, E, LA, R). The consequence operation CnS is a
Tarski consequence C in the language L of the system S and the system
DS = (L, CnS )
Proof
By definition 4.14, the consequence operation CnS : 2E −→ 2E is given
by a formula CnS (Γ) = {E ∈ E : Γ `S E}. We have to show that for any
Γ, Γ0 , Γ1 , Γ2 , Γ3 ∈ 2F conditions (t1) - (t4) of the definition 4.14 hold. The
reflexivity condition (t1) becomes Γ ⊆ CnS (Γ). Let E ∈ Γ. The one element
sequence E is a proof of E from Γ, hence we proved that E ∈ C(Γ) and (t1)
holds. To prove the transitivity condition (t2) assume now that Γ1 ⊆ Γ2 . Let
E ∈ CnS (Γ1 ). It means that Γ1 `S E, i.e E has a formal proof from Γ1 , but
Γ1 ⊆ Γ2 , hence this proof also is a proof from Γ2 , and E ∈ CnS (Γ2 ). This
proves that CnS (Γ1 ) ⊆ CnS (Γ2 ) and the condition (t2) holds. Let now E ∈ Γ1
and Γ1 ⊆ CnS (Γ2 ), so E ∈ CnS (Γ) 2. Let E1 , . . . , En be a formal proof of E
from Γ2 . But Γ2 ⊆ CnS (Γ3 ). It means that any expression from Γ2 has a formal
proof from Γ3 . In particular, all expression in the proof E1 , . . . , En that belong
to Γ2 have their formal proofs from Γ3 . Replacing all these expressions by their
proofs from Γ3 we obtain a proof of E from Γ3 . This proves that Γ1 ⊆ CnS Γ3
and the transitivity condition (t3) holds. Let now E ∈ CnS Γ. This means that
E has a proof E1 , . . . , En from Γ. The set Γ0 = {E1 , . . . , En } is obviously a
169
finite subset of Γ and E ∈ CnS Γ0 and (t4) holds.
170
4.1.2 Syntactic Consistency
Formal theories play crucial role in mathematics and were historically defined
for classical first order logic and consequently for other first and higher order
logics. They are routinely called first order theories. We will discuss them in
more detail in chapter 10 dealing formally with classical predicate logic. First
order theories are hence based on proof systems S with a predicate (first order)
language L. We will call them for short first order proof systems.
We can and we sometimes consider formal theories based on propositional log-
ics, i.e. based on proof systems with language L being propositional. We will
call them propositional theories.
Given a proof system S = (L, E, LA, R). We build (define) a formal theory
based on S as follows.
1. We select a certain finite subset SA of expressions of S, disjoint with the
logical axioms LA of S, i.e. such that LA ∩ SA = ∅.. The set SA is called a set
of specific axioms of the formal theory based on S.
The set ESA of all expressions of the language LSA provable from the set specific
axioms SA (and logical axioms LA) i.e. the set
T(SA) = {E ∈ ESA : SA `S E } (4.17)
171
is called the set of all theorems of the theory T .
If the set SA of specific axioms of T is empty, then the theory T is, by definition,
identified with the system S, i.e. T = S = (L, E, LA, R).
SA 6 `S E;
Observe that the definition 4.13 has purely syntactic meaning. It also reflexes
our intuition what proper provability should mean. it says that a formal the-
ory T based on a proof system S is consistent only when it does not prove all
expressions (formulas in particular cases) of LSA . The theory T such that it
proves everything stated in LSA obviously should be, and its defined as incon-
sistent. In particular, we have the following syntactic definition of consistency-
inconsistency for any proof system S.
4.2 Semantics
We define formally a semantics for a given proof system S = (L, E, LA, R) by
specifying the semantic links of all its components as follows.
172
proof system S, and for any expression E ∈ E there is a formula A ∈ F,
such that E≡M A. For example, in the automated theorem proving system
RS presented in chapter 6 the expressions are finite sequences of formulas of
L = L¬,∩,∪,⇒ . We extend our classical semantics for L to the set F ∗ of all finite
sequences of formulas as follows: for any v : V AR −→ {F, T } and any ∆ ∈ F ∗ ,
∆ = A1 , A2 , ..An , v ∗ (∆) = v ∗ (A1 , A2 , ..An ) = v ∗ (A1 ) ∪ v ∗ (A2 ) ∪ .... ∪ v ∗ (An ),
i.e. ∆ ≡ (A1 ∪ A2 ∪ ... ∪ An ). Sometimes, like in case of Resolution based proof
systems we have also to prove a semantic equivalency of a given formula A of L
with some set EA of expressions (sets of clauses ) representing the formula A.
Given a semantics M for L and its extension to the set E of all expressions. We
extend the notion of tautology to the set L of expressions and write |=M E to
denote that the expression E ∈ E is a tautology under semantics M. We denote
TM = {E ∈ E : |=M E}
While designing a proof system S we want the logical axioms LA to be a subset
of expressions that are tautologies of under the semantics M, i.e.
LA ⊆ TM .
We can, and we often do, invent proof systems with languages without yet es-
tablished semantics. In this case the logical axioms LA serve as description of
properties of tautologies under a future semantics yet to be built. We want to
choose as logical axioms of a proof system S are not only tautologies under an
already known semantics M, but they can also guide us how to define a seman-
tics when it is yet unknown.
173
In case of a propositional language LCON and an extensional semantics M the
M models M are defined in terms of the truth assignment v : V AR −→ LV ,
where LV is the set of logical values with a distinguished value T. The general
definition 4.15 becomes a following definition for a propositional language L and
its extensional semantics M.
A rule of inference be sound under different semantics, but also rules of inference
can be sound under one semantics and not sound under the other.
174
Consider the rule (r2).
The rule (r2) is sound under classical and L by straightforward eveluation.
Assume now v : V AR −→ {F, ⊥, T } be any truth assignment, such that
v |=M ¬¬A, i.e. such that v ∗ (¬¬A) = T under H semantics. We have that
v ∗ (¬¬A) = ¬¬v ∗ (A) = T if and only if ¬v ∗ (A) = F if and only if v ∗ (A) = T or
v ∗ (A) =⊥. This proves that that it is possible to have v |=M ¬¬A and v 6 |=H A,
i.e. that (r2) is not sound.
P1 ∩ P2 ∩ . . . ∩ Pm ≡M C. (4.23)
Example 4.4
Given a propositional language L{¬,∪,⇒} . Consider two rules of inference:
A; B A
(r1) and (r2) .
(A ∪ ¬B) ¬¬A
Both rules (r1) and (r2) are sound under classical and H semantics. The rule
(r2) is strongly and s-strongly sound under classical semantics but is not strongly
sound under H semantics. The rule (r1) in not strongly sound under either
semantics.
175
under both semantics.. Take now v such that v(A) = T and v(B) = F , we get
v ∗ ((A ∪ ¬B)) = F ∪ T = T .This proves that v |= (A ∪ ¬B) and v|=H (A ∪ ¬B).
Obviously v 6|= {A, B} and v 6 |=H {A, B}. This proves that (r1) in not strongly
sound under either semantics.
Consider (r2). It is strongly sound under classical semantic. By (8.77) and the
fact that A 𠪪A (r2) is s-strongly sound and so is strongly sound.
(r2) is sound under H semantics. Assume A = T . We evaluate (in shorthand
notation) ¬¬A = ¬¬T = ¬F = T . (r2) is not strongly sound under H seman-
tics. Take v such that v ∗ (A) =⊥, then v ∗ (¬¬A) = ¬¬ ⊥= ¬F = T . This
proves that there is v such that v |= ¬¬A and v 6|= A and (4.22) does not hold
and so (r2) is not strongly sound.
This also proves that A 6≡H ¬¬A, i.e. (r2) is not s-strongly sound (4.22).
Now we are ready to define a notion of a sound and strongly sound proof system.
Strongly sound proof systems play a role in constructive proofs of completeness
theorem. This is why we introduced and singled them out here.
Here is a simple but an important fact about the notion of strong soundness.
The strong soundness of proof systems guarantee a correctness of of construc-
tive proofs of completeness theorems for automated theorem proving systems
introduced and examined in chapter 6 and in chapter 10.
Proof
Consider a rule
P1 ; P2 ; .... ; Pm
(r) .
C
176
By definition 4.17, (r) is sound and additionally the following implication holds
for any M structures M,
holds for all M structures M. But M 6|=M {P1 , P2 , .Pm } if and only if there
is 1 ≤ i ≤ m such that M 6|=M Pi . This ends the proof.
where
(A ⇒ B)
(r) .
(B ⇒ (A ⇒ B))
S is sound, but not strongly sound under classical and L semantics. It is not
sound under H semantics.
C1. Both axioms are basic classical tautologies. Hence to prove that first axiom
is L tautology we we have to verify only the case (shorthand notation) A =⊥.
But ¬¬ ⊥⇒⊥= ¬ ⊥⇒⊥=⊥⇒⊥= T and we proved |=L (¬¬A ⇒ A). Observe
that (A ⇒ (¬A ⇒ B)) =⊥ if and only if A = T and (¬A ⇒ B) =⊥ if and
only if (¬T ⇒ B) =⊥ if and only if (F ⇒ B) =⊥, what is impossible undef
L semantics. Hence |=L (A ⇒ (¬A ⇒ B)). We prove, as in example 4.3, that
6|=H (¬¬A ⇒ A), and hence S is not sound under H semantics.
C2. The rule (r) is not strongly sound under neither classical nor L semantic.
Let v : V AR −→ {F, ⊥, T } be any truth assignment, such that v ∗ (B) = F
and v ∗ (A) = F . Obviously v |= (B ⇒ (A ⇒ B)) and v |=L (B ⇒ (A ⇒ B)),
but v 6|= (A ⇒ B) and v 6|=L (A ⇒ B).
Nevertheless, (r) is sound under the both semantics by example 4.3.
177
Given a predicate proof system S and a semantics M.
The following holds.
PS ⊆ TM , (4.24)
if `S E then |=M E.
The next step in developing a logic is to answer next necessary and a difficult
question: Given a proof system S, about which we know that all it proves it
true (tautology)with respect to a given semantics. Can we prove all we know to
be true (all tautologies) with respect to the given semantics?
178
4.3 Exercises and Examples
Exercise 4.4
Given a proof system:
(A ⇒ B)
S = (L{¬,⇒} , E = F LA = {(A ⇒ A), (A ⇒ (¬A ⇒ B))}, (r) ).
(B ⇒ (A ⇒ B))
1. Prove that S is sound, but not strongly sound under classical semantics.
2. Prove that S is not sound under K semantics.
3. Write a formal proof in S with 2 applications of the rule (r).
Solution
Parts 1 and 2. In order to prove 1. and 2. we have to verify conditions 1., 2.
and bf 2.’ of definition 8.27. Observe that both axioms of S are basic classical
tautologies. Consider the rule of inference of S.
(A ⇒ B)
(r) .
(B ⇒ (A ⇒ B))
Take any v such that v ∗ ((A ⇒ B))) = T . We evaluate logical value of the
conclusion under the truth assignment v as follows.
v ∗ (B ⇒ (A ⇒ B)) = v ∗ (B) ⇒ T = T
for any B and any value of v ∗ (B). This proves that S is sound under classical
semantics. S is not strongly sound as (A ⇒ B) 6≡ (B ⇒ (A ⇒ B)).
System S is not sound under K semantics because axiom (A ⇒ A) is not a K
semantics tautology.
Part 3. There are many solutions, i.e. one can construct many forrmal proofs.
Here is one of them. For example, one of the formal proofs is a sequence
A1 , A2 , A3 , where
A1 = (A ⇒ A)
(Axiom)
A2 = (A ⇒ (A ⇒ A))
Rule (r) application 1 for A = A, B = A.
A3 = ((A ⇒ A) ⇒ (A ⇒ (A ⇒ A)))
Rule (r) application 2 for A = A, B = (A ⇒ A).
Exercise 4.5
Prove, by constructing a formal proof that
`S ((¬A ⇒ B) ⇒ (A ⇒ (¬A ⇒ B))),
where S is the proof system from Exercise 5.9.
179
Solution
Required formal proof is a sequence A1 , A2 , where
A1 = (A ⇒ (¬A ⇒ B))
Axiom
A2 = ((¬A ⇒ B) ⇒ (A ⇒ (¬A ⇒ B)))
Rule (r) application for A = A, B = (¬A ⇒ B).
Observe that we needed only one application of the rule (r). One more applica-
tion of the rule (r) to A2 gives another solution to Exercise 5.9, namely a proof
A1 , A2 , A3 for A1 , A2 defined above and
A3 = ((A ⇒ (¬A ⇒ B)) ⇒ (¬A ⇒ B) ⇒ (A ⇒ (¬A ⇒ B)))
Rule (r) application for A = (¬A ⇒ B) and B = (A ⇒ (¬A ⇒ B)).
Exercise 4.6
Given a proof system:
(A ⇒ B)
S = (L{∪,⇒} , F, LA = {A1, A2}, (r) ),
(A ⇒ (A ⇒ B))
Solution
Part 1. Axioms of S are basic classical tautologies. The proof (in shorthand
notation) of soundness of the rule of inference is the following. Assume (A ⇒
B) = T . Hence the logical value of conclusion is (A ⇒ (A ⇒ B)) = (A ⇒ T ) =
T for all A. S is not sound under K semantics. Let’s take truth assignment
such that A =⊥, B =⊥. The logical value of axiom A1 is as follows.
(A ⇒ (A ∪ B)) = (⊥⇒ (⊥ ∪ ⊥)) =⊥ and 6 |=K (A ⇒ (A ∪ B)). Observe that
the v such that A =⊥, B =⊥ is not the only v that makes A1 6= T , i.e. proves
that 6 |=K A1.
(A ⇒ (A ∪ B)) 6= T if and only if (A ⇒ (A ∪ B)) = F or (A ⇒ (A ∪ B)) =⊥.
The first case is impossible because A1 is a classical tautology. Consider the
second case. (A ⇒ (A ∪ B)) =⊥ in two cases. c1. A =⊥ and (A ∪ B) = F , i.e.
(⊥ ∪B) = F , what is impossible. c2. A = T and (A ∪ B) =⊥, i.e. (T ∪ B) =⊥,
what is impossible. c3. A =⊥ and (A∪B) =⊥, i.e. (⊥ ∪B) =⊥. This is possible
for B =⊥ or B = F , i.e when A =⊥, B =⊥ or A =⊥, B = F . From the above
observation we get a second solution. S is not sound under K semantics. Axiom
A1 is not K semantics tautology. There are exactly two truth assignments v,
180
such that v 6|= A1. One is, as defined in the first solution, namely A =⊥, B =⊥.
The second is A =⊥, B = F .
Part 2. The formal proof B1 , B2 , B3 is as follows.
B1 = (A ⇒ (A ∪ B))
Axiom
B2 = (A ⇒ (A ⇒ (A ∪ B)))
Rule (r) application for A = A and B = (A ∪ B)
B3 = (A ⇒ (A ⇒ (A ⇒ (A ∪ B))))
Rule (r) application for A = A and B = (A ⇒ (A ∪ B)).
Part 3. The formal proof B1 , B2 is as follows.
B1 = (A ⇒ (B ⇒ A))
Axiom
B2 = (A ⇒ (A ⇒ (B ⇒ A))).
Rule (r) application for A = A and B = (B ⇒ A).
Exercise 4.7
Let S be the following proof system:
A ;B A ; (A ∪ B)
(r1) , (r2) .
(A ∪ ¬B) B
`S ¬(A ⇒ (A ∪ B)).
Solution
Part 1. The system is not sound. Take any v such that v ∗ (A) = T and
v ∗ (B) = F . The premiss (A∪B of the rule (r2) is T under v, and the conclusion
under v is v ∗ (B) = F .
Part 2. The proof of ¬(A ⇒ (A ∪ B)) is as follows.
B1 : (A ⇒ (A ∪ B)),
axiom
B2 : (A ⇒ (A ∪ B)),
axiom
181
B3 : ((A ⇒ (A ∪ B)) ∪ ¬(A ⇒ (A ∪ B))),
rule (r1) application to B1 and B2
B4 : ¬(A ⇒ (A ∪ B)),
rule (r2) application to B1 and B3 ).
Part 3. System S is not sound, so existence of a proof does not guarantee that
what we proved is a tautology. Moreover, the proof of ¬(A ⇒ (A ∪ B)) used
rule (r2) that is not sound.
Exercise 4.8
Create a 3 valued extensional semantics M for the language
L{¬, L, ∪, ⇒} by defining the connectives ¬, ∪, ⇒ on a set {F, ⊥, T } of logical
values. You must follow the following assumptions a1, a2.
a1 The third logical value value is intermediate between truth and falsity, i.e.
the set of logical values is ordered as follows: F <⊥< T.
a2 The value T is the designated value. The semantics has to model a situation
in which one ”likes” only truth; i.e. the connective L must be such that LT = T ,
L ⊥= F, and LF = F . The connectives ¬, ∪, ⇒ can be defined as you wish,
but you have to define them in such a way to make sure that
|=M (LA ∪ ¬LA).
Solution
Here is a simple M semantics. We define the logical connectives by writing
functions defining connectives in form of the truth tables.
M Semantics
L F ⊥ T ¬ F ⊥ T
F F T T F F
∩ F ⊥ T ∪ F ⊥ T ⇒ F ⊥ T
F F F F F F ⊥ T F T T T
⊥ F ⊥ ⊥ ⊥ ⊥ T T ⊥ T ⊥ T
T F ⊥ T T T T T T F F T
We verify whether the condition s3 is satisfied, i.e. whether |=LK (LA ∪ ¬LA)
by simple evaluation. Let v : V AR −→ {F, ⊥, T } be any truth assignment. For
any formula A, v ∗ (A) ∈ {F, ⊥, T } and LF ∪ ¬LF = LF ∪ ¬LF = F ∪ ¬F ∪ T =
T, L ⊥ ∪¬L ⊥= F ∪ ¬F = F ∪ T = T, LT ∪ ¬LT = T ∪ ¬T = F ∪ T = T.
We verify whether |=M (LA ∪ ¬LA) by simple evaluation. Let v : V AR −→
{F, ⊥, T } be any truth assignment. For any formula A, v ∗ (A) ∈ {F, ⊥, T } and
LF ∪ ¬LF = LF ∪ ¬LF = F ∪ ¬F ∪ T = T, L ⊥ ∪¬L ⊥= F ∪ ¬F = F ∪ T =
T, LT ∪ ¬LT = T ∪ ¬T = F ∪ T = T.
182
Exercise 4.9
Let S be the following proof system
where the logical axioms A1, A2 and rules of inference (r1), (r2) defined for any
formulas A, B ∈ F as follows.
A1 (LA ∪ ¬LA),
A2 (A ⇒ LA),
A ;B A
(r1) , (r2) .
(A ∪ B) L(A ⇒ B)
Please, write comments how each step of the proof was obtained
2. Verify whether the system S is M-sound funder the semantics M you have
developed in Exercise 4.8. You can use shorthand notation.
3. If the system S is not sound/ sound under your semantics M then re-
define the connectives in a way that such obtained new semantics N would make
S sound/not sound
Solution
Part 1. Here is the formal proof B1 , B2 , B3 , B4 .
B1 : (La ∪ ¬La), axiom A1 for A = a,
B2 : L((La ∪ ¬La) ⇒ b), rule (r2) for B = b applied to B1 ,
B3 : (Lb ∪ ¬LAb), axiom A1 for A= b,
B4 : ((Lb ∪ ¬Lb) ∪ L((La ∪ ¬La) ⇒ b)), rule (r1) applied to B3 and B2 .
Part 2. Observe that both logical axioms of S are M tautologies. A1 is M
tautology by definition of the semantics, A2 is M tautology by direct evaluation.
Rule (r1) is sound because when A = T and B = T we get A ∪ B = T ∪ T = T .
Rule (r2) is not sound because when A = T and B = F (or B =⊥ ) we get
L(A ⇒ B) = L(T ⇒ F ) = LF = F (or L(T ⇒⊥) = L ⊥= F ).
Part 3. In order to make the rule (r2) sound while preserving the soundness
of axioms A1, A2 we have to modify only the definition of implication. Here is
the N semantics implication
⇒ F ⊥ T
F T T T
⊥ T ⊥ T
T T T T
183
Observe that it would be hard to convince anybody to use our sound proof
system S, as it would be hard to convince anybody to adopt our N semantics.
184
We assume that l1 ≤ l2 ≤ ... ≤ lm , i.e. Lm is totally ordered by a
certain relation ≤ with l1 , lm being smallest and greatest elements, re-
spectively. We denote l1 = F, lm = T and call them (total) False and
Truth, respectively. For example, when m = 2, L2 = {F, T }, F ≤ T .
Semantics S2 is called a classical semantics if the connectives are de-
fined as x ∪ y = max{x, y}, x ∩ y = min{x, y}, ¬T = F, ¬F = T,
and x ⇒ y = ¬x ∪ y, for any x, y ∈ L2 .
Let V AR be a set of propositional variables of L and let Sm be any m-
valued semantics for L. A truth assignment v : V AR −→ Lm is called
a Sm model for a formula A of L if and only if v ∗ (A) = T and logical
value v ∗ (A) is evaluated accordingly to the semantics Sm . We denote is
symbolically as v|=Sm A.
A ;(A⇒B)
Let S = (L, F, {A1, A2, A3}, M P B ) be a proof system with
logical axioms:
A1: (A ⇒ (B ⇒ A)),
A2: ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
A3: ((¬B ⇒ ¬A) ⇒ ((¬B ⇒ A) ⇒ B)).
(i) Prove that S is sound under S2 classical semantics.
(ii) Define your own S2 semantics under which S is not sound.
(iii) Define your own S4 semantics under which S is sound and other S4
semantics under which S is not sound.
(iv) Define your own Sn semantics such that S is sound for all for 2 ≤
n ≤ m.
(v) Show, by construction a formal proof, that `S (A ⇒ A).
185
186
Chapter 5
The Hilbert proof systems are systems based on a language with implication
and contain a Modus Ponens rule as a rule of inference. They are usually called
Hilbert style formalizations. We will call them here Hilbert style proof systems,
or Hilbert systems, for short.
Modus Ponens is probably the oldest of all known rules of inference as it was
already known to the Stoics (3rd century B.C.). It is also considered as the
most natural to our intuitive thinking and the proof systems containing it as
the inference rule play a special role in logic. The Hilbert proof systems put
major emphasis on logical axioms, keeping the rules of inference to minimum,
often in propositional case, admitting only Modus Ponens, as the sole inference
rule.
There are many proof systems that describe classical propositional logic, i.e.
that are complete proof systems with the respect to the classical semantics.
We present here, after Elliott Mendelson’s book Introduction to Mathematical
Logic (1987), a Hilbert proof system for the classical propositional logic and
discuss two ways of proving the Completeness Theorem for it.
Any proof of the Completeness Theorem consists always of two parts. First we
have show that all formulas that have a proof are tautologies. This implication
is also called a Soundness Theorem, or soundness part of the Completeness
Theorem. The second implication says: if a formula is a tautology then it has a
proof. This alone is sometimes called a Completeness Theorem (on assumption
that the system is sound). Traditionally it is called a completeness part of the
187
Completeness Theorem.
The proof of the soundness part is standard. We concentrate here on the com-
pleteness part of the Completeness Theorem and present two proofs of it.
The first proof is based on the one presented in the Mendelson’s book Introduc-
tion to Mathematical Logic (1987). It is is a straightforward constrictive proof
that shows how one can use the assumption that a formula A is a tautology
in order to construct its formal proof. It is hence called a proof - construction
method. It is a beautiful proof
The second proof is non-constrictive. Its strength and importance lies in a fact
that the methods it uses can be applied to the proof of completeness for classical
predicate logic. We will discuss and apply them in Chapter 9.
It proves the completeness part of the Completeness Theorem by proving the
converse implication to it. It shows how one can deduce that a formula A is
not a tautology from the fact that it does not have a proof. It is hence called a
counter-model construction proof.
Both proofs of the Completeness Theorem relay on the Deduction Theorem and
so it is the first theorem we are going to prove.
We consider first a very simple Hilbert proof system based on a language with
implication as the only connective, with two logical axioms (axiom schemas)
which characterize the implication, and with Modus Ponens as a sole rule of
inference. We call it a Hilbert system H1 and define it as follows.
where
A1 (A ⇒ (B ⇒ A)),
A2 ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
(MP) is the following rule of inference, called Modus Ponens
A ; (A ⇒ B)
(M P ) ,
B
188
The formal proof of (A ⇒ A) in H1 is a sequence
B1 , B2 , B3 , B4 , B5 (5.2)
as defined below.
B1 = ((A ⇒ ((A ⇒ A) ⇒ A)) ⇒ ((A ⇒ (A ⇒ A)) ⇒ (A ⇒ A))),
axiom A2 for A = A, B = (A ⇒ A), and C = A
B2 = (A ⇒ ((A ⇒ A) ⇒ A)),
axiom A1 for A = A, B = (A ⇒ A)
B3 = ((A ⇒ (A ⇒ A)) ⇒ (A ⇒ A))),
MP application to B1 and B2
B4 = (A ⇒ (A ⇒ A)),
axiom A1 for A = A, B = A
B5 = (A ⇒ A)
MP application to B3 and B4
Fact 5.1
For any A ∈ F,
`H1 (A ⇒ A)
and the sequence 5.2 constitutes its formal proof.
It is easy to see that the above proof wasn’t constructed automatically. The
main step in its construction was the choice of a proper form (substitution) of
logical axioms to start with, and to continue the proof with. This choice is far
from obvious for un-experienced prover and impossible for a machine, as the
number of possible substitutions is infinite.
Observe that the systems S1 − S4 from the previous Chapter 4 had inference
rules such that it was possible to ”reverse” their use; to use them in the reverse
manner in order to search for proofs, and we were able to do so in a blind,
fully automatic way. We were able to conduct an argument of the type: if
this formula has a proof the only way to construct it is from such and such
formulas by the means of one of the inference rules, and that formula can be
found automatically. We called proof systems with such property syntactically
decidable and defined them formally as follows.
189
called syntactically semi- decidable. If additionally there is an effective
method of deciding that if a proof of E in S not found, it does not exist, the
system S is called syntactically decidable. Otherwise S is syntactically
undecidable.
We will argue now, that one can’t apply the above argument to the proof search
in Hilbert proof systems as they which contain Modus Ponens as an inference
rule.
A general procedure for searching for proofs in a proof system S can be stated
is as follows. Given an expression B of the system S. If it has a proof, it
must be conclusion of the inference rule. Let’s say it is a rule r. We find its
premisses, with B being the conclusion, i.e. we evaluate r−1 (B). If all premisses
are axioms, the proof is found. Otherwise we repeat the procedure for any non-
axiom premiss.
Search for proof in Hilbert Systems must involve the Modus Ponens. The rule
says: given two formulas A and (A ⇒ B) we can conclude a formula B. Assume
now that we have a formula B and want to find its proof. If it is an axiom, we
have the proof: the formula itself. If it is not an axiom, it had to be obtained
by the application of the Modus Ponens rule, to certain two formulas A and
(A ⇒ B). But there is infinitely many of formulas A and (A ⇒ B). I.e. for any
B, the inverse image of B under the rule M P , M P −1 (B) is countably infinit
Obviously, we have the following.
Fact 5.2
Any Hilbert proof system is not syntactically decidable, in particular, the system
H1 is not syntactically decidable.
190
Fact 5.3
The proof system H1 is sound, but not complete under the classical semantics.
E1 , E2 , . . . En
E1 ∈ LA ∪ Γ, En = E
to denote that the expression E has a proof (is provable) from Γ in S and we
write Γ ` E, when the system S is fixed.
When the set of hypothesis Γ is a finite set and Γ = {B1 , B2 , ..., Bn }, then we
write
B1 , B2 , ..., Bn `S E
instead of {B1 , B2 , ..., Bn } `S E. The case when Γ is an empty set i.e. when
Γ = ∅ is a special one. By the definition of a proof of E from Γ, ∅ `S E means
that in the proof of E only logical axioms LA of S were used. We hence write
as we did before
`S E
191
Definition 5.3 (Consequence in S)
Given a proof system S = (L, F, LA, R) and a set Γ ⊆ F. Any formula A ∈ F
provable from Γ, i.e. such that
Γ `S A
The following are simple, but very important properties of the notion of conse-
quence.
Proof
The properties follow directly from the definition 5.2 and their proofs are left
to the reader as an exercise.
The monotonicity property represents the fact that if a formula A is provable
from a set Γ of premisses (hypotheses), then if we add still more premisses, A
is still provable. It hence is often stated as follows,
Exercise 5.1
Construct a proof in H1 of a formula (A ⇒ C) from the set of hypotheses
Γ = {(A ⇒ B), (B ⇒ C)}. I.e. show that
192
Solution
The required formal proof is a sequence
B1 , B2 , .....B7 (5.4)
such that
B1 = (B ⇒ C),
hypothesis
B2 = (A ⇒ B),
hypothesis
B3 = ((B ⇒ C) ⇒ (A ⇒ (B ⇒ C))),
axiom A1 for A = (B ⇒ C), B = A
B4 = (A ⇒ (B ⇒ C))
B1 , B3 and MP
B5 = ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
axiom A2
B6 = ((A ⇒ B) ⇒ (A ⇒ C)),
B5 and B4 and MP
B7 = (A ⇒ C).
B2 and B6 and MP
Exercise 5.2
Show, by constructing a formal proof that A `H1 (A ⇒ A).
Solution
The required formal proof is a sequence
B1 , B2 , B3 (5.5)
such that
B1 = A,
hypothesis
B2 = (A ⇒ (A ⇒ A)),
axiom A1 for B = A,
B3 = (A ⇒ A)
B1 , B2 and MP.
We can further simplify the task of constructing formal proofs in H1 by the use
of the following Deduction Theorem.
193
In mathematical arguments, one often assumes a statement A on the assumption
(hypothesis) of some other statement B and then concludes that we have proved
the implication ”if A, then B”. This reasoning is justified by the following
theorem, called a Deduction Theorem. It was first formulated and proved for a
certain Hilbert proof system S for the classical propositional logic by Herbrand
in 1930 in a form stated below.
if A `S B, then `S (A ⇒ B).
We are going to prove now that for our system H1 is strong enough to prove
the Herbrand Deduction Theorem for it. In fact we formulate and prove a more
general version of the Theorem 5.2.
To formulate it we introduce the following notation. We write Γ, A `S B for Γ ∪
{A}`S B, and in general we write Γ, A1 , A2 , ..., An `S B for Γ∪{A1 , A2 , ..., An }`S B.
We are now going to prove the following.
In particular,
A `H1 B if and only if `H1 (A ⇒ B).
Proof
We use we use the symbol ` instead of `H1 . for simplicity.
Part 1
We first prove the ”if” part:
If Γ, A ` B then Γ ` (A ⇒ B).
B1 , B2 , ..., Bn (5.6)
194
Hence, in particular case, when i = n, we will obtain that also
Γ ` (A ⇒ B).
Case 2. B1 = A.
When B1 = A, then to prove Γ ` (A ⇒ B) means to prove Γ ` (A ⇒ A). This
holds by the monotonicity of the consequence in H1 (Fact 5.4), and the fact
that we have proved (Fact 5.1) that `(A ⇒ A). The above cases conclude the
proof of the Base case i = 1.
Inductive step
Assume that Γ `(A ⇒ Bk ) for all k < i, we will show that using this fact we
can conclude that also Γ `(A ⇒ Bi ).
Consider a formula Bi in the sequence 5.6. By the definition, Bi ∈ {A1, A2} ∪
Γ ∪ {A} or Bi follows by MP from certain Bj , Bm such that j < m < i. We
have to consider again two cases.
Case 1. Bi ∈ {A1, A2} ∪ Γ ∪ {A}.
The proof of (A ⇒ Bi ) from Γ in this case is obtained from the proof of the
Base Step for i = 1 by replacement B1 by Bi and will be omitted here as a
straightforward repetition.
Case 2. Bi is a conclusion of MP.
If Bi is a conclusion of MP, then we must have two formulas Bj , Bm in the
sequence 5.6 such that j < i, m < i, j 6= m and
Bj ; Bm
(M P ) .
Bi
By the inductive assumption, the formulas Bj , Bm are such that
195
Γ ` (A ⇒ Bj ) (5.7)
and
Γ ` (A ⇒ Bm ). (5.8)
Moreover, by the definition of the Modus Ponens rule, the formula Bm has to
have a form (Bj ⇒ Bi ), i.e. Bm = (Bj ⇒ Bi ), and the the inductive assumption
(5.8) can be re-written as follows.
is a substitution of the axiom schema A2 and hence has a proof in our system.
By the monotonicity of the consequence (5.3), it also has a proof from the set
Γ, i.e.
Γ ` ((A ⇒ (Bj ⇒ Bi )) ⇒ ((A ⇒ Bj ) ⇒ (A ⇒ Bi ))). (5.10)
Applying the rule MP to formulas (5.10) and (5.9,) i.e. performing the following
Applying again the rule MP to formulas 5.7 and 5.11, i.e. performing the
following
(A ⇒ Bj ) ; ((A ⇒ Bj ) ⇒ (A ⇒ Bi ))
(M P )
(A ⇒ Bi )
we get that
Γ `(A ⇒ Bi )
what ends the proof of the Inductive Step. By the mathematical induction
principle, we hence have proved that Γ `(A ⇒ Bj ) for all i such that 1 ≤ i ≤ n.
In particular it is true for i = n, what means for Bn = B and we have proved
that
Γ ` (A ⇒ B).
196
This ends the proof of the Part 1.
Part 2
The proof of the inverse implication
if Γ ` (A ⇒ B) then Γ, A `B
The proof of the following useful lemma provides a good example of multiple
applications of the Deduction Theorem 5.3.
Lemma 5.1
For any A, B, C ∈ F,
Proof of (a).
Deduction theorem says:
(A ⇒ B), (B ⇒ C) `H1 (A ⇒ C) if and only if (A ⇒ B), (B ⇒ C), A `H1 C.
We construct a formal proof
B1 , B2 , B3 , B4 , B5
B1 = (A ⇒ B),
hypothesis
B2 = (B ⇒ C),
hypothesis
B3 = A,
hypothesis
B4 = B,
B1 , B3 and MP
197
B5 = C.
B2 , B4 and MP
Proof of (b).
By Deduction Theorem,
B1 , B2 , B3 , B4 , B5 , B6 , B7
B1 = (A ⇒ (B ⇒ C)),
hypothesis
B2 = B,
hypothesis
B3 = ((B ⇒ (A ⇒ B)),
A1 for A = B, B = A
B4 = (A ⇒ B),
B2 , B3 and MP
B5 = ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
axiomA2
B6 = ((A ⇒ B) ⇒ (A ⇒ C)),
B1 , B5 and MP
B7 = (A ⇒ C).
Hilbert System H2
The proof system H1 is sound and strong enough to admit the Deduction Theo-
rem, but is not completeas proved in Fact 5.3. We define now a proof system H2
that is complete with respect to classical semantics. The proof of Completeness
Theorem for H2 is to be presented in the next section.
198
H2 is defined as follows.
A ; (A ⇒ B)
H2 = ( L{¬, ⇒} , F, A1, A2, A3, M P ), (5.12)
B
where for any formulas A, B, C ∈ F of L{¬, ⇒} we define
A1 (A ⇒ (B ⇒ A)),
A2 ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
A3 ((¬B ⇒ ¬A) ⇒ ((¬B ⇒ A) ⇒ B))),
Observation 5.1 Here are some simple, straightforward facts about the proof
system H2 .
1. The language of H2 is obtained from the language of H1 by adding the
connective ¬ to it.
2. H2 is obtained from H1 by adding axiom to it the axiom A3 that characterizes
negation.
3. The use of axioms A1, A2 in the proof of Deduction Theorem 5.3 for H1 is
independent of the negation connective ¬ added to the language of H1 .
4. The proof of Deduction Theorem 5.3 for the system H1 can be repeated as it
is for the system H2 .
In particular,
A `H2 B if and only if `H2 (A ⇒ B).
Observe that for the same reason the Lemma 5.1 holds also for H2 . It is a very
i useful lemma for creating proofs in H2 so we re-state it for it here.
Lemma 5.2
For any A, B, C ∈ F,
199
We know that the axioms A1, A2 are tautologies and the Modus Ponens rule is
sound. We get by simple verification that |= A3, hence the proof system H2 is
sound, and the following holds.
The soundness theorem proves that the system ”produces” only tautologies. We
show, in the next chapter, that our proof system H2 ”produces” not only tau-
tologies, but that all tautologies are provable in it. This is called a completeness
theorem for classical logic.
We present here some examples of formal proofs in H2 . There are two reasons for
presenting them. First reason is that all formulas we prove here to be provable
play a crucial role in the proof of Completeness Theorem for H2 , or are needed
to find formal proofs of those needed. The second reason is that they provide a
”training” ground for a reader to learn how to develop formal proofs. For this
second reason we write some proofs in a full detail and we leave some others for
the reader to complete in a way explained in the following example.
We write, were needed ` instead of `H2 .
Example 5.1
We prove that
`H2 (¬¬B ⇒ B) (5.13)
by constructing its formal proof B1 , . . . , B5 , B6 as follows.
200
B1 = ((¬B ⇒ ¬¬B) ⇒ ((¬B ⇒ ¬B) ⇒ B)),
B2 = ((¬B ⇒ ¬B) ⇒ ((¬B ⇒ ¬¬B) ⇒ B)),
B3 = ¬B ⇒ ¬B),
B4 = ((¬B ⇒ ¬¬B) ⇒ B),
B5 = ¬¬B ⇒ (¬B ⇒ ¬¬B)),
B6 = (¬¬B ⇒ B).
Exercise 5.3
Complete the proof B1 , . . . , B5 , B6 of (8.3) by providing comments how each
step of the proof was obtained.
Solution
The proof of (8.3) with comments complementing it is as follows.
B1 = ((¬B ⇒ ¬¬B) ⇒ ((¬B ⇒ ¬B) ⇒ B)),
axiom A3 for A = ¬B, B = B
B3 = (¬B ⇒ ¬B),
Fact 5.1 for A = ¬B
B6 = (¬¬B ⇒ B)
B4 , B5 and Lemma 5.2 a for A = ¬¬B, B = (¬B ⇒ ¬¬B), C = B.
Lemma 5.2 application is:
(¬¬B ⇒ (¬B ⇒ ¬¬B)), ((¬B ⇒ ¬¬B) ⇒ B) ` (¬¬B ⇒ B)
Remark 5.1
Observe that in In step B2 , B3 , B5 , B6 of the proof B1 , . . . , B5 , B6 we call
previously proved results and use their results as a part of our proof. We can
201
insert previously constructed formal proofs of the results we call upon into our
formal proof.
We repeat our procedure by replacing the step B2 by its formal proof as defined
in the proof of the Lemma 5.1 b, and continue the process for all other steps
which involved application of Lemma 5.2 until we get a full formal proof from
the axioms of H2 only.
Usually we don’t need to do it, but it is important to remember that it always
can be done, if we wished to take time and space to do so.
202
Example 5.2
We prove that
`H2 (B ⇒ ¬¬B) (5.14)
by constructing its formal proof B1 , . . . , B5 as follows.
Exercise 5.4
Complete the proof B1 , . . . , B5 of (8.17) by providing comments how each
step of the proof was obtained.
Solution
The proof of (8.17) with comments complementing it is as follows.
B4 = (B ⇒ (¬¬¬B ⇒ B)),
axiom A1 for A = B, B = ¬¬¬B
B5 = (B ⇒ ¬¬B),
B3 , B4 and Lemma 5.2a for A = B, B = (¬¬¬B ⇒ B), C = ¬¬B, i.e.
Example 5.3
We prove that
`H2 (¬A ⇒ (A ⇒ B)) (5.15)
by constructing its formal proof B1 , . . . , B12 as follows.
203
B1 = ¬A,
B2 = A,
B3 = (A ⇒ (¬B ⇒ A)),
B4 = (¬A ⇒ (¬B ⇒ ¬A)),
B5 = (¬B ⇒ A),
B6 = (¬B ⇒ ¬A),
B7 = ((¬B ⇒ ¬A) ⇒ ((¬B ⇒ A) ⇒ B)),
B8 = ((¬B ⇒ A) ⇒ B),
B9 = B,
B10 = ¬A, A ` B,
B11 = ¬A ` (A ⇒ B),
B12 = (¬A ⇒ (A ⇒ B)).
Example 5.4
We prove that
`H2 ((¬B ⇒ ¬A) ⇒ (A ⇒ B)) (5.16)
by constructing its formal proof B1 , . . . , B7 as follows. Here are consecutive
steps
B1 = (¬B ⇒ ¬A),
B2 = ((¬B ⇒ ¬A) ⇒ ((¬B ⇒ A) ⇒ B)),
B3 = (A ⇒ (¬B ⇒ A)),
B4 = ((¬B ⇒ A) ⇒ B),
B5 = (A ⇒ B),
B6 = (¬B ⇒ ¬A) ` (A ⇒ B),
B7 = ((¬B ⇒ ¬A) ⇒ (A ⇒ B)).
Example 5.5
We prove that
`H2 ((A ⇒ B) ⇒ (¬B ⇒ ¬A)) (5.17)
by constructing its formal proof B1 , . . . , B9 as follows. Here are consecutive
steps
204
B1 = (A ⇒ B),
B2 = (¬¬A ⇒ A),
B3 = (¬¬A ⇒ B),
B4 = (B ⇒ ¬¬B),
B5 = (¬¬A ⇒ ¬¬B),
B7 = (¬B ⇒ ¬A),
B8 = (A ⇒ B) ` (¬B ⇒ ¬A),
Exercise 5.5
Complete the proof B1 , . . . , B9 of (5.17) by providing comments how each
step of the proof was obtained.
Solution
The proof of (5.17) with comments complementing it is as follows.
B1 = (A ⇒ B),
hypothesis
B2 = (¬¬A ⇒ A),
Example 5.10 for B = A
B3 = (¬¬A ⇒ B),
Lemma 5.2 a for A = ¬¬A, B = A, C = B
B4 = (B ⇒ ¬¬B),
Example 5.11
B5 = (¬¬A ⇒ ¬¬B),
Lemma 5.2 a for A = ¬¬A, B = B, C = ¬¬B
B7 = (¬B ⇒ ¬A),
B5 , B6 and MP
B8 = (A ⇒ B) ` (¬B ⇒ ¬A),
B1 − B7
205
B9 = ((A ⇒ B) ⇒ (¬B ⇒ ¬A)).
Deduction Theorem 5.31
Example 5.6
We prove that
`H2 ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)) (5.18)
by constructing its formal proof B1 , . . . , B12 as follows. Here are consecutive
steps.
B1 = (A ⇒ B),
B2 = (¬A ⇒ B),
B3 = ((A ⇒ B) ⇒ (¬B ⇒ ¬A)),
B4 = (¬B ⇒ ¬A),
B5 = ((¬A ⇒ B) ⇒ (¬B ⇒ ¬¬A)),
B6 = (¬B ⇒ ¬¬A),,
B7 = ((¬B ⇒ ¬¬A) ⇒ ((¬B ⇒ ¬A) ⇒ B))),
B8 = ((¬B ⇒ ¬A) ⇒ B),
B9 = B,
B10 = (A ⇒ B), (¬A ⇒ B) ` B,
B11 = (A ⇒ B) ` ((¬A ⇒ B) ⇒ B),
B12 = ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)).
Exercise 5.6
Complete the proof B1 , . . . , B12 of (5.18) by providing comments how each
step of the proof was obtained.
Solution
The proof of (5.18) with comments complementing it is as follows.
B1 = (A ⇒ B),
hypothesis
B2 = (¬A ⇒ B),
hypothesis
B3 = ((A ⇒ B) ⇒ (¬B ⇒ ¬A)),
Example 5.5
206
B4 = (¬B ⇒ ¬A),
B1 , B3 and MP
B5 = ((¬A ⇒ B) ⇒ (¬B ⇒ ¬¬A))
Example 5.5 for A = ¬A, B = B
B6 = (¬B ⇒ ¬¬A),
B2 , B5 and MP
B7 = ((¬B ⇒ ¬¬A) ⇒ ((¬B ⇒ ¬A) ⇒ B))),
axiom A3 for B = B, A = ¬A
B8 = ((¬B ⇒ ¬A) ⇒ B),
B6 , B7 and MP
B9 = B,
B4 , B8 and MP
B10 = (A ⇒ B), (¬A ⇒ B)`H2 B,
B1 − B9
B11 = (A ⇒ B) ` ((¬A ⇒ B) ⇒ B),
Deduction Theorem 5.31
B12 = ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)). Deduction Theorem 5.31
Example 5.7
We prove that
`H2 ((¬A ⇒ A) ⇒ A) (5.19)
by constructing its formal proof B1 , B2 , B3 as follows. Here are consecutive
steps.
Exercise 5.7
Complete the proof B1 , B2 , B3 of (5.19) by providing comments how each
step of the proof was obtained.
Solution
The proof of (5.19) with comments complementing it is as follows.
207
B2 = (¬A ⇒ ¬A),
Lemma 5.2 for A = ¬A
B3 = ((¬A ⇒ A) ⇒ A)).
B1 , B2 and MP
The above Examples 5.10 - 5.7 and the Fact 5.1 provide a proof of the following
lemma.
Lemma 5.3
For any formulas A, B, CinF of the system H2 the following holds.
1. `H2 (A ⇒ A);
2. `H2 (¬¬B ⇒ B);
3. `H2 (B ⇒ ¬¬B);
4. `H2 (¬A ⇒ (A ⇒ B));
5. `H2 ((¬B ⇒ ¬A) ⇒ (A ⇒ B));
6. `H2 ((A ⇒ B) ⇒ (¬B ⇒ ¬A));
7. `H2 (A ⇒ (¬B ⇒ (¬(A ⇒ B)));
8. `H2 ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B));
9. `H2 ((¬A ⇒ A) ⇒ A.
The set of provable formulas from the above Lemma 5.3 includes a set of provable
formulas needed, with H2 axioms to execute two proofs of the Completeness
Theorem 5.6 for H2 . These two proofs represent two very different methods of
proving Completeness Theorem.
Proof One, the first proof of the Completeness Theorem 5.6 presented here is
very elegant and simple, but is applicable only to the classical propositional
208
logic. Methods it uses are specific to a propositional language L{¬, ⇒} and
the proof system H2 . Nevertheless, it can be adopted and extended to other
classical propositional languages L{¬, ∪, ⇒} , L{¬, ∩, ∪,⇒} , L{¬, ∩, ∪,⇒,⇔} , and
proof systems based on them. We do so by adding appropriate new logical
axioms to the logical axioms of H2 (section 5.2.1). Such obtained proof systems
are called extentions of the system H2 . It means that one can think about the
system H2 , i.e. an axiomatization given by set {A1, A2, A3} of logical axioms
of H2 , and its language L{¬, ⇒} as in a sense, a ”minimal one” for classical
propositional logic and its languages that contain implication.
Proof One, i.e. the methods of carrying it, can’t be extended to the classical
predicate logic, not to mention variety of non-classical logics. Hence we present,
in the next section5.3 another, more general proof, called Proof Two, that can.
We have already proved the Soundness Theorem 5.5 for H2 , so in order to prove
the Completeness Theorem 5.6 we need to prove only the completeness part of
the completeness theorem, i.e. the following implication.
For any formula A of H2 ,
if |= A, then `S A. (5.20)
In order to prove (5.20), i.e. to prove that any tautology has a formal proof in
H2 , we need first to present one definition and prove one lemma stated below.
We write ` A instead of `H2 A, as the system H2 is fixed.
Definition 5.4
Let A be a formula and b1 , b2 , ..., bn be all propositional variables that occur in A.
Let v be variable assignment v : V AR −→ {T, F }. We define, for A, b1 , b2 , ..., bn
and v a corresponding formulas A0 , B1 , B2 , ..., Bn as follows:
if v ∗ (A) = T
A
A0 =
¬A if v ∗ (A) = F
bi if v(bi ) = T
Bi =
¬bi if v(bi ) = F
for i = 1, 2, ..., n.
Example 5.8
Let A be a formula
(a ⇒ ¬b) (5.21)
and let v be such that
v(a) = T, v(b) = F. (5.22)
209
In this case b1 = a, b2 = b, and v ∗ (A) = v ∗ (a ⇒ ¬b) = v(a) ⇒ ¬v(b)=
T ⇒ ¬F = T. The corresponding A0 , B1 , B2 are: A0 = A (as v ∗ (A) = T ),
B1 = a (as v(a) = T ), B2 = ¬b (as v(b) = F ).
Exercise 5.8
Let A be a formula ((¬a ⇒ ¬b) ⇒ c) and let v be such that v(a) = T, v(b) =
F, v(c) = F.
Evaluate A0 , B1 , ...Bn as defined by the definition 5.4.
Solution
In this case n = 3 and b1 = a, b2 = b, b3 = c, and v ∗ (A) = v ∗ ((¬a ⇒ ¬b) ⇒ c)
=((¬v(a) ⇒ ¬v(b)) ⇒ v(c)) = ((¬T ⇒ ¬F ) ⇒ F ) = (T ⇒ F ) = F . The
corresponding A0 , B1 , B2 , B2 are: A0 = ¬A = ¬((¬a ⇒ ¬b) ⇒ c) (as v ∗ (A) =
F ), B1 = a (as v(a) = T ), B2 = ¬b (as v(b) = F ). B3 = ¬c (as v(c) = F ).
The lemma stated below describes a method of transforming a semantic notion
of a tautology into a syntactic notion of provability. It defines, for any formula
A and a variable assignment v a corresponding deducibility relation `.
B1 , B2 , ..., Bn ` A0 . (5.23)
The Main Lemma 5.4 states: for any formula A and a variable assignment v, if
0
A , B1 , B2 , ..., Bn are corresponding formulas defined by Definition 5.4, then
B1 , B2 , ..., Bn ` A0 .
210
Case: n = 0
In the case that n = 0 A is atomic and so consists of a single propositional
variable, say a. We have to cases to consider, v ∗ (A) = T or v ∗ (A) = F .
Clearly, if v ∗ (A) = T then we A0 = A = a, B1 = a, and a ` a holds by the
Deduction Theorem and 11.15. I.e. ` (a ⇒ a) holds by ??). Applying
the the Deduction Theorem we get a ` a.
If v ∗ (A) = F then we A0 = ¬A = ¬a, B1 = ¬a, and ` (¬a ⇒ ¬a) holds
by Lemma 5.3. Applying the the Deduction Theorem we get ¬a ` ¬a. So
the lemma holds for the case n = 0.
Now assume that the lemma holds for any A with j < n logical connectives
(any A of the degree j < n). The goal is to prove that it holds for A with the
degree n.
There are several sub-cases to deal with.
Case: A is ¬A1
If A is of the form ¬A1 then A1 has less then n connectives and by the
0
inductive assumption we have the formulas A1 , B1 , B2 , ..., Bn corre-
sponding to the A1 and the propositional variables b1 , b2 , ..., bn in A1 , as
defined by the definition 5.4, such that
0
B1 , B2 , ..., Bn ` A1 . (5.24)
Observe, that the formulas A and ¬A1 have the same propositional vari-
ables, so the corresponding formulas B1 , B2 , ..., Bn are the same for
both of them. We are going to show that the inductive assumption (5.24)
allows us to prove that the lemma holds for A, ie. that
0
B1 , B2 , ..., Bn ` A .
211
Case: A is (A1 ⇒ A2 )
If A is of the form (A1 ⇒ A2 ) then A1 and A2 have less than n connectives.
A = A(b1 , ... bn ) so there are some subsequences c1 , ..., ck and d1 , ...dm ,
for k, m ≤ n, of the sequence b1 , ..., bn such that A1 = A1 (c1 , ..., ck ) and
A2 = A(d1 , ...dm ). A1 and A2 have less than n connectives and so by
the inductive assumption we have appropriate formulas C1 , ..., Ck and
0 0
D1 , ...Dm such that C1 , C2 , . . . , Ck ` A1 and D1 , D2 , . . . , Dm ` A2 .
The formulas C1 , C2 , ..., Ck and D1 , D2 , ..., Dm are subsequences of for-
mulas B1 , B2 , ..., Bn corresponding to the propositional variables in A.
0
Hence by monotonicity we have also that have B1 , B2 , ..., Bn ` A1 and
0
B1 , B2 , ..., Bn ` A2 , where B1 , B2 , ..., Bn are formulas corresponding to
the propositional variables in A.
Now we have the following sub-cases to consider.
With that we have covered all cases and, by induction on n, the proof of the
lemma is complete.
212
Proof Assume that |= A. Let b1 , b2 , ..., bn be all propositional variables that
occur in A, i.e. A = A(b1 , b2 , ..., bn ).
Let v : V AR → {T, F } be any variable assignment, and
By the Main Lemma 5.4 and the assumption that |= A any v ∈ VA defines
formulas B1 , B2 , ..., Bn such that
B1 , B2 , ..., Bn ` A. (5.28)
Step 1: elimination of Bn .
Observe that by definition 5.4, each Bi is bi or ¬bi depending on the
choice of v ∈ VA . In particular Bn = bn or Bn = ¬bn . We choose two
truth assignments v1 6= v2 ∈ VA such that
B1 , B2 , ..., Bn−1 , bn ` A.
B1 , B2 , ...Bn−1 , ¬bn ` A.
213
By Lemma 5.3 of the formula ` ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)). Hence for for
A = bn , B = A we have that
Applying Modus Ponens twice to the above property (5.32) and properties
(5.30), (5.31) we get that
We have eliminated Bn .
w1 |{b1 , ..., bn−2 } = w2 |{b1 , ..., bn−2 } = v1 |{b1 , ..., bn−2 } = v2 |{b1 , ..., bn−2 }
(5.34)
and w1 (bn−1 ) = T and w2 (bn−1 ) = F .
As before we apply Main Lemma, Deduction Theorem, monotonicity,
proper substitutions of the formula ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)),
and Modus Ponens twice and eliminate Bn−1 just as we eliminated Bn .
` A.
Observe that our proof of the fact that ` A is a constructive one. Moreover, we
have used in it only Main Lemma 5.4 and Deduction Theorem 5.3, and both of
them have fully constructive proofs. So we can always reconstruct all steps in
proofs which use the Main Lemma 5.4and Deduction Theorem 5.3 back to the
original axioms of H2 . The same applies to the proofs that use the formulas
proved in H2 that are stated in Lemma 5.3.
It means that for any A ∈ F, such that |= A, the set VA of all v restricted to
A provides us a method of a construction of the formal proof of A in H2 from
its axioms A1, A2, A3 only. .
214
5.2.1 Examples
Example 5.10
As an example of how the Proof One of the Completeness Theorem works, we
consider a following tautology
|= (a ⇒ (¬a ⇒ b))
` (a ⇒ (¬a ⇒ b)).
a, ¬b ` (a ⇒ (¬a ⇒ b)).
Applying the Deduction Theorem 5.3 to the cases above we have that
D1 (Cases 1 and 2)
a ` (b ⇒ (a ⇒ (¬a ⇒ b))),
a ` (¬b ⇒ (a ⇒ (¬a ⇒ b))),
D2 (Cases 2 and 3)
¬a ` (b ⇒ (a ⇒ (¬a ⇒ b))),
¬a ` (¬b ⇒ (a ⇒ (¬a ⇒ b))).
215
By the monotonicity and the proper substitution of formula
((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B))
the provable by Lemma 5.3, we have that
a ` ((b ⇒ (a ⇒ (¬a ⇒ b))) ⇒ ((¬b ⇒ (a ⇒ (¬a ⇒ b))) ⇒ (a ⇒ (¬a ⇒ b))),
¬a ` ((b ⇒ (a ⇒ (¬a ⇒ b))) ⇒ ((¬b ⇒ (a ⇒ (¬a ⇒ b))) ⇒ (a ⇒ (¬a ⇒ b))).
Applying Modus Ponens twice to D1, D2 and these above, respectively, gives
us
a ` (a ⇒ (¬a ⇒ b)) and
¬a ` (a ⇒ (¬a ⇒ b)).
Applying the Deduction Theorem 5.3 to the above we obtain
D3 ` (a ⇒ (a ⇒ (¬a ⇒ b))),
D4 ` (¬a ⇒ (a ⇒ (¬a ⇒ b))).
Example 5.11
The Proof One of Completeness Theorem defines a method of efficiently com-
bining v ∈ VA as defined in (5.27), while constructing the proof of A. Let’s
consider the following tautology A = A(a, b, c)
((¬a ⇒ b) ⇒ (¬(¬a ⇒ b) ⇒ c).
We present bellow all steps of Proof One as applied to A.
By the Main Lemma 5.4 and the assumption that |= A(a, b, c) any v ∈ VA
defines formulas Ba , Bb , Bc such that
Ba , Bb , Bc ` A. (5.37)
216
Step 1: elimination of Bc .
Observe that by definition 5.4, Bc is c or ¬c depending on the choice of
v ∈ VA . We choose two truth assignments v1 =6 v2 ∈ VA such that
Ba , Bb , c ` A.
Ba , Bb ` (c ⇒ A). (5.39)
Ba , Bb , ¬c ` A.
Applying Modus Ponens twice to the above property (5.41) and properties
(5.39), (5.40) we get that
Ba , Bb ` A. (5.42)
and hence we have eliminated Bc .
217
Case 1: w1 (b) = T , by definition 5.4 Bb = b. By the property (5.43), assump-
tion that |= A, and the Main Lemma 5.4 applied to w1
Ba , b ` A.
Ba ` (b ⇒ A). (5.44)
Ba , ¬b ` A.
Applying Modus Ponens twice to the above property (5.46) and properties
(5.44), (5.45) we get that
Ba ` A. (5.47)
and hence we have eliminated Bb .
a ` A.
` (a ⇒ A). (5.49)
218
Case 2: g2 (a) = F hence by definition 5.4 Ba = ¬a. By the property (5.48),
assumption that |= A, and the Main Lemma 5.4 applied to g2
¬a ` A.
Applying Modus Ponens twice to the above property (5.51) and properties
(5.49), (5.50) we get that
` A. (5.52)
and hence we have eliminated Ba , Bb and Bc and constructed the proof of A.
We prove now the completeness part of the Completeness Theorem 5.6 for H2
by proving that the opposite implication:
if |= A, then ` A
We will show now how one can define of a counter-model for A from the fact
that A is not provable. This means that we deduce that a formula A is not
219
a tautology from the fact that it does not have a proof. We hence call it a a
counter-model existence method.
The definition of the counter-model for any non-provable A is much more general
(and less constructive) then in the case of the Proof One in section 5.2. It can
be generalized to the case of predicate logic, and many of non-classical logics;
propositional and predicate. It is hence a much more general method then the
first one and this is the reason we present it here.
We remind that 6|= A means that there is a truth assignment v : V AR −→
{T, F }, such that v ∗ (A) 6= T , i.e. in classical semantics, such that that v ∗ (A) =
F . Such v is called a counter-model for A, hence the proof provides a counter-
model construction method.
Since we assume in (8.16) that A does not have a proof in H2 (6` A) the method
uses this information in order to show that A is not a tautology, i.e. to define v
such that v ∗ (A) = F . We also have to prove that all steps in that method are
correct. This is done in the following steps.
Step 1: Definition of ∆∗
We use the information 6` A to define a special set ∆∗ ⊆ F, such that
¬A ∈ ∆∗ .
Step 2: Counter - model definition
We define the truth assignment v : V AR −→ {T, F } as follows:
if ∆∗ ` a
T
v(a) =
F if ∆∗ ` ¬a.
if ∆∗ ` B
T
v ∗ (B) =
F if ∆∗ ` ¬B.
The definition and the properties of the set ∆∗ , and hence the Step 1, are the
most essential for the proof. The other steps have mainly technical character.
The main notions involved in the Step 1 (definition of ∆∗ ) are: consistent set,
complete set and a consistent complete extension of a set. We are going now to
introduce them and to prove some essential facts about them.
220
There exist two definitions of consistency; semantical and syntactical. The
semantical one uses definition the notion of a model and says, in plain English:
In our Proof Two of the Completeness Theorem we use assumption that a given
formula A does not have a proof to deduce that A is not a tautology. We hence
use the following syntactical definition of consistency.
Consistent set
We say that a set ∆ ⊆ F of formulas is consistent if and only if there
is no a formula A ∈ F such that
Inconsistent set
A set ∆ ⊆ F is inconsistent if and only if there is a formula A ∈ F such
that ∆ ` A and ∆ ` ¬A.
(i) ∆ is consistent,
Proof The implications: (i) implies (ii) and vice-versa are proved by showing
the corresponding opposite implications. I.e. to establish the equivalence of (i)
and (ii), we first show that not (ii) implies not (i), and then that not (i)
implies not (ii).
Case 1
Assume that not (ii). It means that for all formulas A ∈ F we have that
∆ ` A. In particular it is true for a certain A = B and A = ¬B and
hence proves that ∆ is inconsistent, i.e. not (i) holds.
221
Case 2
Assume that not (i), i.e that ∆ is inconsistent. Then there is a formula
A such that ∆ ` A and ∆ ` ¬A. Let B be any formula. Since
(¬A ⇒ (A ⇒ B)) is provable in H2 by Lemma 5.3, hence by applying
Modus Ponens twice and by detaching from it ¬A first, and A next, we
obtain a formal proof of B from the set ∆, so that ∆ ` B for any formula
B. Thus not (ii).
(i) ∆ is inconsistent,
(ii) for all formulas A ∈ F, ∆ ` A.
Lemma 5.7
For every set ∆ of formulas and for every formula A ∈ F, ∆ ` A if and only
if there is a finite subset ∆0 ⊆ ∆ such that ∆0 ` A.
Proof
If ∆0 ` A for a certain ∆0 ⊆ ∆, then by the monotonicity of the consequence,
also ∆ ` A. Assume now that ∆ ` A and let A1 , A2 , ..., An be a formal
proof of A from ∆. Let ∆0 = {A1 , A2 , ..., An } ∩ ∆. Obviously, ∆0 is finite and
A1 , A2 , ..., An is a formal proof of A from ∆0 .
Proof
If ∆ is inconsistent, then for some formula A, ∆ ` A and ∆ ` ¬A. By above
Lemma 5.7, there are finite subsets ∆1 and ∆2 of ∆ such that ∆1 ` A and
∆2 ` ¬A. By monotonicity, the union ∆1 ∪ ∆2 is a finite subset of ∆, such
that ∆1 ∪ ∆2 ` A and ∆1 ∪ ∆2 ` ¬A. Hence ∆1 ∪ ∆2 is a finite inconsistent
subset of ∆. The second implication is the opposite to the one just proved and
222
hence also holds.
The following lemma links the notion of non-provability and consistency. It will
be used as an important step in our proof of the Completeness Theorem.
Lemma 5.8
For any formula A ∈ F, if 6` A, then the set {¬A} is consistent.
Proof
If {¬A} is inconsistent, then by the Inconsistency Condition 5.6 we have {¬A} `
A. This and the Deduction Theorem 5.3 imply ` (¬A ⇒ A). Applying the
Modus Ponens rule to ` (¬A ⇒ A) a formula ((¬A ⇒ A) ⇒ A), provable by
LemmaH2lemma, we get that ` A, contrary to the assumption of the lemma.
Complete and Incomplete Sets
Another important notion, is that of a complete set of formulas. Complete sets,
as defined here are sometimes called maximal, but we use the first name for
them. They are defined as follows.
Complete set
A set ∆ of formulas is called complete if for every formula A ∈ F,
∆ ` A or ∆ ` ¬A. (5.55)
(i) ∆ is complete,
(ii) for every formula A ∈ F, if ∆ 6` A, then the set ∆ ∪ {A} is inconsistent.
Proof
We consider two cases. We show that (i) implies (ii) and vice-versa, that (ii)
also implies (i).
Case 1
Assume that (i) and that for every formula A ∈ F, ∆ 6` A, we have to
show that in this case ∆ ∪ {A} is inconsistent. But if ∆ 6` A, then from
the definition of complete set and assumption that ∆ is complete set, we
get that ∆ ` ¬A. By the monotonicity of the consequence we have that
223
∆ ∪ {A} ` ¬A as well. Since, by formula 11.15 we have ` (A ⇒ A), by
monotonicity ∆ ` (A ⇒ A) and by Deduction Theorem ∆ ∪ {A} ` A.
This proves that ∆ ∪ {A} is inconsistent. Hence (ii) holds.
Case 2
Assume that (ii). Let A be any formula. We want to show that the
condition: ∆ ` A or ∆ ` ¬A is satisfied. If ∆ ` ¬A, then the
condition is obviously satisfied.
If, on other hand, ∆ 6` ¬A, then we are going to show now that it must
be , under the assumption of (ii), that ∆ ` A, i.e. that (i) holds.
Assume that ∆ 6` ¬A, then by (ii), the set ∆ ∪ {¬A} is inconsistent.
It means, by the Consistency Condition 5.5, that ∆ ∪ {¬A} ` A.
By the Deduction Theorem 5.3, this implies that ∆ ` (¬A ⇒ A).
Since ((¬A ⇒ A) ⇒ A) is provable in H2 (Lemma 5.3), by monotonicity
∆ ` ((¬A ⇒ A) ⇒ A). Detaching (¬A ⇒ A), we obtain that ∆ ` A,
what ends the proof that (i) holds.
Incomplete set
A set ∆ of formulas is called incomplete if it is not complete, i.e. if there
exists a formula A ∈ F such that
(i) ∆ is incomplete,
(ii) there is formula A ∈ F such that ∆ 6` A and the set ∆ ∪ {A} is consistent.
Extensions
224
A set ∆∗ of formulas is called an extension of a set ∆ of formulas if the
following condition holds.
{A ∈ F : ∆ ` A} ⊆ {A ∈ F : ∆∗ ` A}. (5.57)
Proof
Assume that the lemma does not hold, i.e. that there is a consistent set ∆,
such that all its consistent extensions are not complete. In particular, as ∆ is
an consistent extension of itself, we have that ∆ is not complete.
The proof consists of a construction of a particular set ∆∗ and proving that it
forms a complete consistent extension of ∆, contrary to the assumption that all
its consistent extensions are not complete.
Construction of ∆∗ .
As we know, the set F of all formulas is enumerable. They can hence be put in
an infinite sequence
Initial Step
In this step we define the sets ∆1 , ∆2 and the formula B1 . We prove that
∆1 and ∆2 are consistent, incomplete extensions of ∆.
We take, as the first set, the set ∆, i.e. we define
∆1 = ∆. (5.59)
225
Let
B1
be the first formula with this property in the sequence (5.58) of all formu-
las; we then define
∆2 = ∆1 ∪ {B1 }. (5.60)
Inductive Step
Suppose that we have defined a sequence
∆1 , ∆2 , ..., ∆n
of incomplete, consistent extensions of ∆, and a sequence
B1 , B2 , ...Bn−1
of formulas, for n ≥ 2.
Since ∆n is incomplete, it follows from the Incomplete Set Condition
Lemma 5.10, that there is a formula B ∈ F such that ∆n 6` B and
the set ∆n ∪ {B} is consistent.
Let
Bn
be the first formula with this property in the sequence (5.58) of all formu-
las.
We then define
By the definition, ∆ ⊆ ∆n ⊆ ∆n+1 and the set ∆n+1 is consistent. Hence ∆n+1
is an incomplete consistent extension of ∆.
By the principle of mathematical induction we have defined an infinite sequence
∆ = ∆1 ⊆ ∆2 ⊆ ..., ⊆ ∆n ⊆ ∆n+1 ⊆ .... (5.62)
such that for all n ∈ N , ∆n is consistent, and moreover, it is an incomplete
consistent extension of ∆.
Moreover, we have also defined a sequence
B1 , B2 , ..., Bn , .... (5.63)
of formulas, such that for all n ∈ N , ∆n 6` Bn , and the set ∆n ∪ {Bn } is
consistent.
Observe that Bn ∈ ∆n+1 for all n ≥ 1.
226
Definition of ∆∗
Now we are ready to define ∆∗ , i.e. we define:
[
∆∗ = ∆n . (5.64)
n∈N
To complete the proof our theorem we have now to prove that ∆∗ is a complete
consistent extension of ∆. Obviously, by the definition, ∆∗ is an extension of
∆. Now we prove (by contradiction) the following.
Fact 5.5
The set ∆∗ is consistent.
Proof
Assume that ∆∗ is inconsistent. By the Finite Inconsistency Theorem 5.7 there
is a finite subset ∆0 of ∆∗ that is inconsistent. By Definition 5.64 have that
[
∆0 = {C1 , ..., Cn } ⊆ ∆n .
n∈N
By the definition, Ci ∈ ∆ki for certain ∆ki in the sequence (5.62) and 1 ≤ i ≥ n.
Hence ∆0 ⊆ ∆m for m = max{k1 , k2 , ..kn }. But all sets of the sequence (5.62)
are consistent. This contradicts the fact that ∆m is inconsistent, as it contains
an inconsistent subset ∆0 . Hence ∆∗ must be consistent.
Fact 5.6
The set ∆∗ is complete.
Proof
Assume that ∆∗ is not complete. By the Incomplete Set Condition Lemma 5.10,
there is a formula B ∈ F such that ∆∗ 6` B and the set ∆∗ ∪ {B} is consistent.
But, by definition (5.64) of ∆∗ , the above condition means that for every n ∈ N ,
∆n 6` B holds and the set ∆n ∪ {B} is consistent.
Since the formula B is one of the formulas of the sequence (5.58) and it would
have to be one of the formulas of the sequence (5.63), i.e. B = Bj for certain
j. Since Bj ∈ ∆j+1 , it proves that B ∈ ∆∗ = n∈N . But this means that
S
∆∗ ` B, contrary to the assumption. This proves that ∆∗ is a complete
consistent extension of ∆ and ends the proof out our lemma.
Now we are ready to prove the completeness theorem for the system H2 .
227
As by assumption our system H2 is sound, we have to prove only the Complete-
ness part of the Completeness Theorem 5.6, i.e we have to show the implication
if |= A, then ` A
for any formula A. We prove it by proving the logically equivalent opposite
implication
if 6` A, then |6 = A.
Definition of v
We define the variable assignment
v : V AR −→ {T, F } (5.66)
as follows:
if ∆∗ ` a
T
v(a) =
F if ∆∗ ` ¬a.
228
Given the above property (5.67) of v (still to be proven), we prove that the v
is in fact, a counter model for any formula A, such that 6` A as follows. Let
A be such that 6` A. By ( 5.65), ¬A ∈ ∆∗ and obviously, ∆∗ ` ¬A. Hence,
by the property (5.67) of v, v ∗ (A) = F , what proves that v is a counter-model
for A and hence ends the proof of the completeness theorem. In order to really
complete the proof we still have to write a proof of the Lemma 5.12.
Assume now that ∆∗ ` ¬A. Then from the fact that ∆∗ is consistent it must
be that ∆∗ 6` A. I.e. ∆∗ 6` ¬C. If so, then ∆∗ ` C, as the set ∆∗ is
complete. Hence by the inductive assumption, that v ∗ (C) = T , and accordingly
Case A = (C ⇒ D). As in the previous case, we assume that the lemma, i.e.
the property (5.67) holds for the formulas C, D and we consider two pos-
sibilities: ∆∗ ` A and ∆∗ ` ¬A.
v ∗ (C) = v ∗ (D) = T,
229
and accordingly
∆∗ 6` (C ⇒ D).
∆∗ ` ¬C.
But this is impossible, since the formula (¬C ⇒ (C ⇒ D)) is provable provable
(Lemma 5.4) and by monotonicity
∆∗ ` (¬C ⇒ (C ⇒ D)).
Lukasiewicz (1929)
where
230
A1 ((¬A ⇒ A) ⇒ A),
A2 (A ⇒ (¬A ⇒ B)),
A3 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))))
for any A, B, C ∈ F.
2. Hilbert and Ackermann (1928)
where
A1 (¬(A ∪ A) ∪ A),
A2 (¬A ∪ (A ∪ B)),
A3 (¬(A ∪ B) ∪ (B ∪ A)),
A4 (¬(¬B ∪ C) ∪ (¬(A ∪ B) ∪ (A ∪ C))),
for any A, B, C ∈ F.
Modus Ponens rule in the language L{¬,∪} has a form
A ; (¬A ∪ B)
(M P ) .
B
In particular,
A `HA B if and only if `HA (¬A ∪ B).
2. Hilbert (1928)
where
A1 (A ⇒ A),
A2 (A ⇒ (B ⇒ A)),
A3 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A4 ((A ⇒ (A ⇒ B)) ⇒ (A ⇒ B)),
231
A5 ((A ⇒ (B ⇒ C)) ⇒ (B ⇒ (A ⇒ C))),
A6 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A7 ((A ∩ B) ⇒ A),
A8 ((A ∩ B) ⇒ B),
A9 ((A ⇒ B) ⇒ ((A ⇒ C) ⇒ (A ⇒ (B ∩ C))),
A10 (A ⇒ (A ∪ B)),
A11 (B ⇒ (A ∪ B)),
A12 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A13 ((A ⇒ B) ⇒ ((A ⇒ ¬B) ⇒ ¬A)),
A14 (¬A ⇒ (A ⇒ B)),
A15 (A ∪ ¬A),
for any A, B, C ∈ F.
Kleene (1952)
where
A1 (A ⇒ (B ⇒ A)),
A2 ((A ⇒ (B ⇒ C)) ⇒ (B ⇒ (A ⇒ C))),
A3 ((A ∩ B) ⇒ A),
A4 ((A ∩ B) ⇒ B),
A5 (A ⇒ (B ⇒ (A ∩ B))),
A6 (A ⇒ (A ∪ B)),
A7 (B ⇒ (A ∪ B)),
A8 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A9 ((A ⇒ B) ⇒ ((A ⇒ ¬B) ⇒ ¬A)),
A10 (¬¬A ⇒ A)
for any A, B, C ∈ F.
Rasiowa-Sikorski (1950)
232
A1 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A2 (A ⇒ (A ∪ B)),
A3 (B ⇒ (A ∪ B)),
A4 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A5 ((A ∩ B) ⇒ A),
A6 ((A ∩ B) ⇒ B),
A7 ((C ⇒ A) ⇒ ((C ⇒ B) ⇒ (C ⇒ (A ∩ B))),
A8 ((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C)),
A9 (((A ∩ B) ⇒ C) ⇒ (A ⇒ (B ⇒ C)),
A10 (A ∩ ¬A) ⇒ B),
A11 ((A ⇒ (A ∩ ¬A)) ⇒ ¬A),
A12 (A ∪ ¬A),
for any A, B, C ∈ F.
Here is the shortest axiomatization for the language L{¬, ⇒} . It contains just
one axiom.
Meredith (1953)
L = ( L{¬, ⇒} , F, A1 M P ), (5.73)
where
A1 ((((((A ⇒ B) ⇒ (¬C ⇒ ¬D)) ⇒ C) ⇒ E)) ⇒ ((E ⇒ A) ⇒ (D ⇒ A))).
where
A1 (((A ↑ (B ↑ C)) ↑ ((D ↑ (D ↑ D)) ↑ ((E ↑ B) ↑ ((A ↑ E) ↑ (A ↑ E)))))).
The rule of inference is (r) is expressed in the language L{↑} as
A ↑ (B ↑ C)
.
A
233
5.5 Exercises
Here are few exercises designed to help the readers with understanding the
notions of completeness, monotonicity of the consequence operation, the role of
the deduction theorem and importance of some basic tautologies.
Let S be any Hilbert proof system
A, (A ⇒ B)
S = (L{∩,∪,⇒,¬} , F, LA, (M P ) ) (5.75)
B
with its set LA of logical axioms such that S is complete under classical se-
mantics.
Let X ⊆ F be any subset of the set F of formulas of the language L{∩,∪,⇒,¬}
of S. We define, as we did in chapter 4, a set Cn(X) of all consequences of
the set X as
Cn(X) = {A ∈ F : X `S A}. (5.76)
Plainly speaking, the set Cn(X) of all consequences of the set X is the set of
all formulas that can be proved in S from the set (LA ∪ X).
Exercise 5.9
1. Prove that for any subsets X, Y of the set F of formulas the following mono-
tonicity property holds.
Solution
1. Let A ∈ F be any formula such that A ∈ Cn(X). By (5.76), we have that
X `S A. This means that A has a formal proof from the set X ∪ LA. But
X ⊆ Y , hence this proof is also a proof from Y ∪ LA, i.e . Y `S A, and hence
A ∈ Cn(Y ). This proves that Cn(X) ⊆ Cn(Y ).
Exercise 5.10
234
Prove that for any set X ⊆ F, the set T ⊆ F of all propositional classical
tautologies of the language L{∩,∪,⇒,¬} of the system S is a subset of Cn(X),
i.e. prove that
T ⊆ Cn(X). (5.78)
2. Do we need the completeness of S to prove that the property (5.78) holds
for S?
Solution
1. The proof system S is complete, so by the completeness theorem we have
that
T = {∈ F : `S A}. (5.79)
By definition (5.76) of the consequence,
{A ∈ F : `S A} = Cn(∅)
Exercise 5.11
Prove that for any formulas A, B ∈ F, and for any set X ⊆ F,
X `S (A ∩ B). (5.81)
235
Similarly, X `S (A ∩ B), by the assumption (5.81), X`S ((A ∩ B) ⇒ B) by
by (??), and so we get X `S B by MP. This proves that A ∈ Cn(X) and
B ∈ Cn(X) and ends the proof of the implication 1.
X `S A, and X `S B. (5.84)
X `S (A ⇒ (B ⇒ (A ∩ B))). (5.85)
Exercise 5.12
Let S be the proof system (5.75). Prove that the Deduction Theorem holds for
S, i.e. prove the following.
For any subset Γ of the set of formulas F of S and for any formulas A, B ∈ F,
Solution
The formulas A1 = (A ⇒ (B ⇒ A)) and A2 = ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒
B) ⇒ (A ⇒ C))) are basic propositional tautologies. By the completeness of S
we have that
Exercise 5.13
Prove that for any A, B ∈ F,
236
1. Proof of the inclusion: Cn({A, B}) ⊆ Cn({(A ∩ B)}).
Assume C ∈ Cn({A, B}), i.e. {A, B}`S C, what we usually write as A, B `S C.
Observe that by exercise 8.7 the Deduction Theorem (theorem 5.3) holds for S.
We apply Deduction Theorem to the assumption A, B `S C twice we get that
the assumption is equivalent to
`S (A ⇒ (B ⇒ C)). (5.88)
We use completeness of S, the fact that the formula (((A ⇒ (B ⇒ C)) ⇒
((A ∩ B) ⇒ C))) is a tautology, and by monotonicity and get that
`S (((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C))). (5.89)
Applying Modus Ponens to the assumption (5.88) and (8.22) we get `S ((A ∩
B) ⇒ C). This is equivalent to (A ∩ B) `S C by Deduction Theorem. We have
proved that C ∈ Cn({(A ∩ B)}).
237
2. A2 = ((a ⇒ (c ⇒ (¬b ⇒ c))) ⇒ ((¬d ⇒ (a ⇒ (¬a ⇒ b))) ⇒ (a ⇒ (¬a ⇒
b))))
v(a) = F, v(b) = F, v(c) = T, v(d) = F
For any of the formulas listed below construct their formal proofs, as described
in the Proof One of the Completeness Theorem. Follow example 5.10, or exam-
ple 5.11.
1. A1 = (¬¬b ⇒ b)
2. A2 = ((a ⇒ b) ⇒ (¬b ⇒ ¬a))
3. A3 = (¬(a ⇒ b) ⇒ ¬(¬b ⇒ ¬a))
4. A4 = (¬(¬(a ⇒ ¬b) ⇒ ¬c) ⇒ ¬(b ⇒ ¬c))
5. A5 = ((a ⇒ (b ⇒ ¬a)) ⇒ (¬(b ⇒ ¬a) ⇒ ¬a)).
238
10. We proved the Completeness Theorem for the proof system H2 based on
the language L{¬,⇒} . Extend the H2 proof system to a proof system S1
based on a language L{¬,⇒,∪} by adding new logical axioms, as we did in
a case of H1 and H2 systems. The added logical axioms must be such that
they allow to adopt the Proof One to S1 , i.e. such that it is a complete
proof system with respect to classical semantics.
11. Repeat the same for the language L{¬,⇒,∩} . Call resulting proof system
S2 .
12. Repeat the same for the language L{¬,⇒,∩,∪} , i.e. extends systems S1 or
S2 to a complete proof system S3 based on the language L{¬,⇒,∩,∪} .
13. Prove Completeness Theorem for the system S3 from the previous prob-
lem.
2. We proved the Completeness Theorem 5.6 for the proof system H2 based
on the language L{¬,⇒} . Extend the H2 proof system to a proof system
S1 based on a language L{¬,⇒,∪} by adding new logical axioms, as we
did in a case of H1 and H2 systems. The added logical axioms must be
such that they allow to adopt the Proof Two to S1 , i.e. such that it is a
complete proof system with respect to classical semantics.
4. Repeat the same for the language L{¬,⇒,∩,∪} , i.e. extends systems S1 or
S2 to a complete proof system S3 based on the language L{¬,⇒,∩,∪} .
Axiomatizations Problems
239
(iii) Prove `HA ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))), for any
A, B, C ∈ F.
(iv) Prove (A ⇒ B), (B ⇒ C) `HA (A ⇒ C), for any A, B, C ∈ F
(v) Prove Deduction Theorem 5.8.
(vi) Prove `HA A if and only if |= A, for any A ∈ F.
2. Let H be Hilbert proof system (??).
(i) Prove `HA ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))), for any
A, B, C ∈ F.
(ii) Prove Deduction Theorem for H.
(ii) Prove Completeness Theorem for H.
3. Let K be Kleene proof system (5.71).
(i) Prove `K (A ⇒ A), for any A ∈ F.
(ii) Prove the following.
For any subset Γ of the set of formulas F of K and for any formulas
A, B ∈ F, Γ, A `K B if and only if Γ `K ((A ⇒ B)).
F = ( F, ∪, ∩, ⇒, ¬ ), (5.92)
240
Prove that the relation (5.95) is an order relation in F/ ≈
(iv) Prove that the relation ≈ defined by (5.94) is a congruence in the
algebra F of formulas defined by (8.5).
2. The algebra LT = ( F/ ≈, ∪, ∩, ⇒, ¬), where the operations ∪, ∩, ⇒
and ¬ are determined by the congruence relation (5.94) i.e.
241
242
Chapter 6
Hilbert style systems are easy to define and admit different proofs of the Com-
pleteness Theorem but they are difficult to use. By humans, not mentioning
computers. Their emphasis is on logical axioms, keeping the rules of inference,
with obligatory Modus Ponens, at a minimum.
Gentzen style proof systems reverse this situation by emphasizing the impor-
tance of inference rules, reducing the role of logical axioms to an absolute mini-
mum. They may be less intuitive then the Hilbert-style systems, but they allow
us to define effective automatic procedures for proof search, what was impossi-
ble in a case of the Hilbert style systems. For this reason they are also called
automated proof systems. They serve as formal models of computing systems
that automate the reasoning process. Building computing systems means pro-
viding an algorithmic description to a formal proof system so that it can be
implemented on a computer to prove theorems in an efficient manner.
The first proof systems of this style was invented by G. Gentzen in 1934, hence
the name. His proof systems for classical and intuitionistic predicate logics intro-
duced special expressions built of formulas called sequents. Hence the Gentzen
style systems using sequents as basic expressions are often called sequent sys-
tems, or Gentzen sequent systems, or simply Gentzen formalizations.
We present here (section 6.5) two Gentzen systems GL and G for classical
243
propositional logic and prove their completeness. We also present a proposi-
tional version of Gentzen original system LK and discuss a proof of Gentzen
Hauptsatz for it. Hauptsatz is literally rendered as the main theorem and is
known as cut-elimination theorem. We prove the equivalency of the cut-free
propositional LK and the complete system G. The Gentzen original formaliza-
tion for intuitionistic propositional logic LI is discussed and presented in chapter
7. The classical and intuitionistic predicate versions are discussed in chapter 9.
Historical importance and lasting influence of Rasiowa and Sikorski work lays
in the fact that they were first to use the proof searching capacity of their proof
system to define a constructive method of proving the completeness theorem
for both propositional and predicate classical logic. We introduce and explain
in detail their method and use it prove the completeness of the RS in section
6.3. We also introduce and discuss two other RS style system RS1 and RS2
in in section 6.4. We also generalize the RS completeness proof method to the
Gentzen sequent systems and prove the completeness of GL and G systems in
section 6.5.1. The completeness proof for proof system RSQ for classical pred-
icate logic is presented in chapter 9.
Language L
Let F denote a set of formulas of L = L{¬,⇒,∪,∩} . The rules of inference of our
system RS operate on finite sequences of formulas, i.e. elements of F ∗ , unlike
on plain formulas F in Hilbert style formalizations.
Expressions E
244
We adopt as the set of expressions E of RS the set F ∗ , i.e. E = F ∗ . We will
denote the expressions of RS, i.e. the finite sequences of formulas by Γ, ∆, Σ,
with indices if necessary.
Semantic Link
The intuitive meaning of a sequence Γ ∈ F ∗ is that the truth assignment v makes
it true if and only if it makes the formula of the form of the disjunction of all
formulas of Γ true. As we know, the disjunction in classical logic is associative
and commutative, i.e., for any formulas A, B, C ∈ F, the formulas (A∪(B ∪C)),
((A ∪ B) ∪ C), (A ∪ (C ∪ B)), ((B ∪ A) ∪ C), (C ∪ (B ∪ A)), (C ∪ (A ∪ B)),
((C ∪ A) ∪ B), etc... are logically equivalent. In particular we write
δ{A,B,C} = A ∪ B ∪ C
A1 , A2 , ..., An (6.1)
Model
A sequence Γ is said to be satisfiable if there is a truth assignment v : V AR −→
{T, F } such that v ∗ (Γ) = T . Such a truth assignment is called a model for Γ.
We denote it as
v |= Γ. (6.3)
Counter- Model
A sequence Γ is said to be falsifiable if there is a truth assignment v, such that
v ∗ (Γ) = F . Such a truth assignment is ] called a counter-model for Γ. We
write it symbolically as
v 6|= Γ. (6.4)
Tautology
245
The sequence Γ is said to be a tautology if v ∗ (Γ) = T for all truth assignments
v : V AR −→ {T, F }. We write it as
|= Γ. (6.5)
Exercise 6.1
Let Γ be a sequence a, (b ∩ a), ¬b, (b ⇒ a).
1. Show that the truth assignment v : V AR −→ {T, F }, such that v(a) = F
and v(b) = T falsifies Γ, i.e. v 6|= Γ.
2. Let Γ be a sequence a, (¬b ∩ a), ¬b, (a ∪ b) and let v be a truth assignment
for which v(a) = T . Prove that v |= Γ.
3. Let Γ be a sequence a, (¬b ∩ a), ¬b, (a ∪ b). Prove that |= Γ.
Solution
1. Γ is the sequence a, (b∩a), ¬b, (b ⇒ a). We eveluate v ∗ (Γ) = v ∗ (δΓ ) = v ∗ (a)∪
v ∗ (b ∩ a) ∪ v ∗ (¬b) ∪ v ∗ (b ⇒ a) = F ∪ (F ∩ T ) ∪ F ∪ (T ⇒ F ) = F ∪ F ∪ F ∪ F = F.
By (6.4) we proved v 6|= Γ.
2. Let Γ be a sequence a, (¬b ∩ a), ¬b, (a ∪ b). We eveluate v ∗ (Γ) = v ∗ (δΓ ) =
v ∗ (a) ∪ v ∗ (¬b ∩ a) ∪ v ∗ (¬b) ∪ v ∗ (a ∪ b) = T ∪ v ∗ (¬b ∩ a) ∪ v ∗ (¬b) ∪ v ∗ (a ∪ b) = T.
By (6.3) we proved v |= Γ.
3. Assume now that Γ is falsifiable i.e. that we have a truth assignment v
for which v ∗ (Γ) = v ∗ (δΓ ) = v ∗ (a) ∪ v ∗ (¬b ∩ a) ∪ v ∗ (¬b) ∪ v ∗ (a ∪ b) = F This is
possible only when (in short-hand notation)
a ∪ (¬b ∩ a) ∪ ¬b ∪ a ∪ b = F,
In order to define the axioms LA and the set of rules of inference of RS we need
to introduce some definitions.
Literals
0
We form a special subset F ⊆ F of formulas, called a set of all literals, which
is defined as follows.
The variables are called positive literals and the elements of the second set of
the above union (6.6) are called negative literals. I.e propositional variables
are called positive literals and the negation of a variable is called a negative
246
literal, a variable or a negation of propositional variable is called a literal.
We denote by
0 0 0
Γ , ∆ , Σ , . . . with indices if necessary, (6.7)
0 0 0
elements of LT ∗ ⊆ F ∗ , i.e. Γ , ∆ , Σ are finite sequences (empty included)
formed out of literals. We call them indecomposable sequences.
We denote by
Γ, ∆, Σ, . . . with indices if necessary, (6.8)
the elements of F ∗ , i.e. Γ, ∆, Σ denote finite sequences (empty included)
formed out of elements of F.
Logical Axioms LA
0 0 0
Γ1 , a, Γ2 , ¬a, Γ3 (6.9)
or of the form
0 0 0
Γ1 , ¬a, Γ2 , a, Γ3 (6.10)
0 0 0
for any variable a ∈ V AR and any sequences Γ1 , Γ2 , Γ3 ∈ LT ∗ of literals.
Semantic Link
Consider axiom (6.9). Directly from the extension of the notion of tautology
to bf RS (6.5), we have that for any truth assignments v : V AR −→ {T, F },
0 0 0 0 0 0 0
v ∗ (Γ1 , ¬a, Γ2 , a, Γ3 ) = v ∗ (Γ1 ) ∪ v ∗ (¬a) ∪ v ∗ (a) ∪ v ∗ (Γ2 , Γ3 ) = v ∗ (Γ1 ) ∪ T ∪
0 0
∗
v (Γ2 , Γ3 ) = T The same applies to the axiom (6.10) We have thus proved the
following.
Fact 6.1
Logical axioms of RS are tautologies.
Rules of inference R
247
All rules of inference of RS are of the form
Γ1 Γ1 ; Γ2
or ,
Γ Γ
Disjunction rules
0 0 0
Γ , A, B, ∆ Γ , ¬A, ∆ : Γ , ¬B, ∆
(∪) 0 , (¬ ∪)
Γ , (A ∪ B), ∆ Γ0 , ¬(A ∪ B), ∆
Conjunction rules
0 0 0
Γ , A, ∆ ; Γ , B, ∆ Γ , ¬A, ¬B, ∆
(∩) , (¬ ∩)
Γ0 , (A ∩ B), ∆ Γ0 , ¬(A ∩ B), ∆
Implication rules
0 0 0
Γ , ¬A, B, ∆ Γ , A, ∆ : Γ , ¬B, ∆
(⇒) 0 , (¬ ⇒)
Γ , (A ⇒ B), ∆ Γ0 , ¬(A ⇒ B), ∆
Negation rule
0
Γ , A, ∆
(¬ ¬)
Γ0 , ¬¬A, ∆
0
where Γ ∈ LT ∗ , ∆ ∈ F ∗ , A, B ∈ F.
248
Formally we define the proof system RS as follows.
defined by (6.11).
`Γ
249
We picture, and write our proof trees with the node on the top, and leafs on
the very bottom, instead of more common way, where the leafs are on the top
and root is on the bottom of the tree.
In particular cases we write our proof trees indicating additionally the name
of the inference rule used at each step of the proof. For example, if the tree-
proof of a given formula A from axioms was obtained by the subsequent use of
the rules (∩), (∪), (∪), (∩), (∪), (¬¬), and (⇒), we represent it as the following
proof tree:
A (conclusion of (⇒))
| (⇒)
conclusion of (¬¬)
| (¬¬)
conclusion of (∪)
| (∪)
conclusion of (∩)
^
(∩)
| (∪) | (∪)
axiom axiom
The proof trees are often called derivation trees and we will use this notion
as well. Remark that the proof trees don’t represent a different definition of
a formal proof. Trees represent a certain visualization of the proofs and any
formal proof in any system can be represented in a tree form.
Example 6.1
Here is a proof tree in RS of the de Morgan law (¬(a ∩ b) ⇒ (¬a ∪ ¬b)).
| (⇒)
¬¬(a ∩ b), (¬a ∪ ¬b)
| (¬ ¬)
(a ∩ b), (¬a ∪ ¬b)
^
(∩)
250
a, (¬a ∪ ¬b) b, (¬a ∪ ¬b)
| (∪) | (∪)
a, ¬a, ¬b b, ¬a, ¬b
To obtain a ”linear ” formal proof of (¬(a ∩ b) ⇒ (¬a ∪ ¬b)) we just write down
the tree as a sequence, starting from the leafs and going up (from left to right)
to the root. The formal proof (with comments) thus obtained is:
a, ¬a, ¬b (axiom)
b, ¬a, ¬b (axiom)
a, (¬a ∪ ¬b) (rule (∪))
b, (¬a ∪ ¬b) (rule (∪))
(a ∩ b), (¬a ∪ ¬b) (rule(∩))
¬¬(a ∩ b), (¬a ∪ ¬b) (rule (¬¬))
(¬(a ∩ b) ⇒ (¬a ∪ ¬b)) (rule (⇒)).
Obviously it should have a proof in RS as we want it to be, and will prove later
to be complete. The search for the proof consists of building a certain tree.
We call it a decomposition tree, to be defined formally later. We proceed as
follows.
Observe that the main connective of A is ⇒. So, if A had a proof in RS it
would have come from the only possible rule used in its last step, namely the
rule (⇒) applied to its premiss, namely a sequence ¬¬(a ∪ b), (¬a ∩ ¬b). So the
last step in the proof of A would look as follows.
| (⇒)
¬¬(a ∪ b), (¬a ∩ ¬b)
251
Now, if the sequence ¬¬(a ∪ b), (¬a ∩ ¬b) (and hence also the formula A) had a
proof in RS its only step at this stage would have been the application of the
rule (¬¬) to a sequence (a ∪ b), (¬a ∩ ¬b). So, if A had a proof, its last two steps
would have been:
(¬(a ∪ b) ⇒ (¬a ∩ ¬b))
| (⇒)
¬¬(a ∪ b), (¬a ∩ ¬b)
| (¬¬)
(a ∪ b), (¬a ∩ ¬b)
Again, if the sequence (a ∪ b), (¬a ∩ ¬b) had a proof in RS its only step at this
stage would have been the application of the rule (∪) to a sequence a, b, (¬a∩¬b).
So, if A had a proof, its last three steps would have been as follows.
| (⇒)
¬¬(a ∪ b), (¬a ∩ ¬b)
| (¬ ¬)
(a ∪ b), (¬a ∩ ¬b)
| (∪)
a, b, (¬a ∩ ¬b)
Now, if the sequence a, b, (¬a ∩ ¬b) had a proof in RS its only step at this stage
would have been the application of the rule (∩) to the sequences a, b, ¬a and
a, b, ¬b as its left and right premisses, respectively. Both sequences are axioms
and the following tree is a proof of A in RS.
| (⇒)
¬¬(a ∪ b), (¬a ∩ ¬b)
| (¬¬)
(a ∪ b), (¬a ∩ ¬b)
| (∪)
a, b, (¬a ∩ ¬b)
^
(∩)
a, b, ¬a a, b, ¬b
252
From the above proof tree of A we construct, if we want, its formal proof, written
in a vertical manner, by writing the two axioms, which form the two premisses of
the rule (∩) one above the other. All other sequences remain the same. I.e. the
following sequence of elements of F ∗ is a formal proof of (¬(a∪b) ⇒ (¬a∩¬b))
in RS.
a, b, ¬b
a, b, ¬a
a, b, (¬a ∩ ¬b)
(a ∪ b), (¬a ∩ ¬b)
¬¬(a ∪ b), (¬a ∩ ¬b)
(¬(a ∪ b) ⇒ (¬a ∩ ¬b))
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c)
^
(∩)
¬c, (a ⇒ c)
(a ⇒ b), (a ⇒ c)
253
Now, if the sequences (a ⇒ b), (a ⇒ c) and ¬c, (a ⇒ c) had proofs in RS, then
their last, and the only steps would have been the the separate application of
the rule (⇒) to the sequences ¬a, b, (a ⇒ c) and ¬c, ¬a, c, respectively. The
sequence ¬c, ¬a, c is an axiom, so we stop the search on this branch. The
sequence ¬a, b, (a ⇒ c) is not an axiom, so the search continues. In this case
we can go one step further: if ¬a, b, (a ⇒ c) had a proof it would have been
only by the application of the rule (⇒) to a sequence ¬a, b, ¬a, c which is not
an axiom and the search ends. The tree generated by this search is called a
decomposition tree and is the following.
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c)
^
(∩)
¬c, (a ⇒ c)
(a ⇒ b), (a ⇒ c)
| (⇒)
| (⇒) ¬c, ¬a, c
¬a, b, (a ⇒ c)
| (⇒)
¬a, b, ¬a, c
254
Introduction of Connectives
The rules of RS are defined in such a way that each of them introduces
a new logical connective, or a negation of a connective to a sequence in
its domain (rules (∪), (⇒), (∩)) or a negation of a new logical connective
(rules (¬ ∪), (¬ ∩), (¬ ⇒), (¬ ¬)).
0
The rule (∪) introduces a new connective ∪ to a sequence Γ , A, B, ∆ and
0
it becomes, after the application of the rule, a sequence Γ , (A ∪ B), ∆.
Hence a name for this rule is (∪).
The rule (¬∪) introduces a negation of a connective, ¬∪ by combining
0 0
sequences Γ , ¬A, ∆ and Γ , ¬B, ∆ into one sequence (conclusion of the
0
rule) Γ , ¬(A ∪ B), ∆. Hence a name for this rule is (¬∪).
The same applies to all remaining rules of RS, hence their names say which
connective, or the negation of which connective has been introduced by
the particular rule.
Decomposition Rules
Building a proof search decomposition tree consists of using the inference
rules in an inverse order; we transform them into rules that transform
a conclusion into its premisses. We call such rules the decomposition
rules. Here are all of RS decomposition rules.
255
Tree Decomposition Rules
(∪) rule
0
Γ , (A ∪ B), ∆
| (∪)
0
Γ , A, B, ∆
(¬ ∪) rule
0
Γ , ¬(A ∪ B), ∆
^
(¬ ∪)
0 0
Γ , ¬A, ∆ Γ , ¬B, ∆
(∩) rule:
0
Γ , (A ∩ B), ∆
^
(∩)
0 0
Γ , A, ∆ Γ , B, ∆
(¬ ∩) rule:
0
Γ , ¬(A ∩ B), ∆
| (¬ ∩)
0
Γ , ¬A, ¬B, ∆
(⇒) rule:
0
Γ , (A ⇒ B), ∆
| (∪)
0
Γ , ¬A, B, ∆
256
(¬ ⇒) rule:
0
Γ , ¬(A ⇒ B), ∆
^
(¬ ⇒)
0 0
Γ , A, ∆ Γ , ¬B, ∆
(¬ ¬) rule:
0
Γ , ¬¬A, ∆
| (¬ ¬)
0
Γ , A, ∆
Observe that we use the same names for the inference and decomposition rules,
as once the we have built the decomposition tree (with use of the decomposition
rules) with all leaves being axioms, it constitutes a proof of A in RS with
branches labeled by the proper inference rules.
Now we still need to introduce few useful definitions and observations.
Definition 6.2
1. A sequence Γ is indecomposable if and only if Γ ∈ LT ∗ .
2. A formula A is decomposable if and only if A ∈ F − LT .
3. A sequence Γ is decomposable if and only if it contains a decomposable
formula.
Directly from the definition 6.8 we have three simple, but important observa-
tions.
Fact 6.2
1. For any decomposable sequence Γ, i.e. for any Γ 6∈ LT ∗ there is exactly
one decomposition rule that can be applied to it. This rule is determined by the
first decomposable formula in Γ, and by the main connective of that formula.
2. If the main connective of the first decomposable formula is ∪, ∩, or ⇒, then
the decomposition rule determined by it is (∪), (∩), or (⇒), respectively.
3. If the main connective of the first decomposable formula is ¬, then the de-
composition rule determined by it is determined by the second connective of the
formula. If the second connective is ∪, ∩, ¬, or ⇒, then corresponding decom-
position rule is (¬∪), (¬∩), (¬¬) and (¬ ⇒).
257
Directly from the Fact 6.2 we we have the following lemma.
We now prove the following Decomposition Tree Theorem 6.1. This Theorem
provides a crucial step in the proof of the Completeness Theorem for RS.
Proof
The tree TΓ is unique by the Unique Decomposition Lemma 6.1. It is finite
because there is a finite number of logical connectives in Γ and all decomposition
rules diminish the number of connectives. If the tree has a non- axiom leaf it is
not a proof by definition. By the its uniqueness it also means that the proof
does not exist.
258
Exercise 6.2
Construct a decomposition tree TA of the following formula A.
Solution
The formula A forms a one element decomposable sequence. The first decom-
position rule used is determined by its main connective. We put a box around
it, to make it more visible. The first and only rule applied is (∪) and we can
write the first segment of our decomposition tree TA :
TA
| (∪)
TA
| (∪)
| (⇒)
The next sequence to decompose is the sequence ¬(a ∪ b), ¬a, (¬a ⇒ ¬c).
The first decomposable formula is ¬(a∪b). Its main connective is ¬, so determine
the appropriate decomposition rule we have to examine next connective, which
is ∪.
The rule determine by this stage of decomposition is (¬ ∪) and now the next
stage of the decomposition tree TA is as follows.
259
TA
| (∪)
((a ∪ b) ⇒ ¬a), (¬a ⇒ ¬c)
| (⇒)
¬ (a ∪ b), ¬a, (¬a ⇒ ¬c)
^
(¬ ∪)
Now we have two decomposable sequences: ¬a, ¬a, (¬a ⇒ ¬c) and ¬b, ¬a, (¬a ⇒
¬c). They both happen to have the same first decomposable formula (¬a ⇒ ¬c).
We decompose it simultenously and obtain the following:
TA
| (∪)
((a ∪ b) ⇒ ¬a), (¬a ⇒ ¬c)
| (⇒)
¬ (a ∪ b), ¬a, (¬a ⇒ ¬c)
^
(¬∪)
| (⇒) | (⇒)
¬a, ¬a, ¬¬a, ¬c ¬b, ¬a, ¬¬a, ¬c
It is easy to see that we need only one more step to complete the process of
constructing the unique decomposition tree of TA , namely, by decomposing the
sequences: ¬a, ¬a, ¬¬a, ¬c and ¬b, ¬a, ¬¬a, ¬c.
The complete decomposition tree TA is:
260
TA
((a ∪ b) ⇒ ¬a) ∪ (¬a ⇒ ¬c))
| (∪)
((a ∪ b) ⇒ ¬a), (¬a ⇒ ¬c)
| (⇒)
¬ (a ∪ b), ¬a, (¬a ⇒ ¬c)
^
(¬∪)
| (⇒) | (⇒)
¬a, ¬a, ¬¬ a, ¬c ¬b, ¬a, ¬¬ a, ¬c
| (¬¬) | (¬¬)
¬a, ¬a, a, ¬c ¬b, ¬a, a, ¬c
Exercise 6.3
Prove that the formula A = (((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c)) is not provable in RS,
i.e.
6 `RS (((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c)).
Solution
We construct the formula A decomposition tree as follows.
TA
(((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c))
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c)
^
(∩)
(a ⇒ b), (a ⇒ c) ¬c, (a ⇒ c)
| (⇒) | (⇒)
¬a, b, (a ⇒ c) ¬c, ¬a, c
| (⇒)
¬a, b, ¬a, c
The above tree TA is unique by the Theorem 6.1 and represents the only possible
search for proof of the formula A = ((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c)) in RS. It has a
non-axiom leaf, hence by Theorem 6.1 the proof of A in RS does not exists.
261
6.3 Strong Soundness and Completeness
Our main goal is to prove the Completeness Theorem for RS. The proof of
completeness presented here is due to Rasiowa and Sikorski, as is the proof
system RS. Their proof, and the proof system was inverted for the classical
predicate logic and was published in 1961. We present their predicate logic proof
system QRS together with the proof of its completeness in chapter 10. Both
completeness proofs, for propositional RS and predicate QRS proof systems,
are constructive as they are based on a direct construction of a counter model
for any unprovable formula. The construction of a counter model for a formula
A uses directly its decomposition tree TA . We call such constructed model a
counter model determined by the tree TA . Both proofs relay heavily of the
notion of a strong soundness. We define it now, adopting Chapter 4 general
definition to our semantics.
We say it less formally that a rule (r) is strongly sound if the conjunction of
its premisses is logically equivalent with the conclusion, i.e.
P1 ∩ P2 ∩ . . . ∩ Pm ≡ C. (6.15)
Proof
The logical axioms (6.9), (6.10) are tautologies by Fact ??. We prove as an
example the strong soundness of two of inference rules: (∪) and (¬∪). Proofs
for all other rules follow the same patterns and are left as an exercise. By
definition 6.4 of strong soundness we have to show the condition (8.77). Written
262
formally it says that we have to show that that if P1 , P2 are premisses of a given
rule and C is its conclusion, then for all truth assignments v : V AR −→ {T, F },
v ∗ (P1 ) = v ∗ (C) in case of one premiss rule, and v ∗ (P1 ) ∩ v ∗ (P2 ) = v ∗ (C), in
case of a two premisses rule. Consider the rule (∪).
0
Γ , A, B, ∆
(∪) 0 .
Γ , (A ∪ B), ∆
By the definition:
0 0 0
v ∗ (Γ , A, B, ∆) = v ∗ (δ{Γ0 ,A,B,∆} ) = v ∗ (Γ ) ∪ v ∗ (A) ∪ v ∗ (B) ∪ v ∗ (∆) = v ∗ (Γ ) ∪
0
v ∗ (A ∪ B) ∪ v ∗ (∆) = v ∗ (δ{Γ0 ,(A∪B),∆} ) = v ∗ (Γ , (A ∪ B), ∆).
By the definition:
0 0 0 0
v ∗ (Γ , ¬A, ∆) ∩ v ∗ (Γ , ¬B, ∆) = (v ∗ (Γ ) ∪ v ∗ (¬A) ∪ v ∗ (∆)) ∩ (v ∗ (Γ ) ∪ v ∗ (¬B) ∪
0 0
v ∗ (∆)) = (v ∗ (Γ , ∆) ∪ v ∗ (¬A)) ∩ (v ∗ (Γ , ∆) ∪ v ∗ (¬B)) = by distributivity =
0 0
(v ∗ (Γ , ∆) ∪ (v ∗ (¬A) ∩ v ∗ (¬B)) = v ∗ (Γ ) ∪ v ∗ (∆) ∪ (v ∗ (¬A ∩ ¬B)) = by the log-
0
ical equivalence of (¬A ∩ ¬B) and ¬(A ∪ B) = v ∗ (δ{Γ0 ,¬(A∪B),∆} = v ∗ (Γ , ¬(A ∪
B), ∆)).
Observe that the strong soundness implies soundness (not only by name!), hence
we have also proved the following.
We have just proved (Theorem 6.2) that all the rules of inference of RS of are
strongly sound, i.e. C ≡ P and C ≡ P1 ∩ P2 . The strong soundness of the rules
means that if at least one of premisses of a rule is false, so is its conclusion. Hence
given a formula A, such that its TA has a branch ending with a non-axiom leaf.
By Strong Soundness Theorem 6.2, any v that make this non-axiom leaf false
also falsifies all sequences on that branch, and hence falsifies the formula A. This
means that any v, such that it falsifies a non-axiom leaf is a counter-model
for A. We have hence proved the following.
263
Given a formula A ∈ F such that its decomposition tree TA contains a non-
axiom leaf LA . Any truth assignment v that falsifies the non-axiom leaf LA is
a counter model for A. We call it a counter-model for A determined by the
decomposition tree TA .
TA
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c)
^
(∩)
(a ⇒ b), (a ⇒ c) ¬c, (a ⇒ c)
| (⇒) | (⇒)
¬a, b, (a ⇒ c) ¬c, ¬a, c
| (⇒)
¬a, b, ¬a, c
The tree TA has a non-axiom leaf LA : ¬a, b, ¬a, c. The truth assignment
v : V AR −→ {T, F } that falsifies the leaf ¬a, b, ¬a, c must be such that
v ∗ (¬a, b, ¬a, c) = v ∗ (¬a)∪v ∗ (b)∪v ∗ (¬a)∪v ∗ (c) = ¬v(a)∪v(b)∪¬v(a)∪v(c) = F,
i.e. v must be such that ¬v(a) ∪ v(b) ∪ ¬v(a) ∪ v(c) = F. We hence get that
v(a) = T, v(b) = F, v(c) = F. By the Counter Model Theorem 6.4, the truth
assignment v determined by the non-axiom leaf also falsifies the formula A, i.e.
we proved that v is a counter model for A and
The Counter Model Theorem 6.4, says that the logical value F determined by
the evaluation a non-axiom leaf ”climbs” the decomposition tree. We picture it
as follows.
TA
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c) = F
^
(∩)
264
(a ⇒ b), (a ⇒ c) = F ¬c, (a ⇒ c)
| (⇒) | (⇒)
¬a, b, (a ⇒ c) = F ¬c, ¬a, c
| (⇒) axiom
¬a, b, ¬a, c = F
Observe that the same counter model construction applies to any other non-
axiom leaf of TA , if exists. The other non-axiom leaf of TA defines another
evaluation of the non- axiom leaf to F that also ”climbs the tree” and hence
defines another counter- model for a formula A. By Counter Model 6.4 all
possible restricted counter-models for A are those determined by its all non-
axioms leaves.
In our case the tree TA has only one non-axiom leaf, and hence the formula
(((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c)) only only one restricted counter model.
Our main goal is to prove the Completeness Theorem for RS. We prove first
the Completeness Theorem for formulas A ∈ F and then we generalize it to any
sequences Γ ∈ F ∗ .
Proof
Case 1. We have already proved the Soundness Theorem 6.3, so we need to
prove only the completeness part of it, namely to prove the implication:
if |= A, then `RS A. (6.16)
We prove instead of the opposite implication:
if 6 `RS A then 6|= A. (6.17)
Assume that A is any formula is such that 6 `RS A. By the Decomposition Tree
Theorem 6.1 the tree TA contains a non-axiom leaf LA . We use the non-axiom
leaf LA to define a truth assignment v : V AR −→ {T, F } which falsifies it
as follows:
F if a appears in LA
v(a) = T if ¬a appears in LA
any value if a does not appear in LA
By the Counter Model Theorem 6.4we have that v also falsifies the formula
A. We proved that
6|= A
265
and it ends the proof of the case 1.
Case 2. Assume that Γ ∈ F ∗ is any sequence such that 6 `RS Γ. But obviously,
`RS Γ if and only if `RS δΓ , where δΓ is any disjunction of all formulas of Γ.
So 6 `RS Γ if and only if 6 `RS δΓ and by already proven Case 1, 6|= δΓ what is
obviously equivalent to 6|= Γ. This ends the proof of Case 2 and Completeness
Theorem.
L = L{¬,⇒,∪,∩} .
266
Disjunction rules
0 0 0
Γ, A, B, ∆ Γ, ¬A, ∆ : Γ, ¬B, ∆
(∪) , (¬ ∪) ,
Γ, (A ∪ B), ∆0 Γ, ¬ (A ∪ B), ∆0
Conjunction rules
0 0 0
Γ, A, ∆ ; Γ, B, ∆ Γ, ¬A, ¬B, ∆
(∩) , (¬ ∩) 0 ,
Γ, (A ∩ B), ∆0 Γ , ¬(A ∩ B), ∆
Implication rules
0 0 0
Γ, ¬A, B, ∆ Γ, A, ∆ : Γ, ¬B, ∆
(⇒) , (¬ ⇒) ,
Γ, (A ⇒ B), ∆0 Γ, ¬ (A ⇒ B), ∆0
Negation rule
0
Γ, A, ∆
(¬ ¬)
Γ, ¬ ¬ A, ∆0
0
where Γ ∈ F ∗ , ∆ ∈ LT ∗ , A, B ∈ F.
Exercise 6.4
Construct a proof in RS1 of a formula
Solution
The decomposition tree below is a proof of A in RS1 as all its leaves are axioms.
TA
| (⇒)
(¬¬(a ∩ b), (¬a ∪ ¬b)
267
| (∪)
¬¬(a ∩ b), ¬a, ¬b
| (¬¬)
(a ∩ b), ¬a, ¬b
^
(∩)
a, ¬a, ¬b b, ¬a, ¬b
Exercise 6.5
Prove that RS1 is strongly sound.
Solution
0
Observe that the system RS1 is obtained from RS by changing the sequence Γ
0
into Γ and the sequence ∆ into ∆ in all of the rules of inference of RS. These
changes do not influence the essence of proof of strong soundness of the rules
0 0
of RS. One has just to replace the sequence Γ by Γ and the sequence ∆ by ∆
in the proof of strong soundness of each rule of RS to obtain a corresponding
proof of strong soundness of corresponding rule of RS1. We do it, for example
for the rule (∪) of RS1. Consider the rule (∪) of RS1:
0
Γ, A, B, ∆
(∪) .
Γ, (A ∪ B), ∆0
We evaluate:
0 0
v ∗ (Γ, A, B, ∆ ) = v ∗ (δ{Γ,A,B,∆0 } ) = v ∗ (Γ) ∪ v ∗ (A) ∪ v ∗ (B) ∪ v ∗ (∆ )
0 0
= v ∗ (Γ) ∪ v ∗ (A ∪ B) ∪ v ∗ (∆ ) = v ∗ (δ{Γ,(A∪B),∆0 } ) = v ∗ (Γ, (A ∪ B), ∆ ).
Exercise 6.6
Define in your own words, for any formula A ∈ F the decomposition tree TA
in RS1.
Solution
The definition of the decomposition tree TA is again, it its essence similar
to the one for RS except for the changes which reflect the differences in the
corresponding rules of inference. We follow now the following steps.
Step 1 Decompose A using a rule defined by its main connective.
Step 2 Traverse resulting sequence Γ on the new node of the tree from right
to left and find the first decomposable formula.
Step 3 Repeat Step 1 and Step 2 until there is no more decomposable
formulas. End of tree construction.
268
Exercise 6.7
Prove the following Completeness Theorem for RS1.
Theorem 6.6
For any formula A ∈ F,
1. `RS1 A if and only if |= A, and for any Γ ∈ F ∗ ,
2. `RS1 Γ if and only if |= Γ.
Solution Part 1.
Observe that directly from the definition of the uniqueness of the decomposition
tree TA we have that the following holds.
Fact 6.3
The decomposition tree TA is a proof if and only if all leaves are axioms and
the proof does not exist otherwise, i.e. we have that 6 `RS1 A if and only if
there is a non- axiom leaf on TA .
The Fact 6.3 together with strong soundness of the rules of inference of RS1
justify the correctness of construction of a counter-model generated by a the
a non- axiom leaf and hence the correctness of the following proof of the
Completeness Theorem.
We prove, as we did in case of RS the implication
Assume that A is any formula such that 6 `RS1 A. By the Fact 6.3 the decom-
position tree TA contains a non-axiom leaf LA . We use the non-axiom leaf LA
and define a truth assignment v which falsifies A, as follows:
F if a appears in LA
v(a) = T if ¬a appears in LA
any value if a does not appear in LA .
Exercise 6.8
269
Construct two decomposition trees in RS2 of the formula
Solution
Here are two out of many more decomposition trees.
T1A
| (⇒)
¬¬(¬a => (a ∩ ¬b)), (¬a ∩ (¬a ∪ ¬b))
| (¬¬)
(¬a => (a ∩ ¬b)), (¬a ∩ (¬a ∪ ¬b))
| (⇒)
¬¬a, (a ∩ ¬b), (¬a ∩ (¬a ∪ ¬b))
| (¬¬)
a, (a ∩ ¬b), (¬a ∩ (¬a ∪ ¬b))
^
(∩)
T2A
| (⇒)
¬¬(¬a => (a ∩ ¬b)), (¬a ∩ (¬a ∪ ¬b))
| (¬¬)
(¬a => (a ∩ ¬b)), (¬a ∩ (¬a ∪ ¬b))
^
(∩)
270
(¬a => (a ∩ ¬b)), ¬a
(¬a => (a ∩ ¬b)), (¬a ∪ ¬b)
| (⇒)
(¬¬a, (a ∩ ¬b)), ¬a | (∪)
| (¬¬) (¬a => (a ∩ ¬b)), ¬a, ¬b
a, (a ∩ ¬b), ¬a | (⇒)
(¬¬a, (a ∩ ¬b), ¬a, ¬b
^
(∩)
| (¬¬)
a, (a ∩ ¬b), ¬a, ¬b
a, a, ¬a a, ¬b, ¬a ^
(∩)
axiom axiom
axiom axiom
Exercise 6.9
Explain why the system RS2 is strongly sound. You can use the strong
soundness of the system RS.
Solution
The only one difference between RS and RS2 is that in RS2 each inference rule
has at the beginning a sequence of any formulas, not only of literals, as in RS.
So there are many ways to apply the decomposition rules while constructing the
decomposition tree, but it does not affect strong soundness, since for all rules
of RS2 premisses and conclusions are still logically equivalent as they were in
RS .
Consider, for example, RS2 rule
Γ, A, B, ∆
(∪) .
Γ, (A ∪ B), ∆
We evaluate v ∗ (Γ, A, B, ∆) = v ∗ (Γ) ∪ v ∗ (A) ∪ v ∗ (B) ∪ v ∗ (∆) = v ∗ (Γ) ∪ v ∗ (A ∪
B) ∪ v ∗ (∆) = v ∗ (Γ, (A ∪ B), ∆). Similarly, as in RS, we show all other rules of
RS2 to be strongly sound, thus RS2 is sound.
Exercise 6.10
Define shortly, in your own words, for any formula A, its decomposition tree
TA in RS2. Justify why your definition is correct. Show that in RS2 the
decomposition tree for some formula A may not be unique.
Solution
Given a formula A. The decomposition tree TA can be defined as follows. It has
271
A as a root. For each node, if there is a rule of RS2 which conclusion has the
same form as node sequence, i.e. there is a decomposition rule to be applied,
then the node has children that are premises of the rule. If the node consists
only of literals (i.e. there is no decomposition rule to be applied), then it does
not have any children. The last statement define a termination condition for
the tree TA .
This definition defines correctly the decomposition tree TA as it identifies and
uses appropriate decomposition rules. Since all rules of inference of RS2 have
a sequence Γ instead of Γ0 as it was in RS, the choice of the decomposition rule
for a node may not unique. For example consider a node (a => b), (b ∪ a).
Γ in the RS2 rules may be a sequence of formulas, not only literals, so for the
node (a => b), (b ∪ a) we can choose as a decomposition rule either (=>) or
(∪). This leads to a non-unique tree.
Exercise 6.11
Prove the following Completeness Theorem for RS2.
Theorem 6.7
For any formula A ∈ F,
1. `RS2 A if and only if |= A, and for any Γ ∈ F ∗ ,
2. `RS2 Γ if and only if |= Γ.
Solution
We need to prove the completeness part only, as the Soundness has been already
proved, i.e. we have to prove the implication (Part 1): for any formula A,
Assume 6`RS2 A. Then every decomposition tree of A has at least one non-
axiom leaf. Otherwise, there would exist a tree with all axiom leaves and it
would be a proof for A. Let TA be a set of all decomposition trees of A. We
choose an arbitrary TA ∈ TA with at least one non-axiom leaf LA . We use the
non-axiom leaf LA to define a truth assignment v which falsifies A, as follows:
F if a appears in LA
v(a) = T if ¬a appears in LA
any value if a does not appear in LA
The value for a sequence that corresponds to the leaf in is F. Since, because of
the strong soundness F ”climbs” the tree, we found a counter-model for A. This
proves that 6|= A. Part 2. proof is identical to the proof in RS case.
Exercise 6.12
272
Write a procedure T REEA such that for any formula A of RS2 it produces its
unique decomposition tree.
Solution
Here is the procedure.
Procedure T REEA (Formula A, Tree T)
{
B = ChoseLef tM ostF ormula(A) // Choose the left most formula that is
not a literal
c = M ainConnective(B) // Find the main connective of B
R = F indRule(c)// Find the rule which conclusion that has this connective
P = P remises(R)// Get the premises for this rule
AddT oT ree(A, P )// add premises as children of A to the tree
For all p in P // go through all premises
T REEA (p, T ) // build subtrees for each premiss
}
Exercise 6.13
Prove completeness of your Procedure T REEA .
Solution
Procedure T REEA provides a unique tree, since it always chooses the most
left indecomposable formula for a choice of a decomposition rule and there is
only one such rule. This procedure is equivalent to RS system, since with the
decomposition rules of RS the most left decomposable formula is always chosen.
The proof RS system is complete, thus this Procedure T REEA is complete.
The system GL presented here is the most similar in its structure to the system
RS (6.18) and hence is the first to be considered. It admits a constructive
273
proof of the Completeness Theorem that is very similar to the proof of the
Completeness Theorem for the system RS.
Language of GL
We adopt a propositional language L = L{∪,∩,⇒,¬} with the set of formulas
denoted by F and we add a new symbol −→ called a Gentzen arrow, to it.
It means we consider formally a new language L1 = L{∪,∩,⇒,¬} ∪ {−→}. As
the next step we build expressions called sequents out of L1 . The sequents are
built out of finite sequences (empty included) of formulas, i.e. elements of F ∗
of L{∪,∩,⇒,¬} , and the additional sign −→.
We denote , as in the RS system, the finite sequences of formulas of of L{∪,∩,⇒,¬}
by Greek capital letters
Γ, ∆, Σ, . . . ,
with indices if necessary. We define a sequent as follows.
Γ −→ ∆
Γ −→ ∆ ≡ (σΓ ⇒ δ∆ ).
Formal semantics
SQ = { Γ −→ ∆ : Γ, ∆ ∈ F ∗ } (6.20)
274
Definition 6.7 For any sequent Γ −→ ∆ ∈ SQ,
v ∗ (Γ −→ ∆) = v ∗ (σΓ ) ⇒ v ∗ (δ∆ ).
Counter- model
The sequent Γ −→ ∆ is falsifiable if there is a truth assignment v, such that
v ∗ (Γ −→ ∆) = F . In this case v is called a counter-model for Γ −→ ∆ and
we write it as
v 6|= Γ −→ ∆.
Tautology
The sequent Γ −→ ∆ is a tautology if v ∗ (Γ −→ ∆) = T for all truth
assignments v : V AR −→ {T, F } and we write
|= Γ −→ ∆.
Example 6.2
Let Γ −→ ∆ be a sequent
a, (b ∩ a) −→ ¬b, (b ⇒ a).
Any truth assignment v, such that v(a) = T and v(b) = T is a model for
Γ −→ ∆, i.e.
|= a, (b ∩ a) −→ ¬b, (b ⇒ a).
275
The Proof System GL
P1 P1 ; P2
or ,
C C
where P1 , P2 and C are sequents. P1 , P2 are called premisses and C is called the
conclusion of the rule of inference. Each rule of inference introduces a new logical
connective to the antecedent or to the succedent of the conclusion sequent. We
denote the rule that introduces the logical connective ◦ to the antecedent of the
conclusion sequent P by (◦ →). The notation (→ ◦) means that the logical
connective is introduced to the succedent of the conclusion sequent P .
As our language contains the connectives: ∩, ∪, ⇒ and ¬, we are going to adopt
the following inference rules: (∩ →) and (→ ∩), (∪ →) and (→ ∪), (⇒→) and
(→⇒), and finally, (¬ →) and (→ ¬).
Definition 6.8
Finite sequences formed out of positive literals i.e. out of propositional vari-
ables are called indecomposable. We denote them as before by
0 0
Γ , ∆ , ...
Axioms of GL
As the axioms of GL we adopt any indecomposable sequent sequent which
contains a positive literal a (variable) that appears on both sides of the sequent
arrow −→, i.e any sequent of the form
Γ0 1 , a, Γ0 2 −→ ∆0 1 , a, ∆0 2 , (6.21)
276
Consider axiom (6.21). Directly from the Definition 6.7 of semantics for bf GL
we evaluate (in shorthand notation), for any truth assignments v : V AR −→
{T, F }, the following (in shorthand notation).
v ∗ (Γ0 1 , a, Γ0 2 −→ ∆0 1 , a, ∆0 2 ) =
Fact 6.4
Logical axioms of GL are tautologies.
Conjunction rules
0 0 0 0
Γ , A, B, Γ −→ ∆ Γ −→ ∆, A, ∆ ; Γ −→ ∆, B, ∆
(∩ →) , (→ ∩) ,
Γ0 , (A ∩ B), Γ −→ ∆0 Γ −→ ∆, (A ∩ B), ∆0
Disjunction rules
0 0 0 0 0
Γ −→ ∆, A, B, ∆ Γ , A, Γ −→ ∆ ; Γ , B, Γ −→ ∆
(→ ∪) , (∪ →) ,
Γ −→ ∆, (A ∪ B), ∆0 Γ0 , (A ∪ B), Γ −→ ∆0
Implication rules
0 0
Γ , A, Γ −→ ∆, B, ∆
(→⇒) 0 ,
Γ , Γ −→ ∆, (A ⇒ B), ∆0
0 0 0 0
Γ , Γ −→ ∆, A, ∆ ; Γ , B, Γ −→ ∆, ∆
(⇒→) ,
Γ0 , (A ⇒ B), Γ −→ ∆, ∆0
Negation rules
0 0 0 0
Γ , Γ −→ ∆, A, ∆ Γ , A, Γ −→ ∆, ∆
(¬ →) , (→ ¬) .
Γ0 , ¬A, Γ −→ ∆, ∆0 Γ0 , Γ −→ ∆, ¬A, ∆0
Formally we define:
GL = (L, SQ, LA, (∪), (¬∪), (∩), (¬∩), (⇒), (¬ ⇒), (¬¬)), (6.23)
277
where SQ = { Γ −→ ∆ : Γ, ∆ ∈ F ∗ }, (∪), (¬∪), (∩), (¬∩), (⇒), (¬ ⇒), (¬¬)
are the inference rules defined above and AL are the logical axioms of the system
defined by the schema (6.21).
We define the notion of a bf formal proof in GL as in any proof system, i.e.,
by a formal proof of a sequent Γ −→ ∆ in the proof system GL we understand
any sequence
Γ1 −→ ∆1 , Γ2 −→ ∆2 , ...., Γn −→ ∆n
of sequents, such that Γ1 −→ ∆1 ∈ AL, Γn −→ ∆n = Γ −→ ∆, and for all i
(1 < i ≤ n) Γi −→ ∆i ∈ AL, or Γi −→ ∆i is a conclusion of one of the inference
rules of GL with all its premisses placed in the sequence Γ1 −→ ∆1 , ....Γi−1 −→
∆i−1 .
We write, as usual, `GL Γ −→ ∆ to denote that Γ −→ ∆ has a formal proof in
GL, or we write simply ` Γ −→ ∆ when the system GL is fixed.
We say that a formula A ∈ F, has a proof in GL and denote it by `GL A if the
sequent −→ A has a proof in GL, i.e. we define:
We write, however, the formal proofs in GL in a form of proof trees rather then
in a form of sequences of sequents.
Proof trees
A proof tree Γ −→ ∆ is a tree TΓ−→∆ satisfying the following conditions:
1. The topmost sequent, i.e the root of TΓ−→∆ is Γ −→ ∆.
2. All leaves are axioms.
3. The nodes are sequents such that each sequent on the tree follows from the
ones immediately preceding it by one of the rules.
| (→⇒)
278
¬(a ∩ b) −→ (¬a ∪ ¬b)
| (→ ∪)
¬(a ∩ b) −→ ¬a, ¬b
| (→ ¬)
b, ¬(a ∩ b) −→ ¬a
| (→ ¬)
b, a, ¬(a ∩ b) −→
| (¬ →)
b, a −→ (a ∩ b)
^
(→ ∩)
b, a −→ a b, a −→ b
Remark 6.1
The proof search in GL (to be defined by the decomposition tree) results are not
always unique; one formula (sequent) can have many proofs.
b −→ ¬a, a b −→ ¬a, b
| (−→ ¬) | (−→ ¬)
b, a −→ a b, a −→ b
279
The process of searching for proofs of a formula A in GL consists, as in the RS
type systems, of building decomposition trees. Their construction is similar to
the one defined for RS type systems and is described intuitiively as follows.
We take a root of a decomposition tree TA a sequent −→ A. For each node, if
there is a rule of GL which conclusion has the same form as the node sequent,
then the node has children that are premises of the rule. If the node consists
only of an indecomposable sequent (built out of variables only), then it does not
have any children. This is a termination condition for the decomposition tree.
We prove that each formula A generates a finite set TA of decomposition trees,
such that the following holds. If there exist a tree TA ∈ TA whose all leaves are
axioms, then tree TA constitutes a proof of A in GL. If all trees in TA have at
least one non-axiom leaf, the proof of A does not exist.
The first step in formally defining a notion of a decomposition tree consists of
transforming the inference rules of GL, as we did in the case of the RS type
systems, into corresponding decomposition rules.
Decomposition rules of GL
Building a proof search decomposition tree consists of using the inference rules
in an inverse order; we transform the inference rules into decomposition rules
by reversing the role of conclusion and its premisses. We call such rules the
decomposition rules. Here are all of GL decomposition rules.
Conjunction decomposition rules
0 0 0
Γ , (A ∩ B), Γ −→ ∆ Γ −→ ∆, (A ∩ B), ∆
(∩ →) , (→ ∩) ,
Γ0 , A, B, Γ −→ ∆0 Γ −→ ∆, A, ∆0 ; Γ −→ ∆, B, ∆0
0 0 0
Γ −→ ∆, (A ∪ B), ∆ Γ , (A ∪ B), Γ −→ ∆
(→ ∪) , (∪ →) 0 ,
Γ −→ ∆, A, B, ∆0 Γ , A, Γ −→ ∆0 ; Γ0 , B, Γ −→ ∆0
0 0
Γ , Γ −→ ∆, (A ⇒ B), ∆
(→⇒) ,
Γ0 , A, Γ −→ ∆, B, ∆0
0 0
Γ , (A ⇒ B), Γ −→ ∆, ∆
(⇒→) ,
Γ0 , Γ −→ ∆, A, ∆0 ; Γ0 , B, Γ −→ ∆, ∆0
280
0 0 0 0
Γ , ¬A, Γ −→ ∆, ∆ Γ , Γ −→ ∆, ¬A, ∆
(¬ →) , (→ ¬) .
Γ0 , Γ −→ ∆, A, ∆0 Γ0 , A, Γ −→ ∆, ∆0
0 0
Γ , (A ∪ B), Γ −→ ∆
^
(∪ →)
0 0 0 0
Γ , A, Γ −→ ∆ Γ , B, Γ −→ ∆
(→ ∩) rule
0
Γ −→ ∆, (A ∩ B), ∆
^
(→ ∩)
0
0
Γ → ∆, B, ∆
Γ −→ ∆, A, ∆
(∩ →) rule
0 0
Γ , (A ∩ B), Γ −→ ∆
| (∩ →)
0 0
Γ , A, B, Γ −→ ∆
(→⇒) rule
0 0
Γ , Γ −→ ∆, (A ⇒ B), ∆
| (→⇒)
0 0
Γ , A, Γ −→ ∆, B, ∆
(⇒→) rule
281
0 0
Γ , (A ⇒ B), Γ −→ ∆, ∆
^
(⇒→)
0 0 0 0
Γ , Γ −→ ∆, A, ∆ Γ , B, Γ −→ ∆, ∆
(¬ →) rule
0 0
Γ , ¬A, Γ −→ ∆, ∆
| (¬ →)
0 0
Γ , Γ −→ ∆, A, ∆
(−→ ¬) rule
0 0
Γ , Γ −→ ∆, ¬A, ∆
| (¬ →)
0 0
Γ , A, Γ −→ ∆, ∆
Observe that we use the same names for the inference and decomposition rules,
as once the we have built a decomposition tree (with use of the decomposition
rules) with all leaves being axioms, it constitutes a proof of A in GL with
branches labeled by the proper inference rules.
We have already defined (definition 6.8) indecomposable sequence as any se-
0 0 0 0
quence Γ −→ ∆ when Γ , ∆ ∈ V AR∗ . In particular, a formula that is not
a positive literal (propositional variable) is called a decomposable formula,
and a sequent Γ −→ ∆ where either Γ or ∆ contains a decomposable formula
is called a decomposable sequent.
By inspecting the domain of the rules we can see that at most two rules could
apply for any given decomposable sequent Γ −→ ∆.
For any decomposable sequent, at most two decomposition rules can be applied
to it. This rule is determined by the first decomposable formula in Γ when we
traverse it from left to right, and by the main connective of that formula, or by
the first decomposable formula in ∆ when we traverse it from the right to left,
and by the main connective of that formula. We hence are now ready to define
a decomposition tree.
282
Decomposition Tree T→A
For each formula A ∈ F, a decomposition tree T→A is a tree build as follows.
Step 1. The sequent −→ A is the root of T→A and for any node Γ −→ ∆ of
the tree we follow the steps below.
Step 2. If Γ −→ ∆ is indecomposable, then Γ −→ ∆ becomes a leaf of the
tree.
Step 3. If Γ −→ ∆ is decomposable, then we pick a decomposition rule that
applies by matching the sequent of the current node with the domain of the
decomposition rule. To do so we proceed as follows.
1. We traverse Γ from left to right to find the first decomposable formula.
Its main connective ◦ identifies a possible decomposition rule (◦ −→). Then
we check if this decomposition rule applies. If it does we put its conclusions
(conclusion) as leaves (leaf).
2. We traverse ∆ from right to left to find the first decomposable formula.
Its main connective ◦ identifies a possible decomposition rule (−→ ◦). Then
we check if this decomposition rule applies. If it does we put its conclusions
(conclusion) as leaves (leaf). 3. If 1. and 2. applies we choose one of the rules.
Step 4. We repeat steps 2 and 3 until we obtain only leaves.
Observation 6.1
The decomposable Γ −→ ∆ is always in the domain in one of the decomposition
rules (◦ −→), (−→ ◦), or in the domain of both. Hence the tree T→A may not
be unique and all possible choices of 3. give all possible decomposition trees.
Exercise 6.14
Prove, by constructing a proper decomposition tree that
283
Solution
By definition,we have that
`GL ((¬a ⇒ b) ⇒ (¬b ⇒ a)) if and only if `GL −→ ((¬a ⇒ b) ⇒ (¬b ⇒ a)).
T→A
| (→⇒)
(¬a ⇒ b) −→ (¬b ⇒ a)
| (→⇒)
¬b, (¬a ⇒ b) −→ a
| (→ ¬)
(¬a ⇒ b) −→ b, a
^
(⇒−→)
−→ ¬a, b, a b −→ b, a
| (→ ¬) axiom
a −→ b, a
axiom
All leaves of the tree are axioms, hence it constitutes a proof in GL.
Exercise 6.15
Prove, by constructing proper decomposition trees that
Solution
Observe that for some formulas A, their decomposition tree T→A in GL may
not be unique. Hence we have to construct all possible decomposition trees to
see that none of them is a proof, i.e. to see that each of them has a non axiom
leaf. We construct the decomposition trees for −→ A as follows.
T1→A
−→ ((a ⇒ b) ⇒ (¬b ⇒ a))
(one choice)
284
| (→⇒)
(a ⇒ b) −→ (¬b ⇒ a)
( first of two choices )
| (→⇒)
¬b, (a ⇒ b) −→ a
(one choice)
| (¬ →)
(a ⇒ b) −→ b, a
(one choice)
^
(⇒−→)
−→ a, b, a b −→ b, a
non − axiom axiom
The tree contains a non- axiom leaf, hence it is not a proof. We have one more
tree to construct.
T2→A
−→ ((a ⇒ b) ⇒ (¬b ⇒ a))
| (→⇒)
(a ⇒ b) −→ (¬b ⇒ a)
(second choice)
^
(⇒−→)
Exercise 6.16
Does the tree below constitute a proof in GL?
285
T→A
| (→ ¬)
¬((¬a ⇒ b) ⇒ (¬b ⇒ a)) −→
| (¬ →)
−→ ((¬a ⇒ b) ⇒ (¬b ⇒ a))
| (→⇒)
(¬a ⇒ b) −→ (¬b ⇒ a)
| (→⇒)
(¬a ⇒ b), ¬b −→ a
| (¬ →)
(¬a ⇒ b) −→ b, a
^
(⇒→)
−→ ¬a, b, a b −→ b, a
| (→ ¬) axiom
a −→ b, a
axiom
Solution
The tree above is not a proof in GL because a decomposition rule used in the
decomposition step below does not exists in GL
(¬a ⇒ b), ¬b −→ a
| (¬ →)
(¬a ⇒ b) −→ b, a.
It is a proof is some system GL1 that has all the rules of GL except its rule
(¬ →)
0 0
Γ , Γ −→ ∆, A, ∆
(¬ →)
Γ0 , ¬A, Γ −→ ∆, ∆0
This rule has to be replaced in by the rule:
0 0
Γ, Γ −→ ∆, A, ∆
(¬ →)1
Γ, ¬A, Γ0 −→ ∆, ∆0
286
6.6 GL Soundness and Completeness
The system GL admits a constructive proof of the Completeness Theorem,
similar to completeness proofs for RS type proof systems (Theorems 9.7, 6.6,
6.7). It also relays on strong soundness property of its inference rules. We are
going to prove that the following holds.
Proof We have already proved (Fact 6.4) that logical axioms of GL are tau-
tologies, so we have to prove now that its rules of inference are strongly sound
(definition 6.4). Proofs of strong soundness of rules of inference of GL are more
involved then the proofs for the RS type rules. We prove as an example the
strong soundness of four of inference rules. Proofs for all other rules follows the
same patterns and is left as an exercise.
By definition 6.4 of strong soundness we have to show the condition (8.77).
Written formally it says that we have to show that that if P1 , P2 are premisses
of a given rule and C is its conclusion, then for all truth assignments v :
V AR −→ {T, F },
v ∗ (P1 ) = v ∗ (C) in case of one premiss rule, and
v ∗ (P1 ) ∩ v ∗ (P2 ) = v ∗ (C), in case of a two premisses rule.
In order to prove it we need additional classical equivalencies listed below. You
can fond a list of most basic classical equivalences in Chapter 3.
0 0
Γ , A, B, Γ −→ ∆
(∩ →)
Γ0 , (A ∩ B), Γ −→ ∆0
0 0 0 0 0
v ∗ (Γ , A, B, Γ −→ ∆ ) = (v ∗ (Γ )∩v ∗ (A)∩v ∗ (B)∩v ∗ (Γ)) ⇒ v ∗ (∆ ) = (v ∗ (Γ )∩
0 0 0
v ∗ (A ∩ B) ∩ v ∗ (Γ)) ⇒ v ∗ (∆ ) = v ∗ (Γ , (A ∩ B), Γ −→ ∆ )
0 0
Γ −→ ∆, A, ∆ ; Γ −→ ∆, B, ∆
(→ ∩)
Γ −→ ∆, (A ∩ B), ∆0
0 0
v ∗ (Γ −→ ∆, A, ∆ ) ∩ v ∗ (Γ −→ ∆, B, ∆ )
0 0
= (v ∗ (Γ) ⇒ v ∗ (∆) ∪ v ∗ (A) ∪ v ∗ (∆ )) ∩ (v ∗ (Γ) ⇒ v ∗ (∆) ∪ v ∗ (B) ∪ v ∗ (∆ ))
287
[we use : ((A ⇒ B) ∩ (A ⇒ C)) ≡ (A ⇒ (B ∩ C))]
0 0
= v ∗ (Γ) ⇒ ((v ∗ (∆) ∪ v ∗ (A) ∪ v ∗ (∆ )) ∩ (v ∗ (∆) ∪ v ∗ (B) ∪ v ∗ (∆ )))
[we use commutativity and distributivity]
0
= v ∗ (Γ) ⇒ (v ∗ (∆) ∪ (v ∗ (A ∩ B)) ∪ v ∗ (∆ ))
0
= v ∗ (Γ −→ ∆, (A ∩ B), ∆ )
0 0 0 0
Γ , A, Γ −→ ∆ ; Γ , B, Γ −→ ∆
(∪ →)
Γ0 , (A ∪ B), Γ −→ ∆0
0 0 0 0
v ∗ (Γ , A, Γ −→ ∆ ) ∩ v ∗ (Γ , B, Γ −→ ∆ )
0 0 0 0
= (v ∗ (Γ ) ∩ v ∗ (A) ∩ v ∗ (Γ)) ⇒ v ∗ (∆ )) ∩ (v ∗ (Γ ) ∩ v ∗ (B) ∩ v ∗ (Γ)) ⇒ v ∗ (∆ ))
[we use: ((A ⇒ C) ∩ (B ⇒ C)) ≡ ((A ∪ B) ⇒ C])
0 0 0
= (v ∗ (Γ ) ∩ v ∗ (A) ∩ v ∗ (Γ)) ∪ (v ∗ (Γ ) ∩ v ∗ (B) ∩ v ∗ (Γ)) ⇒ v ∗ (∆ )
0 0 0
= ((v ∗ (Γ ) ∩ v ∗ (Γ)) ∩ v ∗ (A)) ∪ ((v ∗ (Γ ) ∩ v ∗ (Γ)) ∩ v ∗ (B)) ⇒ v ∗ (∆ )
[we use commutativity and distributivity]
0 0
= ((v ∗ (Γ ) ∩ v ∗ (Γ)) ∩ (v ∗ (A ∪ B)) ⇒ v ∗ (∆ )
0 0
= v ∗ (Γ , (A ∪ B), Γ −→ ∆ )
0 0
Γ , A, Γ −→ ∆, ∆
(→ ¬)
Γ0 , Γ −→ ∆, ¬A, ∆0
0 0 0 0
v ∗ (Γ , A, Γ −→ ∆, ∆ ) = v ∗ (Γ ) ∩ v ∗ (A) ∩ v ∗ (Γ) ⇒ v ∗ (∆) ∪ v ∗ (∆ )
0 0
= (v ∗ (Γ ) ∩ v ∗ (Γ)) ∩ v ∗ (A) ⇒ v ∗ (∆) ∪ v ∗ (∆ )
[we use: ((A ∩ B) ⇒ C) ≡ (A ⇒ (¬B ∪ C))]
0 0 0
= (v ∗ (Γ ) ∩ v ∗ (Γ)) ⇒ ¬v ∗ (A) ∪ v ∗ (∆) ∪ v ∗ (∆ ) = (v ∗ (Γ ) ∩ v ∗ (Γ)) ⇒ v ∗ (∆) ∪
0
v ∗ (¬A) ∪ v ∗ (∆ )
0 0
= v ∗ (Γ , Γ −→ ∆, ¬A, ∆ )
The above shows the premises and conclusions are logically equivalent, therefore
the rules of inference are strongly sound. It ends the proof.
Observe that the strong soundness implies soundness (not only by name!), hence
we have also proved the following
We know by theorem 6.8 that all the rules of inference of GL of are strongly
sound. The strong soundness of the rules means that if at least one of premisses
of a rule is false, so is its conclusion. Hence given a sequent Γ −→ ∆ ∈ SQ, such
288
that its decomposition tree TΓ−→∆ has a branch ending with a non-axiom leaf.
It means that any truth assignment v that make this non-axiom leaf false also
falsifies all sequences on that branch, and hence falsifies the sequent Γ −→ ∆. In
particular, given a sequent −→ A and its tree T−→A , any v, such that falsifies its
a non-axiom leaf is a counter-model for A. We have hence proved the following.
Exercise 6.17
Prove, by constructing a counter-model determined by decomposition tree that
Solution
We construct the decomposition tree for the formula A : ((b ⇒ a) ⇒ (¬b ⇒ a))
as follows.
T→A
| (→⇒)
(b ⇒ a) −→ (¬b ⇒ a)
| (→⇒)
¬b, (b ⇒ a) −→ a
| (¬ →)
(b ⇒ a) −→ b, a
^
(⇒−→)
−→ b, b, a a −→ b, a
non − axiom axiom
289
The non-axiom leaf LA we want to falsify is −→ b, b, a. Let v : V AR −→ {T, F }
be a truth assignment. By definition 6.7 of semantic for GL we have that
v ∗ (LA ) = v ∗ (−→ b, b, a) = (T ⇒ v(b) ∪ v(b) ∪ v(a)). Hence v ∗ (−→ b, b, a) = F
if and only if (T ⇒ v(b) ∪ v(b) ∪ v(a)) = F if and only if v(b) = v(a) = F .
The Theorem 6.10, says that the logical value F determined by the evaluation
a non-axiom leaf LA ”climbs” the decomposition tree. We picture it as follows.
T→A
| (→⇒)
(b ⇒ a) −→ (¬b ⇒ a) F
| (→⇒)
¬b, (b ⇒ a) −→ a F
| (¬ →)
(b ⇒ a) −→ b, a F
^
(⇒−→)
−→ b, b, a F a −→ b, a
non − axiom axiom
Our main goal is to prove the Completeness Theorem for RS. We prove first
the Completeness Theorem for formulas A ∈ F and then we generalize it to any
sequences Γ ∈ F ∗ .
290
Proof
We have already proved the Soundness Theorem 6.9, so we need to prove only
the completeness part of it, namely to prove the implication:
if a appears in Γ0
T
v(a) = F if a appears in ∆0
any value if a does not appear in Γ0 → ∆0
The proof system G is in its structure the most similar to the proof system RS2
defined by (6.19).
It is obtained from in the same way is a proof system obtained from GL by
0 0
changing the indecomposable sequences Γ , ∆ into any sequences Σ, Λ ∈ F ∗
in all of the rules of inference of GL.
The logical axioms LA remain the same; i.e. the components of G are as
follows.
Axioms of G
As the axioms of GL we adopt any indecomposable sequent which contains a
positive literal a (variable) that appears on both sides of the sequent arrow −→,
i.e any sequent of the form
Γ0 1 , a, Γ0 2 −→ ∆0 1 , a, ∆0 2 , (6.28)
291
Inference Rules of G (6.29)
Conjunction rules
Σ, A, B, Γ −→ Λ
(∩ →) ,
Σ, (A ∩ B), Γ −→ Λ
Γ −→ ∆, A, Λ ; Γ −→ ∆, B, Λ
(→ ∩) ,
Γ −→ ∆, (A ∩ B), Λ
Disjunction rules
Γ −→ ∆, A, B, Λ
(→ ∪) ,
Γ −→ ∆, (A ∪ B), Λ
Σ, A, Γ −→ Λ ; Σ, B, Γ −→ Λ
(∪ →) ,
Σ, (A ∪ B), Γ −→ Λ
Implication rules
Σ, A, Γ −→ ∆, B, Λ
(→⇒) ,
Σ, Γ −→ ∆, (A ⇒ B), Λ
Σ, Γ −→ ∆, A, Λ ; Σ, B, Γ −→ ∆, Λ
(⇒→) ,
Σ, (A ⇒ B), Γ −→ ∆, Λ
Negation rules
Σ, Γ −→ ∆, A, Λ Σ, A, Γ −→ ∆, Λ
(¬ →) , (→ ¬) ,
Σ, ¬A, Γ −→ ∆, Λ Σ, Γ −→ ∆, ¬A, Λ
where Γ, ∆, Σ. Λ ∈ F ∗ .
Exercise 6.18 Follow the example of the GL system and adopt all needed
definitions and proofs to prove the completeness of the system G.
Solution
We leave it to the reader to fill in details .In particular, one has to accomplish
the steps below.
1. Explain why the system G is strongly sound. You can use the strong
soundness of the system GL .
2. Prove, as an example, a strong soundness of 4 rules of G.
3. Prove the the following Strong Soundness Theorem for G.
292
Theorem 6.12
The proof system G is strongly sound.
4. Define shortly, in your own words, for any formula A ∈ F, its decomposition
tree T→A in G.
5. Extend your definition to a decomposition tree TΓ→∆ .
6. Prove that for any Γ → ∆ ∈ SQ, the decomposition tree TΓ→∆ are finite.
7. Give an example of formulas A, B ∈ F such that that T→A is unigue and
T→B is not.
8. Prove the following Counter Model Theorem for G.
Theorem 6.13
Given a sequent Γ −→ ∆, such that its decomposition tree TΓ−→∆ contains a
non- axiom leaf LA . Any truth assignment v that falsifies the non-axiom leaf
LA is a counter model for Γ −→ ∆.
293
For the system LI he proved only the adequacy of LI system for intituitionistic
logic since the semantics for the intuitionistic logic didn’t yet exist. He used
the acceptance of the Heying intuitionistic axiom system as the definition of the
intuitionistic logic and proved that any formula provable in the Heyting system
is also provable in LI.
Observe that by presence of the cut rule, Gentzen LK, LI systems are also
a Hilbert system. What distinguishes it from all other known Hilbert proof
systems is the fact that the cut rule could be eliminated from it.
This is Gentzen famous Hauptzatz Theorem, also called Cut Elimination The-
orem. The elimination of the cut rule and the structure of other rules makes
it possible to define effective automatic procedures for proof search, what is
impossible in a case of the Hilbert style systems.
Gentzen, in his proof of Hauptzatz Theorem, developed a powerful technique
adaptable to other logics. We present it here in classical propositional case and
show how to adapt it to the intuitionistic case.
Gentzen proof is purely syntactical. It defines a constructive method of trans-
formation of any formal proof (derivation) of a sequent Γ −→ ∆ that uses a
cut rule (and other rules) into its proof without use of the cut rule. Hence the
English name Cut Elimination Theorem.
The completeness (with respect to algebraic semantics defined in chapter 7) of
the cut free system LI follows directly from LI Hauptzatz Theorem 6.22 and the
intuitionistic completeness theorem (chapter 7). The proof is a straightforward
adaptation of the proof of cut free LK Completeness Theorem 6.23 and is left
as a homework exercise in chapter 7.
Rasiowa and Sikorski method of proving completeness theorem by constructing
counter-models on the decomposition trees is a semantical equivalence to purely
syntactical Gentzen proof of cut elimination. It is relatively simple, powerful
and easy to understand. It was the reason it was first to be presented here. But
it is more difficult and sometimes impossible to apply (generalize) to many non-
classical logics then Gentzen cut elimination method. Moreover the Gentzen
method is more flexible and in this sense more general and powerful. This is
why we preset it here.
Language L
The language is the same as the in case of GL, namely
L = L{¬,∩,∪,⇒} .
Expressions
294
The set of all expressions E is, as before, the set
SQ = {Γ −→ ∆ : Γ, ∆ ∈ F ∗ } (6.31)
of all sequents.
Logical Axioms
There is only one logical axiom, namely
A −→ A,
Rules of Inference
There are two groups of rules of inference and they are defined are as follows.
Cut Rule
295
Γ −→ ∆, A ; A, Σ −→ Θ
(cut) .
Γ, Σ −→ ∆, Θ
The formula A is called a cut formula.
Conjunction
A, Γ −→ ∆
(∩ →)1 ,
(A ∩ B), Γ −→ ∆
B, Γ −→ ∆
(∩ →)2 ,
(A ∩ B), Γ −→ ∆
Γ −→ ∆, A ; Γ −→ ∆, B
(→ ∩) .
Γ −→ ∆, (A ∩ B)
Disjunction
Γ −→ ∆, A
(→ ∪)1 ,
Γ −→ ∆, (A ∪ B)
Γ −→ ∆, B
(→ ∪)2 ,
Γ −→ ∆, (A ∪ B)
A, Γ −→ ∆ ; B, Γ −→ ∆
(∪ →) .
(A ∪ B), Γ −→ ∆
Implication
A, Γ −→ ∆, B
(→⇒) ,
Γ −→ ∆, (A ⇒ B)
Γ −→ ∆, A ; B, Γ −→ ∆
(⇒→) .
(A ⇒ B), Γ −→ ∆
Negation
Γ −→ ∆, A
(¬ →) ,
¬A, Γ −→ ∆
A, Γ −→ ∆
(→ ¬) .
Γ −→ ∆, ¬A
296
Definition 6.10 (Intuitionistic System LI)
We define the intuitionistic Gentzen system LI as
LK = (L, ISQ, AL, I-Structural Rules, I- Cut Rule, I- Logical Rules),
where ISQ is the following subset of the set SQ of all sequents (6.31)
ISQ = {Γ −→ ∆ : ∆ consists of at most one formula }. (6.32)
The set ISQ is called the set of all intuitionistic sequents.
The I-Structural Rules, I- Cut Rule, I- Logical Rules are the LK rules restricted
to the set ISQ (6.32) of the intuitionistic sequents.
Classical System LK
The proofs are often called derivations. In particular, Gentzen, in his work
used the term derivation we will use this notion as well. This is why we denote
the proof trees by D (for derivation).
Finding derivations D in LK are is a more complex process, as the logical rules
are different, then in GL and G. Proofs rely strongly on use of the Structural
Rules. Even if we find a derivation that does not involve the Cut rule, the
Structural rules are usually present. For example, a derivation of Excluded
Middle (A ∪ ¬A) formula B in LK is as follows.
297
D
−→ (A ∪ ¬A)
| (→ contr)
−→ (A ∪ ¬A), (A ∪ ¬A)
| (→ ∪)1
−→ (A ∪ ¬A), A
| (→ exch)
−→ A, (A ∪ ¬A)
| (→ ∪)1
−→ A, ¬A
| (→ ¬)
A −→ A
axiom
Here is as yet another example a proof P ( also cut free) of the de Morgan Law
(¬(A ∩ B) ⇒ (¬A ∪ ¬B)).
P
−→ (¬(A ∩ B) ⇒ (¬A ∪ ¬B))
| (→⇒)
(¬(A ∩ B) −→ (¬A ∪ ¬B))
| (→ ¬)
−→ (¬A ∪ ¬B), (A ∩ B)
^
(⇒−→)
| (→ exch) | (→ exch)
−→ A, (¬A ∪ ¬B) −→ B, (¬A ∪ ¬B)
| (→ ∪)1 | (→ ∪)1
−→ A, ¬A −→ B, ¬B
| (→ ¬) B −→ B
A −→ A axiom
axiom
298
Observe that the Logical Rules are similar in their structure to the rules of the
system G and hence admit the same proof of their soundness.
The rules (→ ∪)1 , (→ ∪)2 and (→ ∪)1 , (→ ∪)2 are not strongly sound as
A 6≡ (A ∩ B), B 6≡ (A ∩ B) and A 6≡ (A ∩ B), B 6≡ (A ∩ B).
All other Logical Rules are strongly sound.
The Contraction and Exchange structural are also strongly sound as for any
formulas A, B ∈ F, A ≡ (A∩A), A ≡ (A∪A) and (A∩B) ≡ (B ∩A), (A∩B) ≡
(B ∩ A). The Weakening rule is sound because (we use shorthand notation) if
a formula (Γ ⇒ ∆) = T then also ((A ∩ Γ) ⇒ ∆)) = T for any logical value of
the formula A. But obviously (Γ ⇒ ∆) 6≡ ((A ∩ Γ) ⇒ ∆)), i.e. the Weakening
rule is not strongly sound.
The Cut rule is sound as the fact (Γ ⇒ (∆ ∪ A)) = T and ((A ∩ Σ) ⇒ Λ) = T
implies that Γ, Σ −→ ∆, Λ. It is not strongly sound. Any truth assignment
such that Γ = T, ∆ = Σ = Λ = A = F proves that (Γ −→ ∆, A) ∩ (A, Σ −→
Λ) 6≡ (Γ, Σ −→ ∆, Λ). Obviously, |= A −→ A.
We have proved that LK is sound and hence the following theorem holds.
if `LK Γ −→ ∆, then |= Γ −→ ∆.
if `LK A, then |= A.
299
Axioms of R (6.34)
The set of logical axioms of the Hilbert style proof system RS for classical
propositional logic all formulas of the forms
A1 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A2 (A ⇒ (A ∪ B)),
A3 (B ⇒ (A ∪ B)),
A4 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A5 ((A ∩ B) ⇒ A),
A6 ((A ∩ B) ⇒ B),
A7 ((C ⇒ A) ⇒ ((C ⇒ B) ⇒ (C ⇒ (A ∩ B))),
A8 ((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C)),
A9 (((A ∩ B) ⇒ C) ⇒ (A ⇒ (B ⇒ C)),
A10 (A ∩ ¬A) ⇒ B),
A11 ((A ⇒ (A ∩ ¬A)) ⇒ ¬A),
A12 (A ∪ ¬A),
where A, B, C ∈ F are any formulas in L = L{¬,∩,∪,⇒} .
We adopt a Modus Ponens
A ; (A ⇒ B)
(M P )
B
as the only inference rule.
We define Hilbert System R as
Theorem 6.16
For any formula A ∈ F,
`R A if and only if |= A.
We leave it as an exercise for the reader to show that all axioms A1 - A12 of the
system R are provable in LK. Moreover, the Modus Ponens is a particular case
300
of the cut rule, for Γ, ∆, Σ empty sequences and Θ containing only one element,
a formula B. We call it also MP rule.
−→ A ; A −→ B
(M P ) .
−→ B
This proves the following.
Theorem 6.17
For any formula A ∈ F,
if `R A, then `LK A.
Directly from the above theorem 6.17, soundness of LK (theorem 6.15) and
completeness of R (theorem 6.16) we get the completeness of LK.
The proof is quite long and involved. We present here its main and most im-
portant steps. To facilitate the proof we introduce a more general form of the
cut rule, called a mix rule defined as follows.
Γ −→ ∆ ; Σ −→ Θ
(mix) , (6.36)
Γ, Σ∗ −→ ∆∗ , Θ
Example 6.3
301
Here are some examples of an applications of the mix rule. Observe that the
mix rule applies, as the cut does, to only one mix formula at the time.
a −→ b, ¬a ; (b ∪ c), b, b, D, b −→
(mix)
a, (b ∪ c), D −→ ¬a
Notice, that every derivation with cut may be transformed into a derivation with
mix by means of a number of weakenings (multiple application of the weakening
rules) and interchanges (multiple application of the exchange rules). Conversely,
every mix may be transformed into a cut derivation by means of a certain num-
ber of preceding exchanges and contractions, though we do not use this fact in
the proof. Observe that cut is a particular case of mix.
Definition 6.11
We say that a derivation DΓ−→∆ of a sequent Γ −→ ∆ has a Property H if
it satisfies the the following conditions.
1. The root Γ −→ ∆ of the derivation DΓ−→∆ is obtained by direct use of
the mix rule, i.e. the mix rule is the last rule of inference used in the proof
(derivation) of Γ −→ ∆.
2. The derivation DΓ−→∆ does not contain any other application of the mix rule,
i.e. the proof (derivation) of Γ −→ ∆ does not contain any other application of
the mix rule.
302
Lemma 6.2 (H lemma)
Any derivation that fulfills the Property H (definition 6.11) may be trans-
formed into a derivation of the same sequent) in which no mix occurs.
Step 3: we use the H lemma 6.2 and to prove the the Hauptzatz as follows.
Hauptzatz proof from H lemma
Let D be any derivation (tree proof). Let Γ −→ ∆ be any node on D such that
its sub-tree DΓ−→∆ has the PropertyH (definition 6.11). By H lemma 6.2 the
sub-tree DΓ−→∆ can be replaced by a tree D∗ Γ−→∆ in which no mix occurs.
The rest of D remains unchanged. We repeat this procedure for each node N,
such that the sub-tree DN has the Property H until every application of mix
rule has systematically been eliminated. This ends the proof of Hauptzatz pro-
vided the H lemma 6.2 has already been proved.
We define now two important notions: degree n and rank r of the derivation
D. Observe that D contains only one application of mix rule, and the mix rule,
contains only one mix formula A. Mix rule used may contain many copies of
A, but there always is only one mix formula. We call is a mix formula of D.
Definition 6.12
Given a derivation tree D with the Property H.
Let A ∈ F be the mix formula of D. The degree n ≥ 0 of A is called the degree
of the derivation D. We write it as degD = degA = n.
Definition 6.13
Given a derivation tree D with the Property H. We define the rank r of D as a
sum of its left rank Lr and right rank Rr of D, i.e.
r = Lr + Rr,
where:
1. the left rank Lr of D in the largest number of consecutive nodes on the branch
of D staring with the node containing the left premiss of the mix rule, such that
each sequent on these nodes contains the mix formula in the succedent;
2. the right rank Rr of D in the largest number of consecutive nodes on the
branch of D staring with the node containing the right premiss of the mix
303
rule, such that each sequent on these nodes contains the mix formula in the
antecedent.
We present some cases and leave similar others to the reader as an exercise.
Observe that first group contains cases that are especially simple in that they
allow the mix to be immediately eliminated. The second group contains the
most important cases since their consideration brings out the basic idea behind
the whole proof, Here we use the induction hypothesis with respect do the degree
of the derivation. We reduce each one of the cases to transformed derivations
of a lower degree.
A, Σ∗ −→ ∆
^
(mix)
A −→ A Σ −→ ∆
A, Σ∗ −→ ∆
304
possibly several exchanges and contractions
Σ −→ ∆
Σ −→ ∆∗ , A
^
(mix)
Σ −→ ∆ A −→ A
Σ −→ ∆∗ , A
Suppose that neither of premisses of mix is an axiom. As the rank r=2 , the
right and left ranks are equal one. This means that in the sequents on the nodes
directly below left premiss of the mix, the mix formula A does not occur in the
succedent; in the sequents on the nodes directly below right premiss of the mix,
the mix formula A does not occur in the antecedent.
In general, if a formula occurs in the antecedent (succedent) of a conclusion of
a rule of inference, it is either obtained by a logical rule or by a contraction rule.
3. The left premiss of the mix rule is the conclusion of a contraction rule
(→ contr). The sub-tree of D containing mix is:
Γ, Σ∗ −→ ∆, Θ
305
^
(mix)
Γ −→ ∆, A Σ −→ Θ
| (→ contr)
Γ −→ ∆
Γ, Σ∗ −→ ∆, Θ
Observe that the whole branch of D that starts with the node Σ −→ Θ disap-
pears. Such obtained D∗ proves the same sequent and contains no mix.
4. The right premiss of the mix rule is the conclusion of a contraction rule
(→ contr). It is a dual case to 3. and is left to the reader.
1. The main connective of the mix formula is ∩, i.e. the mix formula is (A ∩ B).
The left premiss of the mix rule is the conclusion of a rule (→ ∩). The right
premiss of the mix rule is the conclusion of a rule (∩ →)1 .
The sub-tree T of D containing mix is:
Γ, Σ −→ ∆, Θ
^
(mix)
Γ −→ ∆, (A ∩ B) (A ∩ B), Σ −→ Θ
^
((→ ∩)) | (∩ →)1
A, Σ −→ Θ
Γ −→ ∆, A Γ −→ ∆, B
306
Γ, Σ −→ ∆, Θ
Γ −→ ∆, A A, Σ −→ Θ
2. The case when the left premiss of the mix rule is the conclusion of a rule
(→ ∩) and right premiss of the mix rule is the conclusion of a rule (∩ →)2
3. The main connective of the mix formula is ∪, i.e. the mix formula is (A ∪ B).
The left premiss of the mix rule is the conclusion of a rule (→ ∪)1 or (→ ∪)2 .
The right premiss of the mix rule is the conclusion of a rule (∪ →)1 . This is to
be dealt with symmetrically to the ∩ cases.
4. The main connective of the mix formula is ¬, i.e. the mix formula is ¬A.
The left premiss of the mix rule is the conclusion of a rule (→ ¬). The right
premiss of the mix rule is the conclusion of a rule (¬ →).
Γ, Σ −→ ∆, Θ
^
(mix)
Γ −→ ∆, ¬A ¬A, Σ −→ Θ
| (→ ¬) | (¬ →)
A, Γ −→ ∆ Σ −→ Θ, A
Γ, Σ −→ ∆, Θ
307
Σ, Γ∗ −→ Θ∗ , ∆
^
(mix)
Σ −→ Θ, A A, Γ −→ ∆
5. The main connective of the mix formula is ⇒, i.e. the mix formula is
(A ⇒ B). The left premiss of the mix rule is the conclusion of a rule ((→⇒).
The right premiss of the mix rule is the conclusion of a rule (⇒→).
Γ, Σ −→ ∆, Θ
^
(mix)
Γ −→ ∆, (A ⇒ B) (A ⇒ B), Σ −→ Θ
^
| (→⇒) ((→ ∩))
A, Γ −→ ∆, B
Σ −→ Θ, A B, Σ −→ Θ,
Γ, Σ −→ ∆, Θ
Σ −→ Θ, A A, Γ, Σ∗ , −→ ∆∗ , Θ
^
(mix)
A, Γ −→ ∆, B B, Σ −→ Θ,
308
The asteriks are, of course, intended as follows: Σ∗ , ∆∗ results from Σ, ∆ by
the omission of all formulas B; Γ∗ , Σ∗∗ , Θ∗ results from Γ, Σ∗ , Θ by the omission
of all formulas A.
We replace T by T∗ in D and obtain D∗ . Now we have two mixes, but both
mix formulas are of a lower degree then n. We first apply the inductive to the
assumption to the lower mix. Thus it can be eliminated. We can then also
eliminate the upper mix. This ends the proof of the case of rank r=2.
Case r > 2.
In the case r = 2, we generally reduced the derivation to one of lower degree.
Now we shall proceed to reduce the derivation to one of the same degree, but
of a lower rank. This allows us to to be able to carry the induction with respect
to the rank r of the derivation.
We use the inductive assuption in all cases except, as before, a case of an axiom
or structural rules. In these cases the mix can be eliminated immediately, as it
was eliminated in the previous case of rank r = 2.
In a case of logical rules we obtain the reduction of the mix of the lemma to
derivations with mix of a lower ranks which consequently can be eleminated by
the inductive assumption. We carry now proofs for two logical rules: (→ ∩)
and (∪ →. The proof for all other rules is similar and is left to the reader.
Also, we consider a case of left rank Lr= 1 and the right rank Rr = r ¿1. The
symmetrical case left rank Lr = r ¿1 1 and the right rank Rr = 1 is left to the
reader as an exercise.
Θ, Γ∗ −→ Σ∗ , ∆, (A ∩ B)
^
(mix)
Θ −→ Σ Γ −→ ∆, (A ∩ B)
^
(→ ∩)
Γ −→ ∆, A Γ −→ ∆, B
309
Θ, Γ∗ −→ Σ∗ , ∆, (A ∩ B)
^
(→ ∩)
Θ, Γ∗ −→ Σ∗ , ∆, A Θ, Γ∗ −→ Σ∗ , ∆, B
^ ^
(mix) (mix)
Θ −→ Σ Γ −→ ∆, A Θ −→ Σ Γ −→ ∆, A
Θ, (A ∪ B)∗ , Γ∗ −→ Σ∗ , ∆, (A ∩ B)
^
(mix)
Θ −→ Σ (A ∪ B)Γ −→ ∆
^
(∪ →)
A, Γ −→ ∆ B, Γ −→ ∆
∗
(A ∪ B) stands either for or for nothing according as (A ∪ B) is unequal or
equal to the mix formula M . The mix formula M certainly occurs in Γ. For
otherwise M would been equal to (A ∪ B) and the right rank Rr would be equal
to 1 contrary to the assumption.
We transform T into T∗ as follows.
Θ, (A ∪ B), Γ∗ −→ Σ∗ , ∆, (A ∩ B)
^
(∪ →)
310
A, Θ, Γ∗ −→ Σ∗ , ∆ B, Θ, Γ∗ −→ Σ∗ , ∆
some weakenings, exchanges some weakenings, exchanges
∗ ∗ ∗
Θ, A , Γ −→ Σ , ∆ Θ, B ∗ , Γ∗ −→ Σ∗ , ∆
^ ^
(mix) (mix)
Θ −→ Σ A, Γ −→ ∆ Θ −→ Σ B, Γ −→ ∆
Now we have two mixes, but both have the right rank Rr = r-1 and both of
them can be eliminated by the inductive assumption. We replace T by T∗ in
D and obtain D∗ . This ends the proof of the Hauptzatz lemma and hence the
proof of the Hauptzatz Theorem 6.19 and Hauptzatz Theorem 6.20.
Let’s denote by LK - c and LI - c the systems LK, LI without the cut rule,
i.e. we put
LK − c = LK − {(cut)}. (6.37)
LI − c = LI − {(cut)}. (6.38)
This is why the cut-free Gentzen systems LK-c and LI -c are just called Gentzen
LK, LI, respectively.
Directly from the Completeness Theorem 6.18 and the Hauptzatz Theorem 6.19
we get that the following.
311
Let G be the Gentzen sequents proof system defined by (6.27). We replace the
logical axiom of G
Γ0 1 , a, Γ0 2 −→ ∆0 1 , a, ∆0 2 ,
where a ∈ V AR is any propositional variable and Γ0 1 , Γ0 2 , ∆0 1 , ∆0 2 ∈ V AR∗
are any indecomposable sequences, by a new logical axiom
Γ1 , A, Γ2 −→ ∆1 , A, ∆2 (6.39)
Theorem 6.24
For any formula A ∈ F,
312
4. Extend your definition T→A in G to a decomposition tree TΓ→∆ .
5. Prove that for any Γ → ∆ ∈ SQ, the decomposition tree TΓ→∆ in G are
finite.
8. Prove strong soundness of rules (→ ∪), (→⇒) in GL. List all logical equiv-
alences used in the proofs.
9. Prove strong soundness of rules (⇒→), (¬ →) in GL. List all logical equiv-
alences used in the proofs.
10. Prove strong soundness of rules (∪ →), (→ ¬), (∩ →) in G. List all logical
equivalences used in the proofs.
11. Prove strong soundness of rules (⇒→), (→ ∪), (⇒→) in G. List all logical
equivalences used in the proofs.
313
(ii) Find proofs of axioms A3, A7, and A11 of the R system (6.34) in LK-c
and in GK, i.e. proofs of formulas (B ⇒ (A ∪ B)), ((C ⇒ A) ⇒ ((C ⇒
B) ⇒ (C ⇒ (A ∩ B))), and ((A ⇒ (A ∩ ¬A)) ⇒ ¬A). Compare your
results.
(iii) Find proofs of axioms A1, A8, and A9 of the R system (6.34) in LK-
c and in GK, i.e. proofs of formulas ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒
C))), ((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C)), and (((A ∩ B) ⇒ C) ⇒ (A ⇒
(B ⇒ C)). Compare your results.
(iv) Find proofs of axioms A1, A5, and A12 of the R system (6.34) in
LK-c and in GK, i.e. proofs of formulas ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒
C))), ((A ∩ B) ⇒ A), and (A ∪ ¬A) . Compare your results.
18. Re- write carefully the proof of the classical Hauptzatz Theorem 6.19 for
the case o the intuitionistic system LI (definition 6.10.
19. Define shortly, in your own words, for any formula A ∈ F, its decomposi-
tion tree TA in LK-c. Is the tree TA always finite?
20. Given a formula A = (¬(a ∩ b) ⇒ (¬a ∪ ¬b))). Construct one infinite and
one infinite decomposition tree for A.
21. Describe major differences in the decomposition trees in LK-c and GK.
22. We have proved that LK-c and GK are equivalent, i.e. that for any
sequent Γ −→ ∆,
The proof was not constructive; it was obtained from that fact that both
systems are complete.
(ii) Describe a constructive procedure of transforming any proof in GK
into a proof in LK-c.
(i) Transform a proof of a formula (A ⇒ (A ∪ B)) in GK into a proof in
LK-c.
(ii) Describe a constructive procedure of transforming any proof in GK
into a proof in LK-c.
(iii) Show that the procedure of elimination of structural rules of LK-c
leads to the rules inference of GK .
314
Chapter 7
Introduction to
Intuitionistic and Modal
Logics
315
that establish the relationship between classical and intuitionistic logics.
Intuitionists’ view-point on the meaning of the basic logical and set theoretical
concepts used in mathematics is different from that of most mathematicians in
their research.
The basic difference lies in the interpretation of the word exists. For exam-
ple, let A(x) be a statement in the arithmetic of natural numbers. For the
mathematicians the sentence
∃xA(x) (7.1)
The intuitionistic semantics I has to be such that one can prove in that also
316
The above means also that intuitionists interpret differently the meaning of
propositional connectives.
Intuitionistic implication
The intuitionistic implication (A ⇒ B) is considered by to be true if there exists
a method by which a proof of B can be deduced from the proof of A. In the
case of the implication
(¬∀x ¬A(x)) ⇒ ∃xA(x))
there is no general method which, from a proof of the sentence (¬∀x ¬A(x)),
permits is to obtain an intuitionistic proof of the sentence ∃xA(x), i.e. to
construct a number n such that A(n) holds, hence we can’t accept it as an
intuitionistic theorem or tautology.
Intuitionistic negation
The negation and the disjunction are also understood differently. The sentence
¬A is considered intuitionistically true if the acceptance of the sentence A leads
to absurdity.
As a result of above understanding of negation and implication we have that in
the intuitionistic logic I
`I (A ⇒ ¬¬A)
but
6 `I (¬¬A ⇒ A).
Intuitionistic disjunction
The intuitionist regards a disjunction (A ∪ B) as true if one of the sentences
A, B is true and there is a method by which it is possible to find out which of
them is true. As a consequence a classical law of excluded middle
(A ∪ ¬A)
is not acceptable by the intuitionists since there is no general method of finding
out, for any given sentence A, whether A or ¬A is true. This means that the
intuitionistic logic must be such that
6 `I (A ∪ ¬A)
and the intuitionistic semantics I has to be such that
6|=I (A ∪ ¬A).
317
Intuitionists’ view of the concept of infinite set also differs from that which is
generally accepted in mathematics. Intuitionists reject the idea of infinite set as
a closed whole. They look upon an infinite set as something which is constantly
in a state of formation. Thus, for example, the set of all natural numbers is
infinite in the sense that to any given finite set of natural numbers it is always
possible to add one more natural number. The notion of the set of all subsets
of the set of all natural numbers is not regarded meaningful. Thus intuitionists
reject the general idea of a set as defined by a modern set theory.
An exact exposition of the basic ideas of intuitionism is outside the range of our
investigations. Our goal is to give a presentation of of the intuitionistic logic,
which is a sort of reflection of intuitionistic ideas formulated as a proof system.
There are many proof systems describing the intuitionistic logic. We define now
a system I with a set of axioms that is due to Rasiowa (1959). We adopted
this axiomatization for two reasons. Firs is that it is the most natural and
appropriate set of axioms to carry the the algebraic proof of the completeness
theorem and the second is that they visibly describe the main difference between
intuitionistic and classical logic. Namely, by adding the only one more axiom
(A ∪ ¬A) we get a (complete) formalization for classical logic. Here are the
components if the proof system I.
Language We adopt a propositional language L = L{∪,∩,⇒,¬} with the set of
formulas denoted by F.
Axioms
A1 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A2 (A ⇒ (A ∪ B)),
A3 (B ⇒ (A ∪ B)),
A4 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A5 ((A ∩ B) ⇒ A),
A6 ((A ∩ B) ⇒ B),
A7 ((C ⇒ A) ⇒ ((C ⇒ B) ⇒ (C ⇒ (A ∩ B))),
A8 ((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C)),
A9 (((A ∩ B) ⇒ C) ⇒ (A ⇒ (B ⇒ C)),
A10 (A ∩ ¬A) ⇒ B),
A11 ((A ⇒ (A ∩ ¬A)) ⇒ ¬A),
318
where A, B, C are any formulas in L.
Rules of inference
We adopt the Modus Ponens rule
A ; (A ⇒ B)
(M P )
B
A proof system
I = ( L, F A1 − A11, (M P ) ), (7.4)
A1 − A11 defined above, is called a Hilbert style formalization for Intuitionistic
propositional logic.
We introduce, as usual, the notion of a formal proof in I and denote by
`I A
The equation (7.5) can serve as the definition of the relative pseudo-complement
a ⇒ b.
Fact 7.1 Every relatively pseudo-complemented lattice (B, ∩, ∪) has the greatest
element, called a unit element and denoted by 1.
319
Proof Observe that a∩x ≤ a for all x, a ∈ B. By (7.5) we have that x ≤ a ⇒ a
for all x ∈ B, i.e. a ⇒ a = 1.
An abstract algebra
B = (B, 1, ⇒, ∩, ∪, ⇒) (7.6)
is said to be a relatively pseudo-complemented lattice if (B, ∩, ∪) is rel-
atively pseudo-complemented lattice with the relative pseudo-complement ⇒
defined by (7.5) and with the unit element 1 (Fact 7.1 ).
Relatively Pseudo-complemented Set Lattices
Consider a topological space X with an interior operation I. Let G(X) be the
class of all open subsets of X and G ∗ (X) be the class of all both dense and open
subsets of X. Then the algebras
¬a = a ⇒ 0 (7.8)
The pseudo - Boolean algebras are also called Heyting algebras to stress their
connection to the intuitionistic logic.
Let X be topological space with an interior operation I. Let G(X) be the class
of all open subsets of X. Then
Y ⇒ Z = I(X − Y ) ∪ Z
320
and ¬ is defined as
B = (B, 1, 0, ⇒, ∩, ∪, ¬),
¬a = a ⇒ 0.
321
Algebraic Models
We say that a formula A is an intuitionistoc tautology if and only if any
pseudo-Boolean algebra (7.7) is a model for A. This kind of models because
their connection to abstract algebras are called algebraic models.
We put it formally as follows.
B |= A
Definition 7.3
A formula A is valid in a pseudo-Boolean algebra B = (B, 1, 0, ⇒, ∩, ∪, ¬),
if and only if v ∗ (A) = 1 holds for all variables assignments v : V AR −→ B.
Directly from definition 7.2 and definition 7.3 we get the following.
Fact 7.2 For any formula A, |=I A if and only if A is valid in all pseudo-
Boolean algebras.
We write now `I A to denote any proof system for the Intuitionistic proposi-
tional logic, and in particular the Hilbert style formalization for Intuitionistic
propositional logic I defined by (7.4).
322
The intuitionistic completeness theorem 7.2 follows also directly from the fol-
lowing general algebraic completeness theorem 7.3 that combines results of of
Mostowski (1958), Rasiowa (1951) and Rasiowa-Sikorski (1957).
(G(X), X, ∅, ∪, ∩, ⇒, ¬)
The following theorem follows from the equivalence of conditions (i) and (iv).
The following classical tautologies are provable in I and hence are also intu-
itionistic tautologies.
(A ⇒ A), (7.12)
(A ⇒ (B ⇒ A)), (7.13)
(A ⇒ (B ⇒ (A ∩ B))), (7.14)
((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))), (7.15)
(A ⇒ ¬¬A), (7.16)
¬(A ∩ ¬A), (7.17)
((¬A ∪ B) ⇒ (A ⇒ B)), (7.18)
(¬(A ∪ B) ⇒ (¬A ∩ ¬B)), (7.19)
323
((¬A ∩ ¬B) ⇒ (¬(A ∪ B)), (7.20)
((¬A ∪ ¬B) ⇒ (¬A ∩ ¬B)), (7.21)
((A ⇒ B) ⇒ (¬B ⇒ ¬A)), (7.22)
((A ⇒ ¬B) ⇒ (B ⇒ ¬A)), (7.23)
(¬¬¬A ⇒ ¬A), (7.24)
(¬A ⇒ ¬¬¬A), (7.25)
(¬¬(A ⇒ B) ⇒ (A ⇒ ¬¬B)), (7.26)
((C ⇒ A) ⇒ ((C ⇒ (A ⇒ B)) ⇒ (C ⇒ B)), (7.27)
(A ∪ ¬A), (7.28)
(¬¬A ⇒ A), (7.29)
((A ⇒ B) ⇒ (¬A ∪ B)), (7.30)
(¬(A ∩ B) ⇒ (¬A ∪ ¬B)), (7.31)
((¬A ⇒ B) ⇒ (¬B ⇒ A)), (7.32)
((¬A ⇒ ¬B) ⇒ (B ⇒ A)), (7.33)
((A ⇒ B) ⇒ A) ⇒ A), (7.34)
The intuitionistic logic has been created as a rival to the classical one. So a
question about the relationship between these two is a natural one. We present
here some examples of tautologies and some historic results about the connection
between the classical and intuitionistic logic.
The first connection is quite obvious. It was proved by Rasiowa and Sikorski in
1964 that by adding the axiom
A12 (A ∪ ¬A)
to the set of axioms of our system I defined by (7.4) we obtain a Hilbert proof
system H that is complete with respect to classical semantics.
This proves the following.
324
Theorem 7.5
Every formula that is derivable intuitionistically is classically derivable, i.e.
if `I A, then ` A,
where we use symbol ` for classical (complete classical proof system) provability.
We write
|= A and ] |=I A
to denote that A is a classical and intuitionistic tautology, respectively.
As both proof systems, I and H are complete under respective semantics, we
can state this as the following relationship between classical and intuitionistic
tautologies.
Theorem 7.6
For any formula A ∈ F,
if |=I A, then |= A.
The next relationship shows how to obtain intuitionistic tautologies from the
classical tautologies and vice versa. It has been proved by Glivenko in 1929 and
independently by Tarski in 1938.
325
Theorem 7.9 (Gödel, 1931)
For any A, B ∈ F, a formula (A ⇒ ¬B) is a classically provable if and only if
it is an intuitionistically provable, i.e.
Theorem 7.11
A formula (A ⇒ ¬B) is a classical tautology if and only if it is an intuitionistic
tautology, i.e.
Theorem 7.12
If a formula A contains no connectives except ∩ and ¬, then A is a classical
tautology if and only if it is an intuitionistic tautology.
This fact was stated without the proof by Gödel in 1931 and proved by Gentzen
in 1935 via his proof system LI which is presented in chapter 6 and discussed
in the next section 7.2.
`I (A ∪ B) if and only if `I A or `I B.
326
We obtain, via the Completeness Theorem 7.2 the following equivalent semantic
version of the above.
Theorem 7.14
A disjunction (A ∪ B) is intuitionistic tautology if and only if either A or B is
intuitionistic tautology, i.e.
L = L{∪,∩,⇒,¬} (7.35)
The set ISQ is called the set of all intuitionistic sequents; the LI sequents.
327
Axioms of LI
As the axioms of LI we adopt any sequent from the set ISQ defined by ( 7.36),
which contains a formula that appears on both sides of the sequent arrow −→,
i.e any sequent of the form
Γ1 , A, Γ2 −→ A, (7.37)
for any formula A ∈ F of the language (7.35) and for any sequences Γ1 , Γ2 ∈ F ∗ .
Inference rules of LI
The set inference rules is divided into two groups: the structural rules and the
logical rules. They are defined as follows.
Structural Rules of LI
Weakening
Γ −→
(→ weak) .
Γ −→ A
A is called the weakening formula.
Contraction
A, A, Γ −→ ∆
(contr →) ,
A, Γ −→ ∆
A is called the contraction formula , ∆ contains at most one formula.
Exchange
Γ1 , A, B, Γ2 −→ ∆
(exchange →) ,
Γ1 , B, A, Γ2 −→ ∆
A, B, Γ −→ ∆ Γ −→ A ; Γ −→ B
(∩ →) , (→ ∩) ,
(A ∩ B), Γ −→ ∆ Γ −→ (A ∩ B)
Γ −→ A Γ −→ B
(→ ∪)1 , (→ ∪)2 ,
Γ −→ (A ∪ B) Γ −→ (A ∪ B)
328
A, Γ −→ ∆ ; B, Γ −→ ∆
(∪ →) ,
(A ∪ B), Γ −→ ∆
A, Γ −→ B Γ −→ A ; B, Γ −→ ∆
(→⇒) , (⇒→) ,
Γ −→ (A ⇒ B) (A ⇒ B), Γ −→ ∆
Γ −→ A A, Γ −→
(¬ →) , (→ ¬) .
¬A, Γ −→ Γ −→ ¬A
Formally we define:
where ISQ is defined by (7.36), Structural rules and Logical rules are the infer-
ence rules defined above, and LA is the axiom defined by the schema (7.37).
We write
`LI Γ −→ ∆
to denote that the sequent Γ −→ ∆ has a proof in LI.
We say that a formula A ∈ F has a proof in LI and write it as
`LI A
329
In particular, for any formula A,
The particular form the theorem 7.16 was stated without the proof by Gödel in
1931. The theorem proved by Gentzen in 1935 via his Hauptzatz Theorem.
Proof
Assume `LI (A ∪ B). This equivalent to `LI −→ (A ∪ B). The last step in
the proof of −→ (A ∪ B)i LI must be the application of the rule (→ ∪)1 to
the sequent −→ A, or the application of the rule (→ ∪)2 to the sequent −→ B.
There is no other possibilities. We have proved that `LI (A ∪ B) implies `LI A
or `LI B. The inverse is obvious by respective applications of rules (→ ∪)1
(→ ∪)2 to −→ A and −→ B.
Search for proofs in LI is a much more complicated process then the one in
classical systems RS or GL defined in chapter 6.
Here, as in any other Gentzen style proof system, proof search procedure consists
of building the decomposition trees.
In RS the decomposition tree TA of any formula A, and hence of any sequence
Γ is always unique.
In GL the ”blind search” defines, for any formula A a finite number of decom-
position trees, but it can be proved that the search can be reduced to examining
only one of them, due to the absence of structural rules.
In LI the structural rules play a vital role in the proof construction and hence,
in the proof search. We consider here a number of examples to show the com-
plexity of the problem of examining possible decomposition trees for a given
formula A. We are going to see that the fact that a given decomposition tree
ends with an axiom leaf does not always imply that the proof does not exist.
It might only imply that our search strategy was not good. Hence the problem
of deciding whether a given formula A does, or does not have a proof in LI
becomes more complex then in the case of Gentzen system for classical logic.
330
Before we define a heuristic method of searching for proof and deciding whether
such a proof exists or not in LI we make some observations.
A, Γ −→ B
(→⇒)
Γ −→ (A ⇒ B)
Γ −→ (A ⇒ B)
(→⇒) .
A, Γ −→ B
Observation 3: we write our proofs in as trees, instead of sequences of expres-
sions, so the proof search process is a process of building a decomposition
tree. To facilitate the process we write, as before, the decomposition rules,
structural rules included in a ”tree ” form.
For example the the above implication decomposition rule is written as follows.
Γ −→ (A ⇒ B)
| (→⇒)
A, Γ −→ B
The two premisses implication rule (⇒→) written as the tree decomposition
rule becomes
(A ⇒ B), Γ −→
^
(⇒→)
Γ −→ A B, Γ −→
331
For example the structural weakening rule is written as the decomposition rule
is written as
Γ −→ A
(→ weak)
Γ −→
Γ −→ A
| (→ weak)
Γ −→
332
Observation 5: the fact that we find a decomposition tree TA with non-axiom
leaf does not mean that 6 `LI A. This is due to the role of structural rules
in LI and will be discussed later in the chapter.
We illustrate the problems arising with proof search procedures, i.e. de-
composition trees construction in the next section 7.2.2 and give a heuristic
proof searching procedure in the section 7.2.3.
We perform proof search and decide the existence of proofs in LI for a given
formula A ∈ F by constructing its decomposition trees TA . We examine here
some examples to show the complexity of the problem.
Remark
In the following and similar examples when building the decomposition trees for
formulas representing general schemas we treat the capital letters A, B, C, D...
as propositional variables, i.e. as indecomposable formulas.
Example 1
Determine whether `LI −→ ((¬A ∩ ¬B) ⇒ ¬(A ∪ B)).
This means that we have to construct some, or all decomposition trees of
If we find a decomposition tree such that all its leaves are axioms, we have a
proof.
If all possible decomposition trees have a non-axiom leaf, proof of A in LI does
not exist.
Consider the following decomposition tree of −→ ((¬A ∩ ¬B) ⇒ ¬(A ∪ B)).
T1
| (−→⇒)
(¬A ∩ ¬B) −→ ¬(A ∪ B)
| (−→ ¬)
(A ∪ B), (¬A ∩ ¬B) −→
| (exch −→)
(¬A ∩ ¬B), (A ∪ B) −→
| (∩ −→)
333
¬A, ¬B, (A ∪ B) −→
| (¬ −→)
¬B, (A ∪ B) −→ A
| (−→ weak)
¬B, (A ∪ B) −→
| (¬ −→)
(A ∪ B) −→ B
^
(∪ −→)
A −→ B B −→ B
non − axiom axiom
The tree T1 has a non-axiom leaf, so it does not constitute a proof in LI. But
this fact does not yet prove that proof doesn’t exist, as the decomposition tree
in LI is not always unique.
Let’s consider now the following tree.
T2
| (−→⇒)
(¬A ∩ ¬B) −→ ¬(A ∪ B)
| (−→ ¬)
(A ∪ B), (¬A ∩ ¬B) −→
| (exch −→)
(¬A ∩ ¬B), (A ∪ B) −→
| (∩ −→)
¬A, ¬B, (A ∪ B) −→
| (exch −→)
¬A, (A ∪ B), ¬B −→
| (exch −→)
(A ∪ B), ¬A, ¬B −→
^
(∪ −→)
334
A, ¬A, ¬B −→ B, ¬A, ¬B −→
| (exch −→) | (exch −→)
¬A, A, ¬B −→ B, ¬B, ¬A −→
| (¬ −→) | (exch −→)
A, ¬B −→ A ¬B, B, ¬A −→
axiom | (¬ −→)
B, ¬A −→ B
axiom
All leaves of T2 are axioms, what proves that T2 is a proof of A and hence we
proved that
`LI ((¬A ∩ ¬B) ⇒ ¬(A ∪ B)).
Example 2
Part 1: Prove that
`LI −→ (A ⇒ ¬¬A),
Solution of Part 1
To prove that
`LI −→ (A ⇒ ¬¬A)
we have to construct some, or all decomposition trees of
−→ (A ⇒ ¬¬A).
−→ (A ⇒ ¬¬A).
| (−→⇒)
A −→ ¬¬A
| (−→ ¬)
¬A, A −→
335
| (¬ −→)
A −→ A
axiom
Solution of Part 2
To prove that
6 `LI −→ (¬¬A ⇒ A)
we have to construct all decomposition trees of (A ⇒ ¬¬A) and show that each
of them has an non-axiom leaf.
Consider the first decomposition tree defined as follows.
T1
−→ (¬¬A ⇒ A)
We use the first tree created to define all other possible decomposition trees by
exploring the alternative search paths as indicated at the nodes of the tree.
336
T1
−→ (¬¬A ⇒ A)
| (−→⇒)
one of 2 choices
¬¬A −→ A
]
| (contr −→)
second of 2 choices
¬¬A, ¬¬A −→ A
| (−→ weak)
f irst of 2 choices
¬¬A, ¬¬A −→
| (¬ −→)
f irst of 2 choices
¬¬A −→ ¬A
| (−→ ¬)
the only choice
A, ¬¬A −→
| (exch −→)
the only choice
¬¬A, A −→
| (−→ ¬)
the only choice
A −→ ¬A
| (−→ ¬)
f irst of 2 choices
A, A −→
indecomposable
non − axiom
337
We can see from the above decomposition trees that the ”blind” construction
of all possible trees only leads to more complicated trees, due to the presence of
structural rules. Observe that the ”blind” application of (contr −→) gives an
infinite number of decomposition trees. To decide that none of them will produce
a proof we need some extra knowledge about patterns of their construction, or
just simply about the number useful of application of structural rules within
the proofs.
In this case we can just make an ”external” observation that the our first tree
T1 is in a sense a minimal one; that all other trees would only complicate this
one in an inessential way, i.e. we will never produce a tree with all axioms
leaves.
One can formulate a deterministic procedure giving a finite number of trees,
but the proof of its correctness require some extra knowledge. We are going to
discuss a motivation and an heuristics for the proof search in the next section.
Within the scope of this book we accept the ”external” explanation for the
heuristics we use as a sufficient solution.
As we can see from the above examples structural rules and especially the
(contr →) rule complicates the proof searching task.
The Gentzen type proof systems RS and GL from chapter don’t contain the
structural rules and are complete with respect to classical semantics, as is the
original Gentzen system LK, which does contain the structural rules. As (via
Completeness Theorem) all three classical proof system RS, GL, LK are equiv-
alent we can say that the structural rules can be eliminated from the system
LK.
A natural question of elimination of structural rules from the intutionistic
Gentzen system LI arizes.
The following example illustrates the negative answer.
Example 3
We know, by the theorem about the connection between classical and intuition-
istic logic (theorem 7.6) and corresponding Completeness Theorems that for any
formula A ∈ F,
|= A if and only if `I ¬¬A,
We have just proved that 6 `LI (¬¬A ⇒ A). Obviously |= (¬¬A ⇒ A), so we
know that ¬¬(¬¬A ⇒ A) must have a proof in LI.
338
We are going to prove that
`LI ¬¬(¬¬A ⇒ A)
and that the structural rule (contr −→) is essential to the existence of its proof,
i.e. that without it the formula ¬¬(¬¬A ⇒ A) is not provable in LI.
The following decomposition tree T is a proof of ¬¬(¬¬A ⇒ A) in LI.
−→ ¬¬(¬¬A ⇒ A)
339
A, ¬(¬¬A ⇒ A) −→
one of 2 choices
| (exch −→)
¬(¬¬A ⇒ A), A −→
one of 3 choices
| (¬ −→)
A −→ (¬¬A ⇒ A)
one of 3 choices
| (−→⇒)
¬¬A, A −→ A
axiom
Assume now that the rule (contr −→) is not available. All possible decomposi-
tion trees are as follows.
T1
−→ ¬¬(¬¬A ⇒ A)
| (−→ ¬)
one of 2 choices
¬(¬¬A ⇒ A) −→
| (¬ −→)
only one choice
−→ (¬¬A ⇒ A)
| (−→⇒)
one of 2 choices
¬¬A −→ A
| (−→ weak)
only one choice
¬¬A −→
| (¬ −→)
only one choice
−→ ¬A
| (−→ ¬)
one of 2 choices
A −→
non − axiom
340
T2
−→ ¬¬(¬¬A ⇒ A)
| (−→ weak)
second of 2 choices
−→
non − axiom
T3
−→ ¬¬(¬¬A ⇒ A)
| (−→ ¬)
¬(¬¬A ⇒ A) −→
| (¬ −→)
−→ (¬¬A ⇒ A)
| (−→ weak)
second of 2 choices
−→
non − axiom
T4
−→ ¬¬(¬¬A ⇒ A)
| (−→ ¬)
¬(¬¬A ⇒ A) −→
| (¬ −→)
−→ (¬¬A ⇒ A)
| (−→⇒)
]
¬¬A −→ A
| (−→ weak)
only one choice
¬¬A −→
| (¬ −→)
341
only one choice
−→ ¬A
| (−→ weak)
second of 2 choices
−→
non − axiom
Before we define a heuristic method of searching for proof in LI let’s make some
additional observations to the observations 1-5 from section 7.2.1.
n1 = ¬¬A, (A ∩ B) −→ B
and
n2 = (A ∩ B), ¬¬A −→ B.
We are going to see that the results of decomposing n1 and n2 differ dramatically.
Let’s decompose the node n1 . Observe that the only way to be able to de-
compose the formula ¬¬A is to use the rule (→ weak) first. The two possible
decomposition trees that starts at the node n1 are as follows.
T1n1
¬¬A, (A ∩ B) −→ B
342
| (→ weak)
¬¬A, (A ∩ B) −→
| (¬ →)
(A ∩ B) −→ ¬A
| (∩ →)
A, B −→ ¬A
| (→ ¬)
A, A, B −→
non − axiom
T2n1
¬¬A, (A ∩ B) −→ B
| (→ weak)
¬¬A, (A ∩ B) −→
| (¬ →)
(A ∩ B) −→ ¬A
| (→ ¬)
A, (A ∩ B) −→
| (∩ →)
A, A, B −→
non − axiom
Let’s now decompose the node n2 . Observe that following our Observation 6
we start by decomposing the formula (A ∩ B) by the use of the rule (∩ →) first.
A decomposition tree that starts at the node n2 is as follows.
T1n2
(A ∩ B), ¬¬A −→ B
| (∩ →)
A, B, ¬¬A −→ B
axiom
343
This proves that the node n2 is provable in LI, i.e.
Of course, we have also that the node n1 is also provable in LI, as one can
obtain the node n2 from it by the use of the rule (exch →).
Observation 8: the use of structural rules are important and necessary while
we search for proofs. Nevertheless we have to use them on the ”must”
basis and set up some guidelines and priorities for their use.
For example, use of weakening rule discharges the weakening formula, and
hence an information that may be essential to the proof. We should use
it only when it is absolutely necessary for the next decomposition steps.
Hence, the use of weakening rule (→ weak) can, and should be restricted
to the cases when it leads to possibility of the use of the negation rule
(¬ →).
This was the case of the decomposition tree T1n1 . We used it as an
necessary step, but still it discharged too much information and we didn’t
get a proof, when proof of the node existed.
In this case the first rule in our search should have been the exchange
rule, followed by the conjunction rule (no information discharge) not the
weakening (discharge of information) followed by negation rule. The full
proof of the node n1 is the following.
T3n1
¬¬A, (A ∩ B) −→ B
| (exch −→)
(A ∩ B), ¬¬A −→ B
| (∩ →)
A, B, ¬¬A −→ B
axiom
For any A ∈ F we construct the set of decomposition trees T→A following the
rules below.
344
Rules for Decomposition Tree Generation
1. Use first logical rules where applicable.
2. Use (exch →) rule to decompose, via logical rules, as many formulas on the
left side of −→ as possible.
3. Use (→ weak) only on a ”must” basis in connection with (¬ →) rule.
4. Use (contr →) rule as the last recourse and only to formulas that contain ¬
or ⇒ as connectives.
5. Let’s call a formula A to which we apply (contr →) rule a contraction
formula.
6. The only contraction formulas are formulas containing ¬ between theirs log-
ical connectives.
7. Within the process of construction of all possible trees use (contr →) rule
only to contraction formulas.
8. Let C be a contraction formula appearing on the node n of the decomposi-
tion tree of T→A . For any contraction formula C, any node n, we apply
(contr →) rule the the formula C at most as many times as the number
of sub-formulas of C.
If we find a tree with all axiom leaves we have a proof, i.e. `LI A and if all
(finite number) trees have a non-axiom leaf we have proved that proof of A does
not exist, i.e. 6 `LI A.
345
such supplementation is provided by the enriched language. For example, modal
logics are enriched by the addition of two new connectives that represent the
meaning of it is necessary that and it is possible that. We use the notation I for
it is necessary that and C for it is possible that. Other notations used are: ∇,
N, L for it is necessary that, and ♦ P, M for it is possible that. The symbols N,
L, P, M or alike, are often used in computer science investigations. The symbols
∇ and were first to be used in modal logic literature, the symbols I, C come
from algebraic and topological interpretation of modal logics. I corresponds to
the interior of the set and C to its closure.
The idea of a modal logic was first formulated by an American philosopher,
C.I. Lewis in 1918. He has proposed yet another interpretation of lasting con-
sequences, of the logical implication. In an attempt to avoid, what some felt,
the paradoxes of semantics for classical implication which accepts as true that a
false sentence implies any sentence he created a notion of a modal truth, which
lead to the notion of modal logic. The idea was to distinguish two sorts of truth:
necessary truth and mere possible (contingent) truth. A possibly true sentence
is one which, though true, could be false. A necessary truth is hence the one
which could not be otherwise; a contingent (possible) truth is one which could.
The distinction between them is a metaphysical one and should not be confused
with the distinction between a priori and a posteriori truths. An a priori
truth is one which can be known independently of experience, and an a poste-
riori truth is one which cannot. Such notions appeal to epistemic considerations
and the whole area of modal logics bristles with philosophical difficulties and
hence the numbers of logics have been created. Unlike the classical connectives,
the modal connectives do not admit of truth-functional interpretation. This was
the reason for which modal logics was first developed as a proof systems, with
intuitive notion of semantics expressed by the set of adopted axioms.
The first semantics, and hence the proofs of the completeness theorems came
some 20 years later. It took yet another 25 years for discovery and development
of the second more general approach to the semantics. These are two established
ways of interpret modal connectives, i.e. to define modal semantics.
The historically first one is due to Mc Kinsey and Tarski (1944, 1946). It is a
topological interpretation that provides a powerful mathematical interpretation
of some of them, namely S4 and S5. It connects the modal notion of necessity
with the topological notion of interior of a set, and the notion of possibility with
the notion of its closure . Our choice of symbols I and C for modal connectives
comes from this interpretation. The topological interpretation powerful as it is,
is less universal in providing models for other modal logics. The most recent
one is due to Kripke (1964). It uses the notion possible world. Roughly, we say
that CA is true if A is true in some possible world, called actual world, and IA
is true if A is true in every possible world.
We present the formal definition later in this chapter, but this intuitive meaning
can be useful in unconvincing ourselves about validity (or sense) of adopted
346
axioms and rules of inference.
As we have already mentioned, modal logics were first developed, as was the
intuitionistic logic, in a form of proof systems only. First Hilbert style modal
proof system was published by Lewis and Langford in 1932. They presented
a formalization for two modal logics, which they called S1 and S2. They also
outlined three other proof systems, called S3, S4, and S5.
In 1933 Gödel worked with Heyting’s ”sentential logic” proof system, what we
are calling now Intuitionistic logic. He considered a particular modal proof sys-
tem and asserted that theorems of Heyting’s ”sentential logic” could be obtained
from it by using a certain translation. His presentation of the discovered proof
system, now known as S4 logic, was particularly elegant.
Since then hundreds of modal logics have been created. There are some standard
texts in the subject. These are, between the others: Hughes and Cresswell
(1969) for philosophical motivation for various modal logics and Intuitionistic
logic, Bowen (1979) for a detailed and uniform study of Kripke models for
modal logics, Segeberg (1971) for excellent classification, and Fitting (1983), for
extended and uniform studies of automated proof methods for classes of modal
logics.
Hilbert Style Modal Proof Systems
We present here Hilbert style formalization for S4 and S5 logics due to Mc
Kinsey and Tarski (1948) and Rasiowa and Sikorski (1964). We also discuss
the relationship between S4 and S5, and between the Intuitionistic logic and S4
modal logic, as first observed by Gödel.
They stress the connection between S4, S5 and topological spaces which consti-
tute models for them. Hence the use of symbols I, C for necessity and possibility,
respectively. The connective I corresponds to the symbol denoting a topological
interior of a set and C to the closure of a set.
Modal Language
We add to the propositional language L{∪,∩,⇒,¬} two extra one argument con-
nectives I and C. I.e. we adopt
L = L{∪,∩,⇒,¬,I,C} (7.39)
as our modal language. We read a formula IA, CA as necessary A and possible
A, respectively.
The language is common to all modal logics. Modal logics differ on a choice of
axioms and rules of inference, when studied as proof systems and on a choice of
semantics.
McKinsey, Tarski (1948)
As modal logics extend the classical logic, any modal logic contains two groups
347
of axioms: classical and modal.
Axioms Group 1: classical axioms
Any modal logic adopts as its classical axioms any complete set of axioms for a
classical propositional logic.
Axioms Group 2: modal axioms
M1 (IA ⇒ A),
M2 (I(A ⇒ B) ⇒ (IA ⇒ IB)),
M3 (IA ⇒ IIA),
M4 (CA ⇒ ICA).
Rules of inference
We adopt the Modus Ponens (M P )
A ; (A ⇒ B)
(M P )
B
and an additional modal rule (I) introduced by Gödel
A
(I)
IA
referred to as necessitation.
We define modal proof systems S4 and S5 as follows.
Fact 7.3
For any formula A ∈ F, if `S4 A, then `S5 A.
348
The equality sign in equations (7.42), and (7.43) means that we replace the
formula in left side of the equation by the formula in the right side anywhere
where the left side (or right side) formula is appears as a sub formula of a
formula of L. In modal logics S4 and S5 the connective C is expressible by ¬
and I, as stated above by (7.43), we hence assume now that the language L
contains only one modal connective I.
Language
L = L{∩,∪,⇒,¬,I} . (7.44)
Fact 7.4
For any formula A ∈ F, if `RS4 A, then `RS5 A.
349
7.3.1 Algebraic Semantics for S4 and S5
The McKinsey, Tarski proof systems (7.40), (7.41), and Rasiowa, Sikorski proof
systems (7.45), (7.45) for modal logics S4, S5 are complete with the respect to
both algebraic topological semantics, and Kripke semantics.
We shortly discuss the topological semantics, and algebraic completeness the-
orems and leave the Kripke semantics for the reader to explore from other,
multiple sources.
The topological semantics was initiated by McKinsey, Tarski’s (1946, 1948) and
consequently developed and examined by many authors into a field of Alge-
braic Logic.They are presented in detail in now classic algebraic logic books:
”Mathematics of Metamathematics”, Rasiowa, Sikorski (1964) and ”An Alge-
braic Approach to Non-Classical Logics”, Rasiowa (1974).
We want to point out that the first idea of a connection between modal propo-
sitional calculus and topology is due to Tang Tsao -Chen, (1938) and Dugunji
(1940).
Here are some basic definitions.
Boolean Algebra
An abstract algebra
B = (B, 1, 0, ⇒, ∩, ∪, ¬) (7.47)
is said to be a Boolean algebra if it is a distributive lattice and every element
a ∈ B has a complement ¬a ∈ B.
Topological Boolean algebra
By a topological Boolean algebra we mean an abstract algebra
It is easy to prove that in in any topological Boolean algebra (7.50) the following
conditions hold for any a, b ∈ B.
350
C(a ∪ b) = Ca ∪ Cb, Ca ∪ a = Ca, CCa = Ca, and C0 = 0. (7.51)
B = (B, 1, 0, ⇒, ∩, ∪, ¬, I, C ).
such that every open element is closed and every closed element is open, i.e.
such that for any a ∈ B
B |= A
351
The formula A is a modal S4 tautology (S5 tautology) and is denoted by
|=S4 A (|=S5 A)
B |= A (B |= A).
Definition 7.6
A formula A is valid in a topological Boolean algebra B = (B, 1, 0, ⇒
, ∩, ∪, ¬, I, C), if and only if v ∗ (A) = 1 holds for all variables assignments
v : V AR −→ B.
Fact 7.5 For any formula A, |=S4 A (|=S5 A) if and only if A is valid in
all topological Boolean algebras (A is valid in all clopen topological Boolean
algebras).
We write `S4 A and `S5 A do denote any proof system for modal S4, S5
logics and in particular the proof systems (7.40), (7.45), and (7.41), (7.46),
respectively.
The completeness for S4 follows directly from the Theorem 7.18. The com-
pleteness for S5 follows from the S4 completeness and Embedding Theorems
7.22, 7.23. It also can be easily proved independently by adopting the Algebraic
Completeness Theorem proof for S4 to clopen topological algebras.
352
(i) `S4 A,
(ii) |=S4 A,
(iii) A is valid in every topological field of sets B(X),
r
(iv) A is valid in every topological Boolean algebra B with at most 22 elements,
where r is the number of all sub formulas of A,
(iv) v ∗ (A) = X for every variable assignment v in the topological field of sets
B(X) of all subsets of a dense-in -itself metric space X 6= ∅ (in particular of an
n-dimensional Euclidean space X).
On S4 derivable disjunction
In a classical logic it is possible for the disjunction (A ∪ B) to be a tautology
when neither A nor B is a tautology. The tautology (A ∪ ¬A) is the simplest
example. This does not hold for the intuitionistic logic. We have a similar
theorem for modal S4 logic, as proved by McKinsey and Tarski.
Theorem 7.19
A disjunction (IA∪IB) is S4 provable if and only if either A or B is S4 provable,
i.e.
`S4 (IA ∪ IB) if and only if `S4 A or `S4 B.
The proof follows directly from the Completeness Theorem 7.18 and the follow-
ing semantical, proof system independent version of the theorem 7.19.
Theorem 7.21
For any formula A ∈ F,
353
Consider a modal language L with both modal connectives, i.e.
L= L{∪,∩,⇒,¬,I,C} .
The above theorem7.21 says that the S4 tautologies form a subset of S5 tau-
tologies. We have even a more powerful relationship, namely the following.
The fist proof of the above embedding theorems was given by Matsumoto in
1955. Provability. Fitting semantical 1983 Ohnishi and Matsumoto 1957/59
Gentzen Methods in Modal Calculi Osaka Mathematical Journal 9.113 -130
As we have said in the introduction, Gödel was the first to consider the connec-
tion between the intuitionistic logic and a logic which was named later S4. His
proof was purely syntactic in its nature, as semantics for neither intuitionistic
logic nor modal logic S4 had not been invented yet.
The algebraic proof of this fact, was first published by McKinsey and Tarski in
1948. We now define the mapping establishing the connection (definition 7.7)
and refer the reader to Rasiowa and Sikorski book ”Mathematics of Metamath-
ematics” for its proof.
Let L be a propositional language of modal logic, as defined by (7.44), i.e the
language
L = L{∩,∪,⇒,¬,I} .
354
Let L0 be a language obtained from L by elimination of the connective I and by
the replacement the negation connective ¬ by the intuitionistic negation, which
we will denote here by a symbol ∼. Such obtained language
L0 = L{∩,∪,⇒,∼} (7.53)
Example
Let A be a formula
((∼ A ∩ ∼ B) ⇒∼ (A ∪ B))
and f be the mapping of definition 7.7. We evaluate f (A) as follows
f ((∼ A ∩ ∼ B) ⇒∼ (A ∪ B)) =
I(f (∼ A ∩ ∼ B) ⇒ f (∼ (A ∪ B)) =
I((f (∼ A) ∩ f (∼ B)) ⇒ f (∼ (A ∪ B)) =
I((I¬f A ∩ I¬f B) ⇒ I¬f (A ∪ B)) =
I((I¬A ∩ I¬B) ⇒ I¬(f A ∪ f B)) =
I((I¬A ∩ I¬B) ⇒ I¬(A ∪ B)).
355
Theorem 7.25
Let f be the Gödel mapping (definition 7.7). For any formula A of L0 ,
where I, S4 denote any proof systems for intuitionistic and and S4 logic, re-
spectively.
In order to establish the connection between the modal logic and classical logic
we consider the Gódel - Tarski mapping (definition ??) between the modal
language L{∩,∪,⇒,¬,I} and its classical sub-language L{¬,∩,∪,⇒} .
Now with every classical formula A we associate a modal formula f (A) defined
by induction on the length of A as follows:
f ((A ∪ B)) = (f (A) ∪ f (B)), f ((A ∩ B)) = (f (A) ∩ f (B)), f (¬A) = I¬f (A).
Theorem 7.26
Let f be the Gödel mapping (definition 7.7) between L{¬,∩,∪,⇒} and L{∩,∪,⇒,¬,I} .
For any formula A of L{¬,∩,∪,⇒} ,
where H, S5 denote any proof systems for classical and and S5 modal logic,
respectively.
356
(i) Show that the 3 element Heyting algebra is a model for all formulas
(7.12) - (7.27).
(ii) Determine for which of the formulas (7.28) - (7.34) the 3 element
Heyting algebra acts as a counter-model.
7. Give the proof of the Glivenko theorem 7.7, i.e. prove that any formula A
is a classically provable if and only if ¬¬A is an intuitionistically provable.
10. Use the Completeness Theorem 7.18 to show that the following proof
system CS4 is a complete proof system for the modal logic S4.
We adopt the modal language L{∪,∩,⇒,¬,I,C} . We adopt, as before, two
groups of axioms: classical and modal.
Group 1: we take any complete set of axioms for a classical propositional
logic. Group 2: the following modal axioms.
C1 (C(A ∪ B) ⇒ (CA ∪ CB)),
C2 (A ⇒ CA),
C3 (CCA ⇒ CA),
C4 C(A ∩ ¬A).
357
Rules of inference: we adopt the Modus Ponens (M P ) and an additional
rule,
(A ⇒ B)
(C) .
(C¬B ⇒ C¬A)
We define the proof system CS4 as follows
11. Evaluate f (A), where f is the Gödel- Tarski mapping (definition 7.7), for
all the formulas A listed below.
(i) (¬A ∪ ¬B) ⇒ (¬A ∩ ¬B))
(ii) ((A ⇒ B) ⇒ (¬B ⇒ ¬A))
(iii) ((A ⇒ ¬B) ⇒ (B ⇒ ¬A))
(iv) (¬¬¬A ⇒ ¬A)
(v) (¬A ⇒ ¬¬¬A)
(vi) (¬¬(A ⇒ B) ⇒ (A ⇒ ¬¬B))
(vii) ((C ⇒ A) ⇒ ((C ⇒ (A ⇒ B)) ⇒ (C ⇒ B))
12. Use the Completeness Theorem 7.18 and Embedding Theorems 7.22,
7.23 to show the following.
(i) For any formula A, `RS4 A, if and only if vdashRS5 I¬I¬A,
where RS4, RS5 are proof system (7.45) and (7.46).
(ii) For any formula A, `S5 A, if and only if |=S4 I¬I¬IA,
where S4, S5 are proof system (7.40) and (7.41).
358
Chapter 8
Classical Predicate
Semantics and Proof
Systems
Definition 8.1
By a predicate language L we understand a triple
359
where A is a predicate alphabet, T, is the set of terms, and F is a set of
formulas.
A = V AR ∪ CON ∪ P AR ∪ Q ∪ P ∪ F ∪ C, (8.2)
Predicate Variables V AR
We assume that we always have a countably infinite set V AR of predicate
variables, called usually variables. We denote variables by x, y, z, ..., with
indices, if necessary, what we often express by writing
V AR = {x1 , x2 , ....}.
Parenthesis P AR
As in the propositional case, we adopt the signs ( and ) for our parenthesis., i.e.
we define the set P AR as
P AR = {(, )}.
360
Quantifiers Q
We adopt two quantifiers; ∀ (for all, the universal quantifier) and ∃ (there exists,
the existential quantifier), i.e. we have the following set of quantifiers
Q = {∀, ∃}.
In a case of the classical logic and the logics that extend it, it is possible to adopt
only one quantifier and to define the other in terms of it and propositional con-
nectives. It is impossible in a case of many non-classical logics, for example the
intuitionistic logic. But even in the case of classical logic two quantifiers express
better the common intuition, so we assume that we have two of them.
Predicate symbols P
Predicate symbols represent relations. We assume that we have an non empty,
finite or countably infinite set bf P of predicate, or relation symbols. We denote
predicate symbols by P, Q, R, ..., with indices, if necessary, what we often express
by writing
P = {P1 , P2 , ...}.
Function symbols F
We assume that we have a finite (may be empty) or countably infinite set F of
function symbols. When the set F is empty we say that we deal with a language
without functional symbols. We denote functional symbols by f, g, h, ..., with
indices, if necessary, what we often express by writing
F = {f1 , f2 , ...}.
Constant symbols C
We also assume that we have a finite (may be empty) or countably infinite set
C of constant symbols. The elements of C are denoted by c, d, e..., with indices,
if necessary, what we often express by writing
C = {c1 , c2 , ...}.
When the set C is empty we say that we deal with a language without constant
symbols.
361
Sometimes the constant symbols are defined as 0-ary function symbols, i.e.
C ⊆ F. We single them out as a separate set for our convenience.
Observe that what distinguishes now one predicate language L form the pother
is the choice of the components CON , and P, F, C of its alphabet A. We
hence will write
L(P, F, C) (8.4)
2. Terms T
The set T of terms of a predicate language L(P, F, C) is defined as follows.
Example 8.1
362
Let f ∈ F, #f = 1, i.e. f is a one place function symbol. Let x, y be predicate
variables, c, d constants, i.e. x, y ∈ V AR, c, d ∈ C. The following expressions
are terms:
Example 8.2
If F = ∅, C = ∅, then the set T of terms consists of variables only, i.e.
T = V AR = {x1 , x2 , ....}.
Remark 8.1
For any predicate language L(P, F, C), the set T of its terms is always non-
empty.
Example 8.3
If f ∈ F, #f = 1, g ∈ F, #g = 2, x, y ∈ V AR, c, d ∈ C, then some of the
terms are the following:
f (g(x, y)), f (g(c, x)), g(f (f (c)), g(x, y)), g(c, g(x, f (c))).
From time to time, the logicians are and we may be informal about how we
write terms. For instance, if we denote a two place function symbol g by +, we
may write x + y instead +(x, y). Because in this case we can think of x + y as
an unofficial way of designating the ”real” term +(x, y), or even g(x, y).
2. Formulas F
Before we define the set of formulas, we need to define one more set; the set of
atomic, or elementary formulas. They are the ”smallest” formulas as were the
propositional variables in the case of propositional languages.
Atomic formulas
An atomic formula of a predicate language L(P, F, C) is any element of the
alphabet A∗ of the form
R(t1 , t2 , ..., tn ),
where R ∈ P, #R = n, i.e. R is n-ary relational symbol and t1 , t2 , ..., tn are
terms. The set of all atomic formulas is denoted by AF and is defines as
363
Example 8.4
Consider a language
L(∅, {P }, ∅),
for #P = 1, i.e. a language without neither functional, nor constant symbols,
and with one, one-place predicate symbol P . The set of atomic formulas contains
all formulas of the form P (x), for x any variable, i.e.
AF = {P (x) : x ∈ V AR}.
Example 8.5
Let now
L = L({f, g}, {R}, {c, d}),
for #f = 1, #g = 2 , #R = 2, i.e. L has two functional symbols: one -place
symbol f and two-place symbol g; one two-place predicate symbol R, and two
constants: c,d. Some of the atomic formulas in this case are the following.
R(c, d), R(x, f (c)), R(f (g(x, y)), f (g(c, x))), R(y, g(c, g(x, f (c)))).
364
4. if A is a formula of L(P, F, C) and x is a variable, then ∀xA, ∃xA are
formulas of L(P, F, C), i.e. if the following recursive condition holds
Example 8.6
Let L be a language with with the set {∩, ∪, ⇒, ¬} of connectives and with
two functional symbols: one -place and one two-place, one two-place predicate
symbol, and two constants. We write L as
R(c, f (d)), ∃xR(x, f (c)), ¬R(x, y), ∀z(∃xR(x, f (c)) ⇒ ¬R(x, y)),
Example 8.7
Let L be a language with with the set {¬, , ♦, ∩, ∪, ⇒} of connectives and P,
F, and C the same as in previous exercise, i.e.
where #f = 1, #g = 2 , #R = 2.
L is now a language of some first order modal logic. Some of the formulas of
L are the following.
♦¬R(c, f (d)), ♦∃xR(x, f (c)), ¬♦R(x, y), ∀z(∃xR(x, f (c)) ⇒ ¬R(x, y)),
365
The formula (∃xR(x, f (c)) ⇒ ¬R(x, y)) is in the scope of ∀z in ∀z(∃xR(x, f (c)) ⇒
¬R(x, y)). Formula ¬♦∃xR(x, y) is in the scope of ∀y in ∀y¬♦∃xR(x, y).
Observe that the formulas ∃yP (y), (∀x(P (x) ⇒ ∃yQ(x, y))) are closed. We
call a close formula a sentence.
Example 8.8
Consider atomic formulas: P (y), Q(x, c), R(z), P1 (g(x, y), z). Here are some
non atomic formulas formed out of them.
1. (P (y) ∪ ¬Q(x, c)) ∈ F. This is an open formula A with two free variables
x,y. We denote A this as formula A(x, y).
2. ∃x(P (y) ∪ ¬Q(x, c)) ∈ F. We write x to denote that x is a bound variable.
The variable y is free. This is a formula B with one free variable y. We denote
B as a formula B(y).
3. ∀y(P (y) ∪ ¬Q(x, c)) ∈ F. The variable y is bound, the variable x is free.
We denote this formula by for example A1 (x).
366
4. ∀y∃x(P (y) ∪ ¬Q(x, c)) ∈ F has no free variables. It is a closed formula
called also a sentence.
Exercise 8.1
Given the following formulas of L:
P (x, f (c, y)), ∃cP (x, f (c, y)), ∀xf (x, P (c, y)), ∃xP (x, f (c, y)) ⇒ ∀yP (x, f (c, y)).
1. Indicate whether they are, or are not well formed formulas of F. For those
which are not in F write a correct formula.
2. For each correct, or corrected formula identify all components: connectives,
quantifiers, predicate and function symbols, and list all its terms.
3. For each formula identify its s free and bound variables. State which are open
and which are closed formulas (sentences), if any.
Solution
Formula A1 = P (x, f (c, y)).
It is a correct atomic formula. P is a 2 argument predicate symbol, f is a
2 argument function symbol, c is a constant. We write it symbolically: P ∈
P, f ∈ F, c ∈ C. It is an open formula with two free variables x,y. We denote
it by A1 (x, y). It has no bound variables.
Formula A2 = ∃cP (x, f (c, y)).
It is a not a correct formula, i.e. ∃cP (x, f (c, y)) 6∈ F. The expression ∃c has no
meaning because c is a constant, not a variable.
The corrected formulas are: B1 = ∃xP (x, f (c, y)), B2 = ∃yP (x, f (c, y)), and
formulas B = ∃zP (z, f (c, y)) for any variable z different then x and y.
None of the correct formulas are open. Variable y is free in B1 = B1 (y), variable
x is free in B2 = B2 (x), both variables x and y are free in all formulas B =
B(x, y). All formulas are nether close, nor open. The terms appearing in any
of them are the same as in A1 = P (x, f (c, y)) and are: x, y, c, f (c, y).
Formula A3 = ∀xf (x, P (c, y)).
It is a not a correct formula, i.e. ∀xf (x, P (c, y)) 6∈ F. The function symbol f in
front f (x, P (c, y)) indicate a term and terms are not formulas. Moreover, the
atomic formula P (c, y) can’t be put inside a term!
Formula A4 = ∃xP (x, f (c, y)) ⇒ ∀yP (x, f (c, y)).
It is a not a correct formula. The correct formula is A = (∃xP (x, f (c, y)) ⇒
∀yP (x, f (c, y))). It has two free variables x and y and we write it as A = A(x, y).
Informally, in the formula P (x, y) both variables x and y are called free vari-
ables. They are not in the scope of any quantifier. The formula of that type
(without quantifiers) is an open formula.
The formal definition of the set of free variables of a formula is the following.
367
Definition 8.4 (Free and Bound Variables)
The set F V (A) of free variables of a formula A is defined by the induction of
the degree of the formula as follows.
Example 8.9 The formulas ∃xQ(c, g(x, d)), ¬∀x(P (x) ⇒ ∃y(R(f (x), y) ∩
¬P (c))) are sentences. The formulas Q(c, g(x, d)), ¬(P (x) ⇒ (R(f (x), y) ∩
¬P (c))) are open formulas. The formulas ∃xQ(c, g(x, y)), ¬(P (x) ⇒ ∃y(R(f (x), y)∩
¬P (c))) are neither sentences nor open formulas. They contain some free and
some bound variables; the variable y is free in the first formula, the variable x
is free in the second.
The definition 8.1 defines a predicate language L = LCON (P, F, C) (8.3) with
its sets of predicate, function and constant symbol possibly countably infinite
sets. We use its most general case with sets of predicate, function and constant
symbol all countably infinite sets for defining all relevant notions concerning
provability and semantics. In particular, we will define in detail the classical
semantics for this most general form of L and prove the completeness theorem
for classical predicate logic based on it.
When we deal with formal theory Th(SA) with a set SA of specific axioms
we restrict the language LCON (P, F, C) to the symbols characteristic for that
theory. We hence introduce the following definition.
368
Definition 8.5
Given a language LCON (P, F, C) = (A, T, F).
Let F0 ⊆ F be a non-empty, finite subset of formulas of L. Denote by P0 , F0 , C0
the sets of all predicate, function, and constant symbols appearing in the formu-
las from the set F0 . The language
LCON (P0 , F0 , C0 )
F0 = {∃xQ(c, g(x, d)), ¬∀x(P (x) ⇒ ∃y(R(f (x), y)∩¬P (e))), ¬(F (a)∩R(y, h(c))}.
where # Q = #R = 2, #P =# F = 1, # g = 2, #f = # h = 1.
to indicate that F V (A) ⊆ {x1 , x2 , ..., xn } without implying that all of x1 , x2 , ..., xn
are actually free in A. This is similar to the practice in algebra of writing
p(x1 , x2 , ..., xn ) for a polynomial p in the variables x1 , x2 , ..., xn without imply-
ing that all of them have nonzero coefficients.
369
This notation is convenient because we can agree to write as
respectively. The term t = f (x, y) is free for x and is not free for y in A. The
term t = f (x, z) is free for x and z in B. The term t = y is not free neither for
x nor for z in A, B.
Example 8.12
Let A be a formula
The term t1 = f (x) is not free for x in A; the term t2 = g(x, z) is free for z
only, term t3 = h(x, y) is free for y only because x occurs as a bound variable
in A; term t4 .
370
We often use logic symbols, while writing mathematical statements. For exam-
ple mathematicians in order to say ”all natural numbers are greater then zero
and some integers are equal 1” often write it as
Some of them, who are more ”logic oriented”, would also write it as
∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1,
or even as
(∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1).
None of the above symbolic statements are formulas of the predicate language
L. These are mathematical statement written with mathematical and logic
symbols. They are written with different degree of ”logical precision”, the last
being, from a logician point of view the most precise.
Observe that the quantifiers in ∀x∈N and ∃y∈Z used in all of them are not
the one used in the predicate language L, which admits only quantifiers ∀x
and ∃y, for any variables x, y ∈ V AR. The quantifiers ∀x∈N , ∃y∈Z are called
quantifiers with restricted domain. The first is restricted to the domain of
natural numbers, the second to the integers. The restriction of the quantifier
domain can, and often is given by more complicated statements. For example
we say ”for all x > 2” and write ∀x>2 , or we say ”exists x > 2 and at same time
x + 2 < 8” and write symbolically ∃(x>2∩x+2<8) .
Our goal now is to correctly ”translate ” mathematical and natural language
statement into formulas of the predicate language L of the classical predicate
logic with the the set {¬.∩, ∪, ⇒} of propositional connectives. We say ”classical
predicate logic” to express that we define all notions for the classical semantics
to be defined formally in the next section 8.2. One can extend these notions
to non-classical logics, but we describe and will talk only about classical case.
We introduce the quantifiers with restricted domain into the classical predicate
logic language by expressing them within the language L as follows.
Definition 8.9
Given a classical predicate logic language
The quantifiers ∀A(x) , ∃A(x) are called quantifiers with restricted domain,
or restricted quantifiers, where A(x) ∈ F is any formula with any free vari-
able x ∈ V AR.
A formula ∀A(x) B(x) is an abbreviation of a formula ∀x(A(x) ⇒ B(x)) ∈ F.
We write it symbolically as
371
A formula ∃A(x) B(x) stands for a formula ∃x(A(x) ∩ B(x)) ∈ F. We write it
symbolically as
∃A(x) B(x) = ∃x(A(x) ∩ B(x)) (8.8)
Exercise 8.2
Given a mathematical statement S written with logical symbols
(∀x∈N x ≥ 0 ∩ ∃y∈Z y = 1)
1. Translate it into a proper logical formula with restricted domain quantifiers
i.e. into a formula of L that uses the restricted domain quantifiers.
2. Translate your restricted domain quantifiers logical formula into a correct
logical formula without restricted domain quantifiers, i.e. into a formula of L.
Solution
We proceed to write this and other similar problems solutions in a sequence of
steps.
372
The statement S becomes a restricted quantifiers formula:
STEP 4. We apply (8.7) and (8.8) to the formula from STEP 3. and obtain a
formula A of L as a representation of the given mathematical statement S.
Short Solution
The basic statements in S are: x ∈ N, x ≥ 0, y ∈ Z, y = 1. The corresponding
atomic formulas of L are: N (x), G(x, c1 ), Z(y), E(y, c2 ), respectively.
The statement S becomes becomes a restricted quantifiers formula
(∀N (x )G(x, c1 ) ∩ ∃Z(y) E(y, c2 )). Applying restricted quantifiers definition 8.9
and transformation rules (8.7), (8.8) we get a following formula A ∈ F
373
as follows.
Example 8.13
Let L be a language with one two-place predicate symbol, two functional symbols:
one -place and one two-place, and two constants, i.e.
L = L({R}, {f, g}, {c, d}, )
where #R = 2, #f = 1, #g = 2, and c, d ∈ C.
We define a structure M = [U, I] as follows. We take as the universe the set
U = {1, 3, 5, 6}. The predicate R is interpreted as ≤, what we write as RI : ≤.
We interpret f as a function fI : {1, 3, 5, 6} −→ {1, 3, 5, 6} such that fI (x) = 5
for all x ∈ {1, 3, 5, 6}, and we put gI : {1, 3, 5, 6} × {1, 3, 5, 6} −→ {1, 3, 5, 6}
such that gI (x, y) = 1 for all x ∈ {1, 3, 5, 6}. The constant c becomes cI = 3,
and dI = 6, what we write as We write the structure M as
M = [{1, 3, 5, 6} ≤, fI , gI , cI = 3, dI = 6]
Exercise 8.3
Given a language
L = L({R}, {g}, ∅, )
where #R = 2, #g = 2 . Define two structures for L, both with infinite universe:
one infinitely countable and one uncountable.
374
Solution
There are many such structures. Here are two of the very simple.
M1 = [N, , ≤, +], where N is the set of natural numbers, and for example
M2 = [R, , ≤, +], where R is the set of real numbers.
s : V AR −→ U (8.9)
2. for any c ∈ C,
sI (c) = cI ;
375
3. for any t1 , t2 , . . . , tn ∈ T, n ≥ 1, f ∈ F, such that #f = n, and for any
term t = f (t1 , t2 , . . . , tn ), we put
i.e. we define
Exercise 8.4
Consider a language
for # P = # R = 2, #f = # g=1, # h= 2.
Let M = [{0, 1}, I], where the interpretation I is defined as follows.
PI = {(0, 0)}, RI = {(0, 0), (1, 1)},
fI (0) = 0, fI (1) = 0, gI (0) = 1, gI (1) = 1, hI is given by a formula
hI (x, y) = x for all (x, y) ∈ {0, 1} × {0, 1}, and cI = 1, dI = 0.
Solution
First we evaluate terms that are variables and constants of L using the formulas
1. and 2. of definition 8.12: sI (x) = s(x), sI (c) = cI , respectively and obtain:
sI (z) = s(z) = 1, sI (y) = s(y) = 1, sI (x) = s(x) = 1, sI (c) = cI = 1, sI (d) =
dI = 2. We use the formula sI (f (t1 , t2 , . . . , tn )) = fI (sI (t1 ), sI (t2 ), . . . , sI (tn )
to evaluate the rest of terms in T0 and obtain:
sI f (c) = fI (sI (c)) = fI (cI )) = fI (1) = 0, sI f (x) = fI (sI (x)) = fI (1) = 0
sI g(z) = gI (sI (c)) = gI (1) = 1, sI (f (g(d)) = fI (sI (f (g(d))) = fI (gI (sI (d))) =
fI (gI (cI )) = fI (gI (1)) = fI (1) = 0,
sI g(f (g(z)) = gI (fI (gI (sI (z)))) = gI (fI (gI (1))) = gI (fI (1)) = gI (0) = 1,
sI (h(c, f (g(d)))) = hI (sI (c), sI (f (g(d)))) = hI (cI , fI (gI (sI (d))))
= hI (1, fI (gI (0))) = hI (1, fI (1)) = hI (1, 0) = 1,
376
sI (h(f (x), g(z))) = hI (fI (sI (x)), gI (sI (x))) = hI (fI (1), gI (1)) = hI (0, 1) = 0.
Example 8.14
Consider a language
L = L({P, R}, {f, h}, ∅ )
for # P = # R = 2, #f = 1, # h= 2.
Let M = [Z, I], where Z is the set on integers and the interpretation I for
elements of F and C is as follows.
fI : Z −→ Z is given by formula f (m) = m + 1 for all m ∈ Z.
hI : Z × Z −→ Z is given by formula f (m, n) = m + n for all m, n ∈ Z.
Let s : V AR −→ Z be any assignment such that s(x) = −5, s(y) = 2 and
t1 , t2 ∈ T be t1 = h(y, f (f (x))) and t2 = h(f (x), h(x, f (y)).
We evaluate:
sI (t1 ) = sI (h(y, f (x)) = hI (sI (y), fI (sI (x))) = +(2, fI (−5)) = 2 − 4 = −2,
sI (t2 ) = sI (h(f (x), h(x, f (y))) = +(fI (−5), +(−5, 3)) = −4 + (−5 + 3) = −6.
t(x1 , x2 , . . . , xn ).
Observation 8.1
For any term t(x1 , x2 , . . . , xn ) ∈ T, any structure M = [U, I] and any assign-
ments s, s0 of L in M, the following holds.
If s(x) = s0 (x) for all x ∈ {x1 , x2 , . . . , xn }, i.e the assignments s, s0 agree on
all variables appearing in t, then, sI (t) = s0 I (t).
377
to denote any assignment s0 : V AR −→ U such that s, s0 agree on all variables
except on x, such that s0 (x) = a, for certain a ∈ U .
Given a first order (predicate) language L = LCON (P, F, C). The satisfaction
relation (M, s) |= A between structures, assignments and formulas of L is de-
fined by induction on the complexity of formulas of L. It is the satisfaction
relation (M, s) |= A that allows us to distinguish one one semantics for a
given L from the other, and consequently one logic from the other. We define
now only a classical satisfaction and the notion of classical predicate tautology.
(M, s) |= A
(ii) A as not atomic formula and has one of connectives of L as the main
connective.
(iii) A as not atomic formula and begins with one of the quantifiers.
(M, s) |= ∃xA if and only if there is s0 such that s, s0 agree on all variables
except on x, and (M, s0 ) |= A,
(M, s) |= ∀xA if and only if for all s0 such that s, s0 agree on all variables
except on x, and (M, s0 ) |= A.
378
Observe that that the truth or falsity of (M, s) |= A depends only on the values
of s(x) for variables x which are actually free in the formula A. This is why we
often write the condition (iii) as
(iii)’ A(x) as not atomic formula (with a free variable x) and begins with one
of the quantifiers.
(M, s) |= ∃xA(x) if and only if there is s0 such that s(y) = s0 (y) for all
y ∈ V AR − {x}, and (M, s0 ) |= A(x),
(M, s) |= ∀xA if and only if for all s0 such that s(y) = s0 (y) for all
y ∈ V AR − {x}, and (M, s0 ) |= A.
Exercise 8.5
For the structures Mi , find assignments si , s0 i (1 ≤ i ≤ 4), such that
The structures Mi are defined as follows (the interpretation I for each of them
is specified only for symbols in the formula Q(x, c), and N denotes the set of
natural numbers.
Solution
Consider M1 = [{1}, QI :=, cI : 1]. Observe that all s : V AR −→ {1} must
are defined by a formula s(x) = 1 for all x ∈ V AR. We evaluate (definition
8.12), sI (x) = 1, sI (c) = cI = 1. By definition 8.13, (M1 , s) |= Q(x, c) if and
only if (sI (x), sI (c)) ∈ QI , i.e. (1, 1) ∈= what is true as 1 = 1. We have
proved
379
Consider M3 = [N, QI :≥, cI : 0]. Let s : V AR −→ N be any assignment, such
that s(x) = 5. We evaluate sI (x) = 5, sI (c) = 0. Observe that the condition
(sI (x), sI (c)) ∈ QI holds as 5 ≥ 0 and
Example 8.15
Let Mi (1 ≤ i ≤ 4) be structures in defined the exercise 8.5 and let corresponding
assignments si be as defined as its solutions.
1. (M1 , s) |= Q(x, c), (M1 , s) |= ∀xQ(x, c), (M1 , s) |= ∃xQ(x, c).
2. (M2 , s) |= Q(x, c), (M2 , s) 6|= ∀xQ(x, c), (M1 , s) |= ∃xQ(x, c).
3. (M3 , s) |= Q(x, c), (M3 , s) |= ∀xQ(x, c), (M3 , s) |= ∃xQ(x, c).
4. (M4 , s) |= Q(x, c), (M4 , s) 6|= ∀xQ(x, c), (M4 , s) |= ∃xQ(x, c).
We define now a very important semantic notion. It has different names: logi-
cal consequence, logical implication, semantic consequence, logical (semantical)
entailment. We use a name logical consequence and define it as follows.
380
Definition 8.15 (Logical Consequence)
For any A, B ∈ F and any set Γ ⊆ F of formulas of L, we say that a formula
B is a logical consequence of a set Γ and write it as Γ |= B, if and only if
all models of the set Γ are models of the formula B.
Fact 8.1
For any formula A of L,
If A is a sentence, then if there s such that (M, s) |= A, then M is a model for
A, i.e. M |= A
381
By the closure of A we mean the formula obtained from A by prefixing in
universal quantifiers all variables the are free in A. If A does not have free
variables (i.e. is a sentence), the closure if A is defined to be A itself.
Example 8.16
Let Q ∈ P, #Q = 2 and c ∈ C Consider formulas
Q(x, c), ∃xQ(x, c), ∀xQ(x, c)
and the structures from exercise 8.5 defined as follows.
M1 = [{1}, QI :=, cI : 1], M2 = [{1, 2}, QI :≤, cI : 1],
M3 = [N, QI :≥, cI : 0], and M4 = [N, QI :≥, cI : 1.]
Directly from example 8.15 and Fact 8.1, we get that:
1. M1 |= Q(x, c), M1 |= ∀xQ(x, c), M1 |= ∃xQ(x, c).
2. M2 6|= Q(x, c), M2 6|= ∀xQ(x, c), M2 |= ∃xQ(x, c).
3. M3 |= Q(x, c), M3 |= ∀xQ(x, c), M3 |= ∃xQ(x, c).
4. M4 6|= Q(x, c), M4 6|= ∀xQ(x, c), M4 |= ∃xQ(x, c).
By the definition 9.3 we have that A is true in M only when the structure M
is a model for A. This is why we use the notation M |= A in both cases.
Obviously, if A is not true in M, then it is false, and vice versa. This proves
correctness of our definition with respect to the intuitive understanding.
We get directly from definition 8.18 and the example 8.16 the following.
382
Example 8.17
Let M1 − M4 be structures defined in example 8.5.
1. Formulas Q(x, c), ∀xQ(x, c), ∃xQ(x, c) are all true in the structures M1
and M3 .
2. Formula ∃xQ(x, c) is also true in M2 and in M3 .
3. Formulas ¬Q(x, c), ¬∀xQ(x, c), ¬∃xQ(x, c) are all false in the structures
M1 and M3 .
4. Formula ¬∃xQ(x, c) is also false in M2 and in M3 .
5. Formulas (Q(x, c) ∩ ¬Q(x, c)), (¬∀xQ(x, c) ∩ ∀xQ(x, c)), and the formula
(∃xQ(x, c) ∩ ¬∃xQ(x, c)) are all false in all structures M1 − M4 .
6. The formula ∀xQ(x, c) is false in a structure M5 = [N, QI :<, cI : 0].
P3. It is not the case that both M |= A and M |= ¬A, i.e. no formula of L
can be both true and false in M, i.e. there is no formula A, such that M |= A
and M =| A.
P4. If M |= A and M |= (A ⇒ B), then M |= B.
P5. (A ⇒ B) is false in M if and only if M |= A and M |= ¬B, i.e.
383
Definition 8.19 (Valid, Tautology)
Given a language L = L{¬,∩,∪,⇒,¬} (P, F, C), a formula A of L. .
A formula A is predicate tautology (is valid) if and only if M |= A for
all structures M = [U, I], i.e. when A is true in all structures M for L.
We write
|= A or |=p A,
to denote that a formula A is predicate tautology (is valid).
We write
|=p A
when there is a need to stress a distinction between propositional and predicate
tautologies, otherwise we will use the symbol |= .
Predicate tautologies are also called laws of quantifiers.
Following the notation T for the set of all propositional tautologies (chapter 5)
we denote by Tp the set of all predicate tautologies, i.e.
Directly from the definition 8.18, the tautology definition 8.19 we get the fol-
lowing basic properties of logical consequence as defined by definition 8.15.
Property 8.2
For any A, B ∈ F and any set Γ ⊆ F of formulas of L,
P1. A |= B if and only if |= (A ⇒ B).
P2. If A |= B and A is true in M, then B is true in M.
P2. If Γ |= B and if all formulas in Γ are true in M, then B is true in M.
Definition 8.20
For any formula A of predicate language L,
A is not a predicate tautology (6|= A) if and only if there is a structure
M = (U, I) for L, such that M 6|= A.
We call such structure M a counter-model for A.
The definition 8.20 says: to prove that a formula A is not a predicate tautology
one has to show a counter- model M = (U, I). It means one has to show a non-
empty set U, define an interpretation I, and an assignment s : V AR −→ U
384
such that (M, s) 6|= A.
Following the notation C for the set of all propositional tautologies (chapter
5) we denote by Cp the set of all predicate contradictions, i.e.
Directly from the definition 8.18 and Property 8.1 we have the folowing duality
property, the same as the one for propositional logic.
Fact 8.2
For any formula A of predicate a language L,
A ∈ Tp if and only if ¬A ∈ Cp ,
A ∈ Cp if and only if ¬A ∈ Tp .
Obviously, the formulas (Q(x, c) ∩ ¬Q(x, c)), (¬∀xQ(x, c) ∩ ∀xQ(x, c)), and the
formula (∃xQ(x, c) ∩ ¬∃xQ(x, c)) defined in example 8.17 are not only false in
the structures M1 − M4 , but are false in all structures M for L. By definition
8.21 they all are predicate contradictions. Observe that they all are substitutions
of propositional contradictions (a ∩ ¬a) or (¬a ∩ a). By the same argument the
formulas (Q(x, c)∪¬Q(x, c)), (¬∀xQ(x, c)∪∀xQ(x, c)), (∃xQ(x, c)∩¬∃xQ(x, c))
are predicate tautologies as they are substitutions of propositional tautologies
(a ∪ ¬a) or (¬a ∪ a).
We put these examples and observations in a following theorems that establish
relationship between propositional and predicate tautologies and contradictions.
385
We write now |=, =| do denote respectively propositional tautologies and con-
tradiction, and |=p , =|p for predicate tautologies and contradictions. We first
formalize and prove (theorem 8.1) the intuitively obvious fact: if a formula A is
a propositional tautology (contradiction), then replacing propositional variables
in A by any formulas of a predicate language we obtain a formula which is a
predicate tautology (contradiction).
Example 8.18
Let consider the following example of a propositional tautology and a proposi-
tional contradiction.
Substituting ∃xP (x, z) for a, and ∀yR(y, z) for b, we obtain, by theorem 8.1,
that
|=p ((∃xP (x, z) ⇒ ∀yR(y, z)) ⇒ (¬∃xP (x, z) ∪ ∀yR(y, z))) and
We put it all in a more formal and more general and precise language as follows.
Given a propositional language L0 = L{¬,∩,∪,⇒,¬} with the set F0 of formulas
and a predicate languageL = L{¬,∩,∪,⇒,¬} (P, F, C) with the set F of formulas.
Let A(a1 , a2 , . . . , an ) ∈ F0 and A1 , A2 , . . . , An ∈ F. We denote by
Theorem 8.1
Given a propositional language L0 with the set F0 of formulas and a predicate
language L with the set F of formulas.
For any A(a1 , a2 , . . . an ) ∈ F0 and any A1 , A2 , . . . An ∈ F the following holds.
1. If |= A(a1 , a2 , . . . , an ), then |=p A(a1 /A1 , a2 /A2 , . . . , an /An ).
2. If =| A(a1 , a2 , . . . , an ), then =|p A(a1 /A1 , a2 /A2 , . . . , an /An ).
386
Fact 8.3
For any formula A(x) of L,
|= (∀x A(x) ⇒ ∃x A(x)).
Proof
Assume that 6|= (∀x A(x) ⇒ ∃x A(x)). By definition 8.20 there is a structure
M = (U, I) and s : V AR −→ U, such that (M, s) 6|= (∀x A(x) ⇒ ∃x A(x)).
By definition 8.13, (M, s) |= ∀x A(x) and (M, s) 6|= ∃x A(x). It means that
(M, s0 ) |= A(x) for all s0 such that s, s0 agree on all variables except on x,
and it is not true that there is s0 such that s, s0 agree on all variables except
on x, and (M, s0 ) |= A(x). This is impossible and this contradiction proves
|= (∀x A(x) ⇒ ∃x A(x)).
Given a set F of formulas of a predicate language L. We denote by OF set of all
open formulas of L, i.e. formulas without quantifiers. We prove that any open
formula in order to be predicate tautologies must be a substitution definied in
theorem 8.1 of a propositional tautology. I.e. we have the following substitution
theorem.
Theorem 8.2
Any open formula A of a predicate language L is a predicate tautology if and
only if it is a substitution of a propositional tautology as defined in theorem 8.1.
Proof
Observe that every open formula from A ∈ OF is a form
B(a1 /A1 , a2 /A2 , . . . , an /An ) for certain propositional formula B(a1 , a2 , . . . an ),
where A1 , A2 , . . . , An are predicate atomic formulas from the set AF as defined
in (8.5). Theorem 8.2 follows directly from the following.
Lemma 8.1
Let σ be a one to one mapping from the set V0 of propositional variables of
propositional language L0 into the set AF of the atomic formulas of the predicate
language L. For any A(a1 , a2 , . . . an ) ∈ F0 ,
|= A(a1 , a2 , . . . , an ), if and only if |=p A(a1 /σ(a1 ), . . . , an /σ(an )).
Proof of lemma
The implication ”if |= A(a1 , a2 , . . . , an ), then
|=p A(a1 /σ(a1 ), . . . , an /σ(an ))” holds as a particular case of theorem 8.2.
We prove now the converse implication by proving its opposite
if 6|= A(a1 , a2 , . . . , an ), then 6|=p A(a1 /σ(a1 ), . . . , an /σ(an )). (8.16)
Assume 6|= A(a1 , a2 , . . . , an ). There exists a truth assignment v : V0 −→ {T, F }
such that v ∗ (A(a1 , a2 , . . . , an )) = F . We construct a counter model M for
387
A(a1 /σ(a1 ), . . . , an /σ(an )) as follows. Let M = [T, I], where T is the set of all
terms of L, and for any c ∈ C, f ∈ F, P ∈ P we put cI = c, fI (t1 , t2 , . . . , tn ) =
f (t1 , t2 , . . . tn ), PI ⊆ T#P .
Let now the s assignment of L in M be an identity, i.e. s : V AR −→ T is
such that s(x) = x for all x ∈ V AR. We extend s to the interpretation of terms
(definition 8.12) as follows.
sI (x) = s(x) = x, sI (c) = cI = c, sI (f (t1 , t2 , . . . tn )) = fI (sI (t1 ), . . . , sI (tn )) =
f (t1 , t2 , . . . tn ), i.e. we have that sI (t) = t for all t ∈ T.
We have that for every atomic formula P (t1 , t2 , . . . tn ) there is exactly one propo-
sitional variable a, such that P (t1 , t2 , . . . tn ) = σ(a). We define now that PI as
follows.
(t1 , t2 , . . . tn ) ∈ PI if and only if P (t1 , t2 , . . . tn ) = σ(a) and v(a) = T .
(t1 , t2 , . . . tn ) 6∈ PI if and only if P (t1 , t2 , . . . tn ) = σ(a) and v(a) = F .
We assumed that v : V0 −→ {T, F } is such that v ∗ (A(a1 , a2 , . . . , an )) = F .
Directly form definition of the assignment s and the interpretation I we have
that ([T, I], s) 6|= A(a1 /σ(a1 ), . . . , an /σ(an )). It end the roof of lemma 8.1 and
hence the proof of theorem 8.2.
Fact 8.4
The converse implication to (8.3) is not a predicate tautology, i.e. there is a
formula A of L. such that
Proof
Observe that to prove (8.17) we have to provide an example of an instance of a
formula A(x) and construct a counter-model M = (U, I) for it. Let A(x) be an
atomic formula P (x, c), for any P ∈ P, #P = 2. The instance is
388
N of natural numbers, as there is n ∈ N , such that n < 3, for example n = 2,
and it is not true that all natural numbers are smaller then 3.
We have to be very careful when we deal with quantifiers with restricted
domain (definition 8.9). We adopt the following definition for restricted domain
quantifiers.
The most basic predicate tautology (8.3) fails when we use the quantifiers with
restricted domain. We show now that
as follows. We take M = (N, I), where N is the set of real numbers and the
interpretation I is defined as QI :<, PI :>, cI : 0. The ”shorthand” solution is
as follows. The formula 8.19) becomes a mathematical statement
This statement is a false in the set N of natural numbers because the statement
n < 0 is false for all natural numbers and F ⇒ B is a true implication for
any logical value of B, so ∀n (n < 0 ⇒ n > 0) is a true statement and
∃n (n < 0 ∩ n > 0) is obviously false in the set N of natural numbers.
The restricted quantifiers law corresponding to the predicate tautology (8.3) is:
389
By definition 8.22 and restricted quantifiers transformations rules (8.7), (8.8)
proving (8.19 ) is means proving
S2 Let t be free for x in A(x). A(t) is a results from A(x) by replacing t for
all free occurrences of x in A(x), i.e. A(t) = A(x/t). Let s : V AR −→ U and
s0 be obtained from s by replacing s(x) by sI (t). We use (8.24) and induction
on the number of connectives and quantifiers in A(x) and prove
S3 Directly from definition 8.13 and (8.25) we get that for any M = (U, I) and
any s : V AR −→ U ,
390
This proves that (∀x A(x) ⇒ A(t)) is a predicate tautology. Observe that a
term x is free for x in A(x), so we also get as a particular case of t = x that
|= (∀x A(x) ⇒ A(x)).
Proof of (8.23) follows from (8.22), theorem 8.1, property 8.1, theorem 8.5,
and definability law (8.49). We carry it as follows. First we observe that by
theorem 8.1 we have that |= ((∀x¬A(x) ⇒ ¬A(t)) ⇒ (A(t) ⇒ ¬∀x¬A(x))) as a
substitution of propositional tautology ((a ⇒ ¬b) ⇒ (b ⇒ ¬a)). By just proved
(8.22) we have that |= (∀x¬A(x) ⇒ ¬A(t)) for A(x) being a formula ¬A(x).
By P 4 in property 8.1, we get |= (A(t) ⇒ ¬∀x¬A(x)). We apply the existential
quantifier definability law (8.49) and equivalence substitution theorem 8.5 and
get |= (A(t) ⇒ ∃x A(x)). This ends the proof of (8.22).
Remark the restrictions in (8.22) and (8.23) are essential. Here is a simple
example explaining why they are needed in (8.22). The example for (8.23) is
similar.
Let A(x) be a formula ¬∀y P (x, y), for P ∈ P. Notice that a term t = y is not
free for y in A(x). Consider (8.22) A(x) = ¬∀y P (x, y) and t = y.
(∀x¬∀y P (x, y) ⇒ ¬∀y P (y, y)), (8.26)
Take M = [N, I] for I such that PI : =. Obviously, M |= ∀x¬∀y P (x, y)
as ∀n¬∀n(n = n) is a true mathematical statement in the set N of natural
numbers. M 6|= ¬∀y P (y, y) as ¬∀n (n = n) is a false statement for n ∈ N .
Hence M is a counter model for for (8.26) and we proved that without the
restriction (8.22) does not hold.
Here are some useful and easy to prove properties of the notion ”t free for x in
A(x)” (definition 8.7).
391
|= ((B(x) ⇒ A) ⇒ (∃xB(x) ⇒ A)), (8.28)
Distributivity 1
For any formulas A(x), B(x), A, B of L, such that A , B does not contain any
free occurrences of x,
The restrictions that the formulas A, B do not contain any free occurrences of
x is essential for both Generalization and Distributivity 1 tautologies.
Here is a simple example explaining why they are needed in (8.29). The re-
laxation of the assumption that A, B do not contain any free occurrences of x
would lead to the following disaster. Let A and B be both atomic formula P(x).
Thus x is free in A and we have the following instance of (8.29).
Observe that ∀x(P (x) ⇒ P (x)) is a predicate tautology. Take M = [N, I] for
I such that PI = ODD, where ODD ⊆ N is the set of odd numbers. Let
s : V AR −→ N . By definition if I, sI (x) ∈ PI if and only if sI (x) ∈ ODD.
Then obviously (M, s) 6|= ∀x P (x) and M = [N, I] is a counter model for (8.29)
as (M, s) |= ∀x(P (x) ⇒ P (x)).
The examples for (8.30), (8.31), and (8.29) similar.
Distributivity 2
For any formulas A(x), B(x) of L,
392
Consider (8.35). We take as A(x), B(x) atomic formulas Q(x, c), P (x, c). The
particular case of (8.35) is now a formula
Take M = [R, I] where R is the set of real numbers, and the interpretation I is
QI :>, PI :<, cI : 0. The particular case formula becomes an obviously false
mathematical statement
Consider (8.36). We take as Let A(x), B(x) be atomic formulas Q(x, c), R(x, c).
The particular case of (8.36 ) is now a formula
Take M = (R, I) where R is the set of real numbers and QI :≥, RI :<, cI : 0.
The particular formula becomes an obviously false mathematical statement
De Morgan
For any formulas A(x), B(x) of L,
We prove (8.38) as an example.The proofs of all other laws are similar. As-
sume that (8.38) does not hold. By definition 8.16 there is M = (U, I) and
s : V AR −→ U, such that (M, s) |= ¬∀x¬A(x)) and (M, s) 6|= ∃x¬A(x).
Consider (M, s) |= ¬∀xA(x). By satisfaction definition 8.13, (M, s) 6|= ∀xA(x).
This holds only if for all s0 , such that s, s0 agree on all variables except on x,
(M, s0 ) 6|= A(x).
Consider (M, s) 6|= ∃x¬A(x). This holds only if there is no s0 , such that
(M, s0 ) |= ¬A(x), i.e. there is no s0 , such that (M, s0 ) 6|= A(x). This means
that for all s0 , (M, s0 ) |= A(x). Contradiction with (M, s0 ) 6|= A(x).
Quantifiers Alternations
For any formula A(x, y) of L,
393
The converse implications to (8.42) is not a predicate tautology. Take as A(x, y)
an atomic formulas R(x, y). Take M = (R, I) where R is the set of real numbers
and RI :< . The instance of (8.42) particular formula becomes a mathematical
statement
(∀y∃x(x < y) ⇒ ∃x∀y(x < y))
that obviously false in the set of real numbers. We proved
The most frequently used laws of quantifiers have a form of a logical equivalence,
symbolically written as ≡. This not a new logical connective. This is a very
useful symbol. It has the same properties as the equality = and can be used in
the same way we use the equality symbol =.
Note that we use the same equivalence symbol ≡ and the tautology symbol
|= for propositional and predicate languages and semantics when there is no
confusion. Formally we define the predicate equivalence as follows.
Definition 8.24
For any formulas A, B ∈ F of the predicate language L,
A ≡ B if and only if |= (A ⇔ B).
394
Directly from the theorem 8.3 and logical equivalence definition 8.23 we get that
the following is true.
Example 8.19
Consider the following propositional logical equivalence:
(a ⇒ b) ≡ (¬a ∪ b).
Substituting ∃xP (x, z) for a, and ∀yR(y, z) for b, we get from theorem 8.4 that
the following equivalence holds:
B1 = A1 (A/B).
If A ≡ B, then A1 ≡ B1 . (8.44)
Directly from the Dictum de Omi (8.22) and the Generalization (??) tautologies
we get the proof of the following theorem 8.6 useful for building new logical
equivalences from the old, known ones.
Theorem 8.6
For any formulas A(x), B(x) of L.
395
Example 8.20
We know from the example 8.19 that the formulas (∃xP (x, z) ⇒ ∀yR(y, z))
and (¬∃xP (x, z) ∪ ∀yR(y, z)) are logically equivalent. We get, as the direct
consequence of the theorem 8.6 the following equivalences:
Theorem 8.4 and theorem 8.6 allow us to use propositional tautologies and
predicate formulas to build predicate equivalences. Here is a simple example.
Exercise 8.6
Prove that for any formulas A(x), B(x) of L
Solution
By the substituting A(x) for a, and any formula B(x) for b, in the propositional
de Morgan Law: ¬(a ∪ b) ≡ (¬a ∩ ¬b), we get via theorem 8.4 that
We know, from the propositional logic, that for any propositional variables a, b,
a ≡ b if and only if ¬a ≡ ¬b. Substituting ∀x¬(A(x) ∪ B(x)) and ∀x(¬A(x) ∩
¬B(x)) for a and b, respectively, we get that
if and only if
¬∀x¬(A(x) ∪ B(x)) ≡ ¬∀x(¬A(x) ∩ ¬B).
But we have proved that ∀x¬(A(x)∪B) ≡ ∀x(¬A(x)∩¬B) holds, so we conclude
that the equivalence (8.45) also holds.
Exercise 8.7
Prove that for any formulas A(x), B of L
Solution
By the substituting A(x) for a, and any formula B for b, in the propositional
de Morgan law: ¬(a ∪ b) ≡ (¬a ∩ ¬b), we get that
396
Applying the theorem 8.4 to the above we obtain that
∀x¬(A(x) ∪ B) ≡ ∀x(¬A(x) ∩ ¬B).
As we can see, it is possible to obtain a fair amount of predicate tautologies
from the propositional tautologies and theorems 8.3, 8.4 and 8.6, but as we
have proved will for never obtain for example the most basic law: (∀xA(x) ⇒
∃xA(x)), any many the most important others.
We concentrate now only on these laws which have a form of a logical equiva-
lence.They are called the equational laws for quantifiers.
Directly from the definition 8.23 and the de Morgan tautologies (8.38)-(8.41)
we get one of the most important equational laws, called also De Morgan Laws.
De Morgan Laws
¬∀xA(x) ≡ ∃x¬A(x) (8.46)
¬∃xA(x) ≡ ∀x¬A(x) (8.47)
Now we will apply them to show that the quantifiers can be defined one by the
other i.e. that the following Definability Laws hold.
Definability Laws
∀xA(x) ≡ ¬∃x¬A(x) (8.48)
∃xA(x) ≡ ¬∀x¬A(x) (8.49)
The law (8.48) is often used as a definition of the universal quantifier in terms
of the existential one (and negation), the law (8.49) as a definition of the
existential quantifier in terms of the universal one (and negation).
Proof of (8.48)
Substituting any formula A(x) for a variable a in the propositional equivalence
a 𠪪a we get by theorem 8.4 that A(x) 𠪪A(x). Applying the theorem 8.6
to the above we obtain ∃xA(x) ≡ ∃x¬¬A(x). By the de Morgan Law (8.46)
∃x¬¬A(x) ≡ ¬∀x¬A(x) and hence ∃xA(x) ≡ ¬∀x¬A(x), what ends the proof.
Proof of (8.49)
We obtain ∀xA(x) ≡ ∀¬¬A(x) in a similar way as above. By the de Morgan
Law (8.47), ∀¬¬A(x) ≡ ¬∃¬A(x) and hence ∀xA(x) ≡ ¬∃¬A(x), what ends
the proof.
Other important equational laws are the following introduction and elimination
laws. We prove later the first two of them. We show that the laws (11.44) -
(10.39) can be deduced from laws (10.32) and (11.43), the de Morgan laws (8.46),
(8.47), definability laws (8.48), (8.49), propositional tautologies and theorems
8.3, 8.4, and theorem 8.5.
Introduction and Elimination Laws
If B is a formula such that B does not contain any free occurrence of x,
then the following logical equivalences hold.
397
∀x(A(x) ∪ B) ≡ (∀xA(x) ∪ B) (8.50)
∀x(A(x) ∩ B) ≡ (∀xA(x) ∩ B) (8.51)
∃x(A(x) ∪ B) ≡ (∃xA(x) ∪ B) (8.52)
∃x(A(x) ∩ B) ≡ (∃xA(x) ∩ B) (8.53)
∀x(A(x) ⇒ B) ≡ (∃xA(x) ⇒ B) (8.54)
∃x(A(x) ⇒ B) ≡ (∀xA(x) ⇒ B) (8.55)
∀x(B ⇒ A(x)) ≡ (B ⇒ ∀xA(x)) (8.56)
∃x(B ⇒ A(x)) ≡ (B ⇒ ∃xA(x)) (8.57)
398
we get, by the theorem 8.4 that ¬(∀x¬A(x) ∩ ¬B) ≡ (¬∀x¬A(x) ∪ ¬¬B). In a
similar way we prove that ¬¬B ≡ B, by the definability law (8.49) ¬∀x¬A(x) ≡
∃xA(x), hence by theorem 8.5 ¬∀x¬A(x) ∪ ¬¬B ≡ (∃xA(x) ∪ B) and finally,
∃x(A(x) ∪ B) ≡ (∃xA(x) ∪ B), what end the proof.
We can write this proof in a shorter, symbolic way as follows:
law 8.49
∃x(A(x) ∪ B) ≡ ¬∀x¬(A(x) ∪ B)
thm 8.3, 8.4
≡ ¬∀x(¬A(x) ∩ ¬B)
law 11.43
≡ ¬(∀x¬A(x) ∩ ¬B)
(8.46), thm 8.5
≡ (¬∀x¬A(x) ∪ ¬¬B)
thm 8.5
≡ (∃xA(x) ∪ B)
Distributivity Laws
Let A(x), B(x) be any formulas with a free variable x.
Law of distributivity of universal quantifier over conjunction
399
Here is a poof of first equality. The proof of the second one is similar and is left
as an exercise.
Example 8.21
The restricted quantifiers version of (11.43) is the following.
400
8.4 Hilbert Proof Systems Soundness and Com-
pleteness
We adopt now general definition from chapter 4 concerning proof systems to
the case of classical first order (predicate) logic.
We refer the reader to chapters 4 and 5 for a great array of example, exercises,
homework problems explaining in a great detail all notions we introduce here
for the predicate case. The examples and exercises we provide here are not
numerous and restricted to the laws of quantifiers.
Given a language L = L{¬,∩,∪,⇒,¬} (P, F, C). Any proof system
Remark 8.2
We use symbols |=p , Tp to stress the fact that we talk about predicate lah=nguage
and classical predicate tautologies.
401
Given an inference rule r ∈ R of the form
P1 ; P2 ; .... ; Pm
(r) ,
C
where P1 .P2 , . . . , Pm , C ∈ F.
(i) We say that the rule (r) is sound if and only if the following condition
holds for all structures M = [U, I] for L.
(ii) The rule (r) is not sound if and only if there is a structure M = [U, I],
such that
M |= {P1 , P2 , .Pm } and M |6 = C. (8.72)
In order to prove that the rule (r) is sound we have to show the implication
(8.71). It means, by definitions 9.3, 8.18, we have to show that that if all pre-
misses of the rule (r) are true in M = [U, I], so is its conclusion. This also
justifies correctness of the definition 8.25; sound rules do preserve the truth-
fulness as it is defined in our semantics.
Exercise 8.8
Prove the soundness of the rule
¬∀xA(x)
(r1) . (8.73)
∃x¬A(x)
Proof
Assume that the soundness condition (8.71) does not hold for for all structures
M = [U, I]. It means we assume that there is a structure M = [U, I], such that
M |= ¬∀xA(x) and M 6|= ∃x¬A(x).
Let M |= ¬∀xA(x). By definition 9.3, for all s : V AR −→ U we have
(M, s) |= ¬∀x¬A(x)). Hence by satisfaction definition 8.13, (M, s) 6|= ∀xA(x).
This holds only if for all s0 , such that s, s0 agree on all variables except on x,
(M, s0 ) 6|= A(x).
Observe that (M, s) 6|= ∃x¬A(x) only if there is no s0 , such that (M, s0 ) |=
¬A(x), i.e. there is no s0 , such that (M, s0 ) 6|= A(x). This means that for all s0 ,
(M, s0 ) |= A(x). Contradiction with (M, s0 ) 6|= A(x).
Exercise 8.9
Prove the soundness of the rule
∀xA(x)
(r2) . (8.74)
∃xA(x)
402
Proof
Assume that the soundness condition (8.71) does not hold for for all structures
M = [U, I]. It means we assume that there is a structure M = [U, I], such that
M |= ∀xA(x) and M 6|= ∃xA(x).
Let M |= ∀xA(x). By definition 9.3, for all s : V AR −→ U we have (M, s) |=
∀x¬A(x)).
By definition 8.13, (M, s) |= ∀x A(x) and (M, s) 6|= ∃x A(x). It means that
(M, s0 ) |= A(x) for all s0 such that s, s0 agree on all variables except on x, and
it is not true that there is s0 such that s, s0 agree on all variables except on x,
and (M, s0 ) |= A(x). This is impossible and this contradiction proves soundness
of (r2).
Exercise 8.10
Prove that the rule
∃xA(x)
(r3) . (8.75)
∀xA(x)
is not sound.
Proof
Observe that to prove that the rule (8.75) is not sound we have to provide an
example of an instance of a formula A(x) and construct prove (ii) of definition
8.25 for it.
Let A(x) be an atomic formula P (x, c), for any P ∈ P, #P = 2. We take as
M = (N, PI :<, cI : 3) for N set of natural numbers. Let s be any assignment
s : V AR −→ N. Obviously (M, s) |= ∃x P (x, c).
Take any s0 such that s0 (x) = 2 and s0 (y) = s(y) for all y ∈ V AR − {x}. We
have (2, 3) ∈ PI , as 2 < 3 and hence there exists s0 that agrees with s on all
variables except on x, and (M, s0 ) |= P (x, c). But (M, s) 6|= ∀x P (x, c) as for
example for s0 such that s0 (x) = 5 and s0 (y) = s(y) for all y ∈ V AR − {x},
(2, 3) 6∈ PI , as 5 6< 3.
This proves that M = (N, PI :<, cI : 3) is a model for (∃x P (x, c) and hence
6|= ∀x A(x)).
The ”shorthand” solution is: the formula (∃x P (x, c) becomes in M = (N, PI :<
, cI : 3) a true mathematical statement (written with logical symbols): ∃n n <
3. The formula (∀x P (x, c) becomes a mathematical frmula ∀n n < 3 which is
an obviously false statement in the set N of natural numbers, as there is n ∈ N ,
such that n < 3, for example n = 2, and it is not true that all natural numbers
are smaller then 3. So the rule (r3) is not sound.
403
An inference rule r ∈ R of the form
P1 ; P2 ; .... ; Pm
(r)
C
is strongly sound if the following condition holds for all structures M = [U, I]
for L.
P1 ∩ P2 ∩ . . . ∩ Pm ≡ C. (8.77)
Example 8.22
The sound rule (8.73)
¬∀xA(x)
(r1)
∃x¬A(x)
is strongly sound by De Morgan Law (8.46).
The sound rule (8.75)
∀xA(x)
(r2)
∃xA(x)
is not strongly sound by exercise 8.10.
We say that the proof system S is sound if the following conditions hold.
(1) LA ⊆ Tp ;
(2) Each rule of inference r ∈ R is sound.
The proof system S is strongly sound if the condition (2) is replaced by the
following condition (2’)
(2’) Each rule of inference r ∈ R is strongly sound under M.
404
When we define (develop) a proof system S our first goal is to make sure that it
a ”sound” one, i.e. that all we prove in it is true. Proving the following theorem
establishes this goal.
PS ⊆ Tp . (8.80)
Proof
Observe that if we have already proven that S is sound as stated in the defi-
nition 8.27, the proof of the implication (8.79) is straightforward mathematical
induction over the length of a proof.
It means that in order to prove the Soundness Theorem 8.7 for a proof system
Sit is enought to verify the two conditions of the definition 8.27 (1) LA ⊆ Tp
and (2) each rule of inference r ∈ R is sound.
We again refer the reader to chapter 4 for detailed examples, exercises and
problems.
As we can see, proving Soundness Theorem 8.7 for any proof system we develop
is indispensable and the proof is quite easy. The next step in developing a logic
(classical predicate logic in our case now) is to answer necessary and a difficult
question: Given a proof system S, about which we know that all it proves it
true (tautology). Can we prove all we know to be true (all tautologies)?
Proving the following theorem establishes this goal.
PS = Tp . (8.82)
405
Part 2: Completeness part of the Completeness Theorem: Tp ⊆ PS .
Proving the Soundness Theorem for S is usually a straightforward and not a very
difficult task. Proving the Completeness part of the Completeness Theorem is
always a crucial and very difficult task. There are many methods and techniques
for doing so, even for classical proof systems (logics) alone. Non-classical logics
often require new sometimes very sophisticated methods. We presented two
proofs of the Completeness Theorem for classical propositional Hilbert style
proof system in chapter 5, and a constructive proofs for automated theorem
proving systems for classical propositional logic the chapter 6.
We present a proof of the Completeness Theorem for predicate (first order) logic
in the next chapter 9.
A1 = R(x, y, g(c, x)), A2 = ∃xP (x, f (x, y)), A3 = ∃dR(x, y, g(c, d)),
A4 = ∀z(f (x, P (c, y)), A5 = ∃yP (x, f (c, y)) ∪ ∀yP (x, f (c, y)).
(a) Indicate whether they are, or are not well formed formulas of F. For
those which are not in F write a correct formula.
(b) For each correct, or corrected formula identify all components: con-
nectives, quantifiers, predicate and function symbols, and list all its
terms.
(c) For each formula identify its s free and bound variables. State which
are open and which are closed formulas (sentences), if any.
(d) Describe a language defined by the set F0 = {A1 , A2 , . . . A5 } of
formulas that are correct or corrected.
406
3. For each of the following formulas (some with restricted quantifiers) write
2 corresponding natural language sentences.
5. Justify that for any formula A ∈ F and any term t ∈ T the following facts
hold.
(a) A closed tern t, i.e. term with no variables is free for any variable x
in A.
(b) A term t is free for any variable in A if none of the variables in t is
bound in A.
(c) Term t = x is free for x in any formula A.
(d) Any term is free for x in A if A contains no free occurrences of x.
407
(d) Birds can fly and if anyone can fly Tweety can.
(e) Anyone who knows logic loves it.
Classical Semantics
5. Show that the following formulas are predicate tautologies for any formulas
A, B in L.
(a) (∀x∀yA(x, y) ⇒ ∀y∀xA(x, y)).
(b) (∃x∃yA(x, y) ⇒ ∃y∃xA(x, y)).
408
(c) (∀x(A(x) ⇒ B(x)) ⇒ (∀xA(x) ⇒ ∀xB(x))).
6. Prove that the following formulas are not predicate tautologies by finding
their proper instances and constructing counter models for them.
7. Prove that the following formulas are predicate tautologies for any for-
mulas A(x), B(x), A, B of L, such that A , B does not contain any free
occurrences of x.
9. Prove that the converse implication to the formulas listed below are pred-
icate tautologies for any formulas A(x), B(x), A, B of L, such that A, B
does not contain any free occurrences of x.
409
410
Chapter 9
There are several quite distinct approaches to the Completeness Theorem, cor-
responding to the ways of thinking about proofs. Within each of the approaches
there are endless variations in exact formulation, corresponding to the choice of
methods we want to use to proof the Completeness Theorem. Different basic
approaches are important, though, for they lead to different applications. We
have presented two of the approaches for the propositional logic: Hilbert style
formalizations (proof systems) in chapter 5, and Gentzen style formalizations
(automated proof systems) in chapter 6. We have also presented for each of the
approaches methods of proving the completeness theorem. Two proofs of com-
pleteness theorem for Hilbert style proof system in chapter 5 and a constructive
proofs for several Gentzen style proof systems in chapter 6.
There are many proofs of the Completeness Theorem for predicate (first order)
logic. We present here in a great detail, a version of Henkin’s proof as included
in a classic Handbook of Mathematical Logic (1977). It contains a method for re-
ducing certain problems of first-order logic back to problems about propositional
logic. We give independent proof of Compactness Theorem 9.1 for propositional
logic. Reduction to Propositional Logic Theorem 9.2, Compactness Theorem 9.3
for first-order logic, Löwenheim-Skolem Theorem 9.4 and Gödel Completeness
Theorem 9.7 fall out of the Henkin method.
We choose this particular proof of completeness of first order logic not only for
it being one of the oldest and most classical, but also for its connection with
the propositional logic. Moreover, the proof of the Compactness Theorem 9.1 is
based on semantical version of syntactical notions and techniques crucial to the
411
second proof of completeness theorem for propositional logic covered in chapter
5 and hence is familiar to the reader.
The set P ⊆ F, called a set of all prime formulas of L plays in the propositional
logic we define the role
Example 9.1
The following are primitive formulas.
R(t1 , t2 ), ∀x(A(x) ⇒ ¬A(x)), (c = c), ∃x(Q(x, y) ∩ ∀yA(y)).
The following are not primitive formulas.
(R(t1 , t2 ) ⇒ (c = c)), (R(t1 , t2 ) ∪ ∀x(A(x) ⇒ ¬A(x)).
412
Let F, P be sets of all formulas and prime formulas (9.1) of L, respectively.
The smallest set P F ⊆ F such that
(i) P ⊆ P F,
(ii)If A, B ∈ P F, then (A ⇒ B), (A ∪ B), (A ∩ B), and ¬A ∈ P F
is called a set of all propositional formulas of the predicate language L.
The set P is called atomic propositional formulas of L.
v : P −→ {T, F }
413
For the sake of simplicity we will often say model, tautology instead propositional
model, propositional tautology for L.
Proof
Assume that S is a consistent set. By definition 9.7, it has a model. Tts model
is also a model for all its subsets, including all finite subsets, and so all its finite
subsets are consistent.
To prove the converse implication, i.e. the nontrivial half of the Compactness
Theorem we write it in a slightly modified form. To do so, we introduce the
following definition.
414
S3 We show that every finitely consistent set S can be extended to a maximal
finitely consistent set S ∗ . I.e we show that for every finitely consistent set S
there is a set S ∗ , such that S ⊆ S ∗ and S ∗ is maximal finitely consistent.
S4 We use S2 and S3 to justify the following reasoning.
Given a finitely consistent set S. We extend it, via construction to be defined in
the step S3 to a maximal finitely consistent set S ∗ . By the S2, S ∗ is consistent
and hence so is the set S.
This ends the proof of the Compactness Theorem 9.1.
Here are the details and proofs needed for completion of steps S1 - S4.
Step S1
We introduce the following definition.
We use notation MFC for maximal finitely consistent set, and FC for the
finitely consistent set.
Step S2
We prove the following MFC lemma 6.3 and the Property 9.1.
Lemma 9.1
Any MFC set is consistent.
Proof
Given a MFC set denoted by S ∗ . We prove its consistency by constructing
model for it, i.e. by constructing a truth assignment v : P −→ {T, F }, such
that for all A ∈ S ∗ , v ∗ (A) = T .
Observe that directly from the definition 9.10 we have the following property of
the the MFC sets.
Definition 9.11
For any MFC set S ∗ , mapping v : P −→ {T, F }, such that
415
if P ∈ S ∗
T
v(P ) =
F 6 S∗
if P ∈
is called a truth assignment defined by S ∗ .
We extend v to v ∗ : P F −→ {T, F } in a usual way.
if A ∈ S ∗
T
v ∗ (A) =
F 6 S∗
if A ∈
if ¬A ∈ S ∗
∗ T
v (¬A) =
F 6 S∗
if ¬A ∈
416
The remaining cases of A = (B ∩ C), A = (B ⇒ C) are similar to the above and
are left to the reader as an exercise.
This end the proof of lemma 9.1 and completes the step S2.
S3: Maximal finitely consistent extension
Given a finitely consistent set S, we construct its maximal finitely consistent
extension S ∗ as follows.
The set of all formulas of L is countable, so is P F. We assume that all propo-
sitional formulas form a one-to-one sequence
We define a chain
S0 ⊆ S1 ⊆ S2 .... ⊆ Sn ⊆ .... (9.3)
of extentions of the set S by
S0 = S;
Sn ∪ {An } if Sn ∪ {An } is finitely consistent
Sn+1 =
Sn ∪ {¬An } otherwise.
We take
[
S∗ = Sn . (9.4)
n∈N
417
Case 2 Sn+1 = Sn ∪ {¬An }. Observe that this can happen only if Sn ∪ {An }
0 0
is not FC, i.e. there is a finite subset Sn ⊆ Sn , such that Sn ∪ {An } is not
consistent.
Suppose now that Sn+1 is not FC. This means that there is a finite subset
00 00
Sn ⊆ Sn , such that Sn ∪ {¬An } is not consistent.
0 00
Take Sn ∪ Sn . It is a finite subset of Sn so is consistent by the inductive
0 00
assumption. Let v be a model of Sn ∪ Sn . Then one of v ∗ (A), v ∗ (¬A)
0
must be T. This contradicts the inconsistency of both Sn ∪ {An } and
0
Sn ∪ {¬An }.
(∃xA(x) ∪ ¬∃xA(x)),
(∀xA(x) ∪ ¬∀xA(x)),
but the following are predicate (first order) tautologies (valid formulas) that are
not propositional tautologies:
∀x(A(x) ∪ ¬A(x)),
(¬∀xA(x) ⇒ ∃x¬A(x)).
The first formula above is just a prime formula, the second is of the form (¬B ⇒
C), for B and C prime.
To stress the difference between the propositional and predicate (first order)
tautologies some books reserve the word tautology for the propositional tau-
tologies alone, using the notion of valid formula for the predicate (first order)
tautologies. We use here both notions, with the preference to predicate tautology
or tautology for short when there is no room for misunderstanding.
418
To make sure that there is no misunderstandings we remind the following defi-
nitions from chapter 8.
Given a first order language L with the set of variables V AR and the set of
formulas F. Let M = [M, I] be a structure for the language L, with the
universe M and the interpretation I and let s : V AR −→ M be an assignment
of L in M . We bring back some basic definitions from Chapter 8
A is satisfied in M
Given a structure M = [M, I], we say that a formula A is satisfied in M
if there is an assignment s : V AR −→ M such that
(M, s) |= A.
A is true in M
Given a structure M = [M, I], we say that a formula A is true in M if
(M, s) |= A for all assignments s : V AR −→ M .
Model M
If A is true in a structure M = [M, I], then M is called a model for A.
We denote it as
M |= A.
|= A.
Case: A is a sentence
If A is a sentence, then the truth or falsity of (M, s) |= A is completely
independent of s. Thus we write
M |= A
419
Predicate and Propositional Models
The relationship between the predicate models that are defined in terms of
structures M = [M, I] and assignments s : V AR −→ M and propositional
models that are defined in terms of truth assignments v : P −→ {T, F } is
established by the following lemma.
Lemma 9.2
Let M = [M, I] be a structure for the language L and let s : V AR −→ M an
assignment in M. There is a truth assignments v : P −→ {T, F } such that for
all formulas A of L,
(M, s) |= A if and only if v ∗ (A) = T.
In particular, for any set S of sentences of L,
if M |= S then S is consistent in sense of propositional logic.
All formulas of S are different prime formulas, S is hence consistent in the sense
of propositional logic and obviously has no predicate (first-order) model.
420
For any free variable or constant of L, i.e for any u, w, ui , wi ∈ (V AR ∪ C),
E1 u = u,
E2 (u = w ⇒ w = u),
E3 ((u1 = u2 ∩ u2 = u3 ) ⇒ u1 = u3 ),
Observe that given any structure M = [M, I]. We have by simple verification
that for all s : V AR −→ M , and for all A ∈ {E1, E2, E3, E4, E5},
(M, s) |= A.
Fact 9.1
All equality axioms are predicate tautologies (valid) of L.
This is why we still call logic with equality axioms added to it, a logic.
Definition 9.13
For any predicate language L = L(P, F,C), the language
for the set C defined by (10.11) and L(C) defined by (9.9) and the construction
described below is called a witnessing expansion of L. We write also
L(C) = L ∪ C.
421
Construction of the witnessing expansion of L
We define the set C of new constants by constructing an infinite sequence
C0 = ∅, L0 = L ∪ C0 = L.
We denote by
A[x]
the fact that the formula A has exactly one free variable and for each such a
formula we introduce a distinct new constant denoted by
cA[x] .
We define
C1 = {cA[x] : A[x] ∈ L0 }, L1 = L ∪ C1 .
and
L(C) = L ∪ C. (9.9)
422
The axiom H1 says:
If ∃xA(x) is true in a structure, choose an element a satisfying A(x) and give
it a new name cA[x] .
The axiom H2 says:
If ∀xA(x) is false, choose a counterexample b and call it by a new name c¬A[x] .
Observe that the quantifiers axioms Q1, Q2 obviously are predicate tautologies.
The Henkin is obviously not true in every L(C)-structure, but we are going to
show that every L -structure can be turned into an L(C)-structure which is a
model of SHenkin . Before we do so we need to introduce two new notions.
0 0
Thus the reduct and the expansion M and M are the same except that M
0
assigns meanings to the symbols in L − L .
Lemma 9.3
Let M = [M, I] be any structure for the language L and let L(C) be the wit-
0 0
nessing expansion of L. There is an expansion M = [M, I ] of M = [M, I]
0
such that M is a model of the set SHenkin
423
0
Proof In order to define the expansion of M to M we have to define the
0
interpretation I for the symbols of the language L(C) = L ∪ C, such that
0
I | L = I. This means that we have to define cI 0 for all c ∈ C. By the
definition, cI 0 ∈ M , so this also means that we have to assign the elements of
M to all constants c ∈ C in such a way that the resulting expansion is a model
for all sentences from SHenkin .
The quantifier axioms (definition 9.15) are predicate tautologies so they are
going to be true regardless, so we have to worry only about the Henkin axioms
(definition 9.14). Observe now that if the lemma 9.3 holds for the Henkin axiom
H1, then it must hold for the axiom H2. Namely, let’s consider the axiom H2:
(A(c¬A[x] ) ⇒ ∀xA(x)).
0 0
Assume that A(c¬A[x] ) is true in the expansion M , i.e. that M |= A(c¬A[x] )
0 0
and that M 6|= ∀xA(x). This means that M |= ¬∀xA(x) and by the de Morgan
0 0
Laws, M |= ∃x¬A(x). But we have assumed that M is a model for H1. In
0 0
particular M |= (∃x¬A(x) ⇒ ¬A(c¬A[x] )), and hence M |= ¬A(c¬A[x] ) and
0 0
this contradicts the assumption that M |= A(c¬A[x] ). Thus if M is a model
for all axioms of the type H1, it is also a model for all axioms of the type H2.
S
We define cI 0 for all c ∈ C = Cn by induction on n. Let n = 1 and cA[x] ∈ C1 .
By definition, C1 = {cA[x] : A[x] ∈ L}. In this case we have that ∃xA(x) ∈ L and
hence the notion M |= ∃xA(x) is well defined, as M = [M, I] is the structure
for the language L.
As we consider arbitrary structure M, there are two possibilities: M |= ∃xA(x)
or M 6|= ∃xA(x).
We define cI 0 , for all c ∈ C1 as follows.
If M |= ∃xA(x), then (M, v 0 ) |= A(x) for certain v 0 (x) = a ∈ M . We set
(cA[x]) )I 0 = a. If M 6|= ∃xA(x), we set (cA[x]) )I 0 arbitrarily.
This makes all the positive Henkin axioms about the cA[x] ∈ C1 true, i.e. M =
(M, I) |= (∃xA(x) ⇒ A(cA[x] )). But once cA[x] ∈ C1 are all interpreted in M ,
0
then the notion M |= A is defined for all formulas A ∈ L ∪ C1 . We carry the
same argument and define cI 0 , for all c ∈ C2 and so on. The inductive step in
the exactly the same way as the one above.
Now we are ready to state and proof a lemma 9.2 that provides the essential
424
step in the proof of the completeness theorem for predicate logic.
(i) S has a model, i.e. there is a structure M = [M, I] for the language L
such that M |= A for all A ∈ S.
(ii) There is a canonical L(C) structure M = [M, I] which is a model for
S, i.e. such that M |= A for all A ∈ S.
(iii) The set S ∪ SHenkin ∪ EQ is consistent in sense of propositional logic,
where EQ denotes the equality axioms E1 − E5.
Proof The implication (ii) → (i) is immediate. The implication (i) → (iii)
follows from lemma 9.3. We have to prove only the implication (iii) → (ii).
Assume that the set S ∪ SHenkin ∪ EQ is consistent in sense of propositional
logic and let v be a truth assignment to the prime sentences of L(C), such that
v ∗ (A) = T for all A ∈ S ∪ SHenkin ∪ EQ. To prove the lemma, we construct a
canonical L(C) structure M = [M, I] such that, for all sentences A of L(C),
425
(1.) Definition of the universe M of M.
In order to define the universe M we first define a relation ≈ on C by
The equality axioms axioms guarantee that the relation (9.13) is equivalence re-
lation on C, i.e. is reflexive, symmetric, and transitive. All axioms are predicate
tautologies, so v(c = d)) = T by axiom E1 and c ≈ c holds for any c ∈ C.
Symmetry condition ” if c ≈ d, then d ≈ c ” holds by axiom E2. Assume c ≈ d,
by definition v(c = d)) = T . By axiom E2
v ∗ (((c = d ∩ d = e) ⇒ c = e)) = T.
v ∗ ((c = d ∩ d = e) ⇒ c = e) = (T ∩ T ⇒ c = e) = (T ⇒ c = e) = T,
(2.) Definition of RI ⊆ M n .
if [c1 ] = [d1 ], [c2 ] = [d2 ], . . . , [cn ] = [dn ] and ([c1 ], [c2 ], . . . , [cn ]) ∈ RI ,
426
By the assumption [c1 ] = [d1 ], . . . , [cn ] = [dn ] we have that v(c1 = d1 ) =
T, . . . , v(cn = dn ) = T . By the assumption ([c1 ], [c2 ], . . . , [cn ]) ∈ RI , we have
that v(R(c1 , ..., cn )) = T . Hence the condition (9.16) becomes
(3.) Definition of fI : M n → M .
Let c1 , c2 , . . . , cn ∈ C and f ∈ F. We claim that there is c ∈ C such that
f (c1 , c2 , . . . , cn ) = c and v(f (c1 , c2 , . . . , cn ) = c) = T .
For consider the formula A(x) given by f (c1 , c2 , . . . , cn ) = x. If v ∗ (∃xA(x)) =
v ∗ (f (c1 , c2 , . . . , cn ) = x) = T , we want to prove v ∗ (A(cA[x] )) = T , i.e.
v(f (c1 , c2 , . . . , cn ) = cA ) = T.
The argument similar to the one used in (2.) proves that fI is well defined.
(4.) Definition of cI ∈ M .
For any c ∈ C we take cI = [c]. If d ∈ C, then an argument similar to that used
on (3.) shows that there is c ∈ C such that v(d = c) = T , i.e. d ≈ c, so we put
dI = [c].
This completes the construction of the canonical structure M = [M, I] and guar-
antees that (9.10) holds for for all atomic propositional sentences (definition
9.2), i.e. we proved that
427
To complete the proof of the Lemma 9.2 we prove that the property (9.10) holds
for the canonical structure M = [M, I] defined above and all other sentences.
We carry the proof by induction on length of formulas. The case of propositional
connectives is trivial. For example, M |= (A ∩ B) if and only if M |= A
and M |= B) ( follows directly from the satisfaction definition) if and only
if v ∗ (A) = T and v ∗ (B) = T (by the induction hypothesis) if and only if
v ∗ (A ∩ B) = T . We proved
M |= (A ∩ B) if and only if v ∗ (A ∩ B) = T,
for all sentences A, B of L(C). The proof for all other connectives is similar.
We prove now the case of a sentence B of the form ∃xA(x), i.e. we want to show
that
M |= ∃xA(x) if and only if v ∗ (∃xA(x)) = T. (9.18)
v ∗ (∃xA(x)) = T . Then there is a c such that v ∗ (A(c) = T , so by induction
hypothesis, M |= A(c) so M |= ∃xA(x).
On the other hand, if v ∗ (∃xA(x)) = F , then by SHenking quantifier axiom Q2
(definition 9.15) we have that v ∗ (A(t)) = F for all closed terms t of L(C). In
particular, for every c ∈ C v ∗ (A(c)) = F . By induction hypothesis, M |=
¬A(c), for all c ∈ C. Since every element of M is denoted by some c ∈ C,
M |= ¬∃xA(x). Thus we proved (9.18).
The proof of the case of a sentence B of the form ∀xA(x) is similar and is left
to the reader.
The Reduction to Propositional Logic Theorem 9.2 provides not only a method
of constructing models of theories out of symbols, but also gives us immediate
proofs of the Compactness Theorem 9.3 for the predicate logic and Lowenheim-
Skolem Theorem 9.4.
Proof
Let S be a set of predicate formulas such that every finite subset S0 of S has
a model. We need to show that S has a model. By the implication (iii) → (i)
of the Theorem 9.2 this is equivalent to proving that S ∪ SHenkin ∪ EQ is
consistent in the sense of propositional logic. By the Compactness Theorem 9.1
for propositional logic of L, it suffices to prove that for every finite subset S0 ⊂ S,
S0 ∪ SHenkin ∪ EQ is consistent, which follows from the hypothesis and the
implication (i) → (iii) of the Reduction to Propositional Logic Theorem 9.2.
428
Let κ be an infinite cardinal and let Γ be a set of at most κ formulas of the first
order language.
If the set S has a model, then there is a model M = [M, I] of S such that
cardM ≤ κ.
Proof Let L be a predicate language with the alphabet A such that card(A) ≤
S card(F) ≤ κ. By the definition of the witnessing expansion L(C)
κ. Obviously,
of L, C = n Cn and for each n, card(Cn ) ≤ κ. So also cardC ≤ κ. Thus any
canonical structure for L(C) has ≤ κ elements. By the implication (i) → (ii) of
the Reduction to Propositional Logic Lemma 9.2 there is a model of S (canonical
structure) with ≤ κ elements.
Language L
The language L of the proof system H is a predicate (first order) language
with equality (definition 9.12). We assume that the sets P, F, C are infinitely
enumerable. We also assume that it has a full set of propositional connectives,
i.e.
L = L{¬,∩,∪,⇒} (P, F, C). (9.19)
Logical Axioms LA
The set LA of logical axioms consists of three groups of axioms: propositional
axioms P A, equality axioms EA, and quantifiers axioms QA. we write it sym-
429
bolically as
LA = {P A, EA, QA}.
For the set P A of propositional axioms we choose any complete set of axioms
for propositional logic with a full set {¬, ∩, ∩, ⇒} of propositional connectives.
In some formalizations, including the one in the Handbook of Mathematical
Logic, Barwise, ed. (1977) we base our system H on, the authors just say for this
group of axioms: ”all tautologies”. They of course mean all predicate formulas
of L that are substitutions of propositional tautologies. This is done for the
need of being able to use freely these predicate substitutions of propositional
tautologies in the proof of completeness theorem for the proof system they
formalize this way.
In this case these tautologies are listed as axioms of the system and hence are
provable in it. This is a convenient approach, but also the one that makes such
a proof system not to be finately axiomatizable.
We avoid the infinite axiomatization by choosing a proper finite set of pred-
icate language version of propositional axioms that is known (proved already
for propositional case) to be complete, i.e. the one in which all propositional
tautologies are provable.
We choose, for name H (Hilbert) and historical sake, the set of Hilbert (1928)
axioms from chapter 5.
.For the set EA of equational axioms we choose the same set (9.5) as in sec-
tion 9.1.1 because they were used in the proof of Reduction to Propositional
Logic Theorem 9.2 and we want to be able to carry this proof within the system
H.
For the set QA of quantifiers axioms we choose the axioms such that the Henkin
set SHenkin axioms Q1, Q2 are their particular cases, so again a proof of the
Reduction to Propositional Logic Theorem 9.2 can be carried within H.
Rules of inference R
There are three inference rules: Modus Ponens (M P ) and two quantifiers rules
(G), (G1), (G2), called Generalization Rules.
where
L = L{¬,∩,∪,⇒} (P, F, C) is predicate (first order) language with equality (defi-
nition 9.12). We assume that the sets P, F, C are infinitely enumerable.
430
F is the set of all well formed formulas of L.
LA is the set of logical axioms and
431
QA is the set of quantifiers axioms.
Q1 (∀xA(x) ⇒ A(t)),
Q2 (A(t) ⇒ ∃xA(x)),
where where t is a term, A(t) is a result of substitution of t for all free occur-
rences of x in A(x), and t is free for x in A(x), i.e. no occurrence of a variable
in t becomes a bound occurrence in A(t).
where
(M P ) is Modus Ponens rule
A ; (A ⇒ B)
(M P ) ,
B
for any formulas A, B ∈ F.
(G), (G1), (G2) are the following quantifiers generalization rules.
A
(G) ,
∀xA
where A ∈ F. In particular we write
A(x)
(G)
∀xA(x)
(B ⇒ A(x))
(G1) ,
(B ⇒ ∀xA(x))
where A(x), B ∈ F, x ∈ V AR, and B is such that x is not free in B.
(A(x) ⇒ B)
(G2) ,
(∃xA(x) ⇒ B)
where A(x), B ∈ F , x ∈ V AR, and B is such that x is not free in B.
432
Definition 9.18 (Proof from Γ)
Let Γ ⊆ F be any set of formulas of L. A proof in H of a formula A ∈ F from
s set Γ of formulas is a sequence
B1 , B2 , . . . Bn
Γ ` A,
`H A
Γ ` A and ` A
433
Theorem 9.5 (Soundness Theorem)
For every formula A ∈ F of the language L of the proof system H,
if ` A, then |= A.
The soundness theorem proves that the proofs in the system ”produce” only
tautologies. We show here, as the next step that our proof system H ”produces”
not only tautologies, but that all tautologies are provable in it.
This is called a completeness theorem for classical predicate (first order logic,
as it all is proven with respect to classical semantics. This is why it is called a
completeness of predicate logic theorem.
The goal is now to prove the completeness part of the following.
We are going to prove the Gödel’ s Theorem 9.6 as a particular case of The-
orem 9.7 that follows. It is its more general, and more modern version. This
version, as well as the method of proving it, was first introduced by Henkin in
1947. It became with its consequent improvements, as classical as the Gödel’s
own. It uses the notion of a logical implication, and some other notions. We
introduce them below.
Directly from the definition 10.16 have that the following hold.
434
Observe, that in order to prove that Γ |= B we have to show that the implication
if M |= Γ then M |= B
Proof
Let B1 , B2 , . . . Bn be a proof of A(x1 , . . . xn ) from Γ and let M be a model of Γ.
We use Fact 9.2 and prove by induction on n, that M |= ∀x1 ∀x2 . . . ∀xn Bn (x1 , . . . xn ),
and hence M |= ∀x1 ∀x2 . . . ∀xn A(x1 , . . . xn ). The converse implication is obvi-
ous.
Fact 9.2 and Lemma 9.4 show that we need to consider only sentences (closed
formulas) of L, since a formula of F is a tautology if and only if its closure is
a tautology and is provable from Γ if and only if its closure is provable from Γ.
This justifies the following generalization of the original Gödel’ s completeness
of predicate logic Theorem 9.6.
Remark
We want to remind the readers that the Reduction Predicate Logic to Proposi-
tional Logic Section 9.1 is an integral and the first part of the proof the Gödel
Completeness Theorem 9.7. We presented it separately for two reasons.
R1. The reduction method and theorems and their proofs are purely semantical
in their nature and hence are independent of the proof system H.
R2. Because of R1. the reduction method can be used/adapted to a proof of
completeness theorem of any other proof system one needs to prove the classical
completeness theorem for. See section ??.
In order to prove it we must formulate it properly so we need to introduce
few new important and classical notion and prove some lemmas needed for the
proof. The first is the notion of consistency.
435
There are two definitions of consistency; semantical and syntactical. The
semantical one uses definition the notion of a model and says, in plain English:
a set of formulas is consistent if it has a model.
The syntactical one uses the notion of provability and says: a set of formulas
is consistent if one can’t prove a contradiction from it.
We have used, in the Proof Two of the Completeness Theorem for propositional
logic (chapter 5) the syntactical definition of consistency. We use here the
following semantical definition.
Lemma 9.5
For any set Γ ⊆ F of formulas of L and any A ∈ F,
if Γ |= A, then the set Γ ∪ {¬A} is inconsistent.
Proof
Assume Γ |= A and Γ ∪ {¬A} is consistent. By definition 9.21 there is a struc-
ture M = [U, I], such that M |= Γ and M |= ¬A, i.e. M 6|= A. This is a
contradiction with Γ |= A.
Now we are going to prove the following Lemma 9.6 that is crucial, together with
the Reduction to Propositional Logic Theorem 9.2 and the above Lemma 9.5 to
the proof of the Completeness Theorem 9.7.
Lemma 9.6
Let Γ be any set of sentences of a language L of Hilbert proof system H.
The following conditions hold For any formulas A, B ∈ F of L.
(i) If Γ ` (A ⇒ B) and Γ ` (¬A ⇒ B), then Γ ` B.
(ii) If Γ ` ((A ⇒ C) ⇒ B), then Γ ` (¬A ⇒ B) and Γ ` (C ⇒ B).
(iii) If x does not appear in B and if Γ ` ((∃yA(y) ⇒ A(x)) ⇒ B), then Γ ` B.
(iv) If x does not appear in B and if Γ ` ((A(x) ⇒ ∀yA(y)) ⇒ B), then Γ ` B.
Proof
(i) Notice that the formula ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)) is a substitution
436
of a propositional tautology, hence by definition of H, is provable in it. By
monotonicity,
Γ ` ((A ⇒ B) ⇒ ((¬A ⇒ B) ⇒ B)).
By assuption Γ ` (A ⇒ B) and Modus Ponens we get
Γ ` ((¬A ⇒ B) ⇒ B).
Proof
We first prove the completeness part (9.24), i.e. we prove the implication
if Γ |= A, then Γ ` A. (9.26)
437
Let M |= Γ. We construct, as a next step, a witnessing expansion language L(C)
of L (definition 9.13). By the Reduction to Propositional Logic Theorem 9.2,
the set Γ ∪ SHenkin ∪ EQ is consistent in a sense of propositional logic in L. The
set SHenkin is a Henkin Set (definition 9.16) and EQ are equality axioms (9.5)
that are also the equality axioms EQ of H.
By the Compactness Theorem 9.1 for propositional logic of L there is a finite
set S0 ⊆ Γ ∪ SHenkin ∪ EQ such that S0 ∪ {¬A} is inconsistent in the sense of
propositional logic.
We list all elements of S0 in a sequence
A1 , A2 , . . . , An , B1 , B2 , . . . , Bm (9.27)
is a propositional tautology.
We now replace each witnessing constant in this sentence by a distinct new
variable and write the result as
438
. We have A0 = A since A has no witnessing constant in it. The result is still a
tautology and hence is provable in H from propositional axioms P A and Modus
Ponens. By monotonicity
. We hence apply parts (iii) ad (iv) of Theorem 9.2 to successively remove all
B1 0 , ...., Bm 0 and obtain the proof of A from S0 .
This ends the proof that Γ ` A and hence the proof of the completeness part
of (9.24).
if Γ ` A, then Γ |= A,
holds for any sentence A of L directly by Fact 9.2, Lemma 9.4, and Theorem 9.5.
The Theorem 9.6, as expressed by (9.25) follows from Fact 9.2, Lemma 9.4 as a
case of (9.24) for Γ = ∅.
This ends the proof of Theorem 9.8 as well as Theorem 9.7, and the proof of
the original Gödel Completeness of Predicate Logic Theorem 9.6.
439
9.3 Deduction Theorem
In mathematical arguments, one often assumes a statement A on the assumption
(hypothesis) of some other statement B and then concludes that we have proved
the implication ”if A, then B”. This reasoning is justified by the following
theorem, called a Deduction Theorem. It was first formulated and proved for a
certain Hilbert proof system S for the classical propositional logic by Herbrand
in 1930 in a form stated below.
if A `S B, then `S (A ⇒ B).
In particular,
A `H1 B if and only if `H1 (A ⇒ B).
A natural question arises: does deduction theorem holds for the predicate logic
in general and for its proof system H we defined here?.
The Theorem 9.10 cannot be carried directly to the predicate logic, but it nev-
ertheless holds with some modifications. Here is where the problem lays.
Proof
Obviously, A(x) ` ∀xA(x) by Generalization rule (G). Let now A(x) be an
atomic formula P (x). By the Completeness Theorem 9.6, ` (P (x) ⇒ ∀xP (x))
if and only if |= (P (x) ⇒ ∀xP (x)). Consider a structure M = [M, I], where
M contains at least two elements c and d. We define PI ⊆ M as a property
that holds only for c, i.e. PI = {c}. Take any assignment of L in M, i.e.
440
s : V AR −→ M . Then (M, s) |= P (x) only when s(x) = c for all x ∈ V AR.
M = [M, I] is a counter model for (P (x) ⇒ ∀xP (x)), as we found s such
(M, s) |= P (x) and obviously (M, s) 6|= ∀xP (x). This proves that Deduction
Theorem fails for A(x) being an atomic formula P (x).
The Fact 9.3 shows that the problem is with application of the generalization rule
(G) to the formula A ∈ Γ. To handle this we introduce, after Mendelson(1987)
the following notion.
Definition 9.22
Let A be one of formulas in Γ and let
B1 , B2 , ..., Bn (9.34)
a deduction (proof of Bn from Γ, together with justification at each step.
We say that the formula Bi depends upon A in the proof (9.34)
if and only if
(1) Bi is A and the justification for Bi is Bi ∈ Γ
or
(2) Bi is justified as direct consequence by MP or (G) of some preceding for-
mulas in the sequence (9.34), where at least one of these preceding formulas
depends upon A.
Here is a deduction
B1 , B2 , . . . , B5 (9.35)
showing that
A, (∀xA ⇒ C).
B1 A, Hyp
B2 ∀xA, B1 , (G)
B3 (∀xA ⇒ C), Hyp
B4 C, MP on B2 , B3
B5 ∀xC. (G)
Observe that the formulas A, C may, or may not have x as a free variable.
Example 9.2
In the derivation (9.35)
B1 depends upon A,
B2 depends upon A,
B3 depends upon (∀xA ⇒ C),
B4 depends upon A and (∀xA ⇒ C),
B5 depends upon A and (∀xA ⇒ C).
441
Lemma 9.7
If B does not depend upon A in a deduction showing that Γ, A ` B, then
Γ ` B.
Proof
Let B1 , B2 , . . . , Bn = B be a deduction of B from Γ, A in which B does not
depend upon A. we prove by Induction that Γ ` B. Assume that Lemma 9.7
holds for all deductions of the length less than n. If B ∈ Γ or B ∈ LA, then
Γ ` B. If B is a direct consequence of two preceding formulas, then, since B
does not depend upon A , neither do theses preceding formulas. By inductive
hypothesis, theses preceding formulas have a proof from Γ alone. Hence so does
B.
Now we are ready to formulate and prove the Deduction Theorem 9.11 for
predicate logic.
Γ, A ` B,
Γ ` (A ⇒ B).
Proof
The proof extends the proof of the Deduction Theorem for propositional logic
from chapter 5. We adopt the propositional proof (for a different proof system)
to the system H and adding the predicate case. For the sake of clarity and
independence we write now the whole proof in all details.
(1) Assume that Γ, A `B, i.e. that we have a formal proof
B1 , B2 , ..., Bn (9.36)
Γ ` (A ⇒ B).
442
The proof of S is conducted by induction on i ( 1 ≤ i ≤ n).
Base Step i = 1.
When i = 1, it means that the formal proof (5.6) contains only one element B1 .
By the definition of the formal proof from Γ ∪ {A}, we have that B1 ∈ LA, or
B1 ∈ Γ, or B1 = A, i.e.
B1 ∈ LA ∪ Γ ∪ {A}.
Here we have two cases.
Case 1. B1 ∈ LA ∪ Γ.
Observe that the formula is a particular case of A2 of H. By assumption B1 ∈
LA ∪ Γ, hence we get the required proof of (A ⇒ B1 ) from Γ by the following
application of the Modus Ponens rule
B1 ; (B1 ⇒ (A ⇒ B1 ))
(M P ) .
(A ⇒ B1 )
Case 2. B1 = A.
When B1 = A, then to prove Γ ` (A ⇒ B) means to prove Γ ` (A ⇒ A). But
(A ⇒ A) ∈ LA (axiom A21of H), i.e. ` (A ⇒ A). By the monotonicity of the
consequence we have that Γ`(A ⇒ A). The above cases conclude the proof of
the Base case i = 1.
Inductive step
Assume that Γ `(A ⇒ Bk ) for all k < i, we will show that using this fact we
can conclude that also Γ `(A ⇒ Bi ).
Consider a formula Bi in the sequence 9.36. By the definition, Bi ∈ LA∪Γ∪{A}
or Bi follows by MP from certain Bj , Bm such that j < m < i. We have to
consider again two cases.
Case 1. Bi ∈ LA ∪ Γ ∪ {A}.
The proof of (A ⇒ Bi ) from Γ in this case is obtained from the proof of the
Base Step for i = 1 by replacement B1 by Bi and will be omitted here as a
straightforward repetition.
Case 2. Bi is a conclusion of MP.
If Bi is a conclusion of MP, then we must have two formulas Bj , Bm in the
sequence 9.36 such that j < i, m < i, j 6= m and
Bj ; Bm
(M P ) .
Bi
By the inductive assumption, the formulas Bj , Bm are such that
Γ ` (A ⇒ Bj ) (9.37)
and
Γ ` (A ⇒ Bm ). (9.38)
443
Moreover, by the definition of the Modus Ponens rule, the formula Bm has to
have a form (Bj ⇒ Bi ), i.e. Bm = (Bj ⇒ Bi ), and the the inductive assumption
(9.38) can be re-written as follows.
Applying the rule MP to formulas (9.40) and (9.39,) i.e. performing the following
Applying again the rule MP to formulas 9.37 and 9.41, i.e. performing the
following
(A ⇒ Bj ) ; ((A ⇒ Bj ) ⇒ (A ⇒ Bi ))
(M P )
(A ⇒ Bi )
we get that
Γ `(A ⇒ Bi ).
Finally, suppose that there is some j < i such that Bi is ∀xBj . By hypothesis
Γ ` Bj and either (i) Bj does not depend upon A or (ii) x is not free variable
in A.
We have two cases (i) and (ii) to consider.
(i) If Bj does not depend upon A , then by Lemma 9.7 Γ ` Bj and, consequently,
by the generalization rule (G), Γ ` ∀xBj . Thus Γ ` Bi .
Now, by hypothesis Γ ` Bj and by axiom A2, ` (Bi ⇒ (A ⇒ Bi )). Applying
MP we get Γ ` A ⇒ Bi ).
(ii) If x is not free variable in A, then, by Completeness Theorem 9.6 and
|= (∀x(A ⇒ Bj ) ⇒ (A ⇒ ∀xBj )) we have that ` (∀x(A ⇒ Bj ) ⇒ (A ⇒ ∀xBj ))
.
444
Since Γ ` A ⇒ Bi ), we get by the generalization rule (G), Γ ` ∀x(A ⇒ Bj ),
and so, by MP, Γ ` A ⇒ ∀xBj ); that is Γ ` A ⇒ Bi ).
This completes the induction and the case (1) holds for i = n.
(2) The proof of the implication
if Γ ` (A ⇒ B) then Γ, A `B
Propositional Axioms
A1 (¬(A ∪ A) ∪ A), A2 (¬A ∪ (A ∪ B)),
A3 (¬(A ∪ B) ∪ (B ∪ A)), A4 (¬(¬B ∪ C) ∪ (¬(A ∪ B) ∪ (A ∪ C))),
445
for any A, B, C, ∈ F.
Quantifiers Axioms
Q1 (¬∀xA(x) ∪ A(x)), Q2 (¬A(x) ∪ ∃xA(x)),
Q2 (¬A(x) ∪ ∃xA(x)),
for any A(x) ∈ F.
Rules of Inference R
(MP) is the Modus Ponens rule. It has, in the language L{¬,∪} , a form
A ; (¬A ∪ B)
(M P ) .
B
A(x1 , x2 , . . . xn )
(SB) ,
A(t1 , t2 , . . . tn )
where A(x1 , x2 , . . . xn ) ∈ F and t1 , t2 , . . . tn ∈ T.
(G1), (G2) are quantifiers generalization rules.
The HA system is usually written now with the use of implication, i.e. as based
on a language L = L{¬,⇒} (P, F, C), i.e. as a proof system
HAI = (L{¬,⇒} (P, F, C), F, LA, R = {(M P ), (SB), (G1), (G2)}), (9.43)
Propositional Axioms
A1 ((A ∪ A) ⇒ A), A2 (A ⇒ (A ∪ B)),
A3 ((A ∪ B) ⇒ (B ∪ A)), A4 ((¬B ∪ C) ⇒ ((A ∪ B) ⇒ (A ∪ C))),
for any A, B, C, ∈ F.
446
Quantifiers Axioms
Q1 (∀xA(x) ⇒ A(x)), Q2 (A(x) ⇒ ∃xA(x)),
for any A(x) ∈ F.
Rules of Inference R
(M P ) is Modus Ponens rule
A ; (A ⇒ B)
(M P ) ,
B
for any formulas A, B ∈ F.
(SB) is a substitution rule
A(x1 , x2 , . . . xn )
(SB) ,
A(t1 , t2 , . . . tn )
where A(x1 , x2 , . . . xn ) ∈ F and t1 , t2 , . . . tn ∈ T.
(G1), (G2) are quantifiers generalization rules.
(B ⇒ A(x)) (A(x) ⇒ B)
(G1) , (G2) ,
(B ⇒ ∀xA(x)) (∃xA(x) ⇒ B)
where A(x), B ∈ F and B is such that x is not free in B.
The form of the quantifiers axioms Q1, Q2, and quantifiers generalization rule
(Q2) is due to Bernays.
2. Mendelson (1987)
Here is the first order logic proof system HM as introduced in the Elliott
Mendelson’s book Introduction to Mathematical Logic, hence the name. (1987).
It is an generalization to the predicate language of the proof system H2 for
propositional logic defined and studied in Chapter 5.
HM = (L{¬,∪} (P, F, C), F, LA, R = {(M P ), (G)}). (9.44)
Propositional Axioms
A1 (A ⇒ (B ⇒ A)),
A2 ((A ⇒ (B ⇒ C)) ⇒ ((A ⇒ B) ⇒ (A ⇒ C))),
A3 ((¬B ⇒ ¬A) ⇒ ((¬B ⇒ A) ⇒ B))),
for any A, B, C, ∈ F.
447
Quantifiers Axioms
Q1 (∀xA(x) ⇒ A(t)),
where where t is a term, A(t) is a result of substitution of t for all free occurrences
of x in A(x), and t is free for x in A(x), i.e. no occurrence of a variable in t
becomes a bound occurrence in A(t).
Q2 (∀x(B ⇒ A(x)) ⇒ (B ⇒ ∀xA(x))), where A(x), B ∈ F and B is such
that x is not free in B.
Rules of Inference R
(M P ) is the Modus Ponens rule
A ; (A ⇒ B)
(M P ) ,
B
for any formulas A, B ∈ F.
(G) is the generalization rule
A(x)
(G) ,
∀xA(x)
Rasiowa-Sikorski (1950)
Rasiowa and Sikorski are the authors of the first algebraic proof of the Gödel
completeness theorem ever given in 1950. Other algebraic proofs were later
given by Rieger, Beth, Los in 1951, and Scott in 1954.
Here is their original axiomatization.
Propositional Axioms
A1 ((A ⇒ B) ⇒ ((B ⇒ C) ⇒ (A ⇒ C))),
A2 (A ⇒ (A ∪ B)),
A3 (B ⇒ (A ∪ B)),
A4 ((A ⇒ C) ⇒ ((B ⇒ C) ⇒ ((A ∪ B) ⇒ C))),
A5 ((A ∩ B) ⇒ A),
A6 ((A ∩ B) ⇒ B),
A7 ((C ⇒ A) ⇒ ((C ⇒ B) ⇒ (C ⇒ (A ∩ B))),
448
A8 ((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C)),
A9 (((A ∩ B) ⇒ C) ⇒ (A ⇒ (B ⇒ C)),
A10 (A ∩ ¬A) ⇒ B),
A11 ((A ⇒ (A ∩ ¬A)) ⇒ ¬A),
A12 (A ∪ ¬A),
for any A, B, C ∈ F.
A ; (A ⇒ B)
(M P ) ,
B
for any formulas A, B ∈ F.
(SB) is a substitution rule
A(x1 , x2 , . . . xn )
(SB) ,
A(t1 , t2 , . . . tn )
where A(x1 , x2 , . . . xn ) ∈ F and t1 , t2 , . . . tn ∈ T.
(B ⇒ A(x)) (A(x) ⇒ B)
(G1) , (G2) ,
(B ⇒ ∀xA(x)) (∃xA(x) ⇒ B)
(B ⇒ ∀xA(x)) ∃x(A(x) ⇒ B)
(G3) , (G4) ,
(B ⇒ A(x)) A(x) ⇒ B)
where A(x), B ∈ F and B is such that x is not free in B.
The algebraic logic starts from purely logical considerations, abstracts from
them, places them into a general algebraic contest, and makes use of other
branches of mathematics such as topology, set theory, and functional analysis.
For example, Rasiowa and Sikorski algebraic generalization of the completeness
theorem for classical predicate logic is the following.
449
Theorem 9.12 (Rasiowa, Sikorski 1950)
For every formula A of the classical predicate calculus S = {L, C} the following
conditions are equivalent
i A is derivable in RS;
5. Given two proof systems S and K we say that S and K are equivalent
and write it as S ≡ K if they have the same sets of of theorems.
Prove that HA ≡ HAI for HA defined by (9.42) and HAI defined by
(9.43)
450
7. Let RSE be a proof system obtained from RS system defined by (9.45)
by changing the language L{¬,∩,∪,⇒} (P, F, C) of RS to the language with
equality (definition 9.12) and adding Eguality Axioms (9.5) to the set LA
of logocal axioms of RS. Prove Completeness Theorem 9.7 for RSE.
8. Prove Deduction Theorem 9.11 for Mendelson (1973) formalization.
9. In the proof of Deduction Theorem 9.11 for the proof system H we used
gthe completeness of H. Write a proof of the Deduction Theorem 9.11 for
H without use of its completeness.
451
452
Chapter 10
We define and discuss here a Rasiowa and Sikorski Gentzen style proof system
QRS for classical predicate logic. The propositional version of it, the RS proof
system, was studied in detail in chapter 6. These both proof systems admit a
constructive proof of completeness theorem. We adopt Rasiowa, Sikorski (1961)
technique of construction a counter model determined by a decomposition tree
to prove QRS completeness theorem 10.4. The proof, presented in section 10.3,
is a generalization of the completeness proofs of RS and other Gentzen style
propositional systems presented in details in chapter 6. We refer the reader to
this chapter as it provides a good introduction to the subject.
The other Gentzen type predicate proof system, including the original Gentzen
proof systems LK, LI for classical and intuitionistic predicate logics are ob-
tained from their propositional versions discussed in detail in chapter 6. It can
be done in a similar way as a generalization of the propositional RS the predicate
QRS system presented here. We leave these generalizations as an exercises for
the reader. That includes also the predicate language version of Gentzen proof
of cut elimination theorem, Hauptzatz (1935). The Hauptzatz proof for the
predicate classical LK and intuitionistic LI systems is easily obtained from the
propositional proof included in chapter6.
There are of course other types of automated proof systems based on different
methods of deduction.
There is a Natural Deduction mentioned by Gentzen in his Hauptzatz paper in
453
1935 and later fully developed by Dag Prawitz (1965). It is now called Prawitz,
or Gentzen-Prawitz Natural Deduction.
There is a Semantic Tableaux deduction method invented by Evert Beth (1955).
It was conequently simplified and further developed by Raymond Smullyan
(1968). It is now often called Smullyan Semantic Tableaux.
Finally, there is a Resolution. The resolution method can be traced back to Davis
and Putnam (1960). Their work is still known as Davis-Putnam method.The
difficulties of their method were eliminated by John Alan Robinson (1965) and
developed into what we call now Robinson Resolution, or just a Resolution.
There are many excellent textbooks covering each of these methods. We recom-
mend Melvin Fitting book First-order logic and automated theorem proving(2nd
ed.). Springer-Verlag(1996) as the one that not only covers all of them but also
discusses their relationships.
The Resolution proof system for propositional or predicate logic operates on
a set of clauses as a basic expressions and uses a resolution rule as the only
rule of inference. In section 10.4 we define and prove correctness of effective
procedures of converting any formula A into a corresponding set of clauses in
both propositional and predicate cases. The correctness of propositional case
is established by theorem 10.5, of predicate case by theorem 10.6. In the first
step of the predicate procedure we define a process of elimination of quantifiers
from the original language. It is called Skolemization of the language and is
presented in section 10.4.1. The correctness of the Skolemization is established
by Skolem theorem 10.11. In the second step of the procedure we show how
convert a quantifiers free formula into logically equivalent set of clauses. It is
presented with a proof of correctness (theorem 10.13) in section 10.4.2.
E = F ∗.
454
We will denote the expressions of QRS, i.e. the finite sequences of formulas by
In order to define the axioms LA and the set of rules of inference of QRS we
need to bring back some notions and to introduce some definitions.
An atomic formula of the predicate language L defined by (10.1) is any
element of A∗ (finite strings over the alphabet of L) of the form
R(t1 , t2 , ..., tn )
We use symbols R, Q, P, ... with indices if necessary to denote the atomic for-
mulas.
Literals
We form a special subset LT ⊆ F of formulas, called a set of all literals, which
is defined as follows.
The atomic formulas (10.2) are called positive literals and the elements of the
second set of the above union (10.3), i.e. the negations of the atomic formulas
are called negative literals.
Indecomposable, Decomposable Formulas
A formula A ∈ F is indecomposable if and only if it is atomic or a negation of
an atomic formula, i.e. an literal. Otherwise A is decomposable.
Now we form finite sequences out of formulas (and, as a special case, out of
literals). We need to distinguish the sequences formed out of literals from the
sequences formed out of other formulas, so we adopt the following definition
and notaions.
Indecomposable, Decomposable Sequences
A sequence Γ is indecomposable if and only if is formed out of indecomposable
formulas only. Otherwise is decomposable.
We denote indecomposable sequences by by
0 0 0
Γ , ∆ , Σ , . . . with indices if necessary. (10.4)
0 0 0
By definition, Γ , ∆ , Σ . . . are finite sequences (empty included) formed out
0 0 0 0 0 0
of literals, i.e Γ , ∆ , Σ Γ , ∆ , Σ ∈ LT ∗ .
455
We denote by
Γ, ∆, Σ, . . . with indices if necessary, (10.5)
the elements of F ∗ , i.e. we denote Γ, ∆, Σ finite sequences (empty included)
formed out of elements of F.
Logical Axioms LA
As the logical axiom of QRS we adopt any sequence of formulas which contains
a and its negation, i.e any sequence of the form
Disjunction rules
0 0 0
Γ , A, B, ∆ Γ , ¬A, ∆ : Γ , ¬B, ∆
(∪) 0 , (¬ ∪)
Γ , (A ∪ B), ∆ Γ0 , ¬(A ∪ B), ∆
Conjunction rules
0 0 0
Γ , A, ∆ ; Γ , B, ∆ Γ , ¬A, ¬B, ∆
(∩) , (¬ ∩)
Γ0 , (A ∩ B), ∆ Γ0 , ¬(A ∩ B), ∆
Implication rules
0 0 0
Γ , ¬A, B, ∆ Γ , A, ∆ : Γ , ¬B, ∆
(⇒) , (¬ ⇒)
Γ0 , (A ⇒ B), ∆ Γ0 , ¬(A ⇒ B), ∆
Negation rule
0
Γ , A, ∆
(¬ ¬) 0
Γ , ¬¬A, ∆
0
where Γ ∈ LT ∗ , ∆ ∈ F ∗ , A, B ∈ F.
456
(∃)
0
Γ , A(t), ∆, ∃xA(x)
Γ0 , ∃xA(x), ∆
where t is an arbitrary term.
(∀)
0
Γ , A(y), ∆
Γ0 , ∀xA(x), ∆
(¬∀)
0
Γ , ∃x¬A(x), ∆
Γ0 , ¬∀xA(x), ∆
(¬∃)
0
Γ , ∀x¬A(x), ∆
Γ0 , ¬∃xA(x), ∆
0
Γ ∈ LT ∗ , ∆ ∈ F ∗ , A, B ∈ F.
Note that A(t), A(y) denotes a formula obtained from A(x) by writing t, y,
respectively, in place of all occurrences of x in A. The variable y in (∀) is called
the eigenvariable. The condition: where y is a free individual variable which
does not appear in any formula in the conclusion is called the eigenvariable
condition.
All occurrences of y in A(y) of the rule (∀) are fully indicated.
The Proof System QRS
Formally we define the proof system QRS as follows.
R = {(∪), (¬∪), (∩), (¬∩), (⇒), (¬ ⇒), (¬¬), (¬∀), (¬∃), (∀), (∃))}
defined by (10.7).
457
By a formal proof of a sequence Γ in the proof system RS we understand
any sequence
Γ1 , Γ2 , .... Γn (10.9)
of sequences of formulas (elements of F ∗ , such that
1. Γ1 ∈ LA, Γn = Γ, and
2. for all i (1 ≤ i ≤ n) Γi ∈ AL, or Γi is a conclusion of one of the inference
rules of QRS with all its premisses placed in the sequence Γ1 , Γ2 , .... Γi−1 .
We write, as usual,
`QRS Γ
We picture, and write our proof trees with the node on the top, and leafs on
the very bottom, instead of more common way, where the leafs are on the top
and root is on the bottom of the tree.
In particular cases, as in the propositional case, we will write our proof- trees
indicating additionally the name of the inference rule used at each step of the
proof. For example, if in a proof of a formula A from axioms (10.6) we use
subsequently the rules
458
we represent the proof as the following tree denoted by TA .
TA
F ormula A
| (⇒)
conclusion of (∀)
| (∀)
conclusion of (¬¬)
| (¬¬)
conclusion of (∩)
^
(∩)
| (∀) | (∪)
axiom conclusion of (∩)
^
(∩)
axiom axiom
459
Step 2. If ∆ is indecomposable or an axiom, then ∆ becomes a leaf of the
tree.
Step 3. If ∆ is decomposable, then we traverse ∆ from left to right to identify
the first decomposable formula B and identify inference rule treated as de-
composition rule determined uniquely by B. We put its left and right premisses
as the left and right leaves, respectively.
Step 4. We repeat steps 2 and 3 until we obtain only leaves or infinite branch.
In particular case when when Γ has only one element, namely a a formula A ∈ F,
we define we call it a decomposition tree of A and denote by TA .
Here is a detailed definition of the decomposition tree for QRS.
Observe, that the condition 10.1 corresponds to the restriction put on the ap-
plication of the rule (∀).
460
If the main connective of Γ, i.e. the main connective of the first formula in Γ
0
which is not an literal, is (∃). In this case Γ is of the form Γ , ∃xA(x), ∆, we
0
write a sequence Γ , A(t), ∆, ∃xA(x) as its child, where the term t has to fulfill
the following condition.
The fact that the sequence 10.11 is one- to - one and the fact that, by the
conditions 10.1 and 10.2, we always chose the first appropriate term (variable)
from this sequence, guarantee that the decomposition process is also unique in
the case of the quantifiers rules (∀) and (∃).
Theorem 10.1
For any formula A ∈ F,
(i) the decomposition tree TA is unique.
(ii) Moreover, the following conditions hold.
1. If TA is finite and all its leaves are axioms, then
`QRS A
and TA is a tree-proof of A in QRS.
2. If TA is finite and contains a non-axiom leaf, or TA is infinite, then
6 `QRS A.
In all the examples below, the formulas A(x), B(x) represent any formula. But
there is no indication about their particular components, so they are treated as
indecomposable formulas.
Example 10.1
The decomposition tree TA of the de Morgan Law
(¬∀xA(x) ⇒ ∃x¬A(x))
is the following.
461
TA
(¬∀xA(x) ⇒ ∃x¬A(x))
| (⇒)
¬¬∀xA(x), ∃x¬A(x)
| (¬¬)
∀xA(x), ∃x¬A(x)
| (∀)
A(x1 ), ∃x¬A(x)
where x1 is a first free variable in the sequence 10.11 such that x1 does not appear in
∀xA(x), ∃x¬A(x)
| (∃)
(¬∀xA(x) ⇒ ∃x¬A(x))
Example 10.2
The decomposition tree TA of
(∀xA(x) ⇒ ∃xA(x))
is the following.
TA
(∀xA(x) ⇒ ∃xA(x))
| (⇒)
462
¬∀xA(x), ∃xA(x)
| (¬∀)
¬∀xA(x), ∃xA(x)
∃x¬A(x), ∃xA(x)
| (∃)
¬A(t1 ), ∃xA(x), ∃x¬A(x)
where t1 is the first term in the sequence 10.11, such that ¬A(t1 ) does not appear on the tree
above ¬A(t1 ), ∃xA(x), ∃x¬A(x)
| (∃)
¬A(t1 ), A(t1 ), ∃x¬A(x), ∃xA(x)
where t1 is the first term in the sequence 10.11, such that A(t1 ) does not appear on the tree
above ¬A(t1 ), A(t1 ), ∃x¬A(x), ∃xA(x)
Axiom
The above tree also ended with the axiom, hence we proved that
`( QRS) (∀xA(x) ⇒ ∃xA(x)).
Example 10.3
The decomposition tree TA of
(∃xA(x) ⇒ ∀xA(x))
is the following.
TA
(∃xA(x) ⇒ ∀xA(x))
| (⇒)
¬∃xA(x), ∀xA(x)
| (¬∃)
∀x¬A(x), ∀xA(x)
| (∀)
¬A(x1 ), ∀xA(x)
where x1 is a first free variable in 10.11 such that x1 does not appear in ∀x¬A(x), ∀xA(x)
| (∀)
¬A(x1 ), A(x2 )
where x2 is a first free variable in 10.11 such that x2 does not appear in ¬A(x1 ), ∀xA(x), the
sequence 10.11 is one-to- one, hence x1 6= x2
Non - axiom
463
The decomposition tree, for any formula A is unique, so we conclude from the
fact that the above tree has a non-axiom branch that
Remark when constructing the following tree TA for the formula ∃xA(x) in
example 10.4 below we adopt on the right branch of the a tree in the the short-
hand notation instead of the repeating a similar reasoning performed on the left
branch.
Example 10.4
The decomposition tree TA of the formula ∃xA(x) is the following.
TA
∃xA(x)
| (∃)
A(t1 ), ∃xA(x)
where t1 is the first term in the sequence 10.11, such that A(t1 ) does not appear on the tree
above A(t1 ), ∃xA(x)
| (∃)
| (∃)
| (∃)
464
Obviously, the above decomposition tree is infinite, what proves that
6` QRS ∃xA(x).
Remark when constructing the following trees TA in examples 10.5, 10.6 adopt,
as we did in the previous example 10.4, the shorthand notation when the rea-
soning is similar to the one presented in the example 10.4.
Example 10.5
The decomposition tree A of the first formula
is the following.
TA
| (⇒)
¬∃x(A(x) ∩ B(x)), (∃xA(x) ∩ ∃xB(x))
| (¬∃)
∀x¬(A(x) ∩ B(x)), (∃xA(x) ∩ ∃xB(x))
| (∀)
¬(A(x1 ) ∩ B(x1 )), (∃xA(x) ∩ ∃xB(x))
where x1 is a first free variable in the sequence 10.11 such that x1 does not appear in
∀x¬(A(x) ∩ B(x)), (∃xA(x) ∩ ∃xB(x))
| (¬∩)
465
¬A(x1 ), ¬B(x1 ), ∃xA(x) ¬A(x1 ), ¬B(x1 ), ∃xB(x)
| (∃) | (∃)
¬A(x1 ), ¬B(x1 ), A(t1 ), ∃xA(x) ¬A(x1 ), ¬B(x1 ), B(t1 ), ∃xB(x)
where t1 is the first term in the sequence 10.11, | (∃)
such that A(t1 ) does not appear on the tree
...
above ¬A(x1 ), ¬B(x1 ), A(t1 ), ∃xA(x) Observe,
that it is possible that t1 = x1 , as A(x1 ) does | (∃)
not appear on the tree above. By the definition
¬A(x1 ), ¬B(x1 ), ...B(x1 ), ∃xB(x)
of the sequence 10.11, x1 is placed somewhere
in it, i.e. x1 = ti , for certain i ≥ 1. It means
that after i applications of the step (∃) in the
decomposition tree, we will get a step:
| (∃)
All leaves of the above tree TA are axioms, what means that we proved
Example 10.6
The decomposition tree of the formula
is the following.
TA
| (⇒)
¬(∃xA(x) ∩ ∃xB(x))∃x(A(x) ∩ B(x))
| (¬∩)
¬∃xA(x), ¬∃xB(x), ∃x(A(x) ∩ B(x))
| (¬∃)
∀x¬A(x), ¬∃xB(x), ∃x(A(x) ∩ B(x))
466
| (∀)
| (¬∃)
| (∀)
By the reasoning similar to the reasonings in the previous examples we get that x1 6= x2
| (∃)
where t1 is the first term in the sequence 10.11, such that (A(t1 ) ∩ B(t1 )) does not appear on
the tree above ¬A(x1 ), ¬B(x2 ), (A(t1 ) ∩ B(t1 )), ∃x(A(x) ∩ B(x)) Observe, that it is possible that
t1 = x1 , as (A(x1 ) ∩ B(x1 )) does not appear on the tree above. By the definition of the
sequence 10.11, x1 is placed somewhere in it, i.e. x1 = ti , for certain i ≥ 1. For simplicity, we
assume that t1 = x1 and get the sequence:
^
(∩)
467
¬A(x1 ), ¬B(x2 ), ¬A(x1 ), ¬B(x2 ),
A(x1 ), ∃x(A(x) ∩ B(x)) B(x1 ), ∃x(A(x) ∩ B(x))
Axiom | (∃)
¬A(x1 ), ¬B(x2 ), B(x1 ),
(A(x2 ) ∩ B(x2 )), ∃x(A(x) ∩ B(x))
where x2 = t2 (x1 6= x2 ) is the
first term in the sequence 10.11, such that
(A(x2 ) ∩ B(x2 )) does not appear on the
tree above ¬A(x1 ), ¬B(x2 ), (B(x1 ), (A(x2 ) ∩
B(x2 )), ∃x(A(x) ∩ B(x)). We assume that t2 =
x2 for the reason of simplicity.
^
(∩)
¬A(x1 ), ¬A(x1 ),
¬B(x2 ), ¬B(x2 ),
B(x1 ), A(x2 ), B(x1 ), B(x2 ),
∃x(A(x) ∩ B(x)) ∃x(A(x) ∩ B(x))
| (∃) Axiom
...
^
(∩)
...
| (∃)
...
| (∃)
inf inite branch
The above decomposition tree TA contains an infinite branch what means that
468
is the proof system QRS. We adopted their proof to propositional case in chap-
ter 6.The completeness proofs, in the propositional case and in predicate case,
are constructive as they are based on a direct construction of a counter model
for any unprovable formula. The construction of the counter model for the un-
provable formula A uses the decomposition tree TA . We call such constructed
counter model a counter model determined by the tree TA . Rasiowa-Sikorski
type of constructive proofs of counter models determined by the tree decompo-
sition trees relay heavily of the notion of a strong soundness. We define it here
(definition 10.8), adopting chapter 4 general definition to our case.
Given a first order language L (10.1) with the set V AR of variables and the set
F of formulas. We define, after chapter 8 a notion of a model and a counter-
model of a formula A of L and then extend it to the the set F ∗ establishing the
semantics for QRS.
469
Directly from the above definition we get the following, simple fact.
Definition 10.5
For any sequence Γ ∈ F ∗ , by
δΓ
we understand any disjunction of all formulas of Γ.
M |= Γ
if and only if
M |= δΓ .
The sequence Γ is a predicate tautology if and only if the formula δΓ is a predicate
tautology, i.e.
|= Γ if and only if |= δΓ .
Our goal now is to prove the completeness theorem for QRS.The correctness of
the proof we present depends on the strong soundness of the rules of inference
of rules of inference defined as follows.
P1 ; P2 ; .... ; Pm
(r)
C
is strongly sound if the following condition holds for and structure M = [M, I]
for L.
M |= {P1 , P2 , .Pm } if and only if M |= C. (10.12)
We say it less formally that a rule (r) is strongly sound if the conjunction of
its premisses is logically equivalent with the conclusion, i.e.
P1 ∩ P2 ∩ . . . ∩ Pm ≡ C. (10.13)
470
Definition 10.8 (Strongly Sound System)
A predicate language (10.1) proof system S = (L, E, LA, R) is strongly sound
if and only if all logical axioms LA are tautologies and all its rules of inference
r ∈ R are strongly sound.
Proof
We have already proved in chapter 6 strong soundness of the propositional rules.
The quantifiers rule are strongly sound by straightforward verification and is left
as an exercise.
The strong soundness property is stronger then soundness property, hence also
the following holds.
if `QRS A, then |= A.
Proof
We have to prove the inverse implication to the soundness theorem 10.3. We
need to prove the formula A case only because the case of a sequence Γ can be
reduced to the formula case. Namely, the disjunction of all formulas in Γ. I.e.
we prove the implication:
if |= A, then `QRS A.
This means that we want prove that for any formula A, unprovability of A in
QRS (6`QRS A ), allows us to define its counter- model. The counter- model
471
is determined, as in the propositional case, by the decomposition tree TA . By
theorem 10.1 each formula A, generates its unique decomposition tree TA and
A has a proof only if this tree is finite and all its end sequences (leaves) are
axioms. Moreover, it says that we have two cases to consider:
(C1) the tree TA is finite and contains a leaf which is not axiom, or
(C2) the tree TA is infinite.
We will show how to construct a counter- model for A in both cases: a counter-
model determined by a non-axiom leaf of the decomposition tree TA , or a
counter- model determined by an infinite branch of TA .
Proof in case (C1): TA is finite and contains a non- axiom leaf.
TA
| (⇒)
¬∃x(P (x) ∩ R(x, y)), ∀x(P (x) ∩ R(x, y))
| (¬∃)
∀x¬(P (x) ∩ R(x, y)), ∀x(P (x) ∩ R(x, y))
| (∀)
¬(P (x1 ) ∩ R(x1 , y)), ∀x(P (x) ∩ R(x, y))
where x1 is a first free variable in 10.11 such that x1 does not appear in
∀x¬(P (x) ∩ R(x, y)), ∀x(P (x) ∩ R(x, y))
| (¬∩)
472
| (∀)
¬P (x1 ), ¬R(x1 , y), (P (x2 ) ∩ R(x2 , y))
where x2 is a first free variable in the sequence 10.11 such that x2 does not appear in
¬P (x1 ), ¬R(x1 , y), ∀x(P (x) ∩ R(x, y)), the sequence 10.11 is one-to- one, hence x1 6= x2
^
(∩)
There are two non-axiom leaves. In order to define a counter-model for (10.14)
determined by the tree TA we need to chose only one of them. Let’s choose the
leaf
LA = ¬P (x1 ), ¬R(x1 , y), P (x2 ). (10.15)
PI (x1 ) is true for x1 , and not true for x2 . RI (x1 , y) is true (holds) holds for x1
any for any y ∈ V AR.
We define the assignment v : V AR −→ T as identity, i.e., we put v(x) = x for
any x ∈ V AR.
Obviously, for such defined structure [M, I] and the assignment v we have that
([T, I], v) |= P (x1 ), ([T, I], v) |= R(x1 , y), and ([T, I], v) 6|= P (x2 ).
473
We hence obtain that
This proves that such defined structure [T, I] is a counter model for a non-axiom
leaf (10.15). By the strong soundness of QRS) (theorem 10.2) the structure
M = [T, I] is also a counter- model for the formula (10.14), i.e. we proved that
LA ⊆ LT ∗ . (10.16)
By definition, the leaf LA contains only atomic formulas and negations of atomic
formulas.
We use the non-axiom leaf LA (10.16) to define a structure M = [M, I] an
assignment v : V AR −→ M , such that (M, v) 6|= A. Such defined structure M
is called a counter - model determined by the tree TA .
M = [M, I] (10.18)
QI ⊆ Tn
is such that QI (t1 , . . . tn ) holds (is true) for terms t1 , . . . tn if and only if the
negation ¬Q(t1 , . . . tn ) of the formula Q(t1 , . . . tn ) appears on the leaf LA and
QI (t1 , . . . tn ) does not hold (is false) for terms t1 , . . . tn otherwise.
3. For any constant c ∈ C, we put cI = c, for any variable x, xI = x.
For any functional symbol f ∈ F, #f = n,
fI : Tn −→ T
474
is identity function, i.e. we put
fI (t1 , . . . tn ) = f (t1 , . . . tn )
for all t1 , . . . tn ∈ T.
4. We define the assignment v : V AR −→ T as identity, i.e. we put for all
x ∈ V AR v(x) = x.
Obviously, for such defined structure [T, I] and the assignment v we have that
This proves that the structure M = [T, I] and assignment v defined by (10.18)
are such that
([T, I], v) 6|= LA .
By the strong soundness (theorem 10.2) of QRS
6|= A.
The case of the infinite tree is similar, even if a little bit more complicated.
Observe first that the rule (∃) is the t rule of inference (decomposition) which
can ”produces” an infinite branch. We first show how to construct the counter-
model in the case of the simplest application of this rule, i.e. in the case of the
formula
∃xP (x)
where P is an one argument relational symbol. All other cases are similar to
this one. The infinite branch BA in this case consists of all elements of the
decomposition tree:
TA
∃xP (x)
| (∃)
P (t1 ), ∃xP (x)
475
where t1 is the first term in the sequence 10.11, such that P (t1 ) does not appear on the tree
above P (t1 ), ∃xP (x)
| (∃)
| (∃)
| (∃)
| (∃)
.....
| (∃)
.....
The infinite branch of TA , written from the top, in oder of appearance of for-
mulas is
BA = {∃xP (x), P (t1 ), A(t2 ), P (t2 ), P (t4 ), .....}
where t1 , t2 , .... is a one - to one sequence (10.11) of all elements of the set T of
all terms.
This means that the infinite branch B contains with the formula ∃xP (x) all its
instances P (t), for all terms t ∈ T.
We define the structure M = [M, I] and valuation v following the definition
10.17. We take as the universe M the set T of all terms, and now in our case
we define PI as follows: PI (t) holds if ¬P (t) ∈ BA and PI (t) does not hold if
P (t) ∈ BA .
It is easy to see that for any formula P (t) ∈ B,
476
C1: General Method
Let A be any formula such that 6`QRS A.
Let TA be an infinite decomposition tree of a formula A. Let BA the infinite
¯
branch of TA , written from the top, in oder of appearance of sequences Γ ∈ F ∗
on it, where Γ0 = A.
QI (t1 , . . . tn ) ∈ LF
and
(2) QI (t1 , . . . tn ) does holds (is true) for terms t1 , . . . tn if and only if
QI (t1 , . . . tn ) 6∈ LF.
M 6|= A. (10.22)
For this purpose we first introduce, for any formula A ∈ F, an inductive defini-
tion of the order ord A of the formula A.
(1) If A ∈ AF, then ord A = 1.
(2) If ord A = n, then ord ¬A = n + 1. (3) If ord A ≤ n and ord B ≤ n,
then ord (A ∪ B) = ord (A ∩ B) = ord (A ⇒ B) = n + 1.
(4) If ord A(x) = n, then ord ∃xA(x) = ord ∀xA(x) = n + 1.
477
We conduct the proof of (10.23) by contradiction. Suppose that (10.23) does
not hold, i.e. assume that
M |= A. (10.23)
Observe that by assumption (10.23) and the definition (10.25), the formula A
is in M F and hence M F 6= ∅.
Let B 0 be a formula in M F such that ord B 0 ≤ ord B for every B ∈ M F. There
exists Γi ∈∈ BA that is of the form Γ0 , B 0 , ∆ with an indecomposable Γ0 .
We have that B 0 can not be of the form
(A ∪ B), ¬(A ∪ B), (A ∩ B), ¬(A ∩ B), (A ⇒ B), ¬(A ⇒ B), ¬¬A, ∀xA(x).
(10.27)
The formula B 0 can’t be of the form
∃xB(x) (10.28)
since then there exists term t and j such that i ≤ j, B 0 (t) appears in Γj and the
formula B(t) satisfies (10.24). Thus B(t) ∈ M F and ordB(t) < ordB 0 . This
contradicts the definition of B 0 .
Since B 0 is not of the form (10.26), (10.27), (10.28), B 0 is indecomposable. Thus
B 0 ∈ LF (10.20), and consequently by (10.21),
M 6|= B 0 .
On the other hand B 0 by definition is in the set M F and hence is one o the
formulas satisfying (10.24), i.e.
M 6|= B 0 .
6|= A.
This ends the proof of the Completeness Theorem 10.4 for QRS.
478
10.4 Skolemization and Clauses
The resolution proof system for propositional and predicate logic operates on a
set of clauses as a basic expressions and uses a resolution rule as the only
rule of inference.
The goal of this part is to define an effective process of transformation of
any formula A of a predicate language L = L{¬,∪,∩,⇒} (P, F, C) into a certain
corresponding set of clauses CA . This is done in two stages.
S1. We convert any formula A of L into an open formula A∗ of a language L∗ by
a process of elimination of quantifiers from the original L. The method is due
to T. Skolem (1920) and is called Skolemization. The resulting formula A∗ is
equisatisfiable with A: it is satisfiable if and only if the original one is satisfiable
(Skolem Theorem 10.11).
The stage S1. is performed as the first step in a Resolution based automated
theorem prover and is described in section 10.4.1.
S2. We define a proof system QRS∗ based on the language L∗ and use it
transform any formula A∗ of L∗ into an logically equivalent set of clauses CA∗
(theorem 10.13).
The final result of stages S1. and S1 is the set CA of clauses corresponding to
the formula A, called a clausal form of A (theorem 10.6.
The transformation process for any propositional formula A into its logically
equivalent set CA of clauses follows directly from the use of the propositional
system RS (theorem 10.5).
.
Definition 10.10
Given a propositional or predicate language L, and a sequence Γ ∈ LT ∗ . A
clause determined by Γ is a set form out of all elements of the sequence Γ
We we denote it by CΓ .
479
Example 10.7
In particular,
1. if Γ1 = a, a, ¬b, c, ¬b, c and Γ2 = ¬b, c, a, then CΓ1 = CΓ2 {a, c, ¬b}.
2. If Γ1 = ¬P (x1 ), ¬R(x1 , y), P (x2 ), ¬P (x1 ), ¬R(x1 , y), P (x2 ) and
Γ2 = ¬P (x1 ), ¬R(x1 , y), P (x2 ), then CΓ1 = CΓ2 {¬P (x1 ), ¬R(x1 , y), P (x2 )}.
The semantics for clauses is basically the same as for the sequences. We define
it as follows.
Proof
Let L = L{¬,∪,∩,⇒} . Given A ∈ F, we use the RS system (chapter 6) to build
the decomposition tree TA . We form clauses out of the leaves of the tree TA ,
i.e. for every leaf L we create a clause CL determined by L (definition 10.10).
We put
CA = {CL : L is a leaf of TA }.
Directly from the strong soundness (10.13) of rules of inference of RS and the
definition 10.13 we get A ≡ CA . This ends the proof for the propositional case.
Consider a decomposition tree of a formula (((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c))
480
TA
| (∪)
((a ⇒ b) ∩ ¬c), (a ⇒ c)
^
(∩)
(a ⇒ b), (a ⇒ c) ¬c, (a ⇒ c)
| (⇒) | (⇒)
¬a, b, (a ⇒ c) ¬c, ¬a, c
| (⇒)
¬a, b, ¬a, c
Example 10.8
For the formula (((a ⇒ b) ∩ ¬c) ∪ (a ⇒ c)) and the tree TA , the leaves are
L1 = ¬a, b, ¬a, c and CL1 = {¬a, b, c} and
L2 = ¬c, ¬a, c and CL2 = {¬c, ¬a, c}. The set of clauses is
Proof
Given a formula A of a language L. The open formula A∗ of the quantifiers
free language L∗ is obtained by the Skolemization process. The effectiveness
and correctness of the process follows from PNF theorem 10.10 and Skolem
theorem 10.11 described in section 10.4.1.
As the next step, we define (section 10.4.2) a proof system QRS∗ (10.43) based
on the quantifiers free language L∗ . The system QRS∗ is a version of the system
QRS (10.8) restricted to its Propositional Rules. At this point we carry the
proof in a similar way to the proof in the propositional case (theorem 10.5).
Namely, for any formula A∗ of L∗ obtained from A of L we construct its the
481
decomposition tree TA∗ . We form clauses out of the leaves of the tree TA∗ , i.e.
for every leaf L we create a clause CL determined by L and we put
Term t is free for x in A(x). Let A(x) ∈ F and t be a term, A(t) be a result
of substituting t for all free occurrences of x in A(x).
We say that t is free for x in A(x), if no occurrence of a variable in t
becomes a bound occurrence in A(t).
denotes the result of replacing all occurrences of the free variables x, x1 , x2 , ..., xn ,
by the terms t, t1 , t2 , ..., tn , respectively, assuming that t, t1 , t2 , ..., tn are free for
x, x1 , x2 , ..., xn , respectively, in A.
The assumption that t is free for x in A(x) while substituting t for x, is
important because otherwise we would distort the meaning of A(t). This is
illustrated by the following example.
Example 10.9
Let t = y and A(x) be
∃y(x 6= y).
Obviously t is not free for y in A. The substitution of t for x produces a formula
A(t) of the form
∃y(y 6= y),
which has a different meaning than ∃y(x 6= y).
482
Here are more examples illustrating the notion: t is free for x in A(x).
Example 10.10
Let A(x) be a formula
(∀yP (x, y) ∩ Q(x, z))
and t be a term f (x, z), i.e. t = f (x, z).
None of the occurrences of the variables x, z of t is bound in A(t), hence we say
that t = f (x, z) is free for x in (∀yP (x, y) ∩ Q(x, z)).
Example 10.11
Let A(x) be a formula
(∀yP (x, y) ∩ Q(x, z))
The term t = f (y, z) is not free for x in A(x) because substituting t = f (y, z)
on a place of x in A(x) we obtain now a formula A(t) of the form
which contain a bound occurrence of the variable y of t (∀yP (f (y, z), y)).
The other occurrence (Q(f (y, z), z)) of y is free, but it is not sufficient, as for
term to be free for x, all occurrences of its variables has to be free in A(t).
Another important notion we will use here is the following notion of similarity
of formulas.
Intuitively, we say that A(x) and A(y) are similar if and only if A(x) and A(y)
are the same except that A(x) has free occurrences of x in exactly those places
where A(y) has free occurrences of y.
Example 10.12
The formulas ∃z(P (x, z) ⇒ Q(x)) and ∃z(P (y, z) ⇒ Q(y)) are similar.
483
Example 10.13
The formulas A(x): ∃z(P (x, z) ⇒ Q(x, y)) and A(x/y): ∃z(P (y, z) ⇒ Q(y, y))
are not similar; y is free for x in A(x), but the formula A(x/y) has a free
occurrence of y.
Example 10.14
The formulas A(x): ∃z(P (x, z) ⇒ Q(x, y)) and A(x/w): ∃z(P (w, z) ⇒ Q(w, y)
are similar; w is free for x in A(x) and the formula A(x/w) has no free
occurrence of w.
Lemma 10.1
For any formula A(x) ∈ F, if A(x) and A(x/y) are similar A(x) ∼ A(y), then
∀xA(x) ≡ ∀yA(y),
∃xA(x) ≡ ∃yA(y).
Directly from lemma 10.1 and replacement theorem 10.7 we get that the follow-
ing theorem holds.
We can now use theorem 10.8 to prove its more general version.
484
Theorem 10.9 (Naming Variables Apart) Every formula A ∈ F is logi-
cally equivalent to one in which all variables are named apart.
We use the above theorems plus the equational laws for quantifiers (10.31) to
prove, as a next step a so called a Prenex Form Theorem 10.10.
In order to do so we first we define an important notion of prenex normal
form of a formula.
Example 10.15
Let A be a formula (P (x, y) ⇒ ¬∃z R(x, y, z)), its closure A0 ≡ A is
∀x∀y(P (x, y) ⇒ ¬∃z R(x, y, z)).
Q1 x1 Q2 x2 ....Qn xn B
Proof
We use theorems 10.7, 10.8, 10.9, theorem 10.15, and the following logical equiv-
alences (10.31) proved in chapter 2.
485
Equational Laws of Quantifiers (10.31)
Exercise 10.1
Find a prenex normal form PNF of a formula A: (∀x(P (x) ⇒ ∃xQ(x)).
Solution
486
Step 2: Pull out Quantifiers
(C ⇒ ∃yQ(y)) ≡ ∃y (C ⇒ Q(y))
Exercise 10.2 Find a prenex normal form PNF formula A0 for the formula
A:
(∃x∀y R(x, y) ⇒ ∀y∃x R(x, y))
Solution
Step 1: Rename Variables Apart
Take a sub- formula B(x, y) : ∀y∃x R(x, y) of A, get B(x/z, y/w) : ∀z∃w R(z, w)
and replace B(x,y) by B(x/z, y/w) in A and get
such that A0 ≡ A. Now we pull quantifiers ∀z∃w in (R(x, y) ⇒ ∀z∃w R(z, w))
and get the prenex normal form PNF formula
We will show now how any formula A in its prenex normal form PNF we can
transformed it into a corresponding open formula A∗ .
487
The open formula A∗ belongs to a richer language then the initial language to
which the formula A belongs. The transformation process adds new constants,
called Skolem constants, and new function symbols, called Skolem function
symbols to the initial language.
The whole process is called the skolemisation of the initial language L, the such
build extension of the initial language is called a Skolem extension of L,.
A = Q1 x1 Q2 x2 . . . Qn xn B(x1 , x2 , . . . xn ) (10.41)
B(x1 , x2 , . . . . , xn ).
Case 2
All quantifiers Qi for 1 ≤ i ≤ n are existential, i.e. the closed formula A is
B(c1 , c2 , . . . . , cn )
488
form its closure instead. We eliminate quantifiers one by one and step by step
depending on first, and consecutive quantifiers.
Consider P1
First quantifier in A is universal, i. e. A is
∀x1 Q2 x2 . . . Qn xn B(x1 , x2 , . . . xn )
We replace A by a formula A1 :
Q2 x2 . . . Qn xn B(x1 , x2 , . . . xn )
We have eliminated the quantifier Q1 in this case.
Consider P2
First quantifier in A is existential, i. e. A is
∃x1 Q2 x2 . . . Qn xn B(x1 , x2 , . . . xn )
We replace A by a formula A1 :
Q2 x2 . . . Qn xn B(b1 , x2 , . . . xn )
where b1 is a new constant symbol added to our original language L. We
call such constant symbol added to the language Skolem constant symbol.
We have eliminated the quantifier Q1 in this case. We have covered all cases
and this ends the Step 1.
Step 2 Elimination of Q2 x2 .
489
Consider P1
First quantifier in A1 is universal, i.e. A1 is
∀x2 Q3 x3 . . . Qn xn B(x1 , x2 , x3 , . . . xn )
Q3 x3 . . . Qn xn B(x1 , x2 , x3 , . . . xn )
Consider P2
First quantifier in A1 is existential, i.e. A1 is
∃x2 Q3 x3 . . . Qn xn B(x1 , x2 , x3 , . . . xn )
Q3 x3 . . . Qn xn B(x1 , f (x1 ), x3 , . . . xn )
Q2 x2 Q3 x3 . . . Qn xn B(b1 , x2 , . . . xn )
Consider P1
First quantifier in A1 is universal, i.e. A1 is
∀x2 Q3 x3 . . . Qn xn B(b1 , x2 , x3 , . . . xn )
Q3 x3 . . . Qn xn B(b1 , x2 , x3 , . . . xn )
∃x2 Q3 x3 . . . Qn xn B(b1 , x2 , x3 , . . . xn )
490
We replace A1 by A2
Q3 x3 . . . Qn xn B(b1 , b2 , x3 , . . . xn )
where b2 6= b1 is a new Skolem constant symbol added to our original language
L.
We have eliminated the quantifier Q2 in this case. We have covered all cases
and this ends the Step 2. Step 3 Elimination of Q3 x3
Let’s now consider, as an example formula A2 from Step 2; P1 i.e. the formula
Q3 x3 . . . Qn xn B(x1 , x2 , x3 , . . . xn )
We have again 2 choices to consider, but will describe only the following.
Step i
At each Step i, for 1 ≤ i ≤ n), we build a binary tree of possibilities:
P1 Qi xi is universal or P2 Qi xi is existential and as result we obtain a
formula Ai with one less quantifier. The elimination process builds a sequence
of formulas
A, A1 , A2 , . . . , An = A∗
where the formula A belongs to our original language
L = L{¬,∪,∩,⇒} (P, F, C),
∗
the formula A belongs to its Skolem extension language (10.42) defined as
follows.
Definition 10.18
The language L∗ obtained from L by the quantifiers elimination procedure (10.40)
is is called a Skolem extension of L.
L∗ = L{¬,∪,∩,⇒} (P, F ∪ SF, C ∪ SC). (10.42)
491
Observe that in the elimination process (10.40) a universal quantifier intro-
duces free variables in the formula B(x1 , x2 , . . . xn ). The elimination of an
existential quantifier that follows universal quantifiers introduces a new func-
tional symbol with number of arguments equal the number of universal quanti-
fiers preceding it.
The resulting is an open formula A∗ of Skolem extension language L∗ . By
PNF theorem 10.10, for any formula A of L its PNF formula (10.41) exists
and is logically equivalent with A. We hence introduce the following definition.
Solution
We eliminate ∀y1 and get a formula A1
∃y2 ∀y3 ∃y4 B(y1 , y2 , y3 , y4 ).
We eliminate ∃y2 by replacing y2 by h(y1 ) where h is a new one argument
functional symbol added to our original language L.
We get a formula A2
∀y3 ∃y4 B(y1 , h(y1 ), y3 , y4 ).
We eliminate ∀y3 and get a formula A3
∃y4 B(y1 , h(y1 ), y3 , y4 ).
We eliminate ∃y4 by replacing y4 by f (y1 , y3 ), where f is a new two argument
functional symbol added to our original language L.
We get a formula A4 that is our resulting open formula A∗
B(y1 , h(y1 ), y3 , f (y1 , y3 )).
Exercise 10.4
Let now A be a PNF formula
∃y1 ∀y2 ∀y3 ∃y4 ∃y5 ∀y6 B(y1 , y2 , y3 , y4 , y4 , y5 , y6 )
Find the Skolem form of A (the formula B(y1 , y2 , y3 , y4 , y4 , y5 ) is quantifiers
free).
492
Solution
We eliminate ∃y1 and get a formula A1
∀y2 ∀y3 ∃y4 ∃y5 ∀y6 B(b1 , y2 , y3 , y4 , y4 , y5 , y6 )
where b1 is a new constant symbol added to our original language L.
We eliminate ∀y2 , f orally3 and get a formulas A2 , A3 ; here is the formula A3
∃y4 ∃y5 ∀y6 B(b1 , y2 , y3 , y4 , y4 , y5 , y6 )
We eliminate ∃y4 and get a formula A4
∃y5 ∀y6 B(b1 , y2 , y3 , g(y2 , y3 ), y5 , y6 )
where g is a new two argument functional symbol added to our original
language L.
We eliminate ∃y5 and get a formula A5
∀y6 B(b1 , y2 , y3 , g(y2 , y3 ), h(y2 , y3 ), y6 )
where h is a new two argument functional symbol added to our original
language L.
We eliminate ∀y6 and get a formula A6 that is the resulting open formula A∗
B(b1 , y2 , y3 , g(y2 , y3 ), h(y2 , y3 ), y6 ).
493
10.4.2 Clausal Form of Formulas
0 0 0
Γ , A, B, ∆ Γ , ¬A, ∆ : Γ , ¬B, ∆
(∪) , (¬ ∪)
Γ0 , (A ∪ B), ∆ Γ0 , ¬(A ∪ B), ∆
0 0 0
Γ , A, ∆ ; Γ , B, ∆ Γ , ¬A, ¬B, ∆
(∩) , (¬ ∩)
Γ0 , (A ∩ B), ∆ Γ0 , ¬(A ∩ B), ∆
0 0 0
Γ , ¬A, B, ∆ Γ , A, ∆ : Γ , ¬B, ∆
(⇒) , (¬ ⇒)
Γ0 , (A ⇒ B), ∆ Γ0 , ¬(A ⇒ B), ∆
0
Γ , A, ∆
(¬ ¬) 0
Γ , ¬¬A, ∆
0
where Γ ∈ LT ∗ , ∆ ∈ OF ∗ , A, B ∈ OF.
494
For any sequence Γ of formulas of L∗ , any structure M = [M, I] for L∗ ,
M |= Γ if and only if M |= δΓ ,
where δΓ denotes a disjunction of all formulas in Γ.
The semantics for clauses is basically the same as for the sequences. We define
it, after definition 10.5, as follows.
Obviously, all rules of QRS∗ are strongly sound (definition 10.7) and theorem
10.2 holds for QRS∗ , i.e. we have the following.
We are going to prove now that any formula A of L∗ can be transformed into
in logically equivalent set of clauses.
Proof
Given A ∈ OF. Here is the two steps procedure. S1. We construct (finite and
unique) decomposition tree TA . S2. We form clauses out of the leaves of the
tree TA , i.e. for every leaf L we create a clause CL determined by L (definition
10.10) and we put
CA = {CL : L is a leaf of TA }.
495
Exercise 10.5
Find he set CA of clauses for the following formula A.
Solution
S1. We construct the decomposition tree for A as follows
TA
| (∪)
(((P (b, f (x)) ⇒ Q(x)) ∪ ¬R(z)), (P (b, f (x)) ∩ R(z))
| (∪)
(P (b, f (x)) ⇒ Q(x)), ¬R(z), (P (b, f (x)) ∩ R(z))
| (⇒)
¬P (b, f (x)), Q(x), ¬R(z), (P (b, f (x)) ∩ R(z))
^
(∩)
¬P (b, f (x)), Q(x), ¬R(z), P (b, f (x)) ¬P (b, f (x)), Q(x), ¬R(z), R(z)
CA = {{¬P (b, f (x)), Q(x), ¬R(z), P (b, f (x)}, {¬P (b, f (x)), Q(x), ¬R(z), R(z)}.
Definition 10.23 Clausal Form Given a formula A of the language L and its
Skolem form A∗ of L∗ . The set CA∗ of clauses such that
A∗ ≡ CA∗
496
Solution
Step 1: We rename variables apart in A and get a formula A0 :
TA∗
| (⇒)
497
¬(R(x, h(x)) ∪ ¬P (x)), ¬R(z, f (x, z))
^
(¬∪)
CA∗ = { {¬R(x, h(x)), ¬R(z, f (x, z)}, {P (x), ¬R(z, f (x, z))} }.
Show that for any formulas A(x), B(x) with a free variable x the following
holds.
Remider: 1. you treat A(x), B(x) as atomic foprmulas, 2.you must trans-
form formulas with restricted domain quantifiers into proper formulas of
L.
(i) `QRS f orallx (A(x) ∩ B(x)) ≡ (∀xA(x) ∩ ∀xB(x))
(ii) `QRS ∃x (A(x) ∪ B(x)) ≡ (∃xA(x) ∪ ∃xB(x)).
(iii) `QRS ¬∀B(x) A(x) ≡ ∃B(x) .
(iv) `QRS ¬∃B(x) A(x) ≡ ∀B(x) ¬A(x).
(v) `QRS ¬∀xA(x) ≡ ∃x¬A(x).
(vi) `QRS ¬∃xA(x) ≡ ∀x¬A(x).
(vii) `QRS (∀x(B(x) ⇒ A(x)) ⇒ (∃x B(x) ⇒ ∃x (B(x) ∩ A(x))))
498
2. Show that for any formulas A(x), B B where B does not contain any free
occurrence of x the following holds.
(i) `QRS ∀x(A(x) ∩ B) ≡ (∀xA(x) ∩ B).
(ii) `QRS ∀x(A(x) ∪ B) ≡ (∀xA(x) ∪ B).
(iii) `QRS ∃x(A(x) ⇒ B) ≡ (∀xA(x).
(iv) `QRS ∃x(A(x) ⇒ B) ≡ (∀xA(x).
3. Prove that following formulas are not provable in QRS.
Remider: you must transform formulas with restricted domain quantifiers
into proper formulas of L.
(i) ∃C(x) (A(x) ∪ B) 6≡ (∃C(x) A(x) ∪ B).
(ii) ∀C(x) (A(x) ∩ B) 6≡ (∀C(x) A(x) ∩ B).
(iii) ∃C(x) (A(x) ⇒ B) 6≡ (∀C(x) A(x) ⇒ B).
(iv) ∃C(x) (B ⇒ A(x)) 6≡ (B ⇒ ∃xA(x)).
4. Prove that following formulas are not provable in QRS.
(i) (∃x ¬A(x) ⇒ ∀x A(x))
(ii) (∀x∃y A(x, y) ⇒ ∃x∀y A(x, y)).
(iii) (∃x∃y A(x, y) ⇒ ∃y A(y, y)).
(iv) (∀x∃y A(x, y) ⇒ ∃y A(y, y)).
(v) (∀x (A(x) ⇒ B(x)) ⇒ (∀x A(x) ⇒ ∃x B(x))).
5. Prove that following formulas are not provable in QRS.
(i) A1 : ∀x¬∃y(P (x, g(y, y)) ∪ P (x, g(g(y, y), d))).
(ii) A2 : (¬∀yP (f (x, y), c) ⇒ (P (x, c) ∪ P (y, c)))
(iii) A3 : ∀x(P (x) ⇒ ∃yQ(x, y)).
(iv)A4 : ∀x¬∃y(P (x) ∩ ¬Q(x, y)).
6. Find counter-models determined by the decomposition trees TAi for the
following formulas Ai , i = 1, 2, 3, 4.
(i) A1 : ∀x¬∃y(Q(x, g(y)) ∪ R(x, f (x, y), c))).
(ii) A2 : (¬∀yR(f (x, y), c) ⇒ (Q(x, c) ∪ Q(y, c)))
(iii) A3 : ∀x(P (x) ⇒ ∃yQ(x, y)).
(iv)A4 : ∀x¬∃y(P (x) ∩ ¬Q(f (x, y))).
7. Find prenex normal form PNF of the following formulas.
Reminder: We assume that the formula A in PNF is always closed. If it
is not closed we form its closure (definition 10.16) instead.
(i) (∀x(P (x) ⇒ ¬∀yP (y)) ⇒ (∃x R(x, y) ⇒ ∃y (R(x, y) ∩ P (y)))).
(ii) ((∀xQ(x) ⇒ (∃xR(x) ∪ ¬∀xQ(x))) ⇒ (¬∃xQ(x) ∩ R(x))).
499
(iii) (∀x R(f (x, y), c) ⇒ (∃xR(f (x, y), c))∩¬R(f (x, y), c)) ⇒ (¬∀x R(f (x, y), c) ⇒
∃x R(f (x, y), c))).
(iv) ((∃R(y) P (y) ⇒ Q(x)) ⇒ (P (y) ⇒ ∃xQ(x)))
8. Find a Skolem form of the following formulas (the formula B(y1 , y2 , y3 , y4 , y4 )
is quantifiers free).
(i) ∀y1 ∀y2 ∀y3 ∃y4 B(y1 , y2 , y3 , y4 , y4 ).
(ii) ∃y1 ∃y2 ∀y3 ∃y4 B(y1 , y2 , y3 , y4 , y4 ).
(iii) ∃y1 ∀y2 ∃y3 ∃y4 B(y1 , y2 , y3 , y4 , y4 ).
(iv) ∀y1 ∀y2 ∃y3 ∃y4 B(y1 , y2 , y3 , y4 , y4 ).
500
Chapter 11
Formal theories play crucial role in mathematics and were historically defined
for classical predicate (first order logic) and consequently for other first and
higher order logics, classical and non-classical.
The idea of formalism in mathematics, which resulted in the concept of formal
theories, or formalized theories, as they are also called. Their concept was
developed in connection with the Hilbert Program. One of the main objects
of the program was to construct a formal theory that would cover the whole
mathematics and to prove its consistency by employing the simplest of logical
means. This part of the program was called the Consistency Program, where a
formal theory is said to be consistent if no formal proof can be carried in that
theory for a formula A and at the same time for its negation ¬A.
In 1930, while still in his twenties Kurt Gödel made a historic announcement:
Hilbert Consistency Program could not be carried out. He justified his claim by
proving his Inconsistency Theorem, called also Second Incompleteness Theorem.
Roughly speaking the theorem states that a proof of the consistency of every
formal theory that contains arithmetic of natural numbers can be carried out
only in mathematical theory which is more comprehensive than the one whose
consistency is to be proved. In particular, a proof of the consistency of formal
(elementary, first order) arithmetic can be carried out only in mathematical
theory which contains the whole arithmetic and also other theorems that do
not belong to arithmetic. It applies to a formal theory that would cover the
whole mathematics because it would obviously contain the arithmetic on natural
numbers. Hence the Hilbert Consistency Program fails.
Gödel’s result concerning the proofs of the consistency of formal mathematical
theories has had a decisive impact on research in properties of formal theories.
501
Instead of looking for direct proofs of inconsistency of mathematical theories,
mathematicians concentrated largely on relative proofs that demonstrate that
a theory under consideration is consistent if a certain other theory, for example
a formal theory of natural numbers, is consistent. All those relative proofs are
rooted in a deep conviction that even though it cannot be proved that the theory
of natural numbers is free of inconsistencies, it is consistent. This conviction
is confirmed by centuries of development of mathematics and experiences of
mathematicians.
A formal theory is called complete if for every sentence (formula without free
variables) of the language of that theory there is a formal proof of it or of its
negation. A formal theory which does not have this property is called incom-
plete. Hence a formal theory is incomplete if there is a sentence A of the
language of that theory, such that neither A nor ¬A are provable in it. Such
sentences are called undecidable in the theory in question or independent of the
theory.
It might seem that one should be able to formalize a theory such as the formal
theory of natural numbers in a way to make it complete, i.e. free of undecidable
(independent) sentences. But it is not the case in view of Gödel’s Incomplete-
ness Theorem. It states that every consistent formal theory which contains
the arithmetic of natural numbers is incomplete. The Inconsistency Theorem
follows from it. This is why the Incompleteness and Inconsistency Theorems
are now called Gödel First Incompleteness Theorem (theorems 11.3, 11.6) and
Gödel Second Incompleteness (theorems 11.4, 11.7), respectively.
The third part of the Hilbert Program posed and was concerned with the problem
of decidability of formal mathematical theories. A formal theory is called de-
cidable if there is a method of determining, in a finite number of steps, whether
any given formula in that theory is its theorem or not. If a theory is decidable
and if the decision algorithm is known, then the study of problems expressible
in the language of the theory reduces to a purely mechanical procedure. In
undecidable theories there is no mechanical procedure. Most of mathematical
theories are undecidable. Gödel proved in 1931 that the arithmetic of of natural
numbers is undecidable.
We discuss the Hilbert Program and Gödel’s Theorems in more details in sec-
tions 11.3.1 and 11.3.2, respectively.
502
Remark 11.1
We consider here only classical formal theories based on a complete classical
Hilbert style proof system. We also assume that its language contains the full
set {¬, ∩, ∪, ⇒} of propositional connectives.
SA ⊆ F and LA ∩ SA = ∅. (11.2)
LSA ⊆ L (11.3)
503
A proof system
T = (L, F, LA, SA, R), (11.4)
is called a formal theory with the set SA of specific axioms.
The language LSA defined by (11.3) is called the language of the theory T .
The theory T (11.4) is based on a complete classical proof system
Definition 11.4
Given a theory T = (L, F, LA, SA, R). We denote by FSA the set of formulas
of the language LSA of T . We denote by T the set all provable formulas in the
theory T , i.e.
T = {B ∈ FSA : SA ` B.} (11.5)
We also write `T B to denote that B ∈ T.
Fact 11.1 The Hilbert style proof system H defined in chapter 9 is a theory
with equality with the set of specific axioms SA = ∅.
504
proof system H for classical predicate logic with a language
L = (L{¬,∩,∪,⇒} (P, F, C).
The first order formal theories are also called Elementary Theories.
T 1. Theory of equality
Language
LT 1 = L{¬,⇒,∪,∩} (P = {P }, F = ∅, C = ∅),
where # P = 2, i.e. P is a two argument predicate. The intended interpretation
of P is equality, so we use the equality symbol = instead of P . We write x = y
instead = (x, y). We write the language of T 1 as follows.
LT 1 = L{¬,⇒,∪,∩} ({=}, ∅, ∅).
Specific Axioms
e1 x = x,
e2 (x = y ⇒ y = x),
e3 (x = y ⇒ (y = z ⇒ x = z)),
for any x, y, z ∈ V AR,
Exercise 11.1
Show that the theory T 1 of equality is a theory with equality of definition 11.5.
Solution
The first to axioms e1, e2 are particular cases of E1, E2. We have only to show
that the axiom E3 is provable in T 1, i.e. that the formula
((x = y ∩ y = z) ⇒ x = z) ∈ T1, (11.7)
where, by (11.5) T1 = {A ∈ F{e1,e2,e3} : {e1, e2, e3} ` A}.
Observe that by definition, T 1 is based on a complete Hilbert style proof system.
A formula
(((A ⇒ (B ⇒ C)) ⇒ ((A ∩ B) ⇒ C))
is a predicate tautology, hence is provable in T 1 for any A, B, C ∈ F{e1,e2,e3} .
In particular its instance for A : x = y, B : y = z, C : x = z is also provable in
T 1 what means that
(((x = y ⇒ (y = z ⇒ x = z)) ⇒ ((x = y ∩ y = z) ⇒ x = z)) ∈ T1. (11.8)
Applying Modus Ponens (MP) to axiom e3 and (11.8), we get that
((x = y ∩ y = z) ⇒ x = z) ∈ T1.
It proves that (11.7) holds and ends the proof.
505
Observation 11.1 We have chosen to write the specific axioms as open for-
mulas. Sometimes it is more convenient to write them as closed formulas (sen-
tences). In this case new axioms will be closures of axioms that were open
formulas.
Specific Axioms
There are two groups of specific axioms: equality and order axioms. We adopt
equality axioms (11.6) to the language LT 3 as follows.
Equality Axioms
For any x, y, z, x1 , x2 , y1 , y2 , ∈ V AR,
e1 x = x,
e2 (x = y ⇒ y = x),
e3 ((x = y ∩ y = z) ⇒ x = z),
e4 ((x1 = y1 ∩ x2 = y2 ) ⇒ (x1 ≤ x2 ⇒ y1 ≤ y2 )).
506
Partial Order Axioms
o1 x ≤ x, (reflexivity)
o2 ((x ≤ y ∩ y ≤ x) ⇒ x = y), (antisymmetry)
o3 ((x ≤ y ∩ y ≤ z) ⇒ x ≤ z), (trasitivity )
where x, y, z ∈ V AR.
The model of T 3 is called a partially ordered structure.
Specific Axioms
For any x, y, z ∈ V AR,
p1 x 6< x, (irreflexivity)
p2 ((x ≤ y ∩ y ≤ z) ⇒ x ≤ z). (trasitivity )
Specific Axioms
We adopt all axioms of theory T 3 of partial order and add the following addi-
tional axiom.
o4 (x ≤ y) ∪ (y ≤ x).
This axiom says that in linearly ordered sets each two elements are comparable.
507
T 6. Theory of Dense Order
Language
LT 6 = L{¬,⇒,∪,∩} ({=, ≤}, ∅, ∅).
Specific Axioms
We adopt all axioms of theory T 5 of linear order and add the following additional
axiom. We write x 6= y for ¬(x = y), i.e. for the formula ¬ = (x, y).
o5 ((x ≤ y ∩ x 6= y) ⇒ ∃z((x ≤ z ∩ x 6= z) ∩ (z ≤ y ∩ z 6= y))).
This axiom says that in linearly ordered sets between any two different elements
there is a third element between them, respective to the order.
T 7. Lattice Theory
Language
Specific Axioms
There are three groups of specific axioms: equality axioms, order axioms, and
lattice axioms. We adopt equality axioms (11.6) to the language LT 7 as follows.
Equality Axioms
For any x, y, z, x1 , x2 , y1 , y2 , ∈ V AR,
e1 x = x,
e2 (x = y ⇒ y = x),
508
e3 ((x = y ∩ y = z) ⇒ x = z),
e4 ((x1 = y1 ∩ x2 = y2 ) ⇒ (x1 ≤ x2 ⇒ y1 ≤ y2 )),
e5 ((x1 = y1 ∩ x2 = y2 ) ⇒ (x1 ∧ x2 ⇒ y1 ∧ y2 )),
e6 ((x1 = y1 ∩ x2 = y2 ) ⇒ (x1 ∨ x2 ⇒ y1 ∨ y2 )).
Remark 11.2
We write ∧ for the lattice functional symbol of intersection in order to better
distinguish it from the conjunction symbol ∩ in the formula.
The same applies to the next axiom e7 that involves lattice functional symbol ∨
for the union and disjunction symbol ∪ in the formula.
Lattice Axioms
For any x, y, z ∈ V AR,
b1 (x ∧ y) = (y ∧ x), (x ∧ y) = (x ∧ y),
b2 (x ∧ (y ∧ z)) = ((x ∧ y) ∧ z), (x ∨ (y ∨ z)) = ((x ∨ y) ∨ z),
b3 (((x ∧ y) ∨ y) = y), ((x ∧ (x ∨ y)) = x).
Specific Axioms
We adopt all axioms of theory T 7 of lattice theory and add the following addi-
tional axiom.
b4 (x ∧ (y ∨ z)) = ((x ∧ y) ∨ (x ∧ z)).
509
where − is one argument function symbol representing algebra complement.
Specific Axioms
We adopt all axioms of theory T 8 of distributive lattices theory and add the
following additional axioms that characterize the algebra complement −.
b5 (((x ∧ −x) ∨ y) = y), (((x ∨ −x) ∧ y) = y).
−1
LT 10 = L{¬,⇒,∪,∩} ({=}, {◦, }, {e}).
Specific Axioms
There are two groups of specific axioms: equality axioms and group axioms. We
adopt equality axioms (11.6) to the language LT 10 as follows.
Equality Axioms
For any x, y, z, x1 , x2 , y1 , y2 , ∈ V AR,
e1 x = x,
e2 (x = y ⇒ y = x),
e3 ((x = y ∩ y = z) ⇒ x = z),
e4 (x = y ⇒ x−1 = y −1 ),
e5 ((x1 = y1 ∩ x2 = y2 ) ⇒ (x1 ◦ x2 ⇒ y1 ◦ y2 )).
Group Axioms
g1 (x ◦ (y ◦ z)) = ((x ◦ y) ◦ z),
g2 (x ◦ e) = x,
510
g3 (x ◦ x−1 ) = e.
−1
LT 11 = L{¬,⇒,∪,∩} ({=}, {◦, }, {e})0
Specific Axioms
We adopt all axioms of theory T 11 of groups and add the following additional
axiom.
g4 (x ◦ y) = (y ◦ x).
Specific Axioms
For any x, y, z, x1 , x2 , y1 , y2 , ∈ V AR,
a1 (x ◦ (y ◦ z)) = ((x ◦ y) ◦ z),
a2 (x ◦ e) = x,
a3 ∀x∃y((x ◦ y) = e),
a4 x = x,
a5 (x = y ⇒ y = x),
a6 (x = y ⇒ (y = z ⇒ x = z)),
a7 (x = y ⇒ (x ◦ z = y ∩ z ◦ x = z ◦ y)).
511
T 13. Theory of Abelian Groups (2)
We adopt the language and all axioms of theory T 12 of groups and add the
following additional axiom.
a8 (x ◦ y) = (y ◦ x).
Observe that what we formally prove in the formal axiomatic theories presented
here represents only fragments of corresponding axiomatic theories developed in
mathematics. For example Group Theory, Boolean Algebras Theory are fields
in mathematics and many theorems developed there, like the Representation
Theorem for Boolean Algebras, and many, many others in other domains can not
be expressed in the languages of respective formal theories. This is a reason why
we also call them elementary theories. For example, we say elementary group
theory to distinguish it from the Group Theory as a lager field of mathematics.
Next to geometry, the theory of natural numbers is the most intuitive and
intuitively known of all branches of mathematics. This is why the first attempts
to formalize mathematics begin with with arithmetic of natural numbers. The
first attempts of axiomatic formalization of arithmetic of natural numbers was
given by Dedekind in 1879 and by Peano in 1889. The Peano formalization
became known as Peano Postulates (axioms) and can be written as follows.
p1 0 is a natural number.
p2 If nis a natural number, there is another number which we denote by n0 .
We call n0 a successor of n. The intuitive meaning of n0 is n + 1.
p3 0 6= n0 , for any natural number n.
p4 If n0 = m0 , then n = m, for any natural numbers n, m.
p5 If W is is a property that may or may not hold for natural numbers, and
if (i) 0 has the property W and (ii) whenever a natural number n has the
property W, then n0 has the property W,
then all natural numbers have the property W.
p5 is called Principle of Induction.
These axioms, together with a certain amount of set theory, are sufficient to
develop not only theory of natural numbers, but also theory of rational and
even real numbers. But they can’t act as a fully formal theory as they include
intuitive notions like ”property” and ”has a property”.
A formal theory of natural numbers based on Peano Postulates is referred in
512
literature as Peano Arithmetic, or simply PA. We present here formalization by
Mendelson (1973) that is included and worked out in smallest details in his book
Intoduction to Mathematical Logic (1987). We refer the reader to this excellent
book for details and further reading.
We additionally assume now that the system H has as one of ts inference rules
a a generalization rule
A(x)
(G) . (11.10)
∀xA(x)
We do so to facilitate use the Mendelson’s book as a supplementary reading to
the material included here and for additional reading for material not covered
here.
Remark 11.3
The Deduction Theorem as proved in chapter 9 holds for the proof system system
H defined by (11.9).
Specific Axioms
513
P1 (x = y ⇒ (x = z ⇒ y = z)),
P2 (x = y ⇒ x0 = y 0 ),
P3 0 6= x0 ,
P4 (x0 = y 0 ⇒ x = y),
P5 x + 0 = x,
P6 x + y 0 = (x + y)0
P7 x · 0 = 0,
P8 x · y 0 = (x · y) + x,
P9 (A(0) ⇒ (∀x(A(x) ⇒ A(x0 ) ⇒ ∀xA(x)))),
for all formulas A(x) of LP A and all x, y, z ∈ V AR.
By definition 11.4, given a theory T , we denote by T the set all provable formulas
in T . In particular, PA denotes the set of all formulas provable in Peano
Arithmetic PA.
514
Theory P A is one of many formalizations of Peano Arithmetic. They all repre-
sent what we call Peano Arithmetic if they have the same set of theorems. We
adopt hence the following definition.
Fact 11.2
Theory CP A is a Peano Arithmetic.
Proof
By definition 11.6 we have to show that PA = CPA. Observe that LCP A = LP A
, so we have to show that for any formula B of LP A ,
515
Both theories are based on the same Hilbert proof system H, so to prove (11.12)
means to prove that
(1) all axioms C1 − C8 of CP A are provable in P A and vice versa,
(2) all axioms P 1 − P 8 of LP A are provable in CP A.
Here are detailed proofs for axioms P1, and C1. The proofs for other axioms
follow the same pattern.
(1) We prove that the axiom C1 ∀x∀y∀z(x = y ⇒ (y = z ⇒ x = z)) is
provable in P A as follows.
Observe that axioms of CPA are closures of respective axioms of P A. Consider
axiom P1: (x = y ⇒ (y = z ⇒ x = z)). As the proof system H has a
generalization rule 11.10
A(x)
(G)
∀xA(x)
as its rule of inference, we obtain a proof B1, B2, B3, B4 of C1 as follows.
B1: (x = y ⇒ (x = z ⇒ y = z)), (axiom P1)
B2: ∀z(x = y ⇒ (x = z ⇒ y = z)), (GA)
B3: ∀y∀z(x = y ⇒ (x = z ⇒ y = z)), (GA)
B4: ∀x∀y∀z(x = y ⇒ (x = z ⇒ y = z)). (axiom C1)
This ends the proof of (1) for axioms P1, and C1.
(2) We prove that the axiom P1 (x = y ⇒ (y = z ⇒ x = z)) of LP A is
provable in CP A as follows.
By H completeness a predicate tautology
where term t is free for x in A(x) is provable in H for any formula A(x) of L and
hence for any formula A(x) of its particular sublanguage LP A . So its particular
case for A(x) = (x = y ⇒ (x = z ⇒ y = z)) and t = x is provable in CP A, i.e.
we have that the formula
is provable in CP A.
We construct a proof B1, B2, B3, B4, B5, B6, B7 of P1 in CP A in as follows.
B1 ∀x∀y∀z(x = y ⇒ (x = z ⇒ y = z)), (axiom C1)
B2 (∀x∀y∀z(x = y ⇒ (x = z ⇒ y = z)) ⇒ ∀y∀z(x = y ⇒ (x = z ⇒ y = z))),
by (11.15)
B3 ∀y∀z(x = y ⇒ (x = z ⇒ y = z)), MP on B1, B2
516
B4 (∀y∀z(x = y ⇒ (x = z ⇒ y = z)) ⇒ ∀z(x = y ⇒ (x = z ⇒ y = z))), by
(11.15)
B5 ∀z(x = y ⇒ (x = z ⇒ y = z)), MP on B3, B4
B6 (∀z(x = y ⇒ (x = z ⇒ y = z)) ⇒ (x = y ⇒ (x = z ⇒ y = z))), by (11.15)
B7 (x = y ⇒ (x = z ⇒ y = z)) MP on B5, B6
This ends the proof of (2) for axioms P1, and C1.
The proofs for other axioms is similar and are left as homework assignment.
Here are some more basic facts about P A.
Fact 11.3
The following formulas are provable in P A for any terms t, s, r of LP A .
P1’ (t = r ⇒ (t = s ⇒ r = s)),
P2’ (t = r ⇒ t0 = r0 ),
P3’ 0 6= t0 ,
P4’ (t0 = r0 ⇒ t = r),
P5’ t + 0 = t,
P6’ t + r0 = (t + r)0
P7’ t · 0 = 0,
P8’ t · r0 = (t · r) + t.
We named the properties as P1’- P8’ to stress the fact that they are generaliza-
tions of axioms P1 - P8 to the set of all terms of LP A .
Proof
We write the proof for P1’ as an example. Proofs of all other formulas follow
the same pattern.
Consider axiom P1: (x = y ⇒ (y = z ⇒ x = z)). By Fact 11.2 its closure
∀x∀y∀z(x = y ⇒ (x = z ⇒ y = z)) is provable in T hP A , i.e.
where term t is free for x in A(x) is provable in H for any formula A(x) of L and
hence for any formula A(x) of its particular sublanguage LP A . So its particular
case for A(x) = ∀y∀z(x = y ⇒ (x = z ⇒ y = z)) the formula (11.15) is provable
in T hP A . Observe that any term t is free for x in this particular A(x). We get
517
that for any term t,
`P A (t = r ⇒ (t = s ⇒ r = s)).
Fact 11.4
The following formulas are provable in P A for any terms t, s, r of LP A .
a1 t = t,
a2 (t = r ⇒ r = t),
a3 (t = r ⇒ (r = s ⇒ t = s)),
a4 (r = t ⇒ (t = s ⇒ r = s)),
a5 (t = r ⇒ (t + s = r + s)),
a6 t = 0 + t.
Proof
We use in the proof Fact 11.2, Fact 11.3, axioms of PA (11.11, and completeness
of the system H. We denote it in the comments. The details of the steps
518
are similar to the proof of Fact 11.3 and is left to the reader as as homework
assignment.
a1 We construct a proof of t = t in CP A in as follows.
B1 t + 0 = t, P5’ in Fact 11.3
B2 (t + 0 = t ⇒ (t + 0 = t ⇒ t = t)), P1’ in Fact 11.3 for t = t + 0, r = t, s = t
B3 (t + 0 = t ⇒ t = t), MP on B1, B2
B4 t = t. MP on B1, B3
a2 We construct a proof of (t = r ⇒ r = t) as follows.
B1 (t = r ⇒ (t = t ⇒ r = t)), P1’ in Fact 11.3 for r = t, s = t
B2 (t = t ⇒ (t = r ⇒ r = t)), B1, tautology
B3 t = r ⇒ r = t. MP on B2, a1
a3 We construct a proof of (t = r ⇒ (r = s ⇒ t = s)) as follows.
B1 (r = t ⇒ (r = s ⇒ t = s)), P1’ in Fact 11.3
B2 t = r ⇒ r = t, a2
B3 (t = r ⇒ r = t). MP on B1, B2
a4 We construct a proof of (r = t ⇒ (t = s ⇒ r = s)) as follows.
B1 (r = t ⇒ (t = s ⇒ r = s)), a3 for t = r, r = t
B2 (t = s ⇒ (r = t ⇒ r = s)), B1, tautology
B3 s = t ⇒ t = s, a2
B4 (s = t ⇒ (r = t ⇒ r = s)), B1, B2, tautology
B5 (r = t ⇒ (t = s ⇒ r = s)). B4, tautology
The proof uses the Deduction Theorem which holds for the proof system H
(Remark 11.3) and so can be use in P A.
We first apply the Induction Rule to A(z) : (x = y ⇒ x + z = y + z) to prove
`P A ∀z(x = y ⇒ x + z = y + z).
519
B2 y + 0 = y, P5’
B3 x = y, Hyp
B4 (x + 0 = x ⇒ (x = y ⇒ x + 0 = y), a3 for t = x + 0, r = x, s = y
B5 (x = y ⇒ x + 0 = y), MP on B1, B4
B6 x + 0 = y, MP on B3, B5
B7 (x+0 = y ⇒ (y+0 = y ⇒ x+0 = y+0), a4 for r = x+0, t = y, s = y = 0
B8 (y + 0 = y ⇒ x + 0 = y + 0), MP on B6, B7
B9 x + 0 = y + 0), MP on B2, B8
B10 (x = y ⇒ x + 0 = y + 0). B1- B9, Deduction Theorem
Thus, `P A A(0).
(ii) We prove that `P A ∀z(A(z) ⇒ A(z 0 ), i.e.
∀z((x = y ⇒ x + z = y + z) ⇒ (x = y ⇒ x + z 0 = y + z 0 )). Here the steps in
the proof.
C1 (x = y ⇒ x + z = y + z), Hyp
C2 x = y, Hyp
C3 x + z = (x + z)0 ,
0
P6’
C4 y + z 0 = (y + z)0 , P6’
C5 x + z = y + z), MP on B1, B2
C6 (x + z = y + z ⇒ (x + z)0 = (y + z)0 ) P2’ for t = x + z, r = y + z,
C7 (x + z)0 = (y + z)0 , MP on B5, B6
0 0
C8 x + z = y + z , a3 substitution and MP on B3, B7
C9 ((x = y ⇒ x + z = y + z) ⇒ x + z 0 = y + z 0 ) B1- B8, Deduction Theorem
This proves ` A(z) ⇒ A(z 0 ).
C10 (((x = y ⇒ x + 0 = y + 0) ⇒ ((x = y ⇒ x + z = y + z) ⇒ x + z 0 =
y + z 0 )) ⇒ ∀z(x = y ⇒ x + z = y + z)), P9 for A(z) : (x = y ⇒ x + z = y + z)
C11 ((x = y ⇒ x + z = y + z) ⇒ x + z 0 = y + z 0 )) ⇒ ∀z(x = y ⇒ x + z = y + z),
MP on C10 and B10
C12 ∀z(x = y ⇒ x + z = y + z), MP on C11 and C9
C13 ∀y∀z(x = y ⇒ x + z = y + z), (GA)
C14 ∀x∀y∀z(x = y ⇒ x + z = y + z), (GA)
Now we repeat here the proof of P1’ of Fact 11.3. We apply it step by step to
520
C14. We eliminate the quantifiers ∀x∀y∀z and replace variables x, y, z by terms
t, r, s using the tautology (11.15) (∀xA(x) ⇒ A(t)) and Modus Ponens. Finally,
we obtain the proof of a5, i.e.
`P A (t = r ⇒ (t + s = r + s)).
Fact 11.5
The following formulas are provable in P A for any terms t, s, r of LP A .
(i) t · (r + s) = (t · r) + (t · s), distributivity
(ii) (r + s) · t = (r · t) + (s · t), distributivity
(iii) (r · t) · s = r · (t · s), associativity of ·
(iv) (t + s = r + s ⇒ t = r), canlcellation law for +
Proof
(i) Prove `P A t · (x + z) = (x · y) + (x · z) by induction on z.
(ii) Prove from (i) and property t · r = r · t.
(iii) Prove `P A (x · y) · z = x · (y · z) by induction on z.
(iv) Prove (x + z = y + z ⇒ x = y) by induction on z
Fact 11.6 The following formulas are provable in P A for any terms t, s of
LP A .
521
1. t + 1 = t0 ,
2. t · 1 = t,
3. t · 2 = t + t,
4. (t + s = 0 ⇒ (t = 0 ∩ s = 0)),
5. (t 6= 0 ⇒ (s · t = 0 ⇒ s = 0)),
Proof
1. Major steps in the proof of t + 1 = t0 in P A are as follows.
The comments at each step explain how to reconstruct the formal proof from
the properties already proven.
B1 t + 00 = (t + 0)0 , P6’
B2 t + 0 = t, P5’
B3 (t + 0)0 = t0 , B2, P2’, MP
0 0
B4 t + 0 = t , B1, B3, Fact 11.4 a3, MP
B5 t + 1 = t0 . B4, abbreviation
2. Major steps in the proof of t · 1 = t in P A are as follows.
B1 t · 00 = t · 0 + t, P8’
B2 t · 0 = 0, P7’
B3 (t · 0) + t = 0 + t, B1, Fact 11.4 a4, MP
B4 t · 00 = 0 + t, B1, B3, Fact 11.4 a3, MP
B5 0 + t = t, Fact 11.4 a3, a6, MP
B6 t · 00 = t, B4, B5, Fact 11.4 a3, MP
B7 t + 1 = t0 , B6, abbreviation
3. Major steps in the proof of t · 2 = t + t in P A are as follows.
B1 t + 10 = (t · 10 ) + t, P8’
B2 t + 1 = t0 , part 2.
B3 (t · 1) + t = t + t, B2, Fact 11.4 a5, MP
B4 t · 10 = t + t, B1, B3, Fact 11.4 a3, MP
B5 t · 2 = t + t, B4, abbreviation
4. We prove (t + s = 0 ⇒ (t = 0 ∩ s = 0)) by the following steps.
(s1) We apply the Principle of Mathematical Induction to A(y) : (x + y = 0 ⇒
522
(x = 0 ∩ y = 0)) and prove
(s3) We repeat here the proof of P1’ of Fact 11.3. We apply it step by step to
(11.22). We eliminate the quantifiers ∀x∀y and replace variables x, y by terms
t, s using the tautology (11.15) (∀xA(x) ⇒ A(t)) and Modus Ponens. Finally,
we obtain the proof of 4., i.e.
`P A (t + s = 0 ⇒ (t = 0 ∩ s = 0)).
Fact 11.7
Let n, m be any natural numbers.
(1) If m 6= n, then m 6= n.
(2) m + n = m + n and m · mn = m · n are provable in P A .
(3) Any model for P A is infinite.
Proof
Assume m 6= n, then m < n or n < m. Assume m < n. Her are major steps in
the formal proof of n 6= m.
The proof uses the Deduction Theorem which holds for the proof system H
(Remark 11.3) and so can be use in P A.
B1 m = n, Hyp
B2 0000000 = 000000 B2 is abbreviation of B2 for m applications of 0 on left side
of equation and n applications of 0 on the right
B3 0 = t0 , for t = n − m − 1
523
The proof of the case n < m is similar and left to the reader.
(2) We use mathematical induction for natural numbers in the metalanguage
with respect to natural number n. Base case: m + 0 is m.
By P3’ m = m + 0, hence m + 0 = m + 0 and the base step holds.
Inductive step. Assume that m + n = m + n is provable. By P2’ and P6’ we
get (m + n)0 = m + (n)0 . But m + (n + 1) is (m + n)0 and n + 1 is (n)0 . Hence,
m + (n + 1) = m + n + 1 and by mathematical induction m + n = m + n is
provable in T hP A , for all n, m. The proof that m · mn = m · n is provable in
P A for all n, m is similar.
(3) By (2), in a model for PA the objects corresponding to numerals must be
distinct. But the set of numerals is infinitely countable, so universe of any model
for PS must contain infinitely countable subset and hence is infinite.
An order relation can be introduced by in PA as follows.
Fact 11.8
For any terms t, r, s of LP A , the following formulas are provable in P A.
o1 t ≤ t,
o2 (t ≤ s ⇒ (s ≤ r ⇒ t ≤ r)),
o3 ((t ≤ s ∩ s ≤ t) ⇒ t = s),
o4 (t ≤ s ⇒ (t + r ≤ s + r)),
o5 (r > 0 ⇒ (t > 0 ⇒ r · t > 0)).
There are several stronger forms of the the Principle of Mathematical Induction
P9 (A(0) ⇒ (∀x(A(x) ⇒ A(x0 ) ⇒ ∀xA(x)))) that are provable in P A. Here is
one of them.
524
Fact 11.9 (Complete Induction)
The following formula, called Complete Induction Principle is provable in P A.
We proved and cited only some of the basic properties corresponding to proper-
ties of arithmetic of natural numbers. There are many more of them, developed
in many Classical Logic textbooks. We refer the reader especially to Mendelson
(1997) that we found the most rigorous and complete. The proofs included here
are more precise and complete versions of the few of the Mendelson’s proofs.
We selected and proved some direct consequences Peano Arithmetic axioms not
only because they are needed as the starting point for a strict development of
the formal theory of arithmetic of natural numbers but also because they are
good examples of how one develops any formal theory.
From this point on one can generally translate onto the language LP A and
prove in the P A the results from any text on elementary number theory. Some
standard results of number theory are proved with the aid of theory of complex
variables and it is often not known whether elementary proofs (or proofs in P A
can be given for such theorems. The statements of some other results of number
theory cannot even be formulated in P A.
Hence a natural question about the strength and expressive powers of P A is a
very important one. We will address it shortly in next section with connection of
the formulation and proofs of Gödel Theorems. Gödel, in order to prove them
developed the huge scientific apparatus which grew into new field of Mathe-
matics of Recursion Theory, and into Theory of Computation with input from
Church and Turing.
We know by Ryll Nardzewski Theorem 11.1 that PA is not finitely axioma-
tizable. We want to bring reader’s attention a finitely axiomatizable proper
sub-theory of PA, RR, that has the same expressive power with respect to the
Gödel Theorems. Here it is, as formalized and discussed in detail in Mendelson’s
book.
525
Specific Axioms
r1 x = x,
r2 (x = y ⇒ y = x),
r3 (x = y ⇒ (y = z ⇒ x = z)),
r4 (x = y ⇒ x0 = y 0 ),
r5 (x = y ⇒ (x + z = y + z ⇒ z + x = z + y)),
r6 (x = y ⇒ (x · z = y · z ⇒ z · x = z · y)),
r7 (x0 = y 0 ⇒ x = y),
r8 0 6= x0 ,
r9 (x 6= 0 ⇒ ∃y x = y 0 ),
r10 x + 0 = x,
r11 x + y 0 = (x + y)0 ,
r12 x · 0 = 0,
r13 x · y 0 = x · y + x,
r14 (y = x · z + p ∩ ((p < x ∩ y < x · q + r) ∩ r < x) ⇒ p = r).
for any x, y, z, p, q, r ∈ V AR,
Axioms r1 - r13 are due to Robinson (1950), hence the name. Axiom r14 is due
to Mendelson (1973). It expresses the uniqueness of remainder. The relation <
is as defined by definition 11.8.
Gödel showed that there are closed formulas of the language LP A of P A that
are neither provable nor disprovable in P A, if P A is consistent. Hence there is
a formula that is true under standard interpretation but is not provable in P A.
We also see that the incompleteness of P A cannot be attributed to omission
of some essential axiom but has deeper underlying causes that apply to other
theories as well. Robinson proved in 1950, that the Gödel Theorems hold his
system RR. In particular RR has the same incompleteness property as P A.
526
Hilbert as a part of the Hilbert Program. They were concerned with notions of
consistency, completeness, and decidability. The answers to Hilbert problems
were given by Gödel in 1930 in a form of his two theorem. They are some of the
most important and influential results in twentieth century mathematics. We
will discuss here these notions and Gödel’s results.
Consider the Peano Arithmetics P A and a structure M = [M, I] for its language
LP A = L{¬,⇒,∪,∩} ({=}, {0 , +, ·}, {0}),
such that the universe M is the set N of natural numbers (nonnegative integers)
and the interpretation I is defined as follows
(1) the constant symbol 0 is interpreted as a natural number 0,
(2) the one argument function symbol 0 (successor) is interpreted as successor
operation (addition of 1) on natural numbers; succ(n) = n + 1,
(3) the two argument function symbols +, · are interpreted as ordinary addition
and multiplication in N,
527
(4) the predicate symbol ”=” is interpreted as equality relation in N.
Observe that if we recognize that the set N of natural numbers with the standard
interpretation, i.e. the structure (11.24) to be a model for P A, then, of course,
P A is consistent (model-consistent). However, semantic methods, involving a
fair amount of set-theoretic reasoning, are regarded by many (and were regarded
as such by Gödel) as too precarious to serve as basis of consistency proofs.
Moreover, we have not proved formally that the axioms of P A are true under
standard interpretation; we only have taken it as intuitively obvious. Hence for
this and other reasons it is common practice to take the model-consistency of
P A as un explicit, unproved assumption and to adopt, after Gödel the following
syntactic definition of consistency.
A∈T and ¬A ∈ T.
`T A and `T ¬A.
528
Observe that the definitions 11.12, 11.13 have purely syntactic meaning. They
express the common intuition what proper provability should mean. They say
that a provability (formal theory) is a good one (consistent) only when one can’t
prove a formula and its negation; and is inconsistent when it is possible to prove
a contradiction in it.
Here is one of basic characterization of consistent theories.
Proof
Let denote by CC the consistency condition in the definition 11.12 and by CT
consistency condition in the theorem 11.2.
1. We prove implication ” if CC, then CT”.
Assume not CT. This means that A ∈ T for all formulas A,
Definition 11.14
A theory T is consistent if and only if T 6= FSA , i.e. there is A of LSA , such
that A 6∈ T.
529
The next important characterization of a formal theory is the one of its complete-
ness understood as the ability of proving or disapproving any of its statements,
provided it is correctly formulated in its language.
`T A or `T ¬A. (11.30)
A∈T or ¬A ∈ T. (11.31)
A 6∈ T and ¬A 6∈ T.
530
seems to be a norm when one wants to really prove Gödel’s results). It is writ-
ten in a very condensed and general way and concentrates on presentation of
modern results. It assumes that readers are already familiar with the traditional
approach so beautifully presented in Mendelson’s book, but I encourage readers
to reach for it, as it is, in its own style a very interesting work.
We also want to bring to readers attention that the introduction to Smorynski’s
chapter contains an excellent discussion of Hilbert Program and its relationship
to Gödel’s results. It gives an explanation why and how devastating Gödel The-
orems were to the optimism reflected in Hilbert’s Consistency and Conservation
Programs.
531
Hilbert’s Consistency Program asks to devise a finitistic means of proving
the consistency of various formal systems encoding abstract reasoning with ideal
statements.
The Consistency Program is a natural outgrowth and successor to the Conser-
vation Program. There are two reasons for this.
R1. Consistency is the assertion that some string of symbols is not provable.
Since derivations are simple combinatorial manipulations, this is a finitistically
meaningful and ought to have a finitistic proof.
R2. Proving a consistency of a formal system encoding the abstract concepts
already establishes the conservation result!
Reason R1 is straightforward. We will discuss R2 as it is particularly important.
Let’s denote by R a formal systems encoding real statements with their finitistic
proofs and by I the ideal system with its abstract reasoning.
Let A be a real statement ∀x(f (x) = g(x)).
Assume `I A. Then there is a derivation d of A in I. But, derivations are
concrete objects and, for some real formula P (x, y) encoding derivations in I,
`R P (d, pAq),
`R P (c, p¬Aq)
whence `R f (x) = g(x), with free variable x, i.e. `R ∀x(f (x) = g(x)).
To make the above argument rigorous, one has to define and explain the basics
of encoding, the assumptions on the formula P (x, y) and to deliver the whole
argument in a formal rigorous way, i.e. to develop rigorously the whole appa-
ratus developed originally by Gödel and needed for the proofs of his theorems.
We bring it here because it clearly invited Hilbert to establish his Consistency
Program. Since Consistency Program was as broad as the general Conservation
Program and, since it was more tractable, Hilbert fixed on it asserting:
”if the arbitrary given axioms do not contradict each other through their conse-
quences, then they are true, then the objects defined through the axioms exist.
That, for me, is the criterion of truth and existence”.
532
The Consistency Program had as its goal the proof, by finitistic means of the
consistence of strong systems. The solution would completely justify the use of
abstract concepts and would repudiate Brouwer and Weyl.
Gödel proved that it couldn’t work.
In 1920, while in his twenties, Kurt Gödel announced that Hilbert’s Consistency
Program could not be carried out. He had proved two theorems which gave
a blow to the Hilbert’s Program but on the other hand changed the face of
mathematics establishing mathematical logic as strong and rapidly developing
discipline.
Loosely stated these theorems are:
6`T ConT ,
Observe that the Second Incompleteness Theorem destroys the Consistency Pro-
gram. It states that R can’t prove its own consistency, so obviously it can’t prove
consistency of I.
Smorynski’s argument that the First Incompleteness Theorem destroys the Con-
servation Program is as follows. The the sentence A is real and is easily seen
to be true. It asserts its own unprovability and is indeed unprovable. Thus the
Conservation Program cannot be carried out and, hence, the same must hold
for the Consistency Program.
M. Detlefsen in the Appendix of his book ”Hilbert Program: An Essay on Math-
ematical Instrumentalism”, Springer, 2013, argues that Smorynski’s argument
is ambiguous, as he doesn’t tell us whether it is unprovability in R or unprov-
ability in I. We recommend to the reader interested a philosophical discussion
of Hilbert Program to read this Appendix, if not the whole book.
533
We will now formulate the Incompleteness Theorems in a more precise formal
way and describe the main ideas behind their proofs.
Observe that that in order to formalize the Incompleteness Theorems one has
first to ”translate” the sentences A and ConT into the language of T . For the
First Incompleteness Theorems 11.3 one needs to ” translate ” a self-referring
sentence ”I am not provable in a theory T”; for the Second Theorem 11.4 the
self-referring sentence is ”I am consistent”.
The assumption in both theorems is that T contains arithmetic means usually
it contains the Peano Arithmetic PA (11.11), or even its sub-theory RR (11.23),
called Robinson System. In this case the final product of such ”translation”
must be a sentence A or sentence ConT of the language LP A of PA, usually
written as
LP A = L({=}, {0 , +, ·}, {0}).
This ”translation” process into the language of some formal system containing
arithmetic is called arithmetization and encoding, or encoding for short. We
define a notion of arithmetization as follows.
An arithmetization of a theory T is a one-to-one function g from the set of
symbols of the language of T, expressions (formulas) of T, and finite sequences
of expressions of T (proofs) into the set of positive integers. The function g
must satisfy the following conditions.
(1) g is effectively computable;
(2) there is an effective procedure that determines whether any given positive
integer n is in the range of g and, if n is in the range of g, the procedure finds
the object x such that g(x) = m.
Arithmetization, i.e. a method of associating numbers with symbols, expres-
sions, and sequences of expressions was originally devised by Gödel in 1931 in
order to arithmetize Peano Arithmetic PA and encode the arithmetization pro-
cess PA in order to formulate and to prove his Incompleteness Theorems 11.3,
11.4.
Functions and relations whose arguments and values are natural numbers are
called the number-theoretic functions and relations.
In order to arithmetize and encode in a formal system we have to
1. associate numbers with symbols symbols of the language of the system, asso-
ciate numbers with expressions, and sequences of expressions of the language of
the system (arithmetization, encoding of basic syntax, and encoding of syntax)
2. replace assertions about the system by number-theoretic statements, and ex-
press these number-theoretic statements within the formal system itself ( arith-
metization,, encoding).
534
We want the number - theoretic function to be representable P A and the pred-
icates to be expressible in PA, i.e. their characteristic functions to be repre-
sentable in P A.
The study of representability of functions in P A leads to the class of number-
theoretic functions that turn out to be of great importance in mathematical
logic, namely the x primitive recursive and recursive functions. Their definition
and study in a form of a Recursion Theory is an important field of mathe-
matics and of computer science which developed out of the Gödel proof of the
Incompleteness Theorems.
We prove that the class of recursive functions is identical with the class of func-
tions representable in PA, i.e. we prove: every recursive function is representable
in PA and every function representable in PA is recursive.
The representability of primitive recursive and recursive functions in S in gen-
eral and in P A in particular plays crucial role in the encoding process and
consequently in the proof of Gödel Theorems.
The details of arithmetization and encoding are as complicated and tedious as
fascinating but are out of scope of our book. We recommend Mendelson’s book
”Introduction to Mathematical Logic”, 4th ed., Chapman & Hall (1997) as the
one with the most comprehensive and detailed presentation.
We assume at this moment that T is some fixed, but for a moment unspecified
consistent formal theory. We also assume that encoding is done in some fixed
theory S and that T contains S, i.e. the language of T is an extension of the
language of S and
S ⊆ T,
i.e. for any formula A,
if `S A, then `T A. (11.35)
We also assume that T and S contain as constants only numerals (definition
11.7)
0, 1, 2, 3, . . . ,
and T contains infinitely countably many functional and predicate symbols.
Usually S is taken to be a formal theory of arithmetic, but sometimes S can be
a weak set theory. But in any case S always contains numerals.
We also assume that theories T and S as defined by (11.34) are such that the
following Principles of Encoding (11.36) hold.
535
The mechanics, conditions and details of encoding for T and S for S being
Peano Arithmetic P A or its sub-theory Robinson Arithmetic RR (11.23) are
beautifully presented in the smallest detail in Mendelson.
The Smorynski’s approach we discuss here covers a larger class of formal theories
and uses a more general and modern approach. We can’t include all details but
we are convinced that at this stage the reader will be able to follow Smorynski’s
chapter in the Encyclopedia. The chapter is very well and clearly written and
is now classical. We wholeheartedly recommend it as a future reading.
We also follow Smorynski approach explaining what is to be encoded, where it
is to be encoded, and which are the most important encoding and provability
conditions needed for the proofs of the Incompleteness Theorems.
We first encode the syntax of T in S.
Since encoding takes place in S, it has a sufficient supply of constants (countably
infinite set of numerals
0, 1, 2, 3, . . . ,
and closed terms to be used as codes.
We assign to each formula A of the language of T a closed term,
pAq
called the code of A. If A(x) is a formula with a free variable x, then the code
pA(x)q is a closed term encoding the formula A(x), with x viewed as a syntactic
object and not as a parameter.
We do it recursively, first we assign codes (unique closed terms from S) to its
basic syntactic objects, i.e. elements of the alphabet of the language of T.
Terms and formulas are finite sequences of these symbols and derivations (formal
proofs) are also finite sequences of formulas. It means that S have to be able
to encode and manipulate finite sequences. We use for such encoding a class
primitive recursive functions and relations. We assume S admits a representation
of these functions and relations and finish encoding syntax.
S will also have to have certain function symbols and we have to be able to
encode them.
1. S must have we functional symbols, neg, impl, etc., corresponding to the
logical connectives and quantifiers, such that, such that, for all formulas A, B
of the language of T,
536
respectively,
`S sub(pA(x)q, ptq) = pA(t)q. (11.37)
Iteratation of sub allows one to define sub3 , sub4 , sub5 , . . . , such that
We define
P rT (y) ⇔ ∃xP rovT (x, y) (11.38)
and obtain a predicate asserting provability.
However, it is not always true
D1 `T A implies `S P rT (pAq).
537
important for convenience of references and the second name is routinely used
in computer science community.
Mendelson (1977) believes that the central idea was first explicitly mentions by
Carnap who pointed out in 1934 that the result was implicit in the work of
Gödel (1931). Gödel was not aware of Carnap work until 1937.
The theorem 11.5 is called Diagonalization Lemma because the argument used
in its proof has some resemblance to the the diagonal arguments used by Cantor
in 1891. He first used it proving that there are infinite sets that can not be put
in one-to-one correspondence with the set on natural numbers. He then used
its generalization in the proof of his famous Cantor Theorem: for every set X,
its set of all subsets has a larger cardinality than X itself (see chapter 1).
In mathematics, a fixed-point theorem is a name of a theorem saying that a
function f under some conditions, will have a at least one fixed point, i.e. a
point x such that f (x) = x.
The theorem 11.5 says that for any formula A in the language of theory T with
one free variable there is a sentence B such that the formula (B ⇔ A(pBq)) is
provable in T .
Intuitively, B is a self-referential sentence saying that B has property A. The
sentence B can be viewed as a fixed point of the operation assigning to each
formula A the sentence A(pBq). Hence the name Fixed Point Theorem.
Theorem 11.5 proves the existence of self-referential sentences in certain formal
theories of natural numbers. These sentences then, in turn, are to be used to
prove Gödel’s Incompleteness Theorems. Here it is.
`S (B ⇔ A(pBq)).
538
B ⇔ C(m) ⇔ A(sub(m, m))
`S (B ⇔ A(pBq)).
Proof
Applying Diagonalization Lemma 11.5 for a formula A(x) being ¬P rT (x), where
P rT (x) is defined by (11.38) we get that there is a sentence G such that
`S (G ⇔ ¬P rT (pGq)).
539
Let T, S be theories defined by (11.34).
Let ConT be a sentence ¬P rT (pCq)), where is C is any contradictory statement.
Then
6`T ConT .
Proof
Let G the Gödel’s sentence of the First Incompleteness Theorem 11.6.
We prove that
`T (ConT ⇔ G) (11.41)
and use it to prove that 6`T ConT . We conduct the proof by contradiction.
Assume `T ConT . By (11.41) `T (ConT ⇔ G), so `T G what contradicts
the First Incompleteness Theorem 11.6.
To complete the proof we have to to prove now (11.41). We know by Logic 11.1
that
`T (G ⇒ ¬P rT (pCq)). (11.42)
`S P rT (p(C ⇒ G)q).
Observe that we by the property (11.40) in the proof of the First Incompleteness
Theorem 11.3 we have
`S (G ⇒ ¬P rT (pGq)). (11.47)
540
We put (11.46) and (11.47) together and get
`T (ConT ⇒ G).
Remark 11.4
By definition 11.3 the theories T, S are based on a complete proof system for
predicate logic and by the monotonicity of classical consequence everything prov-
able there is provable in T, S. In particular all predicate tautologies are provable
in T and in S.
541
Logic 11.1
Given a complete proof system H, for any formulas A, B of the language of H,
Proof
1. We prove implication if ` (A ⇔ B), then ` (A ⇒ B) and ` (B ⇒ A).
Directly from provability of a tautology ((A ⇔ B) ⇒ ((A ⇒ B) ∩ (B ⇒ A))),
assumption ` (A ⇔ B), and MP we get ` ((A ⇒ B) ∩ (B ⇒ A)). Consequently,
from ` ((A ⇒ B) ∩ (B ⇒ A)), provability of tautologies ((A ∩ B) ⇒ A), ((A ∩
B) ⇒ B) and MP applied twice we get ` (A ⇒ B), ` (B ⇒ A).
2. We prove implication if ` (A ⇒ B) and ` (B ⇒ A), then ` (A ⇔ B).
Directly from provability of tautology ((A ⇒ B) ⇒ ((B ⇒ A) ⇒ (A ⇔ B))),
assumption ` (A ⇒ B), ` (B ⇒ A), MP applied twice we get ` (A ⇔ B).
Logic 11.2 Given a complete proof system H, for any formulas A, B of the
language of H,
Logic 11.3
Given a complete proof system H, for any formulas A, B of the language of H,
Logic 11.4
Given a complete proof system H, for any formulas A, B of the language of H,
542
Observation 11.2
We proved, a part of proof of the Second Incompleteness Theorem 11.7 the equiv-
alence (11.41) which says that the self-referential Gödel sentence G which asserts
its own unprovability is equivalent to the sentence asserting consistency. Hence,
the sentence G is unique up to provable equivalence (11.41) and we can say that
G is the sentence that asserts its own unprovability.
ω-consistency
We used, in the part (ii) of the First Incompleteness Theorem 11.6, an additional
assumption that `T P rT (pGq) implies `T G, instead of a habitual assumption
of ω-consistency.
The concept of ω-consistency was introduced by Gödel for purpose of stating
assumption needed for the proof of his First Incompleteness Theorem 11.3.
The modern researchers proved that the assuption of the ω-consistency can
be replaced, as we did, by other more general better suited for new proofs
conditions.
Informally, we say that T is ω- consistent if the following two conditions are
not satisfied for any formula A:
(i) `T ∃xA(x);
(ii) `T ¬A(n) for every natural number n.
Formally, ω-consistency can be represented (in varying degrees of generality) by
(modification of) the following formula
543
set T rM defines a model M of U :
c0 , c1 , c2 . . . ,
A0 , A1 , A2 , . . . (11.53)
in this augmented language and defines a complete theory by staring with U and
adding at each step n a sentence An , or ¬An according to whether An is con-
sistent with what has been chosen before or not.
The construction is then described within PA. Assuming ConU one can also
prove that the construction never terminates. The resulting set of sentences
forms a complete theory which by axioms (11.52) forms a model of U. Inspection
shows that the truth definition T rM of type ∆2 .
544
Proof
Assume PA is complete. Then, since PA is true, `P A ConP A and we can apply
the completeness theorem 11.8 to obtain a formula T rM which gives a truth
definition for the model of PA. Observe that once P A is complete we have that
P rP A is T rM . We choose G by
`P A (G ⇔ ¬T rM (pGq)). (11.54)
Property 11.1
The sentence GS of the First Incompleteness Theorem 11.6 asserting its own
provability is
(i) unique up to provable equivalence (Observation 11.2);
(ii) the sentence is Π1 and hence true.
The sentence G of the First Incompleteness Theorem 11.6 asserting its own
falsity in the model constructed is
(iii) not unique - for the following implication holds
(iv) the sentence is ∆2 (theorem 11.8, and, by (iii) there is no obvious way od
deciding its truth or falsity.
Georg Kreisler was the first to present a model- theoretic proof of the
following.
545
Then we show, for any presentation of the Henkin proof construction (as given
by encoding, the enumeration of sentences (11.53) . . . etc.) there is a number m
such that, for any model N of P A, the sequence of models determined by the
given presentations must stop after fewer then m steps with a model in which
ConP A is false.
2. Prove the case of axioms P2, C2 and axioms P23, C3 of the Fact 11.2.
3. Prove Fact 11.2 in case of axioms P5, C5 and axioms P8, C8 of the Fact
11.2.
6. Follow the definition 11.8 and prove the following formulas pre provable
in PA for ant terms t, r, s.
(i) t 6< t.
(ii) (t < s ⇒ (s < r ⇒ t < r)).
(iii) (0 < 1), (1 < 2), (2 < 3), (3 < 4), . . . .
(iv) 0 ≤ t.
(v) t ≤ t.
(vi) (t ≤ r ∪ r ≤ t).
(vii) (t ≤ r ⇒ (r ≤ t ⇒ t = r)).
7. Follow the definition 11.8 and prove the following formulas pre provable
in PA for ant terms t, r, s.
(i) (t ≤ s ⇒ (s ≤ r ⇒ t ≤ r)),
(ii) (t ≤ s ⇒ (t + r ≤ s + r)),
(ii) (r > 0 ⇒ (t > 0 ⇒ r · t > 0)).
546
9. Let RR be the Robinson System (11.23). Let n, m be any natural numbers.
Prove the following holds in RR.
(i) If m 6= n, then m 6= n.
(ii) m + n = m + n and m · mn = m · n are provable in RR .
(ii) Any model for RR is infinite.
(i) Write down a detailed proof of correctness of the last part of reasoning:
”But, if R proves consistency of I, we have
whence `R f (x) = g(x), with free variable x, i.e. `R ∀x(f (x) = g(x)).”
(ii) List, prove and use proper Logic Properties similar to properties Logic
11.1 - Logic 11.4 in the proof of Theorem 11.7.
547